|
Metadata-Version: 2.1 |
|
Name: pynvml |
|
Version: 11.5.0 |
|
Summary: Python Bindings for the NVIDIA Management Library |
|
Home-page: http://www.nvidia.com/ |
|
Author: NVIDIA Corporation |
|
Author-email: [email protected] |
|
License: BSD |
|
Classifier: Development Status :: 5 - Production/Stable |
|
Classifier: Intended Audience :: Developers |
|
Classifier: Intended Audience :: System Administrators |
|
Classifier: License :: OSI Approved :: BSD License |
|
Classifier: Operating System :: Microsoft :: Windows |
|
Classifier: Operating System :: POSIX :: Linux |
|
Classifier: Programming Language :: Python |
|
Classifier: Topic :: Software Development :: Libraries :: Python Modules |
|
Classifier: Topic :: System :: Hardware |
|
Classifier: Topic :: System :: Systems Administration |
|
Requires-Python: >=3.6 |
|
Description-Content-Type: text/markdown |
|
License-File: LICENSE.txt |
|
|
|
Python bindings to the NVIDIA Management Library |
|
================================================ |
|
|
|
Provides a Python interface to GPU management and monitoring functions. |
|
|
|
This is a wrapper around the NVML library. |
|
For information about the NVML library, see the NVML developer page |
|
http://developer.nvidia.com/nvidia-management-library-nvml |
|
|
|
As of version 11.0.0, the NVML-wrappers used in pynvml are identical |
|
to those published through [nvidia-ml-py](https://pypi.org/project/nvidia-ml-py/). |
|
|
|
Note that this file can be run with 'python -m doctest -v README.txt' |
|
although the results are system dependent |
|
|
|
Requires |
|
-------- |
|
Python 3, or an earlier version with the ctypes module. |
|
|
|
Installation |
|
------------ |
|
|
|
pip install . |
|
|
|
Usage |
|
----- |
|
|
|
You can use the lower level nvml bindings |
|
|
|
```python |
|
>>> from pynvml import * |
|
>>> nvmlInit() |
|
>>> print("Driver Version:", nvmlSystemGetDriverVersion()) |
|
Driver Version: 410.00 |
|
>>> deviceCount = nvmlDeviceGetCount() |
|
>>> for i in range(deviceCount): |
|
... handle = nvmlDeviceGetHandleByIndex(i) |
|
... print("Device", i, ":", nvmlDeviceGetName(handle)) |
|
... |
|
Device 0 : Tesla V100 |
|
|
|
>>> nvmlShutdown() |
|
``` |
|
|
|
Or the higher level nvidia_smi API |
|
|
|
```python |
|
from pynvml.smi import nvidia_smi |
|
nvsmi = nvidia_smi.getInstance() |
|
nvsmi.DeviceQuery('memory.free, memory.total') |
|
``` |
|
|
|
```python |
|
from pynvml.smi import nvidia_smi |
|
nvsmi = nvidia_smi.getInstance() |
|
print(nvsmi.DeviceQuery('--help-query-gpu'), end='\n') |
|
``` |
|
|
|
Functions |
|
--------- |
|
Python methods wrap NVML functions, implemented in a C shared library. |
|
Each function's use is the same with the following exceptions: |
|
|
|
- Instead of returning error codes, failing error codes are raised as |
|
Python exceptions. |
|
|
|
```python |
|
>>> try: |
|
... nvmlDeviceGetCount() |
|
... except NVMLError as error: |
|
... print(error) |
|
... |
|
Uninitialized |
|
``` |
|
|
|
- C function output parameters are returned from the corresponding |
|
Python function left to right. |
|
|
|
```c |
|
nvmlReturn_t nvmlDeviceGetEccMode(nvmlDevice_t device, |
|
nvmlEnableState_t *current, |
|
nvmlEnableState_t *pending); |
|
``` |
|
|
|
```python |
|
>>> nvmlInit() |
|
>>> handle = nvmlDeviceGetHandleByIndex(0) |
|
>>> (current, pending) = nvmlDeviceGetEccMode(handle) |
|
``` |
|
|
|
- C structs are converted into Python classes. |
|
|
|
```c |
|
nvmlReturn_t DECLDIR nvmlDeviceGetMemoryInfo(nvmlDevice_t device, |
|
nvmlMemory_t *memory); |
|
typedef struct nvmlMemory_st { |
|
unsigned long long total; |
|
unsigned long long free; |
|
unsigned long long used; |
|
} nvmlMemory_t; |
|
``` |
|
|
|
```python |
|
>>> info = nvmlDeviceGetMemoryInfo(handle) |
|
>>> print "Total memory:", info.total |
|
Total memory: 5636292608 |
|
>>> print "Free memory:", info.free |
|
Free memory: 5578420224 |
|
>>> print "Used memory:", info.used |
|
Used memory: 57872384 |
|
``` |
|
|
|
- Python handles string buffer creation. |
|
|
|
```c |
|
nvmlReturn_t nvmlSystemGetDriverVersion(char* version, |
|
unsigned int length); |
|
``` |
|
|
|
```python |
|
>>> version = nvmlSystemGetDriverVersion(); |
|
>>> nvmlShutdown() |
|
``` |
|
|
|
For usage information see the NVML documentation. |
|
|
|
Variables |
|
--------- |
|
|
|
All meaningful NVML constants and enums are exposed in Python. |
|
|
|
The NVML_VALUE_NOT_AVAILABLE constant is not used. Instead None is mapped to the field. |
|
|
|
NVML Permissions |
|
---------------- |
|
|
|
Many of the `pynvml` wrappers assume that the underlying NVIDIA Management Library (NVML) API can be used without admin/root privileges. However, it is certainly possible for the system permissions to prevent pynvml from querying GPU performance counters. For example: |
|
|
|
``` |
|
$ nvidia-smi nvlink -g 0 |
|
GPU 0: Tesla V100-SXM2-32GB (UUID: GPU-96ab329d-7a1f-73a8-a9b7-18b4b2855f92) |
|
NVML: Unable to get the NvLink link utilization counter control for link 0: Insufficient Permissions |
|
``` |
|
|
|
A simple way to check the permissions status is to look for `RmProfilingAdminOnly` in the driver `params` file (Note that `RmProfilingAdminOnly == 1` means that admin/sudo access is required): |
|
|
|
``` |
|
$ cat /proc/driver/nvidia/params | grep RmProfilingAdminOnly |
|
RmProfilingAdminOnly: 1 |
|
``` |
|
|
|
For more information on setting/unsetting the relevant admin privileges, see [these notes](https://developer.nvidia.com/nvidia-development-tools-solutions-ERR_NVGPUCTRPERM-permission-issue-performance-counters) on resolving `ERR_NVGPUCTRPERM` errors. |
|
|
|
|
|
Release Notes |
|
------------- |
|
|
|
- Version 2.285.0 |
|
- Added new functions for NVML 2.285. See NVML documentation for more information. |
|
- Ported to support Python 3.0 and Python 2.0 syntax. |
|
- Added nvidia_smi.py tool as a sample app. |
|
- Version 3.295.0 |
|
- Added new functions for NVML 3.295. See NVML documentation for more information. |
|
- Updated nvidia_smi.py tool |
|
- Includes additional error handling |
|
- Version 4.304.0 |
|
- Added new functions for NVML 4.304. See NVML documentation for more information. |
|
- Updated nvidia_smi.py tool |
|
- Version 4.304.3 |
|
- Fixing nvmlUnitGetDeviceCount bug |
|
- Version 5.319.0 |
|
- Added new functions for NVML 5.319. See NVML documentation for more information. |
|
- Version 6.340.0 |
|
- Added new functions for NVML 6.340. See NVML documentation for more information. |
|
- Version 7.346.0 |
|
- Added new functions for NVML 7.346. See NVML documentation for more information. |
|
- Version 7.352.0 |
|
- Added new functions for NVML 7.352. See NVML documentation for more information. |
|
- Version 8.0.0 |
|
- Refactor code to a nvidia_smi singleton class |
|
- Added DeviceQuery that returns a dictionary of (name, value). |
|
- Added filter parameters on DeviceQuery to match query api in nvidia-smi |
|
- Added filter parameters on XmlDeviceQuery to match query api in nvidia-smi |
|
- Added integer enumeration for filter strings to reduce overhead for performance monitoring. |
|
- Added loop(filter) method with async and callback support |
|
- Version 8.0.1 |
|
- Restructuring directories into two packages (pynvml and nvidia_smi) |
|
- Adding initial tests for both packages |
|
- Some name-convention cleanup in pynvml |
|
- Version 8.0.2 |
|
- Added NVLink function wrappers for pynvml module |
|
- Version 8.0.3 |
|
- Added versioneer |
|
- Fixed nvmlDeviceGetNvLinkUtilizationCounter bug |
|
- Version 8.0.4 |
|
- Added nvmlDeviceGetTotalEnergyConsumption |
|
- Added notes about NVML permissions |
|
- Fixed version-check testing |
|
- Version 11.0.0 |
|
- Updated nvml.py to CUDA 11 |
|
- Updated smi.py DeviceQuery to R460 |
|
- Aligned nvml.py with latest nvidia-ml-py deployment |
|
- Version 11.4.0 |
|
- Updated nvml.py to CUDA 11.4 |
|
- Updated smi.py NVML_BRAND_NAMES |
|
- Aligned nvml.py with latest nvidia-ml-py deployment (11.495.46) |
|
- Version 11.4.1 |
|
- Fix comma bugs in nvml.py |
|
- Version 11.5.0 |
|
- Updated nvml.py to support CUDA 11.5 and CUDA 12 |
|
- Aligned with latest nvidia-ml-py deployment (11.525.84) |
|
|