peacock-data-public-datasets-idc-cronscript
/
venv
/lib
/python3.10
/site-packages
/psutil
/__init__.py
# -*- coding: utf-8 -*- | |
# Copyright (c) 2009, Giampaolo Rodola'. All rights reserved. | |
# Use of this source code is governed by a BSD-style license that can be | |
# found in the LICENSE file. | |
"""psutil is a cross-platform library for retrieving information on | |
running processes and system utilization (CPU, memory, disks, network, | |
sensors) in Python. Supported platforms: | |
- Linux | |
- Windows | |
- macOS | |
- FreeBSD | |
- OpenBSD | |
- NetBSD | |
- Sun Solaris | |
- AIX | |
Works with Python versions 2.7 and 3.6+. | |
""" | |
from __future__ import division | |
import collections | |
import contextlib | |
import datetime | |
import functools | |
import os | |
import signal | |
import subprocess | |
import sys | |
import threading | |
import time | |
try: | |
import pwd | |
except ImportError: | |
pwd = None | |
from . import _common | |
from ._common import AIX | |
from ._common import BSD | |
from ._common import CONN_CLOSE | |
from ._common import CONN_CLOSE_WAIT | |
from ._common import CONN_CLOSING | |
from ._common import CONN_ESTABLISHED | |
from ._common import CONN_FIN_WAIT1 | |
from ._common import CONN_FIN_WAIT2 | |
from ._common import CONN_LAST_ACK | |
from ._common import CONN_LISTEN | |
from ._common import CONN_NONE | |
from ._common import CONN_SYN_RECV | |
from ._common import CONN_SYN_SENT | |
from ._common import CONN_TIME_WAIT | |
from ._common import FREEBSD # NOQA | |
from ._common import LINUX | |
from ._common import MACOS | |
from ._common import NETBSD # NOQA | |
from ._common import NIC_DUPLEX_FULL | |
from ._common import NIC_DUPLEX_HALF | |
from ._common import NIC_DUPLEX_UNKNOWN | |
from ._common import OPENBSD # NOQA | |
from ._common import OSX # deprecated alias | |
from ._common import POSIX # NOQA | |
from ._common import POWER_TIME_UNKNOWN | |
from ._common import POWER_TIME_UNLIMITED | |
from ._common import STATUS_DEAD | |
from ._common import STATUS_DISK_SLEEP | |
from ._common import STATUS_IDLE | |
from ._common import STATUS_LOCKED | |
from ._common import STATUS_PARKED | |
from ._common import STATUS_RUNNING | |
from ._common import STATUS_SLEEPING | |
from ._common import STATUS_STOPPED | |
from ._common import STATUS_TRACING_STOP | |
from ._common import STATUS_WAITING | |
from ._common import STATUS_WAKING | |
from ._common import STATUS_ZOMBIE | |
from ._common import SUNOS | |
from ._common import WINDOWS | |
from ._common import AccessDenied | |
from ._common import Error | |
from ._common import NoSuchProcess | |
from ._common import TimeoutExpired | |
from ._common import ZombieProcess | |
from ._common import memoize_when_activated | |
from ._common import wrap_numbers as _wrap_numbers | |
from ._compat import PY3 as _PY3 | |
from ._compat import PermissionError | |
from ._compat import ProcessLookupError | |
from ._compat import SubprocessTimeoutExpired as _SubprocessTimeoutExpired | |
from ._compat import long | |
if LINUX: | |
# This is public API and it will be retrieved from _pslinux.py | |
# via sys.modules. | |
PROCFS_PATH = "/proc" | |
from . import _pslinux as _psplatform | |
from ._pslinux import IOPRIO_CLASS_BE # NOQA | |
from ._pslinux import IOPRIO_CLASS_IDLE # NOQA | |
from ._pslinux import IOPRIO_CLASS_NONE # NOQA | |
from ._pslinux import IOPRIO_CLASS_RT # NOQA | |
elif WINDOWS: | |
from . import _pswindows as _psplatform | |
from ._psutil_windows import ABOVE_NORMAL_PRIORITY_CLASS # NOQA | |
from ._psutil_windows import BELOW_NORMAL_PRIORITY_CLASS # NOQA | |
from ._psutil_windows import HIGH_PRIORITY_CLASS # NOQA | |
from ._psutil_windows import IDLE_PRIORITY_CLASS # NOQA | |
from ._psutil_windows import NORMAL_PRIORITY_CLASS # NOQA | |
from ._psutil_windows import REALTIME_PRIORITY_CLASS # NOQA | |
from ._pswindows import CONN_DELETE_TCB # NOQA | |
from ._pswindows import IOPRIO_HIGH # NOQA | |
from ._pswindows import IOPRIO_LOW # NOQA | |
from ._pswindows import IOPRIO_NORMAL # NOQA | |
from ._pswindows import IOPRIO_VERYLOW # NOQA | |
elif MACOS: | |
from . import _psosx as _psplatform | |
elif BSD: | |
from . import _psbsd as _psplatform | |
elif SUNOS: | |
from . import _pssunos as _psplatform | |
from ._pssunos import CONN_BOUND # NOQA | |
from ._pssunos import CONN_IDLE # NOQA | |
# This is public writable API which is read from _pslinux.py and | |
# _pssunos.py via sys.modules. | |
PROCFS_PATH = "/proc" | |
elif AIX: | |
from . import _psaix as _psplatform | |
# This is public API and it will be retrieved from _pslinux.py | |
# via sys.modules. | |
PROCFS_PATH = "/proc" | |
else: # pragma: no cover | |
raise NotImplementedError('platform %s is not supported' % sys.platform) | |
# fmt: off | |
__all__ = [ | |
# exceptions | |
"Error", "NoSuchProcess", "ZombieProcess", "AccessDenied", | |
"TimeoutExpired", | |
# constants | |
"version_info", "__version__", | |
"STATUS_RUNNING", "STATUS_IDLE", "STATUS_SLEEPING", "STATUS_DISK_SLEEP", | |
"STATUS_STOPPED", "STATUS_TRACING_STOP", "STATUS_ZOMBIE", "STATUS_DEAD", | |
"STATUS_WAKING", "STATUS_LOCKED", "STATUS_WAITING", "STATUS_LOCKED", | |
"STATUS_PARKED", | |
"CONN_ESTABLISHED", "CONN_SYN_SENT", "CONN_SYN_RECV", "CONN_FIN_WAIT1", | |
"CONN_FIN_WAIT2", "CONN_TIME_WAIT", "CONN_CLOSE", "CONN_CLOSE_WAIT", | |
"CONN_LAST_ACK", "CONN_LISTEN", "CONN_CLOSING", "CONN_NONE", | |
# "CONN_IDLE", "CONN_BOUND", | |
"AF_LINK", | |
"NIC_DUPLEX_FULL", "NIC_DUPLEX_HALF", "NIC_DUPLEX_UNKNOWN", | |
"POWER_TIME_UNKNOWN", "POWER_TIME_UNLIMITED", | |
"BSD", "FREEBSD", "LINUX", "NETBSD", "OPENBSD", "MACOS", "OSX", "POSIX", | |
"SUNOS", "WINDOWS", "AIX", | |
# "RLIM_INFINITY", "RLIMIT_AS", "RLIMIT_CORE", "RLIMIT_CPU", "RLIMIT_DATA", | |
# "RLIMIT_FSIZE", "RLIMIT_LOCKS", "RLIMIT_MEMLOCK", "RLIMIT_NOFILE", | |
# "RLIMIT_NPROC", "RLIMIT_RSS", "RLIMIT_STACK", "RLIMIT_MSGQUEUE", | |
# "RLIMIT_NICE", "RLIMIT_RTPRIO", "RLIMIT_RTTIME", "RLIMIT_SIGPENDING", | |
# classes | |
"Process", "Popen", | |
# functions | |
"pid_exists", "pids", "process_iter", "wait_procs", # proc | |
"virtual_memory", "swap_memory", # memory | |
"cpu_times", "cpu_percent", "cpu_times_percent", "cpu_count", # cpu | |
"cpu_stats", # "cpu_freq", "getloadavg" | |
"net_io_counters", "net_connections", "net_if_addrs", # network | |
"net_if_stats", | |
"disk_io_counters", "disk_partitions", "disk_usage", # disk | |
# "sensors_temperatures", "sensors_battery", "sensors_fans" # sensors | |
"users", "boot_time", # others | |
] | |
# fmt: on | |
__all__.extend(_psplatform.__extra__all__) | |
# Linux, FreeBSD | |
if hasattr(_psplatform.Process, "rlimit"): | |
# Populate global namespace with RLIM* constants. | |
from . import _psutil_posix | |
_globals = globals() | |
_name = None | |
for _name in dir(_psutil_posix): | |
if _name.startswith('RLIM') and _name.isupper(): | |
_globals[_name] = getattr(_psutil_posix, _name) | |
__all__.append(_name) | |
del _globals, _name | |
AF_LINK = _psplatform.AF_LINK | |
__author__ = "Giampaolo Rodola'" | |
__version__ = "5.9.8" | |
version_info = tuple([int(num) for num in __version__.split('.')]) | |
_timer = getattr(time, 'monotonic', time.time) | |
_TOTAL_PHYMEM = None | |
_LOWEST_PID = None | |
_SENTINEL = object() | |
# Sanity check in case the user messed up with psutil installation | |
# or did something weird with sys.path. In this case we might end | |
# up importing a python module using a C extension module which | |
# was compiled for a different version of psutil. | |
# We want to prevent that by failing sooner rather than later. | |
# See: https://github.com/giampaolo/psutil/issues/564 | |
if int(__version__.replace('.', '')) != getattr( | |
_psplatform.cext, 'version', None | |
): | |
msg = "version conflict: %r C extension " % _psplatform.cext.__file__ | |
msg += "module was built for another version of psutil" | |
if hasattr(_psplatform.cext, 'version'): | |
msg += " (%s instead of %s)" % ( | |
'.'.join([x for x in str(_psplatform.cext.version)]), | |
__version__, | |
) | |
else: | |
msg += " (different than %s)" % __version__ | |
msg += "; you may try to 'pip uninstall psutil', manually remove %s" % ( | |
getattr( | |
_psplatform.cext, | |
"__file__", | |
"the existing psutil install directory", | |
) | |
) | |
msg += " or clean the virtual env somehow, then reinstall" | |
raise ImportError(msg) | |
# ===================================================================== | |
# --- Utils | |
# ===================================================================== | |
if hasattr(_psplatform, 'ppid_map'): | |
# Faster version (Windows and Linux). | |
_ppid_map = _psplatform.ppid_map | |
else: # pragma: no cover | |
def _ppid_map(): | |
"""Return a {pid: ppid, ...} dict for all running processes in | |
one shot. Used to speed up Process.children(). | |
""" | |
ret = {} | |
for pid in pids(): | |
try: | |
ret[pid] = _psplatform.Process(pid).ppid() | |
except (NoSuchProcess, ZombieProcess): | |
pass | |
return ret | |
def _pprint_secs(secs): | |
"""Format seconds in a human readable form.""" | |
now = time.time() | |
secs_ago = int(now - secs) | |
fmt = "%H:%M:%S" if secs_ago < 60 * 60 * 24 else "%Y-%m-%d %H:%M:%S" | |
return datetime.datetime.fromtimestamp(secs).strftime(fmt) | |
# ===================================================================== | |
# --- Process class | |
# ===================================================================== | |
class Process(object): # noqa: UP004 | |
"""Represents an OS process with the given PID. | |
If PID is omitted current process PID (os.getpid()) is used. | |
Raise NoSuchProcess if PID does not exist. | |
Note that most of the methods of this class do not make sure | |
the PID of the process being queried has been reused over time. | |
That means you might end up retrieving an information referring | |
to another process in case the original one this instance | |
refers to is gone in the meantime. | |
The only exceptions for which process identity is pre-emptively | |
checked and guaranteed are: | |
- parent() | |
- children() | |
- nice() (set) | |
- ionice() (set) | |
- rlimit() (set) | |
- cpu_affinity (set) | |
- suspend() | |
- resume() | |
- send_signal() | |
- terminate() | |
- kill() | |
To prevent this problem for all other methods you can: | |
- use is_running() before querying the process | |
- if you're continuously iterating over a set of Process | |
instances use process_iter() which pre-emptively checks | |
process identity for every yielded instance | |
""" | |
def __init__(self, pid=None): | |
self._init(pid) | |
def _init(self, pid, _ignore_nsp=False): | |
if pid is None: | |
pid = os.getpid() | |
else: | |
if not _PY3 and not isinstance(pid, (int, long)): | |
msg = "pid must be an integer (got %r)" % pid | |
raise TypeError(msg) | |
if pid < 0: | |
msg = "pid must be a positive integer (got %s)" % pid | |
raise ValueError(msg) | |
try: | |
_psplatform.cext.check_pid_range(pid) | |
except OverflowError: | |
msg = "process PID out of range (got %s)" % pid | |
raise NoSuchProcess(pid, msg=msg) | |
self._pid = pid | |
self._name = None | |
self._exe = None | |
self._create_time = None | |
self._gone = False | |
self._pid_reused = False | |
self._hash = None | |
self._lock = threading.RLock() | |
# used for caching on Windows only (on POSIX ppid may change) | |
self._ppid = None | |
# platform-specific modules define an _psplatform.Process | |
# implementation class | |
self._proc = _psplatform.Process(pid) | |
self._last_sys_cpu_times = None | |
self._last_proc_cpu_times = None | |
self._exitcode = _SENTINEL | |
# cache creation time for later use in is_running() method | |
try: | |
self.create_time() | |
except AccessDenied: | |
# We should never get here as AFAIK we're able to get | |
# process creation time on all platforms even as a | |
# limited user. | |
pass | |
except ZombieProcess: | |
# Zombies can still be queried by this class (although | |
# not always) and pids() return them so just go on. | |
pass | |
except NoSuchProcess: | |
if not _ignore_nsp: | |
msg = "process PID not found" | |
raise NoSuchProcess(pid, msg=msg) | |
else: | |
self._gone = True | |
# This pair is supposed to identify a Process instance | |
# univocally over time (the PID alone is not enough as | |
# it might refer to a process whose PID has been reused). | |
# This will be used later in __eq__() and is_running(). | |
self._ident = (self.pid, self._create_time) | |
def __str__(self): | |
info = collections.OrderedDict() | |
info["pid"] = self.pid | |
if self._name: | |
info['name'] = self._name | |
with self.oneshot(): | |
try: | |
info["name"] = self.name() | |
info["status"] = self.status() | |
except ZombieProcess: | |
info["status"] = "zombie" | |
except NoSuchProcess: | |
info["status"] = "terminated" | |
except AccessDenied: | |
pass | |
if self._exitcode not in (_SENTINEL, None): | |
info["exitcode"] = self._exitcode | |
if self._create_time is not None: | |
info['started'] = _pprint_secs(self._create_time) | |
return "%s.%s(%s)" % ( | |
self.__class__.__module__, | |
self.__class__.__name__, | |
", ".join(["%s=%r" % (k, v) for k, v in info.items()]), | |
) | |
__repr__ = __str__ | |
def __eq__(self, other): | |
# Test for equality with another Process object based | |
# on PID and creation time. | |
if not isinstance(other, Process): | |
return NotImplemented | |
if OPENBSD or NETBSD: # pragma: no cover | |
# Zombie processes on Open/NetBSD have a creation time of | |
# 0.0. This covers the case when a process started normally | |
# (so it has a ctime), then it turned into a zombie. It's | |
# important to do this because is_running() depends on | |
# __eq__. | |
pid1, ctime1 = self._ident | |
pid2, ctime2 = other._ident | |
if pid1 == pid2: | |
if ctime1 and not ctime2: | |
try: | |
return self.status() == STATUS_ZOMBIE | |
except Error: | |
pass | |
return self._ident == other._ident | |
def __ne__(self, other): | |
return not self == other | |
def __hash__(self): | |
if self._hash is None: | |
self._hash = hash(self._ident) | |
return self._hash | |
def _raise_if_pid_reused(self): | |
"""Raises NoSuchProcess in case process PID has been reused.""" | |
if not self.is_running() and self._pid_reused: | |
# We may directly raise NSP in here already if PID is just | |
# not running, but I prefer NSP to be raised naturally by | |
# the actual Process API call. This way unit tests will tell | |
# us if the API is broken (aka don't raise NSP when it | |
# should). We also remain consistent with all other "get" | |
# APIs which don't use _raise_if_pid_reused(). | |
msg = "process no longer exists and its PID has been reused" | |
raise NoSuchProcess(self.pid, self._name, msg=msg) | |
def pid(self): | |
"""The process PID.""" | |
return self._pid | |
# --- utility methods | |
def oneshot(self): | |
"""Utility context manager which considerably speeds up the | |
retrieval of multiple process information at the same time. | |
Internally different process info (e.g. name, ppid, uids, | |
gids, ...) may be fetched by using the same routine, but | |
only one information is returned and the others are discarded. | |
When using this context manager the internal routine is | |
executed once (in the example below on name()) and the | |
other info are cached. | |
The cache is cleared when exiting the context manager block. | |
The advice is to use this every time you retrieve more than | |
one information about the process. If you're lucky, you'll | |
get a hell of a speedup. | |
>>> import psutil | |
>>> p = psutil.Process() | |
>>> with p.oneshot(): | |
... p.name() # collect multiple info | |
... p.cpu_times() # return cached value | |
... p.cpu_percent() # return cached value | |
... p.create_time() # return cached value | |
... | |
>>> | |
""" | |
with self._lock: | |
if hasattr(self, "_cache"): | |
# NOOP: this covers the use case where the user enters the | |
# context twice: | |
# | |
# >>> with p.oneshot(): | |
# ... with p.oneshot(): | |
# ... | |
# | |
# Also, since as_dict() internally uses oneshot() | |
# I expect that the code below will be a pretty common | |
# "mistake" that the user will make, so let's guard | |
# against that: | |
# | |
# >>> with p.oneshot(): | |
# ... p.as_dict() | |
# ... | |
yield | |
else: | |
try: | |
# cached in case cpu_percent() is used | |
self.cpu_times.cache_activate(self) | |
# cached in case memory_percent() is used | |
self.memory_info.cache_activate(self) | |
# cached in case parent() is used | |
self.ppid.cache_activate(self) | |
# cached in case username() is used | |
if POSIX: | |
self.uids.cache_activate(self) | |
# specific implementation cache | |
self._proc.oneshot_enter() | |
yield | |
finally: | |
self.cpu_times.cache_deactivate(self) | |
self.memory_info.cache_deactivate(self) | |
self.ppid.cache_deactivate(self) | |
if POSIX: | |
self.uids.cache_deactivate(self) | |
self._proc.oneshot_exit() | |
def as_dict(self, attrs=None, ad_value=None): | |
"""Utility method returning process information as a | |
hashable dictionary. | |
If *attrs* is specified it must be a list of strings | |
reflecting available Process class' attribute names | |
(e.g. ['cpu_times', 'name']) else all public (read | |
only) attributes are assumed. | |
*ad_value* is the value which gets assigned in case | |
AccessDenied or ZombieProcess exception is raised when | |
retrieving that particular process information. | |
""" | |
valid_names = _as_dict_attrnames | |
if attrs is not None: | |
if not isinstance(attrs, (list, tuple, set, frozenset)): | |
msg = "invalid attrs type %s" % type(attrs) | |
raise TypeError(msg) | |
attrs = set(attrs) | |
invalid_names = attrs - valid_names | |
if invalid_names: | |
msg = "invalid attr name%s %s" % ( | |
"s" if len(invalid_names) > 1 else "", | |
", ".join(map(repr, invalid_names)), | |
) | |
raise ValueError(msg) | |
retdict = {} | |
ls = attrs or valid_names | |
with self.oneshot(): | |
for name in ls: | |
try: | |
if name == 'pid': | |
ret = self.pid | |
else: | |
meth = getattr(self, name) | |
ret = meth() | |
except (AccessDenied, ZombieProcess): | |
ret = ad_value | |
except NotImplementedError: | |
# in case of not implemented functionality (may happen | |
# on old or exotic systems) we want to crash only if | |
# the user explicitly asked for that particular attr | |
if attrs: | |
raise | |
continue | |
retdict[name] = ret | |
return retdict | |
def parent(self): | |
"""Return the parent process as a Process object pre-emptively | |
checking whether PID has been reused. | |
If no parent is known return None. | |
""" | |
lowest_pid = _LOWEST_PID if _LOWEST_PID is not None else pids()[0] | |
if self.pid == lowest_pid: | |
return None | |
ppid = self.ppid() | |
if ppid is not None: | |
ctime = self.create_time() | |
try: | |
parent = Process(ppid) | |
if parent.create_time() <= ctime: | |
return parent | |
# ...else ppid has been reused by another process | |
except NoSuchProcess: | |
pass | |
def parents(self): | |
"""Return the parents of this process as a list of Process | |
instances. If no parents are known return an empty list. | |
""" | |
parents = [] | |
proc = self.parent() | |
while proc is not None: | |
parents.append(proc) | |
proc = proc.parent() | |
return parents | |
def is_running(self): | |
"""Return whether this process is running. | |
It also checks if PID has been reused by another process in | |
which case return False. | |
""" | |
if self._gone or self._pid_reused: | |
return False | |
try: | |
# Checking if PID is alive is not enough as the PID might | |
# have been reused by another process: we also want to | |
# verify process identity. | |
# Process identity / uniqueness over time is guaranteed by | |
# (PID + creation time) and that is verified in __eq__. | |
self._pid_reused = self != Process(self.pid) | |
return not self._pid_reused | |
except ZombieProcess: | |
# We should never get here as it's already handled in | |
# Process.__init__; here just for extra safety. | |
return True | |
except NoSuchProcess: | |
self._gone = True | |
return False | |
# --- actual API | |
def ppid(self): | |
"""The process parent PID. | |
On Windows the return value is cached after first call. | |
""" | |
# On POSIX we don't want to cache the ppid as it may unexpectedly | |
# change to 1 (init) in case this process turns into a zombie: | |
# https://github.com/giampaolo/psutil/issues/321 | |
# http://stackoverflow.com/questions/356722/ | |
# XXX should we check creation time here rather than in | |
# Process.parent()? | |
self._raise_if_pid_reused() | |
if POSIX: | |
return self._proc.ppid() | |
else: # pragma: no cover | |
self._ppid = self._ppid or self._proc.ppid() | |
return self._ppid | |
def name(self): | |
"""The process name. The return value is cached after first call.""" | |
# Process name is only cached on Windows as on POSIX it may | |
# change, see: | |
# https://github.com/giampaolo/psutil/issues/692 | |
if WINDOWS and self._name is not None: | |
return self._name | |
name = self._proc.name() | |
if POSIX and len(name) >= 15: | |
# On UNIX the name gets truncated to the first 15 characters. | |
# If it matches the first part of the cmdline we return that | |
# one instead because it's usually more explicative. | |
# Examples are "gnome-keyring-d" vs. "gnome-keyring-daemon". | |
try: | |
cmdline = self.cmdline() | |
except (AccessDenied, ZombieProcess): | |
# Just pass and return the truncated name: it's better | |
# than nothing. Note: there are actual cases where a | |
# zombie process can return a name() but not a | |
# cmdline(), see: | |
# https://github.com/giampaolo/psutil/issues/2239 | |
pass | |
else: | |
if cmdline: | |
extended_name = os.path.basename(cmdline[0]) | |
if extended_name.startswith(name): | |
name = extended_name | |
self._name = name | |
self._proc._name = name | |
return name | |
def exe(self): | |
"""The process executable as an absolute path. | |
May also be an empty string. | |
The return value is cached after first call. | |
""" | |
def guess_it(fallback): | |
# try to guess exe from cmdline[0] in absence of a native | |
# exe representation | |
cmdline = self.cmdline() | |
if cmdline and hasattr(os, 'access') and hasattr(os, 'X_OK'): | |
exe = cmdline[0] # the possible exe | |
# Attempt to guess only in case of an absolute path. | |
# It is not safe otherwise as the process might have | |
# changed cwd. | |
if ( | |
os.path.isabs(exe) | |
and os.path.isfile(exe) | |
and os.access(exe, os.X_OK) | |
): | |
return exe | |
if isinstance(fallback, AccessDenied): | |
raise fallback | |
return fallback | |
if self._exe is None: | |
try: | |
exe = self._proc.exe() | |
except AccessDenied as err: | |
return guess_it(fallback=err) | |
else: | |
if not exe: | |
# underlying implementation can legitimately return an | |
# empty string; if that's the case we don't want to | |
# raise AD while guessing from the cmdline | |
try: | |
exe = guess_it(fallback=exe) | |
except AccessDenied: | |
pass | |
self._exe = exe | |
return self._exe | |
def cmdline(self): | |
"""The command line this process has been called with.""" | |
return self._proc.cmdline() | |
def status(self): | |
"""The process current status as a STATUS_* constant.""" | |
try: | |
return self._proc.status() | |
except ZombieProcess: | |
return STATUS_ZOMBIE | |
def username(self): | |
"""The name of the user that owns the process. | |
On UNIX this is calculated by using *real* process uid. | |
""" | |
if POSIX: | |
if pwd is None: | |
# might happen if python was installed from sources | |
msg = "requires pwd module shipped with standard python" | |
raise ImportError(msg) | |
real_uid = self.uids().real | |
try: | |
return pwd.getpwuid(real_uid).pw_name | |
except KeyError: | |
# the uid can't be resolved by the system | |
return str(real_uid) | |
else: | |
return self._proc.username() | |
def create_time(self): | |
"""The process creation time as a floating point number | |
expressed in seconds since the epoch. | |
The return value is cached after first call. | |
""" | |
if self._create_time is None: | |
self._create_time = self._proc.create_time() | |
return self._create_time | |
def cwd(self): | |
"""Process current working directory as an absolute path.""" | |
return self._proc.cwd() | |
def nice(self, value=None): | |
"""Get or set process niceness (priority).""" | |
if value is None: | |
return self._proc.nice_get() | |
else: | |
self._raise_if_pid_reused() | |
self._proc.nice_set(value) | |
if POSIX: | |
def uids(self): | |
"""Return process UIDs as a (real, effective, saved) | |
namedtuple. | |
""" | |
return self._proc.uids() | |
def gids(self): | |
"""Return process GIDs as a (real, effective, saved) | |
namedtuple. | |
""" | |
return self._proc.gids() | |
def terminal(self): | |
"""The terminal associated with this process, if any, | |
else None. | |
""" | |
return self._proc.terminal() | |
def num_fds(self): | |
"""Return the number of file descriptors opened by this | |
process (POSIX only). | |
""" | |
return self._proc.num_fds() | |
# Linux, BSD, AIX and Windows only | |
if hasattr(_psplatform.Process, "io_counters"): | |
def io_counters(self): | |
"""Return process I/O statistics as a | |
(read_count, write_count, read_bytes, write_bytes) | |
namedtuple. | |
Those are the number of read/write calls performed and the | |
amount of bytes read and written by the process. | |
""" | |
return self._proc.io_counters() | |
# Linux and Windows | |
if hasattr(_psplatform.Process, "ionice_get"): | |
def ionice(self, ioclass=None, value=None): | |
"""Get or set process I/O niceness (priority). | |
On Linux *ioclass* is one of the IOPRIO_CLASS_* constants. | |
*value* is a number which goes from 0 to 7. The higher the | |
value, the lower the I/O priority of the process. | |
On Windows only *ioclass* is used and it can be set to 2 | |
(normal), 1 (low) or 0 (very low). | |
Available on Linux and Windows > Vista only. | |
""" | |
if ioclass is None: | |
if value is not None: | |
msg = "'ioclass' argument must be specified" | |
raise ValueError(msg) | |
return self._proc.ionice_get() | |
else: | |
self._raise_if_pid_reused() | |
return self._proc.ionice_set(ioclass, value) | |
# Linux / FreeBSD only | |
if hasattr(_psplatform.Process, "rlimit"): | |
def rlimit(self, resource, limits=None): | |
"""Get or set process resource limits as a (soft, hard) | |
tuple. | |
*resource* is one of the RLIMIT_* constants. | |
*limits* is supposed to be a (soft, hard) tuple. | |
See "man prlimit" for further info. | |
Available on Linux and FreeBSD only. | |
""" | |
if limits is not None: | |
self._raise_if_pid_reused() | |
return self._proc.rlimit(resource, limits) | |
# Windows, Linux and FreeBSD only | |
if hasattr(_psplatform.Process, "cpu_affinity_get"): | |
def cpu_affinity(self, cpus=None): | |
"""Get or set process CPU affinity. | |
If specified, *cpus* must be a list of CPUs for which you | |
want to set the affinity (e.g. [0, 1]). | |
If an empty list is passed, all egible CPUs are assumed | |
(and set). | |
(Windows, Linux and BSD only). | |
""" | |
if cpus is None: | |
return sorted(set(self._proc.cpu_affinity_get())) | |
else: | |
self._raise_if_pid_reused() | |
if not cpus: | |
if hasattr(self._proc, "_get_eligible_cpus"): | |
cpus = self._proc._get_eligible_cpus() | |
else: | |
cpus = tuple(range(len(cpu_times(percpu=True)))) | |
self._proc.cpu_affinity_set(list(set(cpus))) | |
# Linux, FreeBSD, SunOS | |
if hasattr(_psplatform.Process, "cpu_num"): | |
def cpu_num(self): | |
"""Return what CPU this process is currently running on. | |
The returned number should be <= psutil.cpu_count() | |
and <= len(psutil.cpu_percent(percpu=True)). | |
It may be used in conjunction with | |
psutil.cpu_percent(percpu=True) to observe the system | |
workload distributed across CPUs. | |
""" | |
return self._proc.cpu_num() | |
# All platforms has it, but maybe not in the future. | |
if hasattr(_psplatform.Process, "environ"): | |
def environ(self): | |
"""The environment variables of the process as a dict. Note: this | |
might not reflect changes made after the process started. | |
""" | |
return self._proc.environ() | |
if WINDOWS: | |
def num_handles(self): | |
"""Return the number of handles opened by this process | |
(Windows only). | |
""" | |
return self._proc.num_handles() | |
def num_ctx_switches(self): | |
"""Return the number of voluntary and involuntary context | |
switches performed by this process. | |
""" | |
return self._proc.num_ctx_switches() | |
def num_threads(self): | |
"""Return the number of threads used by this process.""" | |
return self._proc.num_threads() | |
if hasattr(_psplatform.Process, "threads"): | |
def threads(self): | |
"""Return threads opened by process as a list of | |
(id, user_time, system_time) namedtuples representing | |
thread id and thread CPU times (user/system). | |
On OpenBSD this method requires root access. | |
""" | |
return self._proc.threads() | |
def children(self, recursive=False): | |
"""Return the children of this process as a list of Process | |
instances, pre-emptively checking whether PID has been reused. | |
If *recursive* is True return all the parent descendants. | |
Example (A == this process): | |
A ─┐ | |
│ | |
├─ B (child) ─┐ | |
│ └─ X (grandchild) ─┐ | |
│ └─ Y (great grandchild) | |
├─ C (child) | |
└─ D (child) | |
>>> import psutil | |
>>> p = psutil.Process() | |
>>> p.children() | |
B, C, D | |
>>> p.children(recursive=True) | |
B, X, Y, C, D | |
Note that in the example above if process X disappears | |
process Y won't be listed as the reference to process A | |
is lost. | |
""" | |
self._raise_if_pid_reused() | |
ppid_map = _ppid_map() | |
ret = [] | |
if not recursive: | |
for pid, ppid in ppid_map.items(): | |
if ppid == self.pid: | |
try: | |
child = Process(pid) | |
# if child happens to be older than its parent | |
# (self) it means child's PID has been reused | |
if self.create_time() <= child.create_time(): | |
ret.append(child) | |
except (NoSuchProcess, ZombieProcess): | |
pass | |
else: | |
# Construct a {pid: [child pids]} dict | |
reverse_ppid_map = collections.defaultdict(list) | |
for pid, ppid in ppid_map.items(): | |
reverse_ppid_map[ppid].append(pid) | |
# Recursively traverse that dict, starting from self.pid, | |
# such that we only call Process() on actual children | |
seen = set() | |
stack = [self.pid] | |
while stack: | |
pid = stack.pop() | |
if pid in seen: | |
# Since pids can be reused while the ppid_map is | |
# constructed, there may be rare instances where | |
# there's a cycle in the recorded process "tree". | |
continue | |
seen.add(pid) | |
for child_pid in reverse_ppid_map[pid]: | |
try: | |
child = Process(child_pid) | |
# if child happens to be older than its parent | |
# (self) it means child's PID has been reused | |
intime = self.create_time() <= child.create_time() | |
if intime: | |
ret.append(child) | |
stack.append(child_pid) | |
except (NoSuchProcess, ZombieProcess): | |
pass | |
return ret | |
def cpu_percent(self, interval=None): | |
"""Return a float representing the current process CPU | |
utilization as a percentage. | |
When *interval* is 0.0 or None (default) compares process times | |
to system CPU times elapsed since last call, returning | |
immediately (non-blocking). That means that the first time | |
this is called it will return a meaningful 0.0 value. | |
When *interval* is > 0.0 compares process times to system CPU | |
times elapsed before and after the interval (blocking). | |
In this case is recommended for accuracy that this function | |
be called with at least 0.1 seconds between calls. | |
A value > 100.0 can be returned in case of processes running | |
multiple threads on different CPU cores. | |
The returned value is explicitly NOT split evenly between | |
all available logical CPUs. This means that a busy loop process | |
running on a system with 2 logical CPUs will be reported as | |
having 100% CPU utilization instead of 50%. | |
Examples: | |
>>> import psutil | |
>>> p = psutil.Process(os.getpid()) | |
>>> # blocking | |
>>> p.cpu_percent(interval=1) | |
2.0 | |
>>> # non-blocking (percentage since last call) | |
>>> p.cpu_percent(interval=None) | |
2.9 | |
>>> | |
""" | |
blocking = interval is not None and interval > 0.0 | |
if interval is not None and interval < 0: | |
msg = "interval is not positive (got %r)" % interval | |
raise ValueError(msg) | |
num_cpus = cpu_count() or 1 | |
def timer(): | |
return _timer() * num_cpus | |
if blocking: | |
st1 = timer() | |
pt1 = self._proc.cpu_times() | |
time.sleep(interval) | |
st2 = timer() | |
pt2 = self._proc.cpu_times() | |
else: | |
st1 = self._last_sys_cpu_times | |
pt1 = self._last_proc_cpu_times | |
st2 = timer() | |
pt2 = self._proc.cpu_times() | |
if st1 is None or pt1 is None: | |
self._last_sys_cpu_times = st2 | |
self._last_proc_cpu_times = pt2 | |
return 0.0 | |
delta_proc = (pt2.user - pt1.user) + (pt2.system - pt1.system) | |
delta_time = st2 - st1 | |
# reset values for next call in case of interval == None | |
self._last_sys_cpu_times = st2 | |
self._last_proc_cpu_times = pt2 | |
try: | |
# This is the utilization split evenly between all CPUs. | |
# E.g. a busy loop process on a 2-CPU-cores system at this | |
# point is reported as 50% instead of 100%. | |
overall_cpus_percent = (delta_proc / delta_time) * 100 | |
except ZeroDivisionError: | |
# interval was too low | |
return 0.0 | |
else: | |
# Note 1: | |
# in order to emulate "top" we multiply the value for the num | |
# of CPU cores. This way the busy process will be reported as | |
# having 100% (or more) usage. | |
# | |
# Note 2: | |
# taskmgr.exe on Windows differs in that it will show 50% | |
# instead. | |
# | |
# Note 3: | |
# a percentage > 100 is legitimate as it can result from a | |
# process with multiple threads running on different CPU | |
# cores (top does the same), see: | |
# http://stackoverflow.com/questions/1032357 | |
# https://github.com/giampaolo/psutil/issues/474 | |
single_cpu_percent = overall_cpus_percent * num_cpus | |
return round(single_cpu_percent, 1) | |
def cpu_times(self): | |
"""Return a (user, system, children_user, children_system) | |
namedtuple representing the accumulated process time, in | |
seconds. | |
This is similar to os.times() but per-process. | |
On macOS and Windows children_user and children_system are | |
always set to 0. | |
""" | |
return self._proc.cpu_times() | |
def memory_info(self): | |
"""Return a namedtuple with variable fields depending on the | |
platform, representing memory information about the process. | |
The "portable" fields available on all platforms are `rss` and `vms`. | |
All numbers are expressed in bytes. | |
""" | |
return self._proc.memory_info() | |
def memory_info_ex(self): | |
return self.memory_info() | |
def memory_full_info(self): | |
"""This method returns the same information as memory_info(), | |
plus, on some platform (Linux, macOS, Windows), also provides | |
additional metrics (USS, PSS and swap). | |
The additional metrics provide a better representation of actual | |
process memory usage. | |
Namely USS is the memory which is unique to a process and which | |
would be freed if the process was terminated right now. | |
It does so by passing through the whole process address. | |
As such it usually requires higher user privileges than | |
memory_info() and is considerably slower. | |
""" | |
return self._proc.memory_full_info() | |
def memory_percent(self, memtype="rss"): | |
"""Compare process memory to total physical system memory and | |
calculate process memory utilization as a percentage. | |
*memtype* argument is a string that dictates what type of | |
process memory you want to compare against (defaults to "rss"). | |
The list of available strings can be obtained like this: | |
>>> psutil.Process().memory_info()._fields | |
('rss', 'vms', 'shared', 'text', 'lib', 'data', 'dirty', 'uss', 'pss') | |
""" | |
valid_types = list(_psplatform.pfullmem._fields) | |
if memtype not in valid_types: | |
msg = "invalid memtype %r; valid types are %r" % ( | |
memtype, | |
tuple(valid_types), | |
) | |
raise ValueError(msg) | |
fun = ( | |
self.memory_info | |
if memtype in _psplatform.pmem._fields | |
else self.memory_full_info | |
) | |
metrics = fun() | |
value = getattr(metrics, memtype) | |
# use cached value if available | |
total_phymem = _TOTAL_PHYMEM or virtual_memory().total | |
if not total_phymem > 0: | |
# we should never get here | |
msg = ( | |
"can't calculate process memory percent because total physical" | |
" system memory is not positive (%r)" % (total_phymem) | |
) | |
raise ValueError(msg) | |
return (value / float(total_phymem)) * 100 | |
if hasattr(_psplatform.Process, "memory_maps"): | |
def memory_maps(self, grouped=True): | |
"""Return process' mapped memory regions as a list of namedtuples | |
whose fields are variable depending on the platform. | |
If *grouped* is True the mapped regions with the same 'path' | |
are grouped together and the different memory fields are summed. | |
If *grouped* is False every mapped region is shown as a single | |
entity and the namedtuple will also include the mapped region's | |
address space ('addr') and permission set ('perms'). | |
""" | |
it = self._proc.memory_maps() | |
if grouped: | |
d = {} | |
for tupl in it: | |
path = tupl[2] | |
nums = tupl[3:] | |
try: | |
d[path] = map(lambda x, y: x + y, d[path], nums) | |
except KeyError: | |
d[path] = nums | |
nt = _psplatform.pmmap_grouped | |
return [nt(path, *d[path]) for path in d] # NOQA | |
else: | |
nt = _psplatform.pmmap_ext | |
return [nt(*x) for x in it] | |
def open_files(self): | |
"""Return files opened by process as a list of | |
(path, fd) namedtuples including the absolute file name | |
and file descriptor number. | |
""" | |
return self._proc.open_files() | |
def connections(self, kind='inet'): | |
"""Return socket connections opened by process as a list of | |
(fd, family, type, laddr, raddr, status) namedtuples. | |
The *kind* parameter filters for connections that match the | |
following criteria: | |
+------------+----------------------------------------------------+ | |
| Kind Value | Connections using | | |
+------------+----------------------------------------------------+ | |
| inet | IPv4 and IPv6 | | |
| inet4 | IPv4 | | |
| inet6 | IPv6 | | |
| tcp | TCP | | |
| tcp4 | TCP over IPv4 | | |
| tcp6 | TCP over IPv6 | | |
| udp | UDP | | |
| udp4 | UDP over IPv4 | | |
| udp6 | UDP over IPv6 | | |
| unix | UNIX socket (both UDP and TCP protocols) | | |
| all | the sum of all the possible families and protocols | | |
+------------+----------------------------------------------------+ | |
""" | |
return self._proc.connections(kind) | |
# --- signals | |
if POSIX: | |
def _send_signal(self, sig): | |
assert not self.pid < 0, self.pid | |
self._raise_if_pid_reused() | |
if self.pid == 0: | |
# see "man 2 kill" | |
msg = ( | |
"preventing sending signal to process with PID 0 as it " | |
"would affect every process in the process group of the " | |
"calling process (os.getpid()) instead of PID 0" | |
) | |
raise ValueError(msg) | |
try: | |
os.kill(self.pid, sig) | |
except ProcessLookupError: | |
if OPENBSD and pid_exists(self.pid): | |
# We do this because os.kill() lies in case of | |
# zombie processes. | |
raise ZombieProcess(self.pid, self._name, self._ppid) | |
else: | |
self._gone = True | |
raise NoSuchProcess(self.pid, self._name) | |
except PermissionError: | |
raise AccessDenied(self.pid, self._name) | |
def send_signal(self, sig): | |
"""Send a signal *sig* to process pre-emptively checking | |
whether PID has been reused (see signal module constants) . | |
On Windows only SIGTERM is valid and is treated as an alias | |
for kill(). | |
""" | |
if POSIX: | |
self._send_signal(sig) | |
else: # pragma: no cover | |
self._raise_if_pid_reused() | |
if sig != signal.SIGTERM and not self.is_running(): | |
msg = "process no longer exists" | |
raise NoSuchProcess(self.pid, self._name, msg=msg) | |
self._proc.send_signal(sig) | |
def suspend(self): | |
"""Suspend process execution with SIGSTOP pre-emptively checking | |
whether PID has been reused. | |
On Windows this has the effect of suspending all process threads. | |
""" | |
if POSIX: | |
self._send_signal(signal.SIGSTOP) | |
else: # pragma: no cover | |
self._raise_if_pid_reused() | |
self._proc.suspend() | |
def resume(self): | |
"""Resume process execution with SIGCONT pre-emptively checking | |
whether PID has been reused. | |
On Windows this has the effect of resuming all process threads. | |
""" | |
if POSIX: | |
self._send_signal(signal.SIGCONT) | |
else: # pragma: no cover | |
self._raise_if_pid_reused() | |
self._proc.resume() | |
def terminate(self): | |
"""Terminate the process with SIGTERM pre-emptively checking | |
whether PID has been reused. | |
On Windows this is an alias for kill(). | |
""" | |
if POSIX: | |
self._send_signal(signal.SIGTERM) | |
else: # pragma: no cover | |
self._raise_if_pid_reused() | |
self._proc.kill() | |
def kill(self): | |
"""Kill the current process with SIGKILL pre-emptively checking | |
whether PID has been reused. | |
""" | |
if POSIX: | |
self._send_signal(signal.SIGKILL) | |
else: # pragma: no cover | |
self._raise_if_pid_reused() | |
self._proc.kill() | |
def wait(self, timeout=None): | |
"""Wait for process to terminate and, if process is a children | |
of os.getpid(), also return its exit code, else None. | |
On Windows there's no such limitation (exit code is always | |
returned). | |
If the process is already terminated immediately return None | |
instead of raising NoSuchProcess. | |
If *timeout* (in seconds) is specified and process is still | |
alive raise TimeoutExpired. | |
To wait for multiple Process(es) use psutil.wait_procs(). | |
""" | |
if timeout is not None and not timeout >= 0: | |
msg = "timeout must be a positive integer" | |
raise ValueError(msg) | |
if self._exitcode is not _SENTINEL: | |
return self._exitcode | |
self._exitcode = self._proc.wait(timeout) | |
return self._exitcode | |
# The valid attr names which can be processed by Process.as_dict(). | |
# fmt: off | |
_as_dict_attrnames = set( | |
[x for x in dir(Process) if not x.startswith('_') and x not in | |
{'send_signal', 'suspend', 'resume', 'terminate', 'kill', 'wait', | |
'is_running', 'as_dict', 'parent', 'parents', 'children', 'rlimit', | |
'memory_info_ex', 'oneshot'}]) | |
# fmt: on | |
# ===================================================================== | |
# --- Popen class | |
# ===================================================================== | |
class Popen(Process): | |
"""Same as subprocess.Popen, but in addition it provides all | |
psutil.Process methods in a single class. | |
For the following methods which are common to both classes, psutil | |
implementation takes precedence: | |
* send_signal() | |
* terminate() | |
* kill() | |
This is done in order to avoid killing another process in case its | |
PID has been reused, fixing BPO-6973. | |
>>> import psutil | |
>>> from subprocess import PIPE | |
>>> p = psutil.Popen(["python", "-c", "print 'hi'"], stdout=PIPE) | |
>>> p.name() | |
'python' | |
>>> p.uids() | |
user(real=1000, effective=1000, saved=1000) | |
>>> p.username() | |
'giampaolo' | |
>>> p.communicate() | |
('hi', None) | |
>>> p.terminate() | |
>>> p.wait(timeout=2) | |
0 | |
>>> | |
""" | |
def __init__(self, *args, **kwargs): | |
# Explicitly avoid to raise NoSuchProcess in case the process | |
# spawned by subprocess.Popen terminates too quickly, see: | |
# https://github.com/giampaolo/psutil/issues/193 | |
self.__subproc = subprocess.Popen(*args, **kwargs) | |
self._init(self.__subproc.pid, _ignore_nsp=True) | |
def __dir__(self): | |
return sorted(set(dir(Popen) + dir(subprocess.Popen))) | |
def __enter__(self): | |
if hasattr(self.__subproc, '__enter__'): | |
self.__subproc.__enter__() | |
return self | |
def __exit__(self, *args, **kwargs): | |
if hasattr(self.__subproc, '__exit__'): | |
return self.__subproc.__exit__(*args, **kwargs) | |
else: | |
if self.stdout: | |
self.stdout.close() | |
if self.stderr: | |
self.stderr.close() | |
try: | |
# Flushing a BufferedWriter may raise an error. | |
if self.stdin: | |
self.stdin.close() | |
finally: | |
# Wait for the process to terminate, to avoid zombies. | |
self.wait() | |
def __getattribute__(self, name): | |
try: | |
return object.__getattribute__(self, name) | |
except AttributeError: | |
try: | |
return object.__getattribute__(self.__subproc, name) | |
except AttributeError: | |
msg = "%s instance has no attribute '%s'" % ( | |
self.__class__.__name__, | |
name, | |
) | |
raise AttributeError(msg) | |
def wait(self, timeout=None): | |
if self.__subproc.returncode is not None: | |
return self.__subproc.returncode | |
ret = super(Popen, self).wait(timeout) # noqa | |
self.__subproc.returncode = ret | |
return ret | |
# ===================================================================== | |
# --- system processes related functions | |
# ===================================================================== | |
def pids(): | |
"""Return a list of current running PIDs.""" | |
global _LOWEST_PID | |
ret = sorted(_psplatform.pids()) | |
_LOWEST_PID = ret[0] | |
return ret | |
def pid_exists(pid): | |
"""Return True if given PID exists in the current process list. | |
This is faster than doing "pid in psutil.pids()" and | |
should be preferred. | |
""" | |
if pid < 0: | |
return False | |
elif pid == 0 and POSIX: | |
# On POSIX we use os.kill() to determine PID existence. | |
# According to "man 2 kill" PID 0 has a special meaning | |
# though: it refers to <<every process in the process | |
# group of the calling process>> and that is not we want | |
# to do here. | |
return pid in pids() | |
else: | |
return _psplatform.pid_exists(pid) | |
_pmap = {} | |
def process_iter(attrs=None, ad_value=None): | |
"""Return a generator yielding a Process instance for all | |
running processes. | |
Every new Process instance is only created once and then cached | |
into an internal table which is updated every time this is used. | |
Cached Process instances are checked for identity so that you're | |
safe in case a PID has been reused by another process, in which | |
case the cached instance is updated. | |
The sorting order in which processes are yielded is based on | |
their PIDs. | |
*attrs* and *ad_value* have the same meaning as in | |
Process.as_dict(). If *attrs* is specified as_dict() is called | |
and the resulting dict is stored as a 'info' attribute attached | |
to returned Process instance. | |
If *attrs* is an empty list it will retrieve all process info | |
(slow). | |
""" | |
global _pmap | |
def add(pid): | |
proc = Process(pid) | |
if attrs is not None: | |
proc.info = proc.as_dict(attrs=attrs, ad_value=ad_value) | |
pmap[proc.pid] = proc | |
return proc | |
def remove(pid): | |
pmap.pop(pid, None) | |
pmap = _pmap.copy() | |
a = set(pids()) | |
b = set(pmap.keys()) | |
new_pids = a - b | |
gone_pids = b - a | |
for pid in gone_pids: | |
remove(pid) | |
try: | |
ls = sorted(list(pmap.items()) + list(dict.fromkeys(new_pids).items())) | |
for pid, proc in ls: | |
try: | |
if proc is None: # new process | |
yield add(pid) | |
else: | |
# use is_running() to check whether PID has been | |
# reused by another process in which case yield a | |
# new Process instance | |
if proc.is_running(): | |
if attrs is not None: | |
proc.info = proc.as_dict( | |
attrs=attrs, ad_value=ad_value | |
) | |
yield proc | |
else: | |
yield add(pid) | |
except NoSuchProcess: | |
remove(pid) | |
except AccessDenied: | |
# Process creation time can't be determined hence there's | |
# no way to tell whether the pid of the cached process | |
# has been reused. Just return the cached version. | |
if proc is None and pid in pmap: | |
try: | |
yield pmap[pid] | |
except KeyError: | |
# If we get here it is likely that 2 threads were | |
# using process_iter(). | |
pass | |
else: | |
raise | |
finally: | |
_pmap = pmap | |
def wait_procs(procs, timeout=None, callback=None): | |
"""Convenience function which waits for a list of processes to | |
terminate. | |
Return a (gone, alive) tuple indicating which processes | |
are gone and which ones are still alive. | |
The gone ones will have a new *returncode* attribute indicating | |
process exit status (may be None). | |
*callback* is a function which gets called every time a process | |
terminates (a Process instance is passed as callback argument). | |
Function will return as soon as all processes terminate or when | |
*timeout* occurs. | |
Differently from Process.wait() it will not raise TimeoutExpired if | |
*timeout* occurs. | |
Typical use case is: | |
- send SIGTERM to a list of processes | |
- give them some time to terminate | |
- send SIGKILL to those ones which are still alive | |
Example: | |
>>> def on_terminate(proc): | |
... print("process {} terminated".format(proc)) | |
... | |
>>> for p in procs: | |
... p.terminate() | |
... | |
>>> gone, alive = wait_procs(procs, timeout=3, callback=on_terminate) | |
>>> for p in alive: | |
... p.kill() | |
""" | |
def check_gone(proc, timeout): | |
try: | |
returncode = proc.wait(timeout=timeout) | |
except TimeoutExpired: | |
pass | |
except _SubprocessTimeoutExpired: | |
pass | |
else: | |
if returncode is not None or not proc.is_running(): | |
# Set new Process instance attribute. | |
proc.returncode = returncode | |
gone.add(proc) | |
if callback is not None: | |
callback(proc) | |
if timeout is not None and not timeout >= 0: | |
msg = "timeout must be a positive integer, got %s" % timeout | |
raise ValueError(msg) | |
gone = set() | |
alive = set(procs) | |
if callback is not None and not callable(callback): | |
msg = "callback %r is not a callable" % callback | |
raise TypeError(msg) | |
if timeout is not None: | |
deadline = _timer() + timeout | |
while alive: | |
if timeout is not None and timeout <= 0: | |
break | |
for proc in alive: | |
# Make sure that every complete iteration (all processes) | |
# will last max 1 sec. | |
# We do this because we don't want to wait too long on a | |
# single process: in case it terminates too late other | |
# processes may disappear in the meantime and their PID | |
# reused. | |
max_timeout = 1.0 / len(alive) | |
if timeout is not None: | |
timeout = min((deadline - _timer()), max_timeout) | |
if timeout <= 0: | |
break | |
check_gone(proc, timeout) | |
else: | |
check_gone(proc, max_timeout) | |
alive = alive - gone | |
if alive: | |
# Last attempt over processes survived so far. | |
# timeout == 0 won't make this function wait any further. | |
for proc in alive: | |
check_gone(proc, 0) | |
alive = alive - gone | |
return (list(gone), list(alive)) | |
# ===================================================================== | |
# --- CPU related functions | |
# ===================================================================== | |
def cpu_count(logical=True): | |
"""Return the number of logical CPUs in the system (same as | |
os.cpu_count() in Python 3.4). | |
If *logical* is False return the number of physical cores only | |
(e.g. hyper thread CPUs are excluded). | |
Return None if undetermined. | |
The return value is cached after first call. | |
If desired cache can be cleared like this: | |
>>> psutil.cpu_count.cache_clear() | |
""" | |
if logical: | |
ret = _psplatform.cpu_count_logical() | |
else: | |
ret = _psplatform.cpu_count_cores() | |
if ret is not None and ret < 1: | |
ret = None | |
return ret | |
def cpu_times(percpu=False): | |
"""Return system-wide CPU times as a namedtuple. | |
Every CPU time represents the seconds the CPU has spent in the | |
given mode. The namedtuple's fields availability varies depending on the | |
platform: | |
- user | |
- system | |
- idle | |
- nice (UNIX) | |
- iowait (Linux) | |
- irq (Linux, FreeBSD) | |
- softirq (Linux) | |
- steal (Linux >= 2.6.11) | |
- guest (Linux >= 2.6.24) | |
- guest_nice (Linux >= 3.2.0) | |
When *percpu* is True return a list of namedtuples for each CPU. | |
First element of the list refers to first CPU, second element | |
to second CPU and so on. | |
The order of the list is consistent across calls. | |
""" | |
if not percpu: | |
return _psplatform.cpu_times() | |
else: | |
return _psplatform.per_cpu_times() | |
try: | |
_last_cpu_times = {threading.current_thread().ident: cpu_times()} | |
except Exception: # noqa: BLE001 | |
# Don't want to crash at import time. | |
_last_cpu_times = {} | |
try: | |
_last_per_cpu_times = { | |
threading.current_thread().ident: cpu_times(percpu=True) | |
} | |
except Exception: # noqa: BLE001 | |
# Don't want to crash at import time. | |
_last_per_cpu_times = {} | |
def _cpu_tot_time(times): | |
"""Given a cpu_time() ntuple calculates the total CPU time | |
(including idle time). | |
""" | |
tot = sum(times) | |
if LINUX: | |
# On Linux guest times are already accounted in "user" or | |
# "nice" times, so we subtract them from total. | |
# Htop does the same. References: | |
# https://github.com/giampaolo/psutil/pull/940 | |
# http://unix.stackexchange.com/questions/178045 | |
# https://github.com/torvalds/linux/blob/ | |
# 447976ef4fd09b1be88b316d1a81553f1aa7cd07/kernel/sched/ | |
# cputime.c#L158 | |
tot -= getattr(times, "guest", 0) # Linux 2.6.24+ | |
tot -= getattr(times, "guest_nice", 0) # Linux 3.2.0+ | |
return tot | |
def _cpu_busy_time(times): | |
"""Given a cpu_time() ntuple calculates the busy CPU time. | |
We do so by subtracting all idle CPU times. | |
""" | |
busy = _cpu_tot_time(times) | |
busy -= times.idle | |
# Linux: "iowait" is time during which the CPU does not do anything | |
# (waits for IO to complete). On Linux IO wait is *not* accounted | |
# in "idle" time so we subtract it. Htop does the same. | |
# References: | |
# https://github.com/torvalds/linux/blob/ | |
# 447976ef4fd09b1be88b316d1a81553f1aa7cd07/kernel/sched/cputime.c#L244 | |
busy -= getattr(times, "iowait", 0) | |
return busy | |
def _cpu_times_deltas(t1, t2): | |
assert t1._fields == t2._fields, (t1, t2) | |
field_deltas = [] | |
for field in _psplatform.scputimes._fields: | |
field_delta = getattr(t2, field) - getattr(t1, field) | |
# CPU times are always supposed to increase over time | |
# or at least remain the same and that's because time | |
# cannot go backwards. | |
# Surprisingly sometimes this might not be the case (at | |
# least on Windows and Linux), see: | |
# https://github.com/giampaolo/psutil/issues/392 | |
# https://github.com/giampaolo/psutil/issues/645 | |
# https://github.com/giampaolo/psutil/issues/1210 | |
# Trim negative deltas to zero to ignore decreasing fields. | |
# top does the same. Reference: | |
# https://gitlab.com/procps-ng/procps/blob/v3.3.12/top/top.c#L5063 | |
field_delta = max(0, field_delta) | |
field_deltas.append(field_delta) | |
return _psplatform.scputimes(*field_deltas) | |
def cpu_percent(interval=None, percpu=False): | |
"""Return a float representing the current system-wide CPU | |
utilization as a percentage. | |
When *interval* is > 0.0 compares system CPU times elapsed before | |
and after the interval (blocking). | |
When *interval* is 0.0 or None compares system CPU times elapsed | |
since last call or module import, returning immediately (non | |
blocking). That means the first time this is called it will | |
return a meaningless 0.0 value which you should ignore. | |
In this case is recommended for accuracy that this function be | |
called with at least 0.1 seconds between calls. | |
When *percpu* is True returns a list of floats representing the | |
utilization as a percentage for each CPU. | |
First element of the list refers to first CPU, second element | |
to second CPU and so on. | |
The order of the list is consistent across calls. | |
Examples: | |
>>> # blocking, system-wide | |
>>> psutil.cpu_percent(interval=1) | |
2.0 | |
>>> | |
>>> # blocking, per-cpu | |
>>> psutil.cpu_percent(interval=1, percpu=True) | |
[2.0, 1.0] | |
>>> | |
>>> # non-blocking (percentage since last call) | |
>>> psutil.cpu_percent(interval=None) | |
2.9 | |
>>> | |
""" | |
tid = threading.current_thread().ident | |
blocking = interval is not None and interval > 0.0 | |
if interval is not None and interval < 0: | |
msg = "interval is not positive (got %r)" % interval | |
raise ValueError(msg) | |
def calculate(t1, t2): | |
times_delta = _cpu_times_deltas(t1, t2) | |
all_delta = _cpu_tot_time(times_delta) | |
busy_delta = _cpu_busy_time(times_delta) | |
try: | |
busy_perc = (busy_delta / all_delta) * 100 | |
except ZeroDivisionError: | |
return 0.0 | |
else: | |
return round(busy_perc, 1) | |
# system-wide usage | |
if not percpu: | |
if blocking: | |
t1 = cpu_times() | |
time.sleep(interval) | |
else: | |
t1 = _last_cpu_times.get(tid) or cpu_times() | |
_last_cpu_times[tid] = cpu_times() | |
return calculate(t1, _last_cpu_times[tid]) | |
# per-cpu usage | |
else: | |
ret = [] | |
if blocking: | |
tot1 = cpu_times(percpu=True) | |
time.sleep(interval) | |
else: | |
tot1 = _last_per_cpu_times.get(tid) or cpu_times(percpu=True) | |
_last_per_cpu_times[tid] = cpu_times(percpu=True) | |
for t1, t2 in zip(tot1, _last_per_cpu_times[tid]): | |
ret.append(calculate(t1, t2)) | |
return ret | |
# Use a separate dict for cpu_times_percent(), so it's independent from | |
# cpu_percent() and they can both be used within the same program. | |
_last_cpu_times_2 = _last_cpu_times.copy() | |
_last_per_cpu_times_2 = _last_per_cpu_times.copy() | |
def cpu_times_percent(interval=None, percpu=False): | |
"""Same as cpu_percent() but provides utilization percentages | |
for each specific CPU time as is returned by cpu_times(). | |
For instance, on Linux we'll get: | |
>>> cpu_times_percent() | |
cpupercent(user=4.8, nice=0.0, system=4.8, idle=90.5, iowait=0.0, | |
irq=0.0, softirq=0.0, steal=0.0, guest=0.0, guest_nice=0.0) | |
>>> | |
*interval* and *percpu* arguments have the same meaning as in | |
cpu_percent(). | |
""" | |
tid = threading.current_thread().ident | |
blocking = interval is not None and interval > 0.0 | |
if interval is not None and interval < 0: | |
msg = "interval is not positive (got %r)" % interval | |
raise ValueError(msg) | |
def calculate(t1, t2): | |
nums = [] | |
times_delta = _cpu_times_deltas(t1, t2) | |
all_delta = _cpu_tot_time(times_delta) | |
# "scale" is the value to multiply each delta with to get percentages. | |
# We use "max" to avoid division by zero (if all_delta is 0, then all | |
# fields are 0 so percentages will be 0 too. all_delta cannot be a | |
# fraction because cpu times are integers) | |
scale = 100.0 / max(1, all_delta) | |
for field_delta in times_delta: | |
field_perc = field_delta * scale | |
field_perc = round(field_perc, 1) | |
# make sure we don't return negative values or values over 100% | |
field_perc = min(max(0.0, field_perc), 100.0) | |
nums.append(field_perc) | |
return _psplatform.scputimes(*nums) | |
# system-wide usage | |
if not percpu: | |
if blocking: | |
t1 = cpu_times() | |
time.sleep(interval) | |
else: | |
t1 = _last_cpu_times_2.get(tid) or cpu_times() | |
_last_cpu_times_2[tid] = cpu_times() | |
return calculate(t1, _last_cpu_times_2[tid]) | |
# per-cpu usage | |
else: | |
ret = [] | |
if blocking: | |
tot1 = cpu_times(percpu=True) | |
time.sleep(interval) | |
else: | |
tot1 = _last_per_cpu_times_2.get(tid) or cpu_times(percpu=True) | |
_last_per_cpu_times_2[tid] = cpu_times(percpu=True) | |
for t1, t2 in zip(tot1, _last_per_cpu_times_2[tid]): | |
ret.append(calculate(t1, t2)) | |
return ret | |
def cpu_stats(): | |
"""Return CPU statistics.""" | |
return _psplatform.cpu_stats() | |
if hasattr(_psplatform, "cpu_freq"): | |
def cpu_freq(percpu=False): | |
"""Return CPU frequency as a namedtuple including current, | |
min and max frequency expressed in Mhz. | |
If *percpu* is True and the system supports per-cpu frequency | |
retrieval (Linux only) a list of frequencies is returned for | |
each CPU. If not a list with one element is returned. | |
""" | |
ret = _psplatform.cpu_freq() | |
if percpu: | |
return ret | |
else: | |
num_cpus = float(len(ret)) | |
if num_cpus == 0: | |
return None | |
elif num_cpus == 1: | |
return ret[0] | |
else: | |
currs, mins, maxs = 0.0, 0.0, 0.0 | |
set_none = False | |
for cpu in ret: | |
currs += cpu.current | |
# On Linux if /proc/cpuinfo is used min/max are set | |
# to None. | |
if LINUX and cpu.min is None: | |
set_none = True | |
continue | |
mins += cpu.min | |
maxs += cpu.max | |
current = currs / num_cpus | |
if set_none: | |
min_ = max_ = None | |
else: | |
min_ = mins / num_cpus | |
max_ = maxs / num_cpus | |
return _common.scpufreq(current, min_, max_) | |
__all__.append("cpu_freq") | |
if hasattr(os, "getloadavg") or hasattr(_psplatform, "getloadavg"): | |
# Perform this hasattr check once on import time to either use the | |
# platform based code or proxy straight from the os module. | |
if hasattr(os, "getloadavg"): | |
getloadavg = os.getloadavg | |
else: | |
getloadavg = _psplatform.getloadavg | |
__all__.append("getloadavg") | |
# ===================================================================== | |
# --- system memory related functions | |
# ===================================================================== | |
def virtual_memory(): | |
"""Return statistics about system memory usage as a namedtuple | |
including the following fields, expressed in bytes: | |
- total: | |
total physical memory available. | |
- available: | |
the memory that can be given instantly to processes without the | |
system going into swap. | |
This is calculated by summing different memory values depending | |
on the platform and it is supposed to be used to monitor actual | |
memory usage in a cross platform fashion. | |
- percent: | |
the percentage usage calculated as (total - available) / total * 100 | |
- used: | |
memory used, calculated differently depending on the platform and | |
designed for informational purposes only: | |
macOS: active + wired | |
BSD: active + wired + cached | |
Linux: total - free | |
- free: | |
memory not being used at all (zeroed) that is readily available; | |
note that this doesn't reflect the actual memory available | |
(use 'available' instead) | |
Platform-specific fields: | |
- active (UNIX): | |
memory currently in use or very recently used, and so it is in RAM. | |
- inactive (UNIX): | |
memory that is marked as not used. | |
- buffers (BSD, Linux): | |
cache for things like file system metadata. | |
- cached (BSD, macOS): | |
cache for various things. | |
- wired (macOS, BSD): | |
memory that is marked to always stay in RAM. It is never moved to disk. | |
- shared (BSD): | |
memory that may be simultaneously accessed by multiple processes. | |
The sum of 'used' and 'available' does not necessarily equal total. | |
On Windows 'available' and 'free' are the same. | |
""" | |
global _TOTAL_PHYMEM | |
ret = _psplatform.virtual_memory() | |
# cached for later use in Process.memory_percent() | |
_TOTAL_PHYMEM = ret.total | |
return ret | |
def swap_memory(): | |
"""Return system swap memory statistics as a namedtuple including | |
the following fields: | |
- total: total swap memory in bytes | |
- used: used swap memory in bytes | |
- free: free swap memory in bytes | |
- percent: the percentage usage | |
- sin: no. of bytes the system has swapped in from disk (cumulative) | |
- sout: no. of bytes the system has swapped out from disk (cumulative) | |
'sin' and 'sout' on Windows are meaningless and always set to 0. | |
""" | |
return _psplatform.swap_memory() | |
# ===================================================================== | |
# --- disks/paritions related functions | |
# ===================================================================== | |
def disk_usage(path): | |
"""Return disk usage statistics about the given *path* as a | |
namedtuple including total, used and free space expressed in bytes | |
plus the percentage usage. | |
""" | |
return _psplatform.disk_usage(path) | |
def disk_partitions(all=False): | |
"""Return mounted partitions as a list of | |
(device, mountpoint, fstype, opts) namedtuple. | |
'opts' field is a raw string separated by commas indicating mount | |
options which may vary depending on the platform. | |
If *all* parameter is False return physical devices only and ignore | |
all others. | |
""" | |
def pathconf(path, name): | |
try: | |
return os.pathconf(path, name) | |
except (OSError, AttributeError): | |
pass | |
ret = _psplatform.disk_partitions(all) | |
if POSIX: | |
new = [] | |
for item in ret: | |
nt = item._replace( | |
maxfile=pathconf(item.mountpoint, 'PC_NAME_MAX'), | |
maxpath=pathconf(item.mountpoint, 'PC_PATH_MAX'), | |
) | |
new.append(nt) | |
return new | |
else: | |
return ret | |
def disk_io_counters(perdisk=False, nowrap=True): | |
"""Return system disk I/O statistics as a namedtuple including | |
the following fields: | |
- read_count: number of reads | |
- write_count: number of writes | |
- read_bytes: number of bytes read | |
- write_bytes: number of bytes written | |
- read_time: time spent reading from disk (in ms) | |
- write_time: time spent writing to disk (in ms) | |
Platform specific: | |
- busy_time: (Linux, FreeBSD) time spent doing actual I/Os (in ms) | |
- read_merged_count (Linux): number of merged reads | |
- write_merged_count (Linux): number of merged writes | |
If *perdisk* is True return the same information for every | |
physical disk installed on the system as a dictionary | |
with partition names as the keys and the namedtuple | |
described above as the values. | |
If *nowrap* is True it detects and adjust the numbers which overflow | |
and wrap (restart from 0) and add "old value" to "new value" so that | |
the returned numbers will always be increasing or remain the same, | |
but never decrease. | |
"disk_io_counters.cache_clear()" can be used to invalidate the | |
cache. | |
On recent Windows versions 'diskperf -y' command may need to be | |
executed first otherwise this function won't find any disk. | |
""" | |
kwargs = dict(perdisk=perdisk) if LINUX else {} | |
rawdict = _psplatform.disk_io_counters(**kwargs) | |
if not rawdict: | |
return {} if perdisk else None | |
if nowrap: | |
rawdict = _wrap_numbers(rawdict, 'psutil.disk_io_counters') | |
nt = getattr(_psplatform, "sdiskio", _common.sdiskio) | |
if perdisk: | |
for disk, fields in rawdict.items(): | |
rawdict[disk] = nt(*fields) | |
return rawdict | |
else: | |
return nt(*(sum(x) for x in zip(*rawdict.values()))) | |
disk_io_counters.cache_clear = functools.partial( | |
_wrap_numbers.cache_clear, 'psutil.disk_io_counters' | |
) | |
disk_io_counters.cache_clear.__doc__ = "Clears nowrap argument cache" | |
# ===================================================================== | |
# --- network related functions | |
# ===================================================================== | |
def net_io_counters(pernic=False, nowrap=True): | |
"""Return network I/O statistics as a namedtuple including | |
the following fields: | |
- bytes_sent: number of bytes sent | |
- bytes_recv: number of bytes received | |
- packets_sent: number of packets sent | |
- packets_recv: number of packets received | |
- errin: total number of errors while receiving | |
- errout: total number of errors while sending | |
- dropin: total number of incoming packets which were dropped | |
- dropout: total number of outgoing packets which were dropped | |
(always 0 on macOS and BSD) | |
If *pernic* is True return the same information for every | |
network interface installed on the system as a dictionary | |
with network interface names as the keys and the namedtuple | |
described above as the values. | |
If *nowrap* is True it detects and adjust the numbers which overflow | |
and wrap (restart from 0) and add "old value" to "new value" so that | |
the returned numbers will always be increasing or remain the same, | |
but never decrease. | |
"net_io_counters.cache_clear()" can be used to invalidate the | |
cache. | |
""" | |
rawdict = _psplatform.net_io_counters() | |
if not rawdict: | |
return {} if pernic else None | |
if nowrap: | |
rawdict = _wrap_numbers(rawdict, 'psutil.net_io_counters') | |
if pernic: | |
for nic, fields in rawdict.items(): | |
rawdict[nic] = _common.snetio(*fields) | |
return rawdict | |
else: | |
return _common.snetio(*[sum(x) for x in zip(*rawdict.values())]) | |
net_io_counters.cache_clear = functools.partial( | |
_wrap_numbers.cache_clear, 'psutil.net_io_counters' | |
) | |
net_io_counters.cache_clear.__doc__ = "Clears nowrap argument cache" | |
def net_connections(kind='inet'): | |
"""Return system-wide socket connections as a list of | |
(fd, family, type, laddr, raddr, status, pid) namedtuples. | |
In case of limited privileges 'fd' and 'pid' may be set to -1 | |
and None respectively. | |
The *kind* parameter filters for connections that fit the | |
following criteria: | |
+------------+----------------------------------------------------+ | |
| Kind Value | Connections using | | |
+------------+----------------------------------------------------+ | |
| inet | IPv4 and IPv6 | | |
| inet4 | IPv4 | | |
| inet6 | IPv6 | | |
| tcp | TCP | | |
| tcp4 | TCP over IPv4 | | |
| tcp6 | TCP over IPv6 | | |
| udp | UDP | | |
| udp4 | UDP over IPv4 | | |
| udp6 | UDP over IPv6 | | |
| unix | UNIX socket (both UDP and TCP protocols) | | |
| all | the sum of all the possible families and protocols | | |
+------------+----------------------------------------------------+ | |
On macOS this function requires root privileges. | |
""" | |
return _psplatform.net_connections(kind) | |
def net_if_addrs(): | |
"""Return the addresses associated to each NIC (network interface | |
card) installed on the system as a dictionary whose keys are the | |
NIC names and value is a list of namedtuples for each address | |
assigned to the NIC. Each namedtuple includes 5 fields: | |
- family: can be either socket.AF_INET, socket.AF_INET6 or | |
psutil.AF_LINK, which refers to a MAC address. | |
- address: is the primary address and it is always set. | |
- netmask: and 'broadcast' and 'ptp' may be None. | |
- ptp: stands for "point to point" and references the | |
destination address on a point to point interface | |
(typically a VPN). | |
- broadcast: and *ptp* are mutually exclusive. | |
Note: you can have more than one address of the same family | |
associated with each interface. | |
""" | |
has_enums = _PY3 | |
if has_enums: | |
import socket | |
rawlist = _psplatform.net_if_addrs() | |
rawlist.sort(key=lambda x: x[1]) # sort by family | |
ret = collections.defaultdict(list) | |
for name, fam, addr, mask, broadcast, ptp in rawlist: | |
if has_enums: | |
try: | |
fam = socket.AddressFamily(fam) | |
except ValueError: | |
if WINDOWS and fam == -1: | |
fam = _psplatform.AF_LINK | |
elif ( | |
hasattr(_psplatform, "AF_LINK") | |
and fam == _psplatform.AF_LINK | |
): | |
# Linux defines AF_LINK as an alias for AF_PACKET. | |
# We re-set the family here so that repr(family) | |
# will show AF_LINK rather than AF_PACKET | |
fam = _psplatform.AF_LINK | |
if fam == _psplatform.AF_LINK: | |
# The underlying C function may return an incomplete MAC | |
# address in which case we fill it with null bytes, see: | |
# https://github.com/giampaolo/psutil/issues/786 | |
separator = ":" if POSIX else "-" | |
while addr.count(separator) < 5: | |
addr += "%s00" % separator | |
ret[name].append(_common.snicaddr(fam, addr, mask, broadcast, ptp)) | |
return dict(ret) | |
def net_if_stats(): | |
"""Return information about each NIC (network interface card) | |
installed on the system as a dictionary whose keys are the | |
NIC names and value is a namedtuple with the following fields: | |
- isup: whether the interface is up (bool) | |
- duplex: can be either NIC_DUPLEX_FULL, NIC_DUPLEX_HALF or | |
NIC_DUPLEX_UNKNOWN | |
- speed: the NIC speed expressed in mega bits (MB); if it can't | |
be determined (e.g. 'localhost') it will be set to 0. | |
- mtu: the maximum transmission unit expressed in bytes. | |
""" | |
return _psplatform.net_if_stats() | |
# ===================================================================== | |
# --- sensors | |
# ===================================================================== | |
# Linux, macOS | |
if hasattr(_psplatform, "sensors_temperatures"): | |
def sensors_temperatures(fahrenheit=False): | |
"""Return hardware temperatures. Each entry is a namedtuple | |
representing a certain hardware sensor (it may be a CPU, an | |
hard disk or something else, depending on the OS and its | |
configuration). | |
All temperatures are expressed in celsius unless *fahrenheit* | |
is set to True. | |
""" | |
def convert(n): | |
if n is not None: | |
return (float(n) * 9 / 5) + 32 if fahrenheit else n | |
ret = collections.defaultdict(list) | |
rawdict = _psplatform.sensors_temperatures() | |
for name, values in rawdict.items(): | |
while values: | |
label, current, high, critical = values.pop(0) | |
current = convert(current) | |
high = convert(high) | |
critical = convert(critical) | |
if high and not critical: | |
critical = high | |
elif critical and not high: | |
high = critical | |
ret[name].append( | |
_common.shwtemp(label, current, high, critical) | |
) | |
return dict(ret) | |
__all__.append("sensors_temperatures") | |
# Linux | |
if hasattr(_psplatform, "sensors_fans"): | |
def sensors_fans(): | |
"""Return fans speed. Each entry is a namedtuple | |
representing a certain hardware sensor. | |
All speed are expressed in RPM (rounds per minute). | |
""" | |
return _psplatform.sensors_fans() | |
__all__.append("sensors_fans") | |
# Linux, Windows, FreeBSD, macOS | |
if hasattr(_psplatform, "sensors_battery"): | |
def sensors_battery(): | |
"""Return battery information. If no battery is installed | |
returns None. | |
- percent: battery power left as a percentage. | |
- secsleft: a rough approximation of how many seconds are left | |
before the battery runs out of power. May be | |
POWER_TIME_UNLIMITED or POWER_TIME_UNLIMITED. | |
- power_plugged: True if the AC power cable is connected. | |
""" | |
return _psplatform.sensors_battery() | |
__all__.append("sensors_battery") | |
# ===================================================================== | |
# --- other system related functions | |
# ===================================================================== | |
def boot_time(): | |
"""Return the system boot time expressed in seconds since the epoch.""" | |
# Note: we are not caching this because it is subject to | |
# system clock updates. | |
return _psplatform.boot_time() | |
def users(): | |
"""Return users currently connected on the system as a list of | |
namedtuples including the following fields. | |
- user: the name of the user | |
- terminal: the tty or pseudo-tty associated with the user, if any. | |
- host: the host name associated with the entry, if any. | |
- started: the creation time as a floating point number expressed in | |
seconds since the epoch. | |
""" | |
return _psplatform.users() | |
# ===================================================================== | |
# --- Windows services | |
# ===================================================================== | |
if WINDOWS: | |
def win_service_iter(): | |
"""Return a generator yielding a WindowsService instance for all | |
Windows services installed. | |
""" | |
return _psplatform.win_service_iter() | |
def win_service_get(name): | |
"""Get a Windows service by *name*. | |
Raise NoSuchProcess if no service with such name exists. | |
""" | |
return _psplatform.win_service_get(name) | |
# ===================================================================== | |
def _set_debug(value): | |
"""Enable or disable PSUTIL_DEBUG option, which prints debugging | |
messages to stderr. | |
""" | |
import psutil._common | |
psutil._common.PSUTIL_DEBUG = bool(value) | |
_psplatform.cext.set_debug(bool(value)) | |
def test(): # pragma: no cover | |
from ._common import bytes2human | |
from ._compat import get_terminal_size | |
today_day = datetime.date.today() | |
# fmt: off | |
templ = "%-10s %5s %5s %7s %7s %5s %6s %6s %6s %s" | |
attrs = ['pid', 'memory_percent', 'name', 'cmdline', 'cpu_times', | |
'create_time', 'memory_info', 'status', 'nice', 'username'] | |
print(templ % ("USER", "PID", "%MEM", "VSZ", "RSS", "NICE", # NOQA | |
"STATUS", "START", "TIME", "CMDLINE")) | |
# fmt: on | |
for p in process_iter(attrs, ad_value=None): | |
if p.info['create_time']: | |
ctime = datetime.datetime.fromtimestamp(p.info['create_time']) | |
if ctime.date() == today_day: | |
ctime = ctime.strftime("%H:%M") | |
else: | |
ctime = ctime.strftime("%b%d") | |
else: | |
ctime = '' | |
if p.info['cpu_times']: | |
cputime = time.strftime( | |
"%M:%S", time.localtime(sum(p.info['cpu_times'])) | |
) | |
else: | |
cputime = '' | |
user = p.info['username'] or '' | |
if not user and POSIX: | |
try: | |
user = p.uids()[0] | |
except Error: | |
pass | |
if user and WINDOWS and '\\' in user: | |
user = user.split('\\')[1] | |
user = user[:9] | |
vms = ( | |
bytes2human(p.info['memory_info'].vms) | |
if p.info['memory_info'] is not None | |
else '' | |
) | |
rss = ( | |
bytes2human(p.info['memory_info'].rss) | |
if p.info['memory_info'] is not None | |
else '' | |
) | |
memp = ( | |
round(p.info['memory_percent'], 1) | |
if p.info['memory_percent'] is not None | |
else '' | |
) | |
nice = int(p.info['nice']) if p.info['nice'] else '' | |
if p.info['cmdline']: | |
cmdline = ' '.join(p.info['cmdline']) | |
else: | |
cmdline = p.info['name'] | |
status = p.info['status'][:5] if p.info['status'] else '' | |
line = templ % ( | |
user[:10], | |
p.info['pid'], | |
memp, | |
vms, | |
rss, | |
nice, | |
status, | |
ctime, | |
cputime, | |
cmdline, | |
) | |
print(line[: get_terminal_size()[0]]) # NOQA | |
del memoize_when_activated, division | |
if sys.version_info[0] < 3: | |
del num, x # noqa | |
if __name__ == "__main__": | |
test() | |