D7net
Home
Console
Upload
information
Create File
Create Folder
About
Tools
:
/
proc
/
thread-self
/
root
/
proc
/
self
/
root
/
lib
/
python3.6
/
site-packages
/
procfs
/
Filename :
procfs.py
back
Copy
#!/usr/bin/python3 # -*- python -*- # -*- coding: utf-8 -*- # SPDX-License-Identifier: GPL-2.0-only # # Copyright (C) 2007-2015 Red Hat, Inc. # import os import platform import re import time from functools import reduce from six.moves import range from procfs.utilist import bitmasklist VERSION = "0.7.3" def is_s390(): """ Return True if running on s390 or s390x """ machine = platform.machine() return bool(re.search('s390', machine)) def process_cmdline(pid_info): """ Returns the process command line, if available in the given `process' class, if not available, falls back to using the comm (short process name) in its pidstat key. """ if pid_info["cmdline"]: return reduce(lambda a, b: a + " %s" % b, pid_info["cmdline"]).strip() try: """ If a pid disappears before we query it, return None """ return pid_info["stat"]["comm"] except: return None class pidstat: """ Provides a dictionary to access the fields in the per process /proc/PID/stat files. One can obtain the available fields by asking for the keys of the dictionary, e.g.: >>> p = procfs.pidstat(1) >>> print p.keys() ['majflt', 'rss', 'cnswap', 'cstime', 'pid', 'session', 'startstack', 'startcode', 'cmajflt', 'blocked', 'exit_signal', 'minflt', 'nswap', 'environ', 'priority', 'state', 'delayacct_blkio_ticks', 'policy', 'rt_priority', 'ppid', 'nice', 'cutime', 'endcode', 'wchan', 'num_threads', 'sigcatch', 'comm', 'stime', 'sigignore', 'tty_nr', 'kstkeip', 'utime', 'tpgid', 'itrealvalue', 'kstkesp', 'rlim', 'signal', 'pgrp', 'flags', 'starttime', 'cminflt', 'vsize', 'processor'] And then access the various process properties using it as a dictionary: >>> print p['comm'] systemd >>> print p['priority'] 20 >>> print p['state'] S Please refer to the 'procfs(5)' man page, by using: $ man 5 procfs To see information for each of the above fields, it is part of the 'man-pages' RPM package. """ # Entries with the same value, the one with a comment after it is the # more recent, having replaced the other name in v4.1-rc kernel times. PF_ALIGNWARN = 0x00000001 PF_STARTING = 0x00000002 PF_EXITING = 0x00000004 PF_EXITPIDONE = 0x00000008 PF_VCPU = 0x00000010 PF_WQ_WORKER = 0x00000020 # /* I'm a workqueue worker */ PF_FORKNOEXEC = 0x00000040 PF_MCE_PROCESS = 0x00000080 # /* process policy on mce errors */ PF_SUPERPRIV = 0x00000100 PF_DUMPCORE = 0x00000200 PF_SIGNALED = 0x00000400 PF_MEMALLOC = 0x00000800 # /* set_user noticed that RLIMIT_NPROC was exceeded */ PF_NPROC_EXCEEDED = 0x00001000 PF_FLUSHER = 0x00001000 PF_USED_MATH = 0x00002000 PF_USED_ASYNC = 0x00004000 # /* used async_schedule*(), used by module init */ PF_NOFREEZE = 0x00008000 PF_FROZEN = 0x00010000 PF_FSTRANS = 0x00020000 PF_KSWAPD = 0x00040000 PF_MEMALLOC_NOIO = 0x00080000 # /* Allocating memory without IO involved */ PF_SWAPOFF = 0x00080000 PF_LESS_THROTTLE = 0x00100000 PF_KTHREAD = 0x00200000 PF_RANDOMIZE = 0x00400000 PF_SWAPWRITE = 0x00800000 PF_SPREAD_PAGE = 0x01000000 PF_SPREAD_SLAB = 0x02000000 PF_THREAD_BOUND = 0x04000000 # /* Userland is not allowed to meddle with cpus_allowed */ PF_NO_SETAFFINITY = 0x04000000 PF_MCE_EARLY = 0x08000000 # /* Early kill for mce process policy */ PF_MEMPOLICY = 0x10000000 PF_MUTEX_TESTER = 0x20000000 PF_FREEZER_SKIP = 0x40000000 PF_FREEZER_NOSIG = 0x80000000 # /* this thread called freeze_processes and should not be frozen */ PF_SUSPEND_TASK = 0x80000000 proc_stat_fields = ["pid", "comm", "state", "ppid", "pgrp", "session", "tty_nr", "tpgid", "flags", "minflt", "cminflt", "majflt", "cmajflt", "utime", "stime", "cutime", "cstime", "priority", "nice", "num_threads", "itrealvalue", "starttime", "vsize", "rss", "rlim", "startcode", "endcode", "startstack", "kstkesp", "kstkeip", "signal", "blocked", "sigignore", "sigcatch", "wchan", "nswap", "cnswap", "exit_signal", "processor", "rt_priority", "policy", "delayacct_blkio_ticks", "environ"] def __init__(self, pid, basedir="/proc"): self.pid = pid try: self.load(basedir) except FileNotFoundError: # The file representing the pid has disappeared # propagate the error to the user to handle raise def __getitem__(self, fieldname): return self.fields[fieldname] def keys(self): return list(self.fields.keys()) def values(self): return list(self.fields.values()) def has_key(self, fieldname): return fieldname in self.fields def items(self): return self.fields def __contains__(self, fieldname): return fieldname in self.fields def load(self, basedir="/proc"): try: f = open(f"{basedir}/{self.pid}/stat") except FileNotFoundError: # The pid has disappeared, propagate the error raise fields = f.readline().strip().split(') ') f.close() fields = fields[0].split(' (') + fields[1].split() self.fields = {} nr_fields = min(len(fields), len(self.proc_stat_fields)) for i in range(nr_fields): attrname = self.proc_stat_fields[i] value = fields[i] if attrname == "comm": self.fields["comm"] = value.strip('()') else: try: self.fields[attrname] = int(value) except: self.fields[attrname] = value def is_bound_to_cpu(self): """ Returns true if this process has a fixed smp affinity mask, not allowing it to be moved to a different set of CPUs. """ return bool(self.fields["flags"] & self.PF_THREAD_BOUND) def process_flags(self): """ Returns a list with all the process flags known, details depend on kernel version, declared in the file include/linux/sched.h in the kernel sources. As of v4.2-rc7 these include (from include/linux/sched.h comments): PF_EXITING Getting shut down PF_EXITPIDONE Pi exit done on shut down PF_VCPU I'm a virtual CPU PF_WQ_WORKER I'm a workqueue worker PF_FORKNOEXEC Forked but didn't exec PF_MCE_PROCESS Process policy on mce errors PF_SUPERPRIV Used super-user privileges PF_DUMPCORE Dumped core PF_SIGNALED Killed by a signal PF_MEMALLOC Allocating memory PF_NPROC_EXCEEDED Set_user noticed that RLIMIT_NPROC was exceeded PF_USED_MATH If unset the fpu must be initialized before use PF_USED_ASYNC Used async_schedule*(), used by module init PF_NOFREEZE This thread should not be frozen PF_FROZEN Frozen for system suspend PF_FSTRANS Inside a filesystem transaction PF_KSWAPD I am kswapd PF_MEMALLOC_NOIO Allocating memory without IO involved PF_LESS_THROTTLE Throttle me less: I clean memory PF_KTHREAD I am a kernel thread PF_RANDOMIZE Randomize virtual address space PF_SWAPWRITE Allowed to write to swap PF_NO_SETAFFINITY Userland is not allowed to meddle with cpus_allowed PF_MCE_EARLY Early kill for mce process policy PF_MUTEX_TESTER Thread belongs to the rt mutex tester PF_FREEZER_SKIP Freezer should not count it as freezable PF_SUSPEND_TASK This thread called freeze_processes and should not be frozen """ sflags = [] for attr in dir(self): if attr[:3] != "PF_": continue value = getattr(self, attr) if value & self.fields["flags"]: sflags.append(attr) return sflags def cannot_set_affinity(self, pid): PF_NO_SETAFFINITY = 0x04000000 try: return bool(int(self.processes[pid]["stat"]["flags"]) & PF_NO_SETAFFINITY) except: return True def cannot_set_thread_affinity(self, pid, tid): PF_NO_SETAFFINITY = 0x04000000 try: return bool(int(self.processes[pid].threads[tid]["stat"]["flags"]) & PF_NO_SETAFFINITY) except: return True class pidstatus: """ Provides a dictionary to access the fields in the per process /proc/PID/status files. This provides additional information about processes and threads to what can be obtained with the procfs.pidstat() class. One can obtain the available fields by asking for the keys of the dictionary, e.g.: >>> import procfs >>> p = procfs.pidstatus(1) >>> print p.keys() ['VmExe', 'CapBnd', 'NSpgid', 'Tgid', 'NSpid', 'VmSize', 'VmPMD', 'ShdPnd', 'State', 'Gid', 'nonvoluntary_ctxt_switches', 'SigIgn', 'VmStk', 'VmData', 'SigCgt', 'CapEff', 'VmPTE', 'Groups', 'NStgid', 'Threads', 'PPid', 'VmHWM', 'NSsid', 'VmSwap', 'Name', 'SigBlk', 'Mems_allowed_list', 'VmPeak', 'Ngid', 'VmLck', 'SigQ', 'VmPin', 'Mems_allowed', 'CapPrm', 'Seccomp', 'VmLib', 'Cpus_allowed', 'Uid', 'SigPnd', 'Pid', 'Cpus_allowed_list', 'TracerPid', 'CapInh', 'voluntary_ctxt_switches', 'VmRSS', 'FDSize'] >>> print p["Pid"] 1 >>> print p["Threads"] 1 >>> print p["VmExe"] 1248 kB >>> print p["Cpus_allowed"] f >>> print p["SigQ"] 0/30698 >>> print p["VmPeak"] 320300 kB >>> Please refer to the 'procfs(5)' man page, by using: $ man 5 procfs To see information for each of the above fields, it is part of the 'man-pages' RPM package. In the man page there will be references to further documentation, like referring to the "getrlimit(2)" man page when explaining the "SigQ" line/field. """ def __init__(self, pid, basedir="/proc"): self.pid = pid self.load(basedir) def __getitem__(self, fieldname): return self.fields[fieldname] def keys(self): return list(self.fields.keys()) def values(self): return list(self.fields.values()) def has_key(self, fieldname): return fieldname in self.fields def items(self): return self.fields def __contains__(self, fieldname): return fieldname in self.fields def load(self, basedir="/proc"): self.fields = {} with open(f"{basedir}/{self.pid}/status") as f: for line in f.readlines(): fields = line.split(":") if len(fields) != 2: continue name = fields[0] value = fields[1].strip() try: self.fields[name] = int(value) except: self.fields[name] = value class process: """ Information about a process with a given pid, provides a dictionary with two entries, instances of different wrappers for /proc/ process related meta files: "stat" and "status", see the documentation for procfs.pidstat and procfs.pidstatus for further info about those classes. """ def __init__(self, pid, basedir="/proc"): self.pid = pid self.basedir = basedir def __getitem__(self, attr): if not hasattr(self, attr): if attr in ("stat", "status"): if attr == "stat": sclass = pidstat else: sclass = pidstatus try: setattr(self, attr, sclass(self.pid, self.basedir)) except FileNotFoundError: # The pid has disappeared, progate the error raise elif attr == "cmdline": self.load_cmdline() elif attr == "threads": self.load_threads() elif attr == "cgroups": self.load_cgroups() elif attr == "environ": self.load_environ() return getattr(self, attr) def has_key(self, attr): return hasattr(self, attr) def __contains__(self, attr): return hasattr(self, attr) def load_cmdline(self): try: with open(f"/proc/{self.pid}/cmdline") as f: self.cmdline = f.readline().strip().split('\0')[:-1] except FileNotFoundError: """ This can happen when a pid disappears """ self.cmdline = None except UnicodeDecodeError: """ TODO - this shouldn't happen, needs to be investigated """ self.cmdline = None def load_threads(self): self.threads = pidstats(f"/proc/{self.pid}/task/") # remove thread leader del self.threads[self.pid] def load_cgroups(self): self.cgroups = "" with open(f"/proc/{self.pid}/cgroup") as f: for line in reversed(f.readlines()): if len(self.cgroups) != 0: self.cgroups = self.cgroups + "," + line[:-1] else: self.cgroups = line[:-1] def load_environ(self): """ Loads the environment variables for this process. The entries then become available via the 'environ' member, or via the 'environ' dict key when accessing as p["environ"]. E.g.: >>> all_processes = procfs.pidstats() >>> firefox_pid = all_processes.find_by_name("firefox") >>> firefox_process = all_processes[firefox_pid[0]] >>> print firefox_process["environ"]["PWD"] /home/acme >>> print len(firefox_process.environ.keys()) 66 >>> print firefox_process["environ"]["SHELL"] /bin/bash >>> print firefox_process["environ"]["USERNAME"] acme >>> print firefox_process["environ"]["HOME"] /home/acme >>> print firefox_process["environ"]["MAIL"] /var/spool/mail/acme >>> """ self.environ = {} with open(f"/proc/{self.pid}/environ") as f: for x in f.readline().split('\0'): if len(x) > 0: y = x.split('=') self.environ[y[0]] = y[1] class pidstats: """ Provides access to all the processes in the system, to get a picture of how many processes there are at any given moment. The entries can be accessed as a dictionary, keyed by pid. Also there are methods to find processes that match a given COMM or regular expression. """ def __init__(self, basedir="/proc"): self.basedir = basedir self.processes = {} self.reload() def __getitem__(self, key): return self.processes[key] def __delitem__(self, key): # not clear on why this can fail, but it can try: del self.processes[key] except: pass def keys(self): return list(self.processes.keys()) def values(self): return list(self.processes.values()) def has_key(self, key): return key in self.processes def items(self): return self.processes def __contains__(self, key): return key in self.processes def reload(self): """ This operation will throw away the current dictionary contents, if any, and read all the pid files from /proc/, instantiating a 'process' instance for each of them. This is a high overhead operation, and should be avoided if the perf python binding can be used to detect when new threads appear and existing ones terminate. In RHEL it is found in the python-perf rpm package. More information about the perf facilities can be found in the 'perf_event_open' man page. """ del self.processes self.processes = {} pids = os.listdir(self.basedir) for spid in pids: try: pid = int(spid) except: continue self.processes[pid] = process(pid, self.basedir) def reload_threads(self): to_remove = [] for pid in list(self.processes.keys()): try: self.processes[pid].load_threads() except OSError: # process vanished, remove it to_remove.append(pid) for pid in to_remove: del self.processes[pid] def find_by_name(self, name): name = name[:15] pids = [] for pid in list(self.processes.keys()): try: if name == self.processes[pid]["stat"]["comm"]: pids.append(pid) except IOError: # We're doing lazy loading of /proc files # So if we get this exception is because the # process vanished, remove it del self.processes[pid] return pids def find_by_regex(self, regex): pids = [] for pid in list(self.processes.keys()): try: if regex.match(self.processes[pid]["stat"]["comm"]): pids.append(pid) except IOError: # We're doing lazy loading of /proc files # So if we get this exception is because the # process vanished, remove it del self.processes[pid] return pids def find_by_cmdline_regex(self, regex): pids = [] for pid in list(self.processes.keys()): try: if regex.match(process_cmdline(self.processes[pid])): pids.append(pid) except IOError: # We're doing lazy loading of /proc files # So if we get this exception is because the # process vanished, remove it del self.processes[pid] return pids def get_per_cpu_rtprios(self, basename): cpu = 0 priorities = "" processed_pids = [] while True: name = f"{basename}/{cpu}" pids = self.find_by_name(name) if not pids or len([n for n in pids if n not in processed_pids]) == 0: break for pid in pids: try: priorities += f'{self.processes[pid]["stat"]["rt_priority"]}' except IOError: # We're doing lazy loading of /proc files # So if we get this exception is because the # process vanished, remove it del self.processes[pid] processed_pids += pids cpu += 1 priorities = priorities.strip(',') return priorities def get_rtprios(self, name): cpu = 0 priorities = "" processed_pids = [] while True: pids = self.find_by_name(name) if not pids or len([n for n in pids if n not in processed_pids]) == 0: break for pid in pids: try: priorities += f'{self.processes[pid]["stat"]["rt_priority"]}' except IOError: # We're doing lazy loading of /proc files # So if we get this exception is because the # process vanished, remove it del self.processes[pid] processed_pids += pids cpu += 1 priorities = priorities.strip(',') return priorities def is_bound_to_cpu(self, pid): """ Checks if a given pid can't have its SMP affinity mask changed. """ return self.processes[pid]["stat"].is_bound_to_cpu() class interrupts: """ Information about IRQs in the system. A dictionary keyed by IRQ number will have as its value another dictionary with "cpu", "type" and "users" keys, with the SMP affinity mask, type of IRQ and the drivers associated with each interrupt. The information comes from the /proc/interrupts file, documented in 'man procfs(5)', for instance, the 'cpu' dict is an array with one entry per CPU present in the sistem, each value being the number of interrupts that took place per CPU. E.g.: >>> import procfs >>> interrupts = procfs.interrupts() >>> thunderbolt_irq = interrupts.find_by_user("thunderbolt") >>> print thunderbolt_irq 34 >>> thunderbolt = interrupts[thunderbolt_irq] >>> print thunderbolt {'affinity': [0, 1, 2, 3], 'type': 'PCI-MSI', 'cpu': [3495, 0, 81, 0], 'users': ['thunderbolt']} >>> """ def __init__(self): self.interrupts = {} self.reload() def __getitem__(self, key): return self.interrupts[str(key)] def keys(self): return list(self.interrupts.keys()) def values(self): return list(self.interrupts.values()) def has_key(self, key): return str(key) in self.interrupts def items(self): return self.interrupts def __contains__(self, key): return str(key) in self.interrupts def reload(self): del self.interrupts self.interrupts = {} with open("/proc/interrupts") as f: for line in f.readlines(): line = line.strip() fields = line.split() if fields[0][:3] == "CPU": self.nr_cpus = len(fields) continue irq = fields[0].strip(":") self.interrupts[irq] = {} self.interrupts[irq] = self.parse_entry(fields[1:], line) try: nirq = int(irq) except: continue self.interrupts[irq]["affinity"] = self.parse_affinity(nirq) def parse_entry(self, fields, line): dict = {} dict["cpu"] = [] dict["cpu"].append(int(fields[0])) nr_fields = len(fields) if nr_fields >= self.nr_cpus: dict["cpu"] += [int(i) for i in fields[1:self.nr_cpus]] if nr_fields > self.nr_cpus: dict["type"] = fields[self.nr_cpus] # look if there are users (interrupts 3 and 4 haven't) if nr_fields > self.nr_cpus + 1: dict["users"] = [a.strip() for a in fields[nr_fields - 1].split(',')] else: dict["users"] = [] return dict def parse_affinity(self, irq): try: with open(f"/proc/irq/{irq}/smp_affinity") as f: line = f.readline() return bitmasklist(line, self.nr_cpus) except IOError: return [0, ] def find_by_user(self, user): """ Looks up a interrupt number by the name of one of its users" E.g.: >>> import procfs >>> interrupts = procfs.interrupts() >>> thunderbolt_irq = interrupts.find_by_user("thunderbolt") >>> print thunderbolt_irq 34 >>> thunderbolt = interrupts[thunderbolt_irq] >>> print thunderbolt {'affinity': [0, 1, 2, 3], 'type': 'PCI-MSI', 'cpu': [3495, 0, 81, 0], 'users': ['thunderbolt']} >>> """ for i in list(self.interrupts.keys()): if "users" in self.interrupts[i] and \ user in self.interrupts[i]["users"]: return i return None def find_by_user_regex(self, regex): """ Looks up a interrupt number by a regex that matches names of its users" E.g.: >>> import procfs >>> import re >>> interrupts = procfs.interrupts() >>> usb_controllers = interrupts.find_by_user_regex(re.compile(".*hcd")) >>> print usb_controllers ['22', '23', '31'] >>> print [ interrupts[irq]["users"] for irq in usb_controllers ] [['ehci_hcd:usb4'], ['ehci_hcd:usb3'], ['xhci_hcd']] >>> """ irqs = [] for i in list(self.interrupts.keys()): if "users" not in self.interrupts[i]: continue for user in self.interrupts[i]["users"]: if regex.match(user): irqs.append(i) break return irqs class cmdline: """ Parses the kernel command line (/proc/cmdline), turning it into a dictionary." Useful to figure out if some kernel boolean knob has been turned on, as well as to find the value associated to other kernel knobs. It can also be used to find out about parameters passed to the init process, such as 'BOOT_IMAGE', etc. E.g.: >>> import procfs >>> kcmd = procfs.cmdline() >>> print kcmd.keys() ['LANG', 'BOOT_IMAGE', 'quiet', 'rhgb', 'rd.lvm.lv', 'ro', 'root'] >>> print kcmd["BOOT_IMAGE"] /vmlinuz-4.3.0-rc1+ >>> """ def __init__(self): self.options = {} self.parse() def parse(self): with open("/proc/cmdline") as f: for option in f.readline().strip().split(): fields = option.split("=") if len(fields) == 1: self.options[fields[0]] = True else: self.options[fields[0]] = fields[1] def __getitem__(self, key): return self.options[key] def keys(self): return list(self.options.keys()) def values(self): return list(self.options.values()) def items(self): return self.options class cpuinfo: """ Dictionary with information about CPUs in the system. Please refer to 'man procfs(5)' for further information about the '/proc/cpuinfo' file, that is the source of the information provided by this class. The 'man lscpu(1)' also has information about a program that uses the '/proc/cpuinfo' file. Using this class one can obtain the number of CPUs in a system: >>> cpus = procfs.cpuinfo() >>> print cpus.nr_cpus 4 It is also possible to figure out aspects of the CPU topology, such as how many CPU physical sockets exists, i.e. groups of CPUs sharing components such as CPU memory caches: >>> print len(cpus.sockets) 1 Additionally dictionary with information common to all CPUs in the system is available: >>> print cpus["model name"] Intel(R) Core(TM) i7-3667U CPU @ 2.00GHz >>> print cpus["cache size"] 4096 KB >>> """ def __init__(self, filename="/proc/cpuinfo"): self.tags = {} self.nr_cpus = 0 self.sockets = [] self.parse(filename) def __getitem__(self, key): return self.tags[key.lower()] def keys(self): return list(self.tags.keys()) def values(self): return list(self.tags.values()) def items(self): return self.tags def parse(self, filename): with open(filename) as f: for line in f.readlines(): line = line.strip() if not line: continue fields = line.split(":") tagname = fields[0].strip().lower() if tagname == "processor": self.nr_cpus += 1 continue if is_s390() and tagname == "cpu number": self.nr_cpus += 1 continue if tagname == "core id": continue self.tags[tagname] = fields[1].strip() if tagname == "physical id": socket_id = self.tags[tagname] if socket_id not in self.sockets: self.sockets.append(socket_id) self.nr_sockets = self.sockets and len(self.sockets) or \ (self.nr_cpus / ("siblings" in self.tags and int(self.tags["siblings"]) or 1)) self.nr_cores = ("cpu cores" in self.tags and int( self.tags["cpu cores"]) or 1) * self.nr_sockets class smaps_lib: """ Representation of an mmap in place for a process. Can be used to figure out which processes have an library mapped, etc. The 'perm' member can be used to figure out executable mmaps, i.e. libraries. The 'vm_start' and 'vm_end' in turn can be used when trying to resolve processor instruction pointer addresses to a symbol name in a library. """ def __init__(self, lines): fields = lines[0].split() self.vm_start, self.vm_end = [int(a, 16) for a in fields[0].split("-")] self.perms = fields[1] self.offset = int(fields[2], 16) self.major, self.minor = fields[3].split(":") self.inode = int(fields[4]) if len(fields) > 5: self.name = fields[5] else: self.name = None self.tags = {} for line in lines[1:]: fields = line.split() tag = fields[0][:-1].lower() try: self.tags[tag] = int(fields[1]) except: # VmFlags are strings self.tags[tag] = fields def __getitem__(self, key): return self.tags[key.lower()] def keys(self): return list(self.tags.keys()) def values(self): return list(self.tags.values()) def items(self): return self.tags class smaps: """ List of libraries mapped by a process. Parses the lines in the /proc/PID/smaps file, that is further documented in the procfs(5) man page. Example: Listing the executable maps for the 'sshd' process: >>> import procfs >>> processes = procfs.pidstats() >>> sshd = processes.find_by_name("sshd") >>> sshd_maps = procfs.smaps(sshd[0]) >>> for i in range(len(sshd_maps)): ... if 'x' in sshd_maps[i].perms: ... print "%s: %s" % (sshd_maps[i].name, sshd_maps[i].perms) ... /usr/sbin/sshd: r-xp /usr/lib64/libnss_files-2.20.so: r-xp /usr/lib64/librt-2.20.so: r-xp /usr/lib64/libkeyutils.so.1.5: r-xp /usr/lib64/libkrb5support.so.0.1: r-xp /usr/lib64/libfreebl3.so: r-xp /usr/lib64/libpthread-2.20.so: r-xp ... """ def __init__(self, pid): self.pid = pid self.entries = [] self.reload() def parse_entry(self, f, line): lines = [] if not line: line = f.readline().strip() if not line: return lines.append(line) while True: line = f.readline() if not line: break line = line.strip() if line.split()[0][-1] == ':': lines.append(line) else: break self.entries.append(smaps_lib(lines)) return line def __len__(self): return len(self.entries) def __getitem__(self, index): return self.entries[index] def reload(self): line = None with open(f"/proc/{self.pid}/smaps") as f: while True: line = self.parse_entry(f, line) if not line: break self.nr_entries = len(self.entries) def find_by_name_fragment(self, fragment): result = [] for i in range(self.nr_entries): if self.entries[i].name and \ self.entries[i].name.find(fragment) >= 0: result.append(self.entries[i]) return result class cpustat: """ CPU statistics, obtained from a line in the '/proc/stat' file, Please refer to 'man procfs(5)' for further information about the '/proc/stat' file, that is the source of the information provided by this class. """ def __init__(self, fields): self.name = fields[0] (self.user, self.nice, self.system, self.idle, self.iowait, self.irq, self.softirq) = [int(i) for i in fields[1:8]] if len(fields) > 7: self.steal = int(fields[7]) if len(fields) > 8: self.guest = int(fields[8]) def __repr__(self): s = f"< user: {self.user}, nice: {self.nice}, system: {self.system}, idle: {self.idle}, iowait: {self.iowait}, irq: {self.irq}, softirq: {self.softirq}" if hasattr(self, 'steal'): s += f", steal: {self.steal}" if hasattr(self, 'guest'): s += f", guest: {self.guest}" return s + ">" class cpusstats: """ Dictionary with information about CPUs in the system. First entry in the dictionary gives an aggregate view of all CPUs, each other entry is about separate CPUs. Please refer to 'man procfs(5)' for further information about the '/proc/stat' file, that is the source of the information provided by this class. """ def __init__(self, filename="/proc/stat"): self.entries = {} self.time = None self.hertz = os.sysconf(2) self.filename = filename self.reload() def __iter__(self): return iter(self.entries) def __getitem__(self, key): return self.entries[key] def __len__(self): return len(list(self.entries.keys())) def keys(self): return list(self.entries.keys()) def values(self): return list(self.entries.values()) def items(self): return self.entries def reload(self): last_entries = self.entries self.entries = {} with open(self.filename) as f: for line in f.readlines(): fields = line.strip().split() if fields[0][:3].lower() != "cpu": continue c = cpustat(fields) if c.name == "cpu": idx = 0 else: idx = int(c.name[3:]) + 1 self.entries[idx] = c last_time = self.time self.time = time.time() if last_entries: delta_sec = self.time - last_time interval_hz = delta_sec * self.hertz for cpu in list(self.entries.keys()): if cpu not in last_entries: curr.usage = 0 continue curr = self.entries[cpu] prev = last_entries[cpu] delta = (curr.user - prev.user) + \ (curr.nice - prev.nice) + \ (curr.system - prev.system) curr.usage = (delta / interval_hz) * 100 curr.usage = min(curr.usage, 100) if __name__ == '__main__': import sys ints = interrupts() for i in list(ints.interrupts.keys()): print(f"{i}: {ints.interrupts[i]}") options = cmdline() for o in list(options.options.keys()): print(f"{o}: {options.options[o]}") cpu = cpuinfo() print(f"\ncpuinfo data: {cpu.nr_cpus} processors") for tag in list(cpu.keys()): print(f"{tag}={cpu[tag]}") print("smaps:\n" + ("-" * 40)) s = smaps(int(sys.argv[1])) for i in range(s.nr_entries): print(f"{s.entries[i].vm_start:#x} {s.entries[i].name}") print("-" * 40) for a in s.find_by_name_fragment(sys.argv[2]): print(a["Size"]) ps = pidstats() print(ps[1]) cs = cpusstats() while True: time.sleep(1) cs.reload() for cpu in cs: print(f"{cpu}: {cs[cpu]}") print("-" * 10)