Files
poky/meta/lib/buildstats.py
Aryaman Gupta 1125adc33b buildstats.bbclass: correct sampling of system stats
The last time of sampling would be updated within the SystemStats class
but not re-recorded into the datastore, leading to multiple samples being
collected in the same second in the sample function of buildstats.py.
Fix this to collect and store only one sample per second within a
certain tolerance to deal with variation in the arrival time.

This fix elimates the spikiness of sampled data, in cases where the difference
between the current and the last sample is taken. Previously, since many
samples per second were recorded, certain types of data would result in a
very small elapsed time and hence a small numerical difference. For example,
the CPU usage from /proc/stat is a running total of usage and taking the
difference between data collected 0.1 seconds apart would result in usage
appearing lower than it actually was.

(From OE-Core rev: 0e2df45ab066bb4ad2c4f8622ee9c1a8ecdea9cb)

Signed-off-by: Aryaman Gupta <aryaman.gupta@windriver.com>
Signed-off-by: Randy MacLeod <randy.macleod@windriver.com>
Signed-off-by: Luca Ceresoli <luca.ceresoli@bootlin.com>
Signed-off-by: Alexandre Belloni <alexandre.belloni@bootlin.com>
Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
2022-06-29 16:16:56 +01:00

213 lines
9.7 KiB
Python

#
# SPDX-License-Identifier: GPL-2.0-only
#
# Implements system state sampling. Called by buildstats.bbclass.
# Because it is a real Python module, it can hold persistent state,
# like open log files and the time of the last sampling.
import time
import re
import bb.event
class SystemStats:
def __init__(self, d):
bn = d.getVar('BUILDNAME')
bsdir = os.path.join(d.getVar('BUILDSTATS_BASE'), bn)
bb.utils.mkdirhier(bsdir)
file_handlers = [('diskstats', self._reduce_diskstats),
('meminfo', self._reduce_meminfo),
('stat', self._reduce_stat)]
# Some hosts like openSUSE have readable /proc/pressure files
# but throw errors when these files are opened. Catch these error
# and ensure that the reduce_proc_pressure directory is not created.
if os.path.exists("/proc/pressure"):
try:
source = open('/proc/pressure/cpu', 'rb')
source.read()
pressuredir = os.path.join(bsdir, 'reduced_proc_pressure')
bb.utils.mkdirhier(pressuredir)
file_handlers.extend([('pressure/cpu', self._reduce_pressure),
('pressure/io', self._reduce_pressure),
('pressure/memory', self._reduce_pressure)])
except Exception:
pass
self.proc_files = []
for filename, handler in (file_handlers):
# The corresponding /proc files might not exist on the host.
# For example, /proc/diskstats is not available in virtualized
# environments like Linux-VServer. Silently skip collecting
# the data.
if os.path.exists(os.path.join('/proc', filename)):
# In practice, this class gets instantiated only once in
# the bitbake cooker process. Therefore 'append' mode is
# not strictly necessary, but using it makes the class
# more robust should two processes ever write
# concurrently.
destfile = os.path.join(bsdir, '%sproc_%s.log' % ('reduced_' if handler else '', filename))
self.proc_files.append((filename, open(destfile, 'ab'), handler))
self.monitor_disk = open(os.path.join(bsdir, 'monitor_disk.log'), 'ab')
# Last time that we sampled /proc data resp. recorded disk monitoring data.
self.last_proc = 0
self.last_disk_monitor = 0
# Minimum number of seconds between recording a sample. This becames relevant when we get
# called very often while many short tasks get started. Sampling during quiet periods
# depends on the heartbeat event, which fires less often.
# By default, the Heartbeat events occur roughly once every second but the actual time
# between these events deviates by a few milliseconds, in most cases. Hence
# pick a somewhat arbitary tolerance such that we sample a large majority
# of the Heartbeat events. This ignores rare events that fall outside the minimum
# and may lead an extra sample in a given second every so often. However, it allows for fairly
# consistent intervals between samples without missing many events.
self.tolerance = 0.01
self.min_seconds = 1.0 - self.tolerance
self.meminfo_regex = re.compile(rb'^(MemTotal|MemFree|Buffers|Cached|SwapTotal|SwapFree):\s*(\d+)')
self.diskstats_regex = re.compile(rb'^([hsv]d.|mtdblock\d|mmcblk\d|cciss/c\d+d\d+.*)$')
self.diskstats_ltime = None
self.diskstats_data = None
self.stat_ltimes = None
# Last time we sampled /proc/pressure. All resources stored in a single dict with the key as filename
self.last_pressure = {"pressure/cpu": None, "pressure/io": None, "pressure/memory": None}
def close(self):
self.monitor_disk.close()
for _, output, _ in self.proc_files:
output.close()
def _reduce_meminfo(self, time, data, filename):
"""
Extracts 'MemTotal', 'MemFree', 'Buffers', 'Cached', 'SwapTotal', 'SwapFree'
and writes their values into a single line, in that order.
"""
values = {}
for line in data.split(b'\n'):
m = self.meminfo_regex.match(line)
if m:
values[m.group(1)] = m.group(2)
if len(values) == 6:
return (time,
b' '.join([values[x] for x in
(b'MemTotal', b'MemFree', b'Buffers', b'Cached', b'SwapTotal', b'SwapFree')]) + b'\n')
def _diskstats_is_relevant_line(self, linetokens):
if len(linetokens) != 14:
return False
disk = linetokens[2]
return self.diskstats_regex.match(disk)
def _reduce_diskstats(self, time, data, filename):
relevant_tokens = filter(self._diskstats_is_relevant_line, map(lambda x: x.split(), data.split(b'\n')))
diskdata = [0] * 3
reduced = None
for tokens in relevant_tokens:
# rsect
diskdata[0] += int(tokens[5])
# wsect
diskdata[1] += int(tokens[9])
# use
diskdata[2] += int(tokens[12])
if self.diskstats_ltime:
# We need to compute information about the time interval
# since the last sampling and record the result as sample
# for that point in the past.
interval = time - self.diskstats_ltime
if interval > 0:
sums = [ a - b for a, b in zip(diskdata, self.diskstats_data) ]
readTput = sums[0] / 2.0 * 100.0 / interval
writeTput = sums[1] / 2.0 * 100.0 / interval
util = float( sums[2] ) / 10 / interval
util = max(0.0, min(1.0, util))
reduced = (self.diskstats_ltime, (readTput, writeTput, util))
self.diskstats_ltime = time
self.diskstats_data = diskdata
return reduced
def _reduce_nop(self, time, data, filename):
return (time, data)
def _reduce_stat(self, time, data, filename):
if not data:
return None
# CPU times {user, nice, system, idle, io_wait, irq, softirq} from first line
tokens = data.split(b'\n', 1)[0].split()
times = [ int(token) for token in tokens[1:] ]
reduced = None
if self.stat_ltimes:
user = float((times[0] + times[1]) - (self.stat_ltimes[0] + self.stat_ltimes[1]))
system = float((times[2] + times[5] + times[6]) - (self.stat_ltimes[2] + self.stat_ltimes[5] + self.stat_ltimes[6]))
idle = float(times[3] - self.stat_ltimes[3])
iowait = float(times[4] - self.stat_ltimes[4])
aSum = max(user + system + idle + iowait, 1)
reduced = (time, (user/aSum, system/aSum, iowait/aSum))
self.stat_ltimes = times
return reduced
def _reduce_pressure(self, time, data, filename):
"""
Return reduced pressure: {avg10, avg60, avg300} and delta total compared to the previous sample
for the cpu, io and memory resources. A common function is used for all 3 resources since the
format of the /proc/pressure file is the same in each case.
"""
if not data:
return None
tokens = data.split(b'\n', 1)[0].split()
avg10 = float(tokens[1].split(b'=')[1])
avg60 = float(tokens[2].split(b'=')[1])
avg300 = float(tokens[3].split(b'=')[1])
total = int(tokens[4].split(b'=')[1])
reduced = None
if self.last_pressure[filename]:
delta = total - self.last_pressure[filename]
reduced = (time, (avg10, avg60, avg300, delta))
self.last_pressure[filename] = total
return reduced
def sample(self, event, force):
"""
Collect and log proc or disk_monitor stats periodically.
Return True if a new sample is collected and hence the value last_proc or last_disk_monitor
is changed.
"""
retval = False
now = time.time()
if (now - self.last_proc > self.min_seconds) or force:
for filename, output, handler in self.proc_files:
with open(os.path.join('/proc', filename), 'rb') as input:
data = input.read()
if handler:
reduced = handler(now, data, filename)
else:
reduced = (now, data)
if reduced:
if isinstance(reduced[1], bytes):
# Use as it is.
data = reduced[1]
else:
# Convert to a single line.
data = (' '.join([str(x) for x in reduced[1]]) + '\n').encode('ascii')
# Unbuffered raw write, less overhead and useful
# in case that we end up with concurrent writes.
os.write(output.fileno(),
('%.0f\n' % reduced[0]).encode('ascii') +
data +
b'\n')
self.last_proc = now
retval = True
if isinstance(event, bb.event.MonitorDiskEvent) and \
((now - self.last_disk_monitor > self.min_seconds) or force):
os.write(self.monitor_disk.fileno(),
('%.0f\n' % now).encode('ascii') +
''.join(['%s: %d\n' % (dev, sample.total_bytes - sample.free_bytes)
for dev, sample in event.disk_usage.items()]).encode('ascii') +
b'\n')
self.last_disk_monitor = now
retval = True
return retval