Source code for autotest.client.base_utils

"""
DO NOT import this file directly - import client/bin/utils.py,
which will mix this in

Convenience functions for use by tests or whomever.

Note that this file is mixed in by utils.py - note very carefully the
precedence order defined there
"""
import os
import shutil
import commands
import pickle
import re
import fnmatch
import logging
import platform
from autotest.client.shared import error, utils, magic


[docs]def grep(pattern, file): """ This is mainly to fix the return code inversion from grep Also handles compressed files. returns 1 if the pattern is present in the file, 0 if not. """ command = 'grep "%s" > /dev/null' % pattern ret = cat_file_to_cmd(file, command, ignore_status=True) return not ret
[docs]def difflist(list1, list2): """returns items in list2 that are not in list1""" diff = [] for x in list2: if x not in list1: diff.append(x) return diff
[docs]def cat_file_to_cmd(file, command, ignore_status=0, return_output=False): """ equivalent to 'cat file | command' but knows to use zcat or bzcat if appropriate """ if not os.path.isfile(file): raise NameError('invalid file %s to cat to command %s' % (file, command)) if return_output: run_cmd = utils.system_output else: run_cmd = utils.system if magic.guess_type(file) == 'application/x-bzip2': cat = 'bzcat' elif magic.guess_type(file) == 'application/x-gzip': cat = 'zcat' else: cat = 'cat' return run_cmd('%s %s | %s' % (cat, file, command), ignore_status=ignore_status)
[docs]def extract_tarball_to_dir(tarball, dir): """ Extract a tarball to a specified directory name instead of whatever the top level of a tarball is - useful for versioned directory names, etc """ if os.path.exists(dir): if os.path.isdir(dir): shutil.rmtree(dir) else: os.remove(dir) pwd = os.getcwd() os.chdir(os.path.dirname(os.path.abspath(dir))) newdir = extract_tarball(tarball) os.rename(newdir, dir) os.chdir(pwd)
[docs]def extract_tarball(tarball): """Returns the directory extracted by the tarball.""" extracted = cat_file_to_cmd(tarball, 'tar xvf - 2>/dev/null', return_output=True).splitlines() dir = None for line in extracted: line = re.sub(r'^./', '', line) if not line or line == '.': continue topdir = line.split('/')[0] if os.path.isdir(topdir): if dir: assert(dir == topdir) else: dir = topdir if dir: return dir else: raise NameError('extracting tarball produced no dir')
[docs]def hash_file(filename, size=None, method="md5"): """ Calculate the hash of filename. If size is not None, limit to first size bytes. Throw exception if something is wrong with filename. Can be also implemented with bash one-liner (assuming size%1024==0): dd if=filename bs=1024 count=size/1024 | sha1sum - :param filename: Path of the file that will have its hash calculated. :param method: Method used to calculate the hash. Supported methods: * md5 * sha1 :return: Hash of the file, if something goes wrong, return None. """ chunksize = 4096 fsize = os.path.getsize(filename) if not size or size > fsize: size = fsize f = open(filename, 'rb') try: hash = utils.hash(method) except ValueError: logging.error("Unknown hash type %s, returning None" % method) while size > 0: if chunksize > size: chunksize = size data = f.read(chunksize) if len(data) == 0: logging.debug("Nothing left to read but size=%d" % size) break hash.update(data) size -= len(data) f.close() return hash.hexdigest()
[docs]def unmap_url_cache(cachedir, url, expected_hash, method="md5"): """ Downloads a file from a URL to a cache directory. If the file is already at the expected position and has the expected hash, let's not download it again. :param cachedir: Directory that might hold a copy of the file we want to download. :param url: URL for the file we want to download. :param expected_hash: Hash string that we expect the file downloaded to have. :param method: Method used to calculate the hash string (md5, sha1). """ # Let's convert cachedir to a canonical path, if it's not already cachedir = os.path.realpath(cachedir) if not os.path.isdir(cachedir): try: os.makedirs(cachedir) except Exception: raise ValueError('Could not create cache directory %s' % cachedir) file_from_url = os.path.basename(url) file_local_path = os.path.join(cachedir, file_from_url) file_hash = None failure_counter = 0 while not file_hash == expected_hash: if os.path.isfile(file_local_path): file_hash = hash_file(file_local_path, method) if file_hash == expected_hash: # File is already at the expected position and ready to go src = file_from_url else: # Let's download the package again, it's corrupted... logging.error("Seems that file %s is corrupted, trying to " "download it again" % file_from_url) src = url failure_counter += 1 else: # File is not there, let's download it src = url if failure_counter > 1: raise EnvironmentError("Consistently failed to download the " "package %s. Aborting further download " "attempts. This might mean either the " "network connection has problems or the " "expected hash string that was determined " "for this file is wrong" % file_from_url) file_path = utils.unmap_url(cachedir, src, cachedir) return file_path
[docs]def force_copy(src, dest): """Replace dest with a new copy of src, even if it exists""" if os.path.isfile(dest): os.remove(dest) if os.path.isdir(dest): dest = os.path.join(dest, os.path.basename(src)) shutil.copyfile(src, dest) return dest
[docs]def file_contains_pattern(file, pattern): """Return true if file contains the specified egrep pattern""" if not os.path.isfile(file): raise NameError('file %s does not exist' % file) cmd_result = utils.run('egrep -q "' + pattern + '" ' + file, ignore_status=True, verbose=False) return not cmd_result.exit_status
[docs]def list_grep(list, pattern): """True if any item in list matches the specified pattern.""" compiled = re.compile(pattern) for line in list: match = compiled.search(line) if (match): return 1 return 0
[docs]def get_os_vendor(): """ Try to guess what's the os vendor. """ logging.warn('utils.get_os_vendor() is deprecated, please use ' 'autotest.client.shared.distro.detect() instead') vendor = 'Unknown' if os.path.isfile('/etc/SuSE-release'): return 'SUSE' issue = '/etc/issue' if not os.path.isfile(issue): return vendor if file_contains_pattern(issue, 'Red Hat'): vendor = 'Red Hat' if file_contains_pattern(issue, 'CentOS'): vendor = 'Red Hat' elif file_contains_pattern(issue, 'Fedora'): vendor = 'Fedora' elif file_contains_pattern(issue, 'SUSE'): vendor = 'SUSE' elif file_contains_pattern(issue, 'Ubuntu'): vendor = 'Ubuntu' elif file_contains_pattern(issue, 'Debian'): vendor = 'Debian' logging.debug("Detected OS vendor: %s", vendor) return vendor
[docs]def get_cc(): try: return os.environ['CC'] except KeyError: return 'gcc'
[docs]def get_vmlinux(): """Return the full path to vmlinux Ahem. This is crap. Pray harder. Bad Martin. """ vmlinux = '/boot/vmlinux-%s' % os.uname()[2] if os.path.isfile(vmlinux): return vmlinux vmlinux = '/lib/modules/%s/build/vmlinux' % os.uname()[2] if os.path.isfile(vmlinux): return vmlinux return None
[docs]def get_systemmap(): """Return the full path to System.map Ahem. This is crap. Pray harder. Bad Martin. """ map = '/boot/System.map-%s' % os.uname()[2] if os.path.isfile(map): return map map = '/lib/modules/%s/build/System.map' % os.uname()[2] if os.path.isfile(map): return map return None
[docs]def get_modules_dir(): """Return the modules dir for the running kernel version""" kernel_version = os.uname()[2] return '/lib/modules/%s/kernel' % kernel_version
[docs]def get_cpu_info(): """ Reads /proc/cpuinfo and returns a list of file lines :returns: `list` of lines from /proc/cpuinfo file :rtype: `list` """ f = open('/proc/cpuinfo', 'r') cpuinfo = f.readlines() f.close() return cpuinfo
[docs]def cpu_has_flags(flags): """ Check if a list of flags are available on current CPU info :param flags: A `list` of cpu flags that must exists on the current CPU. :type flags: `list` :returns: `bool` True if all the flags were found or False if not :rtype: `list` """ cpu_info = get_cpu_info() if not isinstance(flags, list): flags = [flags] for flag in flags: if not list_grep(cpu_info, '.*%s.*' % flag): return False return True
[docs]def get_cpu_vendor_name(): """ Get the current cpu vendor name :returns: string 'intel' or 'amd' or 'power7' depending on the current CPU architecture. :rtype: `string` """ vendors_map = { 'intel': ("GenuineIntel", ), 'amd': ("AMD", ), 'power7': ("POWER7", ) } cpu_info = get_cpu_info() for vendor, identifiers in vendors_map.items(): for identifier in identifiers: if list_grep(cpu_info, identifier): return vendor return None
[docs]def get_cpu_arch(): """Work out which CPU architecture we're running on""" f = open('/proc/cpuinfo', 'r') cpuinfo = f.readlines() f.close() if list_grep(cpuinfo, '^cpu.*(RS64|POWER3|Broadband Engine)'): return 'power' elif list_grep(cpuinfo, '^cpu.*POWER4'): return 'power4' elif list_grep(cpuinfo, '^cpu.*POWER5'): return 'power5' elif list_grep(cpuinfo, '^cpu.*POWER6'): return 'power6' elif list_grep(cpuinfo, '^cpu.*POWER7'): return 'power7' elif list_grep(cpuinfo, '^cpu.*PPC970'): return 'power970' elif list_grep(cpuinfo, 'ARM'): return 'arm' elif list_grep(cpuinfo, '^flags.*:.* lm .*'): return 'x86_64' else: return 'i386'
[docs]def get_current_kernel_arch(): """Get the machine architecture""" # this returns platform.uname()[4] return platform.machine()
[docs]def get_file_arch(filename): # -L means follow symlinks file_data = utils.system_output('file -L ' + filename) if file_data.count('80386'): return 'i386' return None
[docs]def count_cpus(): """number of CPUs in the local machine according to /proc/cpuinfo""" f = file('/proc/cpuinfo', 'r') cpus = 0 for line in f.readlines(): if line.lower().startswith('processor'): cpus += 1 return cpus
[docs]def sysctl(key, value=None): """Generic implementation of sysctl, to read and write. :param key: A location under /proc/sys :param value: If not None, a value to write into the sysctl. :return: The single-line sysctl value as a string. """ path = '/proc/sys/%s' % key if value is not None: utils.write_one_line(path, str(value)) return utils.read_one_line(path)
[docs]def sysctl_kernel(key, value=None): """(Very) partial implementation of sysctl, for kernel params""" if value is not None: # write utils.write_one_line('/proc/sys/kernel/%s' % key, str(value)) else: # read out = utils.read_one_line('/proc/sys/kernel/%s' % key) return int(re.search(r'\d+', out).group(0))
def _convert_exit_status(sts): if os.WIFSIGNALED(sts): return -os.WTERMSIG(sts) elif os.WIFEXITED(sts): return os.WEXITSTATUS(sts) else: # impossible? raise RuntimeError("Unknown exit status %d!" % sts)
[docs]def where_art_thy_filehandles(): """Dump the current list of filehandles""" os.system("ls -l /proc/%d/fd >> /dev/tty" % os.getpid())
[docs]def dump_object(object): """Dump an object's attributes and methods kind of like dir() """ for item in object.__dict__.iteritems(): print item try: (key, value) = item dump_object(value) except Exception: continue
[docs]def environ(env_key): """return the requested environment variable, or '' if unset""" if (os.environ.has_key(env_key)): return os.environ[env_key] else: return ''
[docs]def prepend_path(newpath, oldpath): """prepend newpath to oldpath""" if (oldpath): return newpath + ':' + oldpath else: return newpath
[docs]def append_path(oldpath, newpath): """append newpath to oldpath""" if (oldpath): return oldpath + ':' + newpath else: return newpath
[docs]def avgtime_print(dir): """ Calculate some benchmarking statistics. Input is a directory containing a file called 'time'. File contains one-per-line results of /usr/bin/time. Output is average Elapsed, User, and System time in seconds, and average CPU percentage. """ f = open(dir + "/time") user = system = elapsed = cpu = count = 0 r = re.compile('([\d\.]*)user ([\d\.]*)system (\d*):([\d\.]*)elapsed (\d*)%CPU') for line in f.readlines(): try: s = r.match(line) user += float(s.group(1)) system += float(s.group(2)) elapsed += (float(s.group(3)) * 60) + float(s.group(4)) cpu += float(s.group(5)) count += 1 except Exception: raise ValueError("badly formatted times") f.close() return "Elapsed: %0.2fs User: %0.2fs System: %0.2fs CPU: %0.0f%%" % \ (elapsed / count, user / count, system / count, cpu / count)
[docs]def running_config(): """ Return path of config file of the currently running kernel """ version = os.uname()[2] for config in ('/proc/config.gz', '/boot/config-%s' % version, '/lib/modules/%s/build/.config' % version): if os.path.isfile(config): return config return None
[docs]def check_for_kernel_feature(feature): config = running_config() if not config: raise TypeError("Can't find kernel config file") if magic.guess_type(config) == 'application/x-gzip': grep = 'zgrep' else: grep = 'grep' grep += ' ^CONFIG_%s= %s' % (feature, config) if not utils.system_output(grep, ignore_status=True): raise ValueError("Kernel doesn't have a %s feature" % (feature))
[docs]def cpu_online_map(): """ Check out the available cpu online map """ cpus = [] for line in open('/proc/cpuinfo', 'r').readlines(): if line.startswith('processor'): cpus.append(line.split()[2]) # grab cpu number return cpus
[docs]def check_glibc_ver(ver): glibc_ver = commands.getoutput('ldd --version').splitlines()[0] glibc_ver = re.search(r'(\d+\.\d+(\.\d+)?)', glibc_ver).group() if utils.compare_versions(glibc_ver, ver) == -1: raise error.TestError("Glibc too old (%s). Glibc >= %s is needed." % (glibc_ver, ver))
[docs]def check_kernel_ver(ver): kernel_ver = os.uname()[2] kv_tmp = re.split(r'[-]', kernel_ver)[0:3] # In compare_versions, if v1 < v2, return value == -1 if utils.compare_versions(kv_tmp[0], ver) == -1: raise error.TestError("Kernel too old (%s). Kernel > %s is needed." % (kernel_ver, ver))
[docs]def human_format(number): # Convert number to kilo / mega / giga format. if number < 1024: return "%d" % number kilo = float(number) / 1024.0 if kilo < 1024: return "%.2fk" % kilo meg = kilo / 1024.0 if meg < 1024: return "%.2fM" % meg gig = meg / 1024.0 return "%.2fG" % gig
[docs]def to_seconds(time_string): """Converts a string in M+:SS.SS format to S+.SS""" elts = time_string.split(':') if len(elts) == 1: return time_string return str(int(elts[0]) * 60 + float(elts[1]))
[docs]def extract_all_time_results(results_string): """Extract user, system, and elapsed times into a list of tuples""" pattern = re.compile(r"(.*?)user (.*?)system (.*?)elapsed") results = [] for result in pattern.findall(results_string): results.append(tuple([to_seconds(elt) for elt in result])) return results
[docs]def pickle_load(filename): return pickle.load(open(filename, 'r')) # Return the kernel version and build timestamp.
[docs]def running_os_release(): return os.uname()[2:4]
[docs]def running_os_ident(): (version, timestamp) = running_os_release() return version + '::' + timestamp
[docs]def running_os_full_version(): (version, timestamp) = running_os_release() return version # much like find . -name 'pattern'
[docs]def locate(pattern, root=os.getcwd()): for path, dirs, files in os.walk(root): for f in files: if fnmatch.fnmatch(f, pattern): yield os.path.abspath(os.path.join(path, f))
[docs]def freespace(path): """Return the disk free space, in bytes""" s = os.statvfs(path) return s.f_bavail * s.f_bsize
[docs]def disk_block_size(path): """Return the disk block size, in bytes""" return os.statvfs(path).f_bsize
[docs]def get_cpu_family(): procinfo = utils.system_output('cat /proc/cpuinfo') CPU_FAMILY_RE = re.compile(r'^cpu family\s+:\s+(\S+)', re.M) matches = CPU_FAMILY_RE.findall(procinfo) if matches: return int(matches[0]) else: raise error.TestError('Could not get valid cpu family data')
[docs]def get_disks(): df_output = utils.system_output('df') disk_re = re.compile(r'^(/dev/hd[a-z]+)3', re.M) return disk_re.findall(df_output)
[docs]def load_module(module_name): # Checks if a module has already been loaded if module_is_loaded(module_name): return False utils.system('/sbin/modprobe ' + module_name) return True
[docs]def unload_module(module_name): """ Removes a module. Handles dependencies. If even then it's not possible to remove one of the modules, it will trhow an error.CmdError exception. :param module_name: Name of the module we want to remove. """ l_raw = utils.system_output("/sbin/lsmod").splitlines() lsmod = [x for x in l_raw if x.split()[0] == module_name] if len(lsmod) > 0: line_parts = lsmod[0].split() if len(line_parts) == 4: submodules = line_parts[3].split(",") for submodule in submodules: unload_module(submodule) utils.system("/sbin/modprobe -r %s" % module_name) logging.info("Module %s unloaded" % module_name) else: logging.info("Module %s is already unloaded" % module_name)
[docs]def module_is_loaded(module_name): module_name = module_name.replace('-', '_') modules = utils.system_output('/sbin/lsmod').splitlines() for module in modules: if module.startswith(module_name) and module[len(module_name)] == ' ': return True return False
[docs]def get_loaded_modules(): lsmod_output = utils.system_output('/sbin/lsmod').splitlines()[1:] return [line.split(None, 1)[0] for line in lsmod_output]
[docs]def get_cpu_vendor(): cpuinfo = open('/proc/cpuinfo').read() vendors = re.findall(r'(?m)^vendor_id\s*:\s*(\S+)\s*$', cpuinfo) for i in xrange(1, len(vendors)): if vendors[i] != vendors[0]: raise error.TestError('multiple cpu vendors found: ' + str(vendors)) return vendors[0]
[docs]def ping_default_gateway(): """Ping the default gateway.""" network = open('/etc/sysconfig/network') m = re.search('GATEWAY=(\S+)', network.read()) if m: gw = m.group(1) cmd = 'ping %s -c 5 > /dev/null' % gw return utils.system(cmd, ignore_status=True) raise error.TestError('Unable to find default gateway')
[docs]def process_is_alive(name_pattern): """ 'pgrep name' misses all python processes and also long process names. 'pgrep -f name' gets all shell commands with name in args. So look only for command whose initial pathname ends with name. Name itself is an egrep pattern, so it can use | etc for variations. """ return utils.system("pgrep -f '^([^ /]*/)*(%s)([ ]|$)'" % name_pattern, ignore_status=True) == 0
[docs]def get_hwclock_seconds(utc=True): """ Return the hardware clock in seconds as a floating point value. Use Coordinated Universal Time if utc is True, local time otherwise. Raise a ValueError if unable to read the hardware clock. """ cmd = '/sbin/hwclock --debug' if utc: cmd += ' --utc' hwclock_output = utils.system_output(cmd, ignore_status=True) match = re.search(r'= ([0-9]+) seconds since .+ (-?[0-9.]+) seconds$', hwclock_output, re.DOTALL) if match: seconds = int(match.group(1)) + float(match.group(2)) logging.debug('hwclock seconds = %f' % seconds) return seconds raise ValueError('Unable to read the hardware clock -- ' + hwclock_output)
[docs]def set_wake_alarm(alarm_time): """ Set the hardware RTC-based wake alarm to 'alarm_time'. """ utils.write_one_line('/sys/class/rtc/rtc0/wakealarm', str(alarm_time))
[docs]def set_power_state(state): """ Set the system power state to 'state'. """ utils.write_one_line('/sys/power/state', state)
[docs]def standby(): """ Power-on suspend (S1) """ set_power_state('standby')
[docs]def suspend_to_ram(): """ Suspend the system to RAM (S3) """ set_power_state('mem')
[docs]def suspend_to_disk(): """ Suspend the system to disk (S4) """ set_power_state('disk')
[docs]def get_cpu_stat(key): """ Get load per cpu from /proc/stat :return: list of values of CPU times """ stats = [] stat_file = open('/proc/stat', 'r') line = stat_file.readline() while line: if line.startswith(key): stats = line.split()[1:] break line = stat_file.readline() return stats
[docs]def get_uptime(): """ :return: return the uptime of system in secs in float in error case return 'None' """ cmd = "/bin/cat /proc/uptime" (status, output) = commands.getstatusoutput(cmd) if status == 0: return output.split()[0] else: return None