mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-28 02:15:21 +07:00
19508c048a
This patch adds platform dependency into the test case 15 (perf_event_attr). It is based on a suggestion from Jiri Olsa. Add a new optional attribute named 'arch' in the [config] section of the test case file. It is a comma separated list of architecture names this test can be executed on. For example: arch = x86_64,alpha,ppc If this attribute is missing the test is executed on any platform. This does not break existing behavior. The values listed for this attribute should be identical to uname -m output. If the list starts with an exclamation mark (!) the comparison is inverted, for example for arch = !s390x,ppc the test is not executed on s390x or ppc platforms. The exclamation mark must be at the beginnning of the list. Here is an example debug output: [root@s35lp76]# fgrep arch tests/attr/test-stat-C2 arch = x86_64,alpha,ppc [root@s35lp76]# PERF_TEST_ATTR=/tmp /usr/bin/python2 ./tests/attr.py \ -d ./tests/attr/ -p ./perf -vvvvv -t test-stat-C1 provides the following output: running './tests/attr//test-stat-C1' test limitation 'x86_64,alpha,ppc' <--- new loading expected events Event event:base-stat fd = 1 group_fd = -1 ..... Here is the output when a test is skipped: [root@s35lp76]# fgrep arch tests/attr/test-stat-C1 arch = !s390x [root@s35lp76]# PERF_TEST_ATTR=/tmp /usr/bin/python2 ./tests/attr.py \ -d ./tests/attr/ -p ./perf -vvvvv -t test-stat-C1 provides the following output: test limitation '!s390x' <--- new skipped [s390x] './tests/attr//test-stat-C1' <--- new The test is skipped with return code 0. Suggested-and-Acked-by: Jiri Olsa <jolsa@redhat.com> Reviewed-by: Arnaldo Carvalho de Melo <acme@kernel.org> Signed-off-by: Thomas Richter <tmricht@linux.vnet.ibm.com> Cc: Hendrik Brueckner <brueckner@linux.vnet.ibm.com> Cc: linux-s390@vger.kernel.org Link: http://lkml.kernel.org/r/20170622073625.86762-1-tmricht@linux.vnet.ibm.com Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
380 lines
11 KiB
Python
380 lines
11 KiB
Python
#! /usr/bin/python
|
|
|
|
import os
|
|
import sys
|
|
import glob
|
|
import optparse
|
|
import tempfile
|
|
import logging
|
|
import shutil
|
|
import ConfigParser
|
|
|
|
class Fail(Exception):
|
|
def __init__(self, test, msg):
|
|
self.msg = msg
|
|
self.test = test
|
|
def getMsg(self):
|
|
return '\'%s\' - %s' % (self.test.path, self.msg)
|
|
|
|
class Notest(Exception):
|
|
def __init__(self, test, arch):
|
|
self.arch = arch
|
|
self.test = test
|
|
def getMsg(self):
|
|
return '[%s] \'%s\'' % (self.arch, self.test.path)
|
|
|
|
class Unsup(Exception):
|
|
def __init__(self, test):
|
|
self.test = test
|
|
def getMsg(self):
|
|
return '\'%s\'' % self.test.path
|
|
|
|
class Event(dict):
|
|
terms = [
|
|
'cpu',
|
|
'flags',
|
|
'type',
|
|
'size',
|
|
'config',
|
|
'sample_period',
|
|
'sample_type',
|
|
'read_format',
|
|
'disabled',
|
|
'inherit',
|
|
'pinned',
|
|
'exclusive',
|
|
'exclude_user',
|
|
'exclude_kernel',
|
|
'exclude_hv',
|
|
'exclude_idle',
|
|
'mmap',
|
|
'comm',
|
|
'freq',
|
|
'inherit_stat',
|
|
'enable_on_exec',
|
|
'task',
|
|
'watermark',
|
|
'precise_ip',
|
|
'mmap_data',
|
|
'sample_id_all',
|
|
'exclude_host',
|
|
'exclude_guest',
|
|
'exclude_callchain_kernel',
|
|
'exclude_callchain_user',
|
|
'wakeup_events',
|
|
'bp_type',
|
|
'config1',
|
|
'config2',
|
|
'branch_sample_type',
|
|
'sample_regs_user',
|
|
'sample_stack_user',
|
|
]
|
|
|
|
def add(self, data):
|
|
for key, val in data:
|
|
log.debug(" %s = %s" % (key, val))
|
|
self[key] = val
|
|
|
|
def __init__(self, name, data, base):
|
|
log.debug(" Event %s" % name);
|
|
self.name = name;
|
|
self.group = ''
|
|
self.add(base)
|
|
self.add(data)
|
|
|
|
def compare_data(self, a, b):
|
|
# Allow multiple values in assignment separated by '|'
|
|
a_list = a.split('|')
|
|
b_list = b.split('|')
|
|
|
|
for a_item in a_list:
|
|
for b_item in b_list:
|
|
if (a_item == b_item):
|
|
return True
|
|
elif (a_item == '*') or (b_item == '*'):
|
|
return True
|
|
|
|
return False
|
|
|
|
def equal(self, other):
|
|
for t in Event.terms:
|
|
log.debug(" [%s] %s %s" % (t, self[t], other[t]));
|
|
if not self.has_key(t) or not other.has_key(t):
|
|
return False
|
|
if not self.compare_data(self[t], other[t]):
|
|
return False
|
|
return True
|
|
|
|
def diff(self, other):
|
|
for t in Event.terms:
|
|
if not self.has_key(t) or not other.has_key(t):
|
|
continue
|
|
if not self.compare_data(self[t], other[t]):
|
|
log.warning("expected %s=%s, got %s" % (t, self[t], other[t]))
|
|
|
|
# Test file description needs to have following sections:
|
|
# [config]
|
|
# - just single instance in file
|
|
# - needs to specify:
|
|
# 'command' - perf command name
|
|
# 'args' - special command arguments
|
|
# 'ret' - expected command return value (0 by default)
|
|
# 'arch' - architecture specific test (optional)
|
|
# comma separated list, ! at the beginning
|
|
# negates it.
|
|
#
|
|
# [eventX:base]
|
|
# - one or multiple instances in file
|
|
# - expected values assignments
|
|
class Test(object):
|
|
def __init__(self, path, options):
|
|
parser = ConfigParser.SafeConfigParser()
|
|
parser.read(path)
|
|
|
|
log.warning("running '%s'" % path)
|
|
|
|
self.path = path
|
|
self.test_dir = options.test_dir
|
|
self.perf = options.perf
|
|
self.command = parser.get('config', 'command')
|
|
self.args = parser.get('config', 'args')
|
|
|
|
try:
|
|
self.ret = parser.get('config', 'ret')
|
|
except:
|
|
self.ret = 0
|
|
|
|
try:
|
|
self.arch = parser.get('config', 'arch')
|
|
log.warning("test limitation '%s'" % self.arch)
|
|
except:
|
|
self.arch = ''
|
|
|
|
self.expect = {}
|
|
self.result = {}
|
|
log.debug(" loading expected events");
|
|
self.load_events(path, self.expect)
|
|
|
|
def is_event(self, name):
|
|
if name.find("event") == -1:
|
|
return False
|
|
else:
|
|
return True
|
|
|
|
def skip_test(self, myarch):
|
|
# If architecture not set always run test
|
|
if self.arch == '':
|
|
# log.warning("test for arch %s is ok" % myarch)
|
|
return False
|
|
|
|
# Allow multiple values in assignment separated by ','
|
|
arch_list = self.arch.split(',')
|
|
|
|
# Handle negated list such as !s390x,ppc
|
|
if arch_list[0][0] == '!':
|
|
arch_list[0] = arch_list[0][1:]
|
|
log.warning("excluded architecture list %s" % arch_list)
|
|
for arch_item in arch_list:
|
|
# log.warning("test for %s arch is %s" % (arch_item, myarch))
|
|
if arch_item == myarch:
|
|
return True
|
|
return False
|
|
|
|
for arch_item in arch_list:
|
|
# log.warning("test for architecture '%s' current '%s'" % (arch_item, myarch))
|
|
if arch_item == myarch:
|
|
return False
|
|
return True
|
|
|
|
def load_events(self, path, events):
|
|
parser_event = ConfigParser.SafeConfigParser()
|
|
parser_event.read(path)
|
|
|
|
# The event record section header contains 'event' word,
|
|
# optionaly followed by ':' allowing to load 'parent
|
|
# event' first as a base
|
|
for section in filter(self.is_event, parser_event.sections()):
|
|
|
|
parser_items = parser_event.items(section);
|
|
base_items = {}
|
|
|
|
# Read parent event if there's any
|
|
if (':' in section):
|
|
base = section[section.index(':') + 1:]
|
|
parser_base = ConfigParser.SafeConfigParser()
|
|
parser_base.read(self.test_dir + '/' + base)
|
|
base_items = parser_base.items('event')
|
|
|
|
e = Event(section, parser_items, base_items)
|
|
events[section] = e
|
|
|
|
def run_cmd(self, tempdir):
|
|
junk1, junk2, junk3, junk4, myarch = (os.uname())
|
|
|
|
if self.skip_test(myarch):
|
|
raise Notest(self, myarch)
|
|
|
|
cmd = "PERF_TEST_ATTR=%s %s %s -o %s/perf.data %s" % (tempdir,
|
|
self.perf, self.command, tempdir, self.args)
|
|
ret = os.WEXITSTATUS(os.system(cmd))
|
|
|
|
log.info(" '%s' ret %d " % (cmd, ret))
|
|
|
|
if ret != int(self.ret):
|
|
raise Unsup(self)
|
|
|
|
def compare(self, expect, result):
|
|
match = {}
|
|
|
|
log.debug(" compare");
|
|
|
|
# For each expected event find all matching
|
|
# events in result. Fail if there's not any.
|
|
for exp_name, exp_event in expect.items():
|
|
exp_list = []
|
|
log.debug(" matching [%s]" % exp_name)
|
|
for res_name, res_event in result.items():
|
|
log.debug(" to [%s]" % res_name)
|
|
if (exp_event.equal(res_event)):
|
|
exp_list.append(res_name)
|
|
log.debug(" ->OK")
|
|
else:
|
|
log.debug(" ->FAIL");
|
|
|
|
log.debug(" match: [%s] matches %s" % (exp_name, str(exp_list)))
|
|
|
|
# we did not any matching event - fail
|
|
if (not exp_list):
|
|
exp_event.diff(res_event)
|
|
raise Fail(self, 'match failure');
|
|
|
|
match[exp_name] = exp_list
|
|
|
|
# For each defined group in the expected events
|
|
# check we match the same group in the result.
|
|
for exp_name, exp_event in expect.items():
|
|
group = exp_event.group
|
|
|
|
if (group == ''):
|
|
continue
|
|
|
|
for res_name in match[exp_name]:
|
|
res_group = result[res_name].group
|
|
if res_group not in match[group]:
|
|
raise Fail(self, 'group failure')
|
|
|
|
log.debug(" group: [%s] matches group leader %s" %
|
|
(exp_name, str(match[group])))
|
|
|
|
log.debug(" matched")
|
|
|
|
def resolve_groups(self, events):
|
|
for name, event in events.items():
|
|
group_fd = event['group_fd'];
|
|
if group_fd == '-1':
|
|
continue;
|
|
|
|
for iname, ievent in events.items():
|
|
if (ievent['fd'] == group_fd):
|
|
event.group = iname
|
|
log.debug('[%s] has group leader [%s]' % (name, iname))
|
|
break;
|
|
|
|
def run(self):
|
|
tempdir = tempfile.mkdtemp();
|
|
|
|
try:
|
|
# run the test script
|
|
self.run_cmd(tempdir);
|
|
|
|
# load events expectation for the test
|
|
log.debug(" loading result events");
|
|
for f in glob.glob(tempdir + '/event*'):
|
|
self.load_events(f, self.result);
|
|
|
|
# resolve group_fd to event names
|
|
self.resolve_groups(self.expect);
|
|
self.resolve_groups(self.result);
|
|
|
|
# do the expectation - results matching - both ways
|
|
self.compare(self.expect, self.result)
|
|
self.compare(self.result, self.expect)
|
|
|
|
finally:
|
|
# cleanup
|
|
shutil.rmtree(tempdir)
|
|
|
|
|
|
def run_tests(options):
|
|
for f in glob.glob(options.test_dir + '/' + options.test):
|
|
try:
|
|
Test(f, options).run()
|
|
except Unsup, obj:
|
|
log.warning("unsupp %s" % obj.getMsg())
|
|
except Notest, obj:
|
|
log.warning("skipped %s" % obj.getMsg())
|
|
|
|
def setup_log(verbose):
|
|
global log
|
|
level = logging.CRITICAL
|
|
|
|
if verbose == 1:
|
|
level = logging.WARNING
|
|
if verbose == 2:
|
|
level = logging.INFO
|
|
if verbose >= 3:
|
|
level = logging.DEBUG
|
|
|
|
log = logging.getLogger('test')
|
|
log.setLevel(level)
|
|
ch = logging.StreamHandler()
|
|
ch.setLevel(level)
|
|
formatter = logging.Formatter('%(message)s')
|
|
ch.setFormatter(formatter)
|
|
log.addHandler(ch)
|
|
|
|
USAGE = '''%s [OPTIONS]
|
|
-d dir # tests dir
|
|
-p path # perf binary
|
|
-t test # single test
|
|
-v # verbose level
|
|
''' % sys.argv[0]
|
|
|
|
def main():
|
|
parser = optparse.OptionParser(usage=USAGE)
|
|
|
|
parser.add_option("-t", "--test",
|
|
action="store", type="string", dest="test")
|
|
parser.add_option("-d", "--test-dir",
|
|
action="store", type="string", dest="test_dir")
|
|
parser.add_option("-p", "--perf",
|
|
action="store", type="string", dest="perf")
|
|
parser.add_option("-v", "--verbose",
|
|
action="count", dest="verbose")
|
|
|
|
options, args = parser.parse_args()
|
|
if args:
|
|
parser.error('FAILED wrong arguments %s' % ' '.join(args))
|
|
return -1
|
|
|
|
setup_log(options.verbose)
|
|
|
|
if not options.test_dir:
|
|
print 'FAILED no -d option specified'
|
|
sys.exit(-1)
|
|
|
|
if not options.test:
|
|
options.test = 'test*'
|
|
|
|
try:
|
|
run_tests(options)
|
|
|
|
except Fail, obj:
|
|
print "FAILED %s" % obj.getMsg();
|
|
sys.exit(-1)
|
|
|
|
sys.exit(0)
|
|
|
|
if __name__ == '__main__':
|
|
main()
|