X-Git-Url: https://git.libre-soc.org/?a=blobdiff_plain;f=tests%2FSConscript;h=1228e78d28e58ee753fe7315b4ef965b4d669c8c;hb=866cfaf9dc596d8547e14bc2133fb962776572a7;hp=5eadce6d42c78e3f414a45e0f749779a1f0fced5;hpb=0850c3dedbedaaa269ff2e28680b91ae3767b163;p=gem5.git diff --git a/tests/SConscript b/tests/SConscript index 5eadce6d4..1228e78d2 100644 --- a/tests/SConscript +++ b/tests/SConscript @@ -1,6 +1,6 @@ # -*- mode:python -*- -# Copyright (c) 2004-2005 The Regents of The University of Michigan +# Copyright (c) 2004-2006 The Regents of The University of Michigan # All rights reserved. # # Redistribution and use in source and binary forms, with or without @@ -25,6 +25,9 @@ # THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +# +# Authors: Steve Reinhardt +# Kevin Lim import os import sys @@ -58,10 +61,12 @@ def check_test(target, source, env): # Exclude m5stats.txt since we will use diff-out on that. Execute(env.subst('diff -ubr ${SOURCES[0].dir} ${SOURCES[1].dir} ' + '-I "^command line:" ' + # for stdout file - '-I "^M5 compiled on" ' + # for stderr file - '-I "^M5 simulation started" ' + # for stderr file + '-I "^M5 compiled " ' + # for stderr file + '-I "^M5 started " ' + # for stderr file + '-I "^M5 executing on " ' + # for stderr file '-I "^Simulation complete at" ' + # for stderr file '-I "^Listening for" ' + # for stderr file + '-I "listening for remote gdb" ' + # for stderr file '--exclude=m5stats.txt --exclude=SCCS ' + '--exclude=${TARGETS[0].file} ' + '> ${TARGETS[0]}', target=target, source=source), None) @@ -136,109 +141,78 @@ def update_test_string(target, source, env): updateAction = env.Action(update_test, update_test_string) -def test_builder(env, category, cpu_list=[], os_list=[], refdir='ref', timeout=15): - """Define a test. - - Args: - category -- string describing test category (e.g., 'quick') - cpu_list -- list of CPUs to runs this test on (blank means all compiled CPUs) - os_list -- list of OSs to run this test on - refdir -- subdirectory containing reference output (default 'ref') - timeout -- test timeout in minutes (only enforced on pool) - - """ - - default_refdir = False - if refdir == 'ref': - default_refdir = True - if len(cpu_list) == 0: - cpu_list = env['CPU_MODELS'] -# if len(os_list) == 0: -# raise RuntimeError, "No OS specified" -# else: -# for test_os in os_list: -# build_cpu_test(env, category, test_os, cpu_list, refdir, timeout) - # Loop through CPU models and generate proper options, ref directories for each - for cpu in cpu_list: - test_os = '' - if cpu == "AtomicSimpleCPU": - cpu_option = ('','atomic/') - elif cpu == "TimingSimpleCPU": - cpu_option = ('--timing','timing/') - elif cpu == "O3CPU": - cpu_option = ('--detailed','detailed/') - else: - raise TypeError, "Unknown CPU model specified" - - if default_refdir: - # Reference stats located in ref/arch/os/cpu or ref/arch/cpu if no OS specified - test_refdir = os.path.join(refdir, env['TARGET_ISA']) - if test_os != '': - test_refdir = os.path.join(test_refdir, test_os) - cpu_refdir = os.path.join(test_refdir, cpu_option[1]) - - ref_stats = os.path.join(cpu_refdir, 'm5stats.txt') - - # base command for running test - base_cmd = '${SOURCES[0]} -d $TARGET.dir ${SOURCES[1]}' - base_cmd = base_cmd + ' ' + cpu_option[0] - # stdout and stderr files - cmd_stdout = '${TARGETS[0]}' - cmd_stderr = '${TARGETS[1]}' - - stdout_string = cpu_option[1] + 'stdout' - stderr_string = cpu_option[1] + 'stderr' - m5stats_string = cpu_option[1] + 'm5stats.txt' - outdiff_string = cpu_option[1] + 'outdiff' - statsdiff_string = cpu_option[1] + 'statsdiff' - status_string = cpu_option[1] + 'status' - - # Prefix test run with batch job submission command if appropriate. - # Output redirection is also different for batch runs. - # Batch command also supports timeout arg (in seconds, not minutes). - if env['BATCH']: - cmd = [env['BATCH_CMD'], '-t', str(timeout * 60), - '-o', cmd_stdout, '-e', cmd_stderr, base_cmd] - else: - cmd = [base_cmd, '>', cmd_stdout, '2>', cmd_stderr] +def test_builder(env, ref_dir): + """Define a test.""" + + (category, name, _ref, isa, opsys, config) = ref_dir.split('/') + assert(_ref == 'ref') + + # target path (where test output goes) is the same except without + # the 'ref' component + tgt_dir = os.path.join(category, name, isa, opsys, config) + + # prepend file name with tgt_dir + def tgt(f): + return os.path.join(tgt_dir, f) + + ref_stats = os.path.join(ref_dir, 'm5stats.txt') + new_stats = tgt('m5stats.txt') + status_file = tgt('status') + + # Base command for running test. We mess around with indirectly + # referring to files via SOURCES and TARGETS so that scons can + # mess with paths all it wants to and we still get the right + # files. + base_cmd = '${SOURCES[0]} -d $TARGET.dir ${SOURCES[1]} %s' % tgt_dir + # stdout and stderr files + cmd_stdout = '${TARGETS[0]}' + cmd_stderr = '${TARGETS[1]}' + + # Prefix test run with batch job submission command if appropriate. + # Output redirection is also different for batch runs. + # Batch command also supports timeout arg (in seconds, not minutes). + timeout = 15 # used to be a param, probably should be again + if env['BATCH']: + cmd = [env['BATCH_CMD'], '-t', str(timeout * 60), + '-o', cmd_stdout, '-e', cmd_stderr, base_cmd] + else: + cmd = [base_cmd, '>', cmd_stdout, '2>', cmd_stderr] - env.Command([stdout_string, stderr_string, m5stats_string], [env.M5Binary, 'run.py'], - ' '.join(cmd)) - - # order of targets is important... see check_test - env.Command([outdiff_string, statsdiff_string, status_string], - [ref_stats, m5stats_string], - testAction) - - # phony target to echo status - if env['update_ref']: - p = env.Command(cpu_option[1] + '_update', [ref_stats, m5stats_string, status_string], - updateAction) - else: - p = env.Command(cpu_option[1] + '_print', [status_string], printAction) - env.AlwaysBuild(p) + env.Command([tgt('stdout'), tgt('stderr'), new_stats], + [env.M5Binary, 'run.py'], ' '.join(cmd)) + + # order of targets is important... see check_test + env.Command([tgt('outdiff'), tgt('statsdiff'), status_file], + [ref_stats, new_stats], + testAction) + + # phony target to echo status + if env['update_ref']: + p = env.Command(tgt('_update'), + [ref_stats, new_stats, status_file], + updateAction) + else: + p = env.Command(tgt('_print'), [status_file], printAction) + + env.AlwaysBuild(p) - env.Tests.setdefault(category, []) - env.Tests[category] += p -# Make test_builder a "wrapper" function. See SCons wiki page at -# http://www.scons.org/cgi-bin/wiki/WrapperFunctions. -SConsEnvironment.Test = test_builder +# Figure out applicable configs based on build type +configs = [] +if env['FULL_SYSTEM']: + if env['TARGET_ISA'] == 'alpha': + if not env['ALPHA_TLASER']: + configs += ['tsunami-simple-atomic', + 'tsunami-simple-timing', + 'tsunami-simple-atomic-dual', + 'tsunami-simple-timing-dual'] +else: + configs += ['simple-atomic', 'simple-timing', 'o3-timing'] cwd = os.getcwd() os.chdir(str(Dir('.').srcdir)) -scripts = glob.glob('*/SConscript') +for config in configs: + dirs = glob.glob('*/*/ref/%s/*/%s' % (env['TARGET_ISA'], config)) + for d in dirs: + test_builder(env, d) os.chdir(cwd) - -for s in scripts: - SConscript(s, exports = 'env', duplicate = False) - -# Set up phony commands for various test categories -allTests = [] -for (key, val) in env.Tests.iteritems(): - env.Command(key, val, env.NoAction) - allTests += val - -# The 'all' target is redundant since just specifying the test -# directory name (e.g., ALPHA_SE/test/opt) has the same effect. -env.Command('all', allTests, env.NoAction)