X-Git-Url: https://git.libre-soc.org/?a=blobdiff_plain;f=SConstruct;h=146b156d183937e15965cf264748fb58e9e88fd9;hb=e2bc2142d4ce6c7cd655939f7430956837343aed;hp=c832fa1b86fd3fea3c4944aea0c6360dedb54aad;hpb=55bfe13705a3eccdffb6846dd87df5f190b04c99;p=gem5.git diff --git a/SConstruct b/SConstruct index c832fa1b8..146b156d1 100755 --- a/SConstruct +++ b/SConstruct @@ -1,5 +1,17 @@ # -*- mode:python -*- +# Copyright (c) 2013, 2015, 2016 ARM Limited +# All rights reserved. +# +# The license below extends only to copyright in the software and shall +# not be construed as granting a license to any other intellectual +# property including but not limited to intellectual property relating +# to a hardware implementation of the functionality of the software +# licensed hereunder. You may use the software subject to the license +# terms below provided that you ensure that this notice is replicated +# unmodified and in its entirety in all distributions of the software, +# modified or unmodified, in source code or in binary form. +# # Copyright (c) 2011 Advanced Micro Devices, Inc. # Copyright (c) 2009 The Hewlett-Packard Development Company # Copyright (c) 2004-2005 The Regents of The University of Michigan @@ -81,16 +93,15 @@ For more details, see: """ raise -# We ensure the python version early because we have stuff that -# requires python 2.4 +# We ensure the python version early because because python-config +# requires python 2.5 try: - EnsurePythonVersion(2, 4) + EnsurePythonVersion(2, 5) except SystemExit, e: print """ You can use a non-default installation of the Python interpreter by -either (1) rearranging your PATH so that scons finds the non-default -'python' first or (2) explicitly invoking an alternative interpreter -on the scons script. +rearranging your PATH so that scons finds the non-default 'python' and +'python-config' first. For more details, see: http://gem5.org/wiki/index.php/Using_a_non-default_Python_installation @@ -98,8 +109,10 @@ For more details, see: raise # Global Python includes +import itertools import os import re +import shutil import subprocess import sys @@ -116,7 +129,7 @@ extra_python_paths = [ Dir('src/python').srcnode().abspath, # gem5 includes Dir('ext/ply').srcnode().abspath, # ply is used by several files ] - + sys.path[1:1] = extra_python_paths from m5.util import compareVersions, readCommand @@ -161,14 +174,36 @@ AddLocalOption('--colors', dest='use_colors', action='store_true', help="Add color to abbreviated scons output") AddLocalOption('--no-colors', dest='use_colors', action='store_false', help="Don't add color to abbreviated scons output") +AddLocalOption('--with-cxx-config', dest='with_cxx_config', + action='store_true', + help="Build with support for C++-based configuration") AddLocalOption('--default', dest='default', type='string', action='store', help='Override which build_opts file to use for defaults') AddLocalOption('--ignore-style', dest='ignore_style', action='store_true', help='Disable style checking hooks') +AddLocalOption('--no-lto', dest='no_lto', action='store_true', + help='Disable Link-Time Optimization for fast') +AddLocalOption('--force-lto', dest='force_lto', action='store_true', + help='Use Link-Time Optimization instead of partial linking' + + ' when the compiler doesn\'t support using them together.') AddLocalOption('--update-ref', dest='update_ref', action='store_true', help='Update test reference outputs') AddLocalOption('--verbose', dest='verbose', action='store_true', help='Print full tool command lines') +AddLocalOption('--without-python', dest='without_python', + action='store_true', + help='Build without Python configuration support') +AddLocalOption('--without-tcmalloc', dest='without_tcmalloc', + action='store_true', + help='Disable linking against tcmalloc') +AddLocalOption('--with-ubsan', dest='with_ubsan', action='store_true', + help='Build with Undefined Behavior Sanitizer if available') +AddLocalOption('--with-asan', dest='with_asan', action='store_true', + help='Build with Address Sanitizer if available') + +if GetOption('no_lto') and GetOption('force_lto'): + print '--no-lto and --force-lto are mutually exclusive' + Exit(1) termcap = get_termcap(GetOption('use_colors')) @@ -177,19 +212,42 @@ termcap = get_termcap(GetOption('use_colors')) # Set up the main build environment. # ######################################################################## + +# export TERM so that clang reports errors in color use_vars = set([ 'AS', 'AR', 'CC', 'CXX', 'HOME', 'LD_LIBRARY_PATH', - 'LIBRARY_PATH', 'PATH', 'PYTHONPATH', 'RANLIB', 'SWIG' ]) + 'LIBRARY_PATH', 'PATH', 'PKG_CONFIG_PATH', 'PROTOC', + 'PYTHONPATH', 'RANLIB', 'TERM' ]) + +use_prefixes = [ + "ASAN_", # address sanitizer symbolizer path and settings + "CCACHE_", # ccache (caching compiler wrapper) configuration + "CCC_", # clang static analyzer configuration + "DISTCC_", # distcc (distributed compiler wrapper) configuration + "INCLUDE_SERVER_", # distcc pump server settings + "M5", # M5 configuration (e.g., path to kernels) + ] use_env = {} -for key,val in os.environ.iteritems(): - if key in use_vars or key.startswith("M5"): +for key,val in sorted(os.environ.iteritems()): + if key in use_vars or \ + any([key.startswith(prefix) for prefix in use_prefixes]): use_env[key] = val -main = Environment(ENV=use_env) +# Tell scons to avoid implicit command dependencies to avoid issues +# with the param wrappes being compiled twice (see +# http://scons.tigris.org/issues/show_bug.cgi?id=2811) +main = Environment(ENV=use_env, IMPLICIT_COMMAND_DEPENDENCIES=0) main.Decider('MD5-timestamp') main.root = Dir(".") # The current directory (where this file lives). main.srcdir = Dir("src") # The source directory +main_dict_keys = main.Dictionary().keys() + +# Check that we have a C/C++ compiler +if not ('CC' in main_dict_keys and 'CXX' in main_dict_keys): + print "No C++ compiler installed (package g++ on Ubuntu and RedHat)" + Exit(1) + # add useful python code PYTHONPATH so it can be used by subprocesses # as well main.AppendENVPath('PYTHONPATH', extra_python_paths) @@ -205,21 +263,40 @@ main.AppendENVPath('PYTHONPATH', extra_python_paths) hgdir = main.root.Dir(".hg") + +style_message = """ +You're missing the gem5 style hook, which automatically checks your code +against the gem5 style rules on %s. +This script will now install the hook in your %s. +Press enter to continue, or ctrl-c to abort: """ + mercurial_style_message = """ You're missing the gem5 style hook, which automatically checks your code -against the gem5 style rules on hg commit and qrefresh commands. This -script will now install the hook in your .hg/hgrc file. +against the gem5 style rules on hg commit and qrefresh commands. +This script will now install the hook in your .hg/hgrc file. +Press enter to continue, or ctrl-c to abort: """ + +git_style_message = """ +You're missing the gem5 style or commit message hook. These hooks help +to ensure that your code follows gem5's style rules on git commit. +This script will now install the hook in your .git/hooks/ directory. +Press enter to continue, or ctrl-c to abort: """ + +mercurial_style_upgrade_message = """ +Your Mercurial style hooks are not up-to-date. This script will now +try to automatically update them. A backup of your hgrc will be saved +in .hg/hgrc.old. Press enter to continue, or ctrl-c to abort: """ mercurial_style_hook = """ # The following lines were automatically added by gem5/SConstruct # to provide the gem5 style-checking hooks [extensions] -style = %s/util/style.py +hgstyle = %s/util/hgstyle.py [hooks] -pretxncommit.style = python:style.check_style -pre-qrefresh.style = python:style.check_style +pretxncommit.style = python:hgstyle.check_style +pre-qrefresh.style = python:hgstyle.check_style # End of SConstruct additions """ % (main.root.abspath) @@ -231,20 +308,58 @@ hook. It is important. """ # Check for style hook and prompt for installation if it's not there. -# Skip this if --ignore-style was specified, there's no .hg dir to -# install a hook in, or there's no interactive terminal to prompt. -if not GetOption('ignore_style') and hgdir.exists() and sys.stdin.isatty(): +# Skip this if --ignore-style was specified, there's no interactive +# terminal to prompt, or no recognized revision control system can be +# found. +ignore_style = GetOption('ignore_style') or not sys.stdin.isatty() + +# Try wire up Mercurial to the style hooks +if not ignore_style and hgdir.exists(): style_hook = True + style_hooks = tuple() + hgrc = hgdir.File('hgrc') + hgrc_old = hgdir.File('hgrc.old') try: from mercurial import ui ui = ui.ui() - ui.readconfig(hgdir.File('hgrc').abspath) - style_hook = ui.config('hooks', 'pretxncommit.style', None) and \ - ui.config('hooks', 'pre-qrefresh.style', None) + ui.readconfig(hgrc.abspath) + style_hooks = (ui.config('hooks', 'pretxncommit.style', None), + ui.config('hooks', 'pre-qrefresh.style', None)) + style_hook = all(style_hooks) + style_extension = ui.config('extensions', 'style', None) except ImportError: print mercurial_lib_not_found - if not style_hook: + if "python:style.check_style" in style_hooks: + # Try to upgrade the style hooks + print mercurial_style_upgrade_message + # continue unless user does ctrl-c/ctrl-d etc. + try: + raw_input() + except: + print "Input exception, exiting scons.\n" + sys.exit(1) + shutil.copyfile(hgrc.abspath, hgrc_old.abspath) + re_style_hook = re.compile(r"^([^=#]+)\.style\s*=\s*([^#\s]+).*") + re_style_extension = re.compile("style\s*=\s*([^#\s]+).*") + old, new = open(hgrc_old.abspath, 'r'), open(hgrc.abspath, 'w') + for l in old: + m_hook = re_style_hook.match(l) + m_ext = re_style_extension.match(l) + if m_hook: + hook, check = m_hook.groups() + if check != "python:style.check_style": + print "Warning: %s.style is using a non-default " \ + "checker: %s" % (hook, check) + if hook not in ("pretxncommit", "pre-qrefresh"): + print "Warning: Updating unknown style hook: %s" % hook + + l = "%s.style = python:hgstyle.check_style\n" % hook + elif m_ext and m_ext.group(1) == style_extension: + l = "hgstyle = %s/util/hgstyle.py\n" % main.root.abspath + + new.write(l) + elif not style_hook: print mercurial_style_message, # continue unless user does ctrl-c/ctrl-d etc. try: @@ -255,13 +370,76 @@ if not GetOption('ignore_style') and hgdir.exists() and sys.stdin.isatty(): hgrc_path = '%s/.hg/hgrc' % main.root.abspath print "Adding style hook to", hgrc_path, "\n" try: - hgrc = open(hgrc_path, 'a') - hgrc.write(mercurial_style_hook) - hgrc.close() + with open(hgrc_path, 'a') as f: + f.write(mercurial_style_hook) except: print "Error updating", hgrc_path sys.exit(1) +def install_git_style_hooks(): + try: + gitdir = Dir(readCommand( + ["git", "rev-parse", "--git-dir"]).strip("\n")) + except Exception, e: + print "Warning: Failed to find git repo directory: %s" % e + return + + git_hooks = gitdir.Dir("hooks") + def hook_exists(hook_name): + hook = git_hooks.File(hook_name) + return hook.exists() + + def hook_install(hook_name, script): + hook = git_hooks.File(hook_name) + if hook.exists(): + print "Warning: Can't install %s, hook already exists." % hook_name + return + + if hook.islink(): + print "Warning: Removing broken symlink for hook %s." % hook_name + os.unlink(hook.get_abspath()) + + if not git_hooks.exists(): + mkdir(git_hooks.get_abspath()) + git_hooks.clear() + + abs_symlink_hooks = git_hooks.islink() and \ + os.path.isabs(os.readlink(git_hooks.get_abspath())) + + # Use a relative symlink if the hooks live in the source directory, + # and the hooks directory is not a symlink to an absolute path. + if hook.is_under(main.root) and not abs_symlink_hooks: + script_path = os.path.relpath( + os.path.realpath(script.get_abspath()), + os.path.realpath(hook.Dir(".").get_abspath())) + else: + script_path = script.get_abspath() + + try: + os.symlink(script_path, hook.get_abspath()) + except: + print "Error updating git %s hook" % hook_name + raise + + if hook_exists("pre-commit") and hook_exists("commit-msg"): + return + + print git_style_message, + try: + raw_input() + except: + print "Input exception, exiting scons.\n" + sys.exit(1) + + git_style_script = File("util/git-pre-commit.py") + git_msg_script = File("ext/git-commit-msg") + + hook_install("pre-commit", git_style_script) + hook_install("commit-msg", git_msg_script) + +# Try to wire up git to the style hooks +if not ignore_style and main.root.Entry(".git").exists(): + install_git_style_hooks() ################################################### # @@ -348,7 +526,7 @@ global_vars = Variables(global_vars_file, args=ARGUMENTS) global_vars.AddVariables( ('CC', 'C compiler', environ.get('CC', main['CC'])), ('CXX', 'C++ compiler', environ.get('CXX', main['CXX'])), - ('SWIG', 'SWIG tool', environ.get('SWIG', main['SWIG'])), + ('PROTOC', 'protoc tool', environ.get('PROTOC', 'protoc')), ('BATCH', 'Use batch pool for build and tests', False), ('BATCH_CMD', 'Batch pool submission command name', 'qdo'), ('M5_BUILD_CACHE', 'Cache built objects in this directory', False), @@ -468,75 +646,209 @@ else: main['CCCOMSTR'] = Transform("CC") main['CXXCOMSTR'] = Transform("CXX") main['ASCOMSTR'] = Transform("AS") - main['SWIGCOMSTR'] = Transform("SWIG") main['ARCOMSTR'] = Transform("AR", 0) main['LINKCOMSTR'] = Transform("LINK", 0) + main['SHLINKCOMSTR'] = Transform("SHLINK", 0) main['RANLIBCOMSTR'] = Transform("RANLIB", 0) main['M4COMSTR'] = Transform("M4") main['SHCCCOMSTR'] = Transform("SHCC") main['SHCXXCOMSTR'] = Transform("SHCXX") Export('MakeAction') +# Initialize the Link-Time Optimization (LTO) flags +main['LTO_CCFLAGS'] = [] +main['LTO_LDFLAGS'] = [] + +# According to the readme, tcmalloc works best if the compiler doesn't +# assume that we're using the builtin malloc and friends. These flags +# are compiler-specific, so we need to set them after we detect which +# compiler we're using. +main['TCMALLOC_CCFLAGS'] = [] + CXX_version = readCommand([main['CXX'],'--version'], exception=False) CXX_V = readCommand([main['CXX'],'-V'], exception=False) main['GCC'] = CXX_version and CXX_version.find('g++') >= 0 -main['SUNCC'] = CXX_V and CXX_V.find('Sun C++') >= 0 -main['ICC'] = CXX_V and CXX_V.find('Intel') >= 0 main['CLANG'] = CXX_version and CXX_version.find('clang') >= 0 -if main['GCC'] + main['SUNCC'] + main['ICC'] + main['CLANG'] > 1: +if main['GCC'] + main['CLANG'] > 1: print 'Error: How can we have two at the same time?' Exit(1) # Set up default C++ compiler flags -if main['GCC']: +if main['GCC'] or main['CLANG']: + # As gcc and clang share many flags, do the common parts here main.Append(CCFLAGS=['-pipe']) main.Append(CCFLAGS=['-fno-strict-aliasing']) - main.Append(CCFLAGS=['-Wall', '-Wno-sign-compare', '-Wundef']) - # Read the GCC version to check for versions with bugs - # Note CCVERSION doesn't work here because it is run with the CC - # before we override it from the command line + # Enable -Wall and -Wextra and then disable the few warnings that + # we consistently violate + main.Append(CCFLAGS=['-Wall', '-Wundef', '-Wextra', + '-Wno-sign-compare', '-Wno-unused-parameter']) + # We always compile using C++11 + main.Append(CXXFLAGS=['-std=c++11']) + if sys.platform.startswith('freebsd'): + main.Append(CCFLAGS=['-I/usr/local/include']) + main.Append(CXXFLAGS=['-I/usr/local/include']) + + main['FILTER_PSHLINKFLAGS'] = lambda x: str(x).replace(' -shared', '') + main['PSHLINKFLAGS'] = main.subst('${FILTER_PSHLINKFLAGS(SHLINKFLAGS)}') + main['PLINKFLAGS'] = main.subst('${LINKFLAGS}') + shared_partial_flags = ['-r', '-nostdlib'] + main.Append(PSHLINKFLAGS=shared_partial_flags) + main.Append(PLINKFLAGS=shared_partial_flags) +else: + print termcap.Yellow + termcap.Bold + 'Error' + termcap.Normal, + print "Don't know what compiler options to use for your compiler." + print termcap.Yellow + ' compiler:' + termcap.Normal, main['CXX'] + print termcap.Yellow + ' version:' + termcap.Normal, + if not CXX_version: + print termcap.Yellow + termcap.Bold + "COMMAND NOT FOUND!" +\ + termcap.Normal + else: + print CXX_version.replace('\n', '') + print " If you're trying to use a compiler other than GCC" + print " or clang, there appears to be something wrong with your" + print " environment." + print " " + print " If you are trying to use a compiler other than those listed" + print " above you will need to ease fix SConstruct and " + print " src/SConscript to support that compiler." + Exit(1) + +if main['GCC']: + # Check for a supported version of gcc. >= 4.8 is chosen for its + # level of c++11 support. See + # http://gcc.gnu.org/projects/cxx0x.html for details. gcc_version = readCommand([main['CXX'], '-dumpversion'], exception=False) + if compareVersions(gcc_version, "4.8") < 0: + print 'Error: gcc version 4.8 or newer required.' + print ' Installed version:', gcc_version + Exit(1) + main['GCC_VERSION'] = gcc_version - if not compareVersions(gcc_version, '4.4.1') or \ - not compareVersions(gcc_version, '4.4.2'): - print 'Info: Tree vectorizer in GCC 4.4.1 & 4.4.2 is buggy, disabling.' - main.Append(CCFLAGS=['-fno-tree-vectorize']) - if compareVersions(gcc_version, '4.6') >= 0: - main.Append(CXXFLAGS=['-std=c++0x']) -elif main['ICC']: - pass #Fix me... add warning flags once we clean up icc warnings -elif main['SUNCC']: - main.Append(CCFLAGS=['-Qoption ccfe']) - main.Append(CCFLAGS=['-features=gcc']) - main.Append(CCFLAGS=['-features=extensions']) - main.Append(CCFLAGS=['-library=stlport4']) - main.Append(CCFLAGS=['-xar']) - #main.Append(CCFLAGS=['-instances=semiexplicit']) + + if compareVersions(gcc_version, '4.9') >= 0: + # Incremental linking with LTO is currently broken in gcc versions + # 4.9 and above. A version where everything works completely hasn't + # yet been identified. + # + # https://gcc.gnu.org/bugzilla/show_bug.cgi?id=67548 + main['BROKEN_INCREMENTAL_LTO'] = True + if compareVersions(gcc_version, '6.0') >= 0: + # gcc versions 6.0 and greater accept an -flinker-output flag which + # selects what type of output the linker should generate. This is + # necessary for incremental lto to work, but is also broken in + # current versions of gcc. It may not be necessary in future + # versions. We add it here since it might be, and as a reminder that + # it exists. It's excluded if lto is being forced. + # + # https://gcc.gnu.org/gcc-6/changes.html + # https://gcc.gnu.org/ml/gcc-patches/2015-11/msg03161.html + # https://gcc.gnu.org/bugzilla/show_bug.cgi?id=69866 + if not GetOption('force_lto'): + main.Append(PSHLINKFLAGS='-flinker-output=rel') + main.Append(PLINKFLAGS='-flinker-output=rel') + + # gcc from version 4.8 and above generates "rep; ret" instructions + # to avoid performance penalties on certain AMD chips. Older + # assemblers detect this as an error, "Error: expecting string + # instruction after `rep'" + as_version_raw = readCommand([main['AS'], '-v', '/dev/null', + '-o', '/dev/null'], + exception=False).split() + + # version strings may contain extra distro-specific + # qualifiers, so play it safe and keep only what comes before + # the first hyphen + as_version = as_version_raw[-1].split('-')[0] if as_version_raw else None + + if not as_version or compareVersions(as_version, "2.23") < 0: + print termcap.Yellow + termcap.Bold + \ + 'Warning: This combination of gcc and binutils have' + \ + ' known incompatibilities.\n' + \ + ' If you encounter build problems, please update ' + \ + 'binutils to 2.23.' + \ + termcap.Normal + + # Make sure we warn if the user has requested to compile with the + # Undefined Benahvior Sanitizer and this version of gcc does not + # support it. + if GetOption('with_ubsan') and \ + compareVersions(gcc_version, '4.9') < 0: + print termcap.Yellow + termcap.Bold + \ + 'Warning: UBSan is only supported using gcc 4.9 and later.' + \ + termcap.Normal + + disable_lto = GetOption('no_lto') + if not disable_lto and main.get('BROKEN_INCREMENTAL_LTO', False) and \ + not GetOption('force_lto'): + print termcap.Yellow + termcap.Bold + \ + 'Warning: Your compiler doesn\'t support incremental linking' + \ + ' and lto at the same time, so lto is being disabled. To force' + \ + ' lto on anyway, use the --force-lto option. That will disable' + \ + ' partial linking.' + \ + termcap.Normal + disable_lto = True + + # Add the appropriate Link-Time Optimization (LTO) flags + # unless LTO is explicitly turned off. Note that these flags + # are only used by the fast target. + if not disable_lto: + # Pass the LTO flag when compiling to produce GIMPLE + # output, we merely create the flags here and only append + # them later + main['LTO_CCFLAGS'] = ['-flto=%d' % GetOption('num_jobs')] + + # Use the same amount of jobs for LTO as we are running + # scons with + main['LTO_LDFLAGS'] = ['-flto=%d' % GetOption('num_jobs')] + + main.Append(TCMALLOC_CCFLAGS=['-fno-builtin-malloc', '-fno-builtin-calloc', + '-fno-builtin-realloc', '-fno-builtin-free']) + + # add option to check for undeclared overrides + if compareVersions(gcc_version, "5.0") > 0: + main.Append(CCFLAGS=['-Wno-error=suggest-override']) + elif main['CLANG']: + # Check for a supported version of clang, >= 3.1 is needed to + # support similar features as gcc 4.8. See + # http://clang.llvm.org/cxx_status.html for details clang_version_re = re.compile(".* version (\d+\.\d+)") - clang_version_match = clang_version_re.match(CXX_version) + clang_version_match = clang_version_re.search(CXX_version) if (clang_version_match): clang_version = clang_version_match.groups()[0] - if compareVersions(clang_version, "2.9") < 0: - print 'Error: clang version 2.9 or newer required.' + if compareVersions(clang_version, "3.1") < 0: + print 'Error: clang version 3.1 or newer required.' print ' Installed version:', clang_version Exit(1) else: print 'Error: Unable to determine clang version.' Exit(1) - main.Append(CCFLAGS=['-pipe']) - main.Append(CCFLAGS=['-fno-strict-aliasing']) - main.Append(CCFLAGS=['-Wall', '-Wno-sign-compare', '-Wundef']) - main.Append(CCFLAGS=['-Wno-tautological-compare']) - main.Append(CCFLAGS=['-Wno-self-assign']) - # Ruby makes frequent use of extraneous parantheses in the printing - # of if-statements - main.Append(CCFLAGS=['-Wno-parentheses']) - - if compareVersions(clang_version, "3") >= 0: - main.Append(CXXFLAGS=['-std=c++0x']) + # clang has a few additional warnings that we disable, extraneous + # parantheses are allowed due to Ruby's printing of the AST, + # finally self assignments are allowed as the generated CPU code + # is relying on this + main.Append(CCFLAGS=['-Wno-parentheses', + '-Wno-self-assign', + # Some versions of libstdc++ (4.8?) seem to + # use struct hash and class hash + # interchangeably. + '-Wno-mismatched-tags', + ]) + + main.Append(TCMALLOC_CCFLAGS=['-fno-builtin']) + + # On Mac OS X/Darwin we need to also use libc++ (part of XCode) as + # opposed to libstdc++, as the later is dated. + if sys.platform == "darwin": + main.Append(CXXFLAGS=['-stdlib=libc++']) + main.Append(LIBS=['c++']) + + # On FreeBSD we need libthr. + if sys.platform.startswith('freebsd'): + main.Append(LIBS=['thr']) + else: print termcap.Yellow + termcap.Bold + 'Error' + termcap.Normal, print "Don't know what compiler options to use for your compiler." @@ -547,7 +859,7 @@ else: termcap.Normal else: print CXX_version.replace('\n', '') - print " If you're trying to use a compiler other than GCC, ICC, SunCC," + print " If you're trying to use a compiler other than GCC" print " or clang, there appears to be something wrong with your" print " environment." print " " @@ -573,74 +885,77 @@ if sys.platform == 'cygwin': # cygwin has some header file issues... main.Append(CCFLAGS=["-Wno-uninitialized"]) -# Check for SWIG -if not main.has_key('SWIG'): - print 'Error: SWIG utility not found.' - print ' Please install (see http://www.swig.org) and retry.' - Exit(1) - -# Check for appropriate SWIG version -swig_version = readCommand([main['SWIG'], '-version'], exception='').split() -# First 3 words should be "SWIG Version x.y.z" -if len(swig_version) < 3 or \ - swig_version[0] != 'SWIG' or swig_version[1] != 'Version': - print 'Error determining SWIG version.' - Exit(1) - -min_swig_version = '1.3.34' -if compareVersions(swig_version[2], min_swig_version) < 0: - print 'Error: SWIG version', min_swig_version, 'or newer required.' - print ' Installed version:', swig_version[2] - Exit(1) - -# Set up SWIG flags & scanner -swig_flags=Split('-c++ -python -modern -templatereduce $_CPPINCFLAGS') -main.Append(SWIGFLAGS=swig_flags) +# Check for the protobuf compiler +protoc_version = readCommand([main['PROTOC'], '--version'], + exception='').split() -# filter out all existing swig scanners, they mess up the dependency -# stuff for some reason -scanners = [] -for scanner in main['SCANNERS']: - skeys = scanner.skeys - if skeys == '.i': - continue - - if isinstance(skeys, (list, tuple)) and '.i' in skeys: - continue - - scanners.append(scanner) - -# add the new swig scanner that we like better -from SCons.Scanner import ClassicCPP as CPPScanner -swig_inc_re = '^[ \t]*[%,#][ \t]*(?:include|import)[ \t]*(<|")([^>"]+)(>|")' -scanners.append(CPPScanner("SwigScan", [ ".i" ], "CPPPATH", swig_inc_re)) - -# replace the scanners list that has what we want -main['SCANNERS'] = scanners - -# Add a custom Check function to the Configure context so that we can -# figure out if the compiler adds leading underscores to global -# variables. This is needed for the autogenerated asm files that we -# use for embedding the python code. -def CheckLeading(context): - context.Message("Checking for leading underscore in global variables...") - # 1) Define a global variable called x from asm so the C compiler - # won't change the symbol at all. - # 2) Declare that variable. - # 3) Use the variable - # - # If the compiler prepends an underscore, this will successfully - # link because the external symbol 'x' will be called '_x' which - # was defined by the asm statement. If the compiler does not - # prepend an underscore, this will not successfully link because - # '_x' will have been defined by assembly, while the C portion of - # the code will be trying to use 'x' - ret = context.TryLink(''' - asm(".globl _x; _x: .byte 0"); - extern int x; - int main() { return x; } - ''', extension=".c") - context.env.Append(LEADING_UNDERSCORE=ret) +# First two words should be "libprotoc x.y.z" +if len(protoc_version) < 2 or protoc_version[0] != 'libprotoc': + print termcap.Yellow + termcap.Bold + \ + 'Warning: Protocol buffer compiler (protoc) not found.\n' + \ + ' Please install protobuf-compiler for tracing support.' + \ + termcap.Normal + main['PROTOC'] = False +else: + # Based on the availability of the compress stream wrappers, + # require 2.1.0 + min_protoc_version = '2.1.0' + if compareVersions(protoc_version[1], min_protoc_version) < 0: + print termcap.Yellow + termcap.Bold + \ + 'Warning: protoc version', min_protoc_version, \ + 'or newer required.\n' + \ + ' Installed version:', protoc_version[1], \ + termcap.Normal + main['PROTOC'] = False + else: + # Attempt to determine the appropriate include path and + # library path using pkg-config, that means we also need to + # check for pkg-config. Note that it is possible to use + # protobuf without the involvement of pkg-config. Later on we + # check go a library config check and at that point the test + # will fail if libprotobuf cannot be found. + if readCommand(['pkg-config', '--version'], exception=''): + try: + # Attempt to establish what linking flags to add for protobuf + # using pkg-config + main.ParseConfig('pkg-config --cflags --libs-only-L protobuf') + except: + print termcap.Yellow + termcap.Bold + \ + 'Warning: pkg-config could not get protobuf flags.' + \ + termcap.Normal + + +# Check for 'timeout' from GNU coreutils. If present, regressions will +# be run with a time limit. We require version 8.13 since we rely on +# support for the '--foreground' option. +if sys.platform.startswith('freebsd'): + timeout_lines = readCommand(['gtimeout', '--version'], + exception='').splitlines() +else: + timeout_lines = readCommand(['timeout', '--version'], + exception='').splitlines() +# Get the first line and tokenize it +timeout_version = timeout_lines[0].split() if timeout_lines else [] +main['TIMEOUT'] = timeout_version and \ + compareVersions(timeout_version[-1], '8.13') >= 0 + +# Add a custom Check function to test for structure members. +def CheckMember(context, include, decl, member, include_quotes="<>"): + context.Message("Checking for member %s in %s..." % + (member, decl)) + text = """ +#include %(header)s +int main(){ + %(decl)s test; + (void)test.%(member)s; + return 0; +}; +""" % { "header" : include_quotes[0] + include + include_quotes[1], + "decl" : decl, + "member" : member, + } + + ret = context.TryCompile(text, extension=".cc") context.Result(ret) return ret @@ -649,11 +964,9 @@ def CheckLeading(context): conf = Configure(main, conf_dir = joinpath(build_root, '.scons_config'), log_file = joinpath(build_root, 'scons_config.log'), - custom_tests = { 'CheckLeading' : CheckLeading }) - -# Check for leading underscores. Don't really need to worry either -# way so don't need to check the return code. -conf.CheckLeading() + custom_tests = { + 'CheckMember' : CheckMember, + }) # Check if we should compile a 64 bit binary on Mac OS X/Darwin try: @@ -687,61 +1000,57 @@ if not conf: conf = NullConf(main) -# Find Python include and library directories for embedding the -# interpreter. For consistency, we will use the same Python -# installation used to run scons (and thus this script). If you want -# to link in an alternate version, see above for instructions on how -# to invoke scons with a different copy of the Python interpreter. -from distutils import sysconfig - -py_getvar = sysconfig.get_config_var - -py_debug = getattr(sys, 'pydebug', False) -py_version = 'python' + py_getvar('VERSION') + (py_debug and "_d" or "") - -py_general_include = sysconfig.get_python_inc() -py_platform_include = sysconfig.get_python_inc(plat_specific=True) -py_includes = [ py_general_include ] -if py_platform_include != py_general_include: - py_includes.append(py_platform_include) - -py_lib_path = [ py_getvar('LIBDIR') ] -# add the prefix/lib/pythonX.Y/config dir, but only if there is no -# shared library in prefix/lib/. -if not py_getvar('Py_ENABLE_SHARED'): - py_lib_path.append(py_getvar('LIBPL')) - -py_libs = [] -for lib in py_getvar('LIBS').split() + py_getvar('SYSLIBS').split(): - if not lib.startswith('-l'): - # Python requires some special flags to link (e.g. -framework - # common on OS X systems), assume appending preserves order - main.Append(LINKFLAGS=[lib]) - else: - lib = lib[2:] - if lib not in py_libs: - py_libs.append(lib) -py_libs.append(py_version) - -main.Append(CPPPATH=py_includes) -main.Append(LIBPATH=py_lib_path) - # Cache build files in the supplied directory. if main['M5_BUILD_CACHE']: print 'Using build cache located at', main['M5_BUILD_CACHE'] CacheDir(main['M5_BUILD_CACHE']) - -# verify that this stuff works -if not conf.CheckHeader('Python.h', '<>'): - print "Error: can't find Python.h header in", py_includes - Exit(1) - -for lib in py_libs: - if not conf.CheckLib(lib): - print "Error: can't find library %s required by python" % lib +main['USE_PYTHON'] = not GetOption('without_python') +if main['USE_PYTHON']: + # Find Python include and library directories for embedding the + # interpreter. We rely on python-config to resolve the appropriate + # includes and linker flags. ParseConfig does not seem to understand + # the more exotic linker flags such as -Xlinker and -export-dynamic so + # we add them explicitly below. If you want to link in an alternate + # version of python, see above for instructions on how to invoke + # scons with the appropriate PATH set. + # + # First we check if python2-config exists, else we use python-config + python_config = readCommand(['which', 'python2-config'], + exception='').strip() + if not os.path.exists(python_config): + python_config = readCommand(['which', 'python-config'], + exception='').strip() + py_includes = readCommand([python_config, '--includes'], + exception='').split() + # Strip the -I from the include folders before adding them to the + # CPPPATH + main.Append(CPPPATH=map(lambda inc: inc[2:], py_includes)) + + # Read the linker flags and split them into libraries and other link + # flags. The libraries are added later through the call the CheckLib. + py_ld_flags = readCommand([python_config, '--ldflags'], + exception='').split() + py_libs = [] + for lib in py_ld_flags: + if not lib.startswith('-l'): + main.Append(LINKFLAGS=[lib]) + else: + lib = lib[2:] + if lib not in py_libs: + py_libs.append(lib) + + # verify that this stuff works + if not conf.CheckHeader('Python.h', '<>'): + print "Error: can't find Python.h header in", py_includes + print "Install Python headers (package python-dev on Ubuntu and RedHat)" Exit(1) + for lib in py_libs: + if not conf.CheckLib(lib): + print "Error: can't find library %s required by python" % lib + Exit(1) + # On Solaris you need to use libsocket for socket ops if not conf.CheckLibWithHeader(None, 'sys/socket.h', 'C++', 'accept(0,0,0);'): if not conf.CheckLibWithHeader('socket', 'sys/socket.h', 'C++', 'accept(0,0,0);'): @@ -756,6 +1065,22 @@ if not conf.CheckLibWithHeader('z', 'zlib.h', 'C++','zlibVersion();'): print ' Please install zlib and try again.' Exit(1) +# If we have the protobuf compiler, also make sure we have the +# development libraries. If the check passes, libprotobuf will be +# automatically added to the LIBS environment variable. After +# this, we can use the HAVE_PROTOBUF flag to determine if we have +# got both protoc and libprotobuf available. +main['HAVE_PROTOBUF'] = main['PROTOC'] and \ + conf.CheckLibWithHeader('protobuf', 'google/protobuf/message.h', + 'C++', 'GOOGLE_PROTOBUF_VERIFY_VERSION;') + +# If we have the compiler but not the library, print another warning. +if main['PROTOC'] and not main['HAVE_PROTOBUF']: + print termcap.Yellow + termcap.Bold + \ + 'Warning: did not find protocol buffer library and/or headers.\n' + \ + ' Please install libprotobuf-dev for tracing support.' + \ + termcap.Normal + # Check for librt. have_posix_clock = \ conf.CheckLibWithHeader(None, 'time.h', 'C', @@ -763,14 +1088,40 @@ have_posix_clock = \ conf.CheckLibWithHeader('rt', 'time.h', 'C', 'clock_nanosleep(0,0,NULL,NULL);') -if conf.CheckLib('tcmalloc_minimal'): - have_tcmalloc = True -else: - have_tcmalloc = False +have_posix_timers = \ + conf.CheckLibWithHeader([None, 'rt'], [ 'time.h', 'signal.h' ], 'C', + 'timer_create(CLOCK_MONOTONIC, NULL, NULL);') + +if not GetOption('without_tcmalloc'): + if conf.CheckLib('tcmalloc'): + main.Append(CCFLAGS=main['TCMALLOC_CCFLAGS']) + elif conf.CheckLib('tcmalloc_minimal'): + main.Append(CCFLAGS=main['TCMALLOC_CCFLAGS']) + else: + print termcap.Yellow + termcap.Bold + \ + "You can get a 12% performance improvement by "\ + "installing tcmalloc (libgoogle-perftools-dev package "\ + "on Ubuntu or RedHat)." + termcap.Normal + + +# Detect back trace implementations. The last implementation in the +# list will be used by default. +backtrace_impls = [ "none" ] + +if conf.CheckLibWithHeader(None, 'execinfo.h', 'C', + 'backtrace_symbols_fd((void*)0, 0, 0);'): + backtrace_impls.append("glibc") +elif conf.CheckLibWithHeader('execinfo', 'execinfo.h', 'C', + 'backtrace_symbols_fd((void*)0, 0, 0);'): + # NetBSD and FreeBSD need libexecinfo. + backtrace_impls.append("glibc") + main.Append(LIBS=['execinfo']) + +if backtrace_impls[-1] == "none": + default_backtrace_impl = "none" print termcap.Yellow + termcap.Bold + \ - "You can get a 12% performance improvement by installing tcmalloc "\ - "(libgoogle-perftools-dev package on Ubuntu or RedHat)." + \ - termcap.Normal + "No suitable back trace implementation found." + \ + termcap.Normal if not have_posix_clock: print "Can't find library for POSIX clocks." @@ -781,6 +1132,61 @@ if not have_fenv: print "Warning: Header file not found." print " This host has no IEEE FP rounding mode control." +# Check if we should enable KVM-based hardware virtualization. The API +# we rely on exists since version 2.6.36 of the kernel, but somehow +# the KVM_API_VERSION does not reflect the change. We test for one of +# the types as a fall back. +have_kvm = conf.CheckHeader('linux/kvm.h', '<>') +if not have_kvm: + print "Info: Compatible header file not found, " \ + "disabling KVM support." + +# Check if the TUN/TAP driver is available. +have_tuntap = conf.CheckHeader('linux/if_tun.h', '<>') +if not have_tuntap: + print "Info: Compatible header file not found." + +# x86 needs support for xsave. We test for the structure here since we +# won't be able to run new tests by the time we know which ISA we're +# targeting. +have_kvm_xsave = conf.CheckTypeSize('struct kvm_xsave', + '#include ') != 0 + +# Check if the requested target ISA is compatible with the host +def is_isa_kvm_compatible(isa): + try: + import platform + host_isa = platform.machine() + except: + print "Warning: Failed to determine host ISA." + return False + + if not have_posix_timers: + print "Warning: Can not enable KVM, host seems to lack support " \ + "for POSIX timers" + return False + + if isa == "arm": + return host_isa in ( "armv7l", "aarch64" ) + elif isa == "x86": + if host_isa != "x86_64": + return False + + if not have_kvm_xsave: + print "KVM on x86 requires xsave support in kernel headers." + return False + + return True + else: + return False + + +# Check if the exclude_host attribute is available. We want this to +# get accurate instruction counts in KVM. +main['HAVE_PERF_ATTR_EXCLUDE_HOST'] = conf.CheckMember( + 'linux/perf_event.h', 'struct perf_event_attr', 'exclude_host') + + ###################################################################### # # Finish the configuration @@ -794,7 +1200,9 @@ main = conf.Finish() # Define the universe of supported ISAs all_isa_list = [ ] +all_gpu_isa_list = [ ] Export('all_isa_list') +Export('all_gpu_isa_list') class CpuModel(object): '''The CpuModel class encapsulates everything the ISA parser needs to @@ -802,17 +1210,10 @@ class CpuModel(object): # Dict of available CPU model objects. Accessible as CpuModel.dict. dict = {} - list = [] - defaults = [] # Constructor. Automatically adds models to CpuModel.dict. - def __init__(self, name, filename, includes, strings, default=False): + def __init__(self, name, default=False): self.name = name # name of model - self.filename = filename # filename for output exec code - self.includes = includes # include files needed in exec file - # The 'strings' dict holds all the per-CPU symbols we can - # substitute into templates etc. - self.strings = strings # This cpu is enabled by default self.default = default @@ -821,7 +1222,6 @@ class CpuModel(object): if name in CpuModel.dict: raise AttributeError, "CpuModel '%s' already registered" % name CpuModel.dict[name] = self - CpuModel.list.append(name) Export('CpuModel') @@ -835,9 +1235,17 @@ Export('sticky_vars') export_vars = [] Export('export_vars') +# For Ruby +all_protocols = [] +Export('all_protocols') +protocol_dirs = [] +Export('protocol_dirs') +slicc_includes = [] +Export('slicc_includes') + # Walk the tree and execute all SConsopts scripts that wil add to the # above variables -if not GetOption('verbose'): +if GetOption('verbose'): print "Reading SConsopts" for bdir in [ base_dir ] + extras_dir_list: if not isdir(bdir): @@ -850,12 +1258,14 @@ for bdir in [ base_dir ] + extras_dir_list: SConscript(joinpath(root, 'SConsopts')) all_isa_list.sort() +all_gpu_isa_list.sort() sticky_vars.AddVariables( EnumVariable('TARGET_ISA', 'Target ISA', 'alpha', all_isa_list), + EnumVariable('TARGET_GPU_ISA', 'Target GPU ISA', 'hsail', all_gpu_isa_list), ListVariable('CPU_MODELS', 'CPU models', sorted(n for n,m in CpuModel.dict.iteritems() if m.default), - sorted(CpuModel.list)), + sorted(CpuModel.dict.keys())), BoolVariable('EFENCE', 'Link with Electric Fence malloc debugger', False), BoolVariable('SS_COMPATIBLE_FP', @@ -867,11 +1277,21 @@ sticky_vars.AddVariables( BoolVariable('USE_POSIX_CLOCK', 'Use POSIX Clocks', have_posix_clock), BoolVariable('USE_FENV', 'Use IEEE mode control', have_fenv), BoolVariable('CP_ANNOTATE', 'Enable critical path annotation capability', False), + BoolVariable('USE_KVM', 'Enable hardware virtualized (KVM) CPU models', have_kvm), + BoolVariable('USE_TUNTAP', + 'Enable using a tap device to bridge to the host network', + have_tuntap), + BoolVariable('BUILD_GPU', 'Build the compute-GPU model', False), + EnumVariable('PROTOCOL', 'Coherence protocol for Ruby', 'None', + all_protocols), + EnumVariable('BACKTRACE_IMPL', 'Post-mortem dump implementation', + backtrace_impls[-1], backtrace_impls) ) # These variables get exported to #defines in config/*.hh (see src/SConscript). -export_vars += ['USE_FENV', 'SS_COMPATIBLE_FP', - 'TARGET_ISA', 'CP_ANNOTATE', 'USE_POSIX_CLOCK' ] +export_vars += ['USE_FENV', 'SS_COMPATIBLE_FP', 'TARGET_ISA', 'TARGET_GPU_ISA', + 'CP_ANNOTATE', 'USE_POSIX_CLOCK', 'USE_KVM', 'USE_TUNTAP', + 'PROTOCOL', 'HAVE_PROTOBUF', 'HAVE_PERF_ATTR_EXCLUDE_HOST'] ################################################### # @@ -914,42 +1334,100 @@ config_builder = Builder(emitter = config_emitter, action = config_action) main.Append(BUILDERS = { 'ConfigFile' : config_builder }) -# libelf build is shared across all configs in the build root. -main.SConscript('ext/libelf/SConscript', - variant_dir = joinpath(build_root, 'libelf')) +################################################### +# +# Builders for static and shared partially linked object files. +# +################################################### -# gzstream build is shared across all configs in the build root. -main.SConscript('ext/gzstream/SConscript', - variant_dir = joinpath(build_root, 'gzstream')) +partial_static_builder = Builder(action=SCons.Defaults.LinkAction, + src_suffix='$OBJSUFFIX', + src_builder=['StaticObject', 'Object'], + LINKFLAGS='$PLINKFLAGS', + LIBS='') + +def partial_shared_emitter(target, source, env): + for tgt in target: + tgt.attributes.shared = 1 + return (target, source) +partial_shared_builder = Builder(action=SCons.Defaults.ShLinkAction, + emitter=partial_shared_emitter, + src_suffix='$SHOBJSUFFIX', + src_builder='SharedObject', + SHLINKFLAGS='$PSHLINKFLAGS', + LIBS='') + +main.Append(BUILDERS = { 'PartialShared' : partial_shared_builder, + 'PartialStatic' : partial_static_builder }) + +# builds in ext are shared across all configs in the build root. +ext_dir = abspath(joinpath(str(main.root), 'ext')) +ext_build_dirs = [] +for root, dirs, files in os.walk(ext_dir): + if 'SConscript' in files: + build_dir = os.path.relpath(root, ext_dir) + ext_build_dirs.append(build_dir) + main.SConscript(joinpath(root, 'SConscript'), + variant_dir=joinpath(build_root, build_dir)) + +main.Prepend(CPPPATH=Dir('ext/pybind11/include/')) ################################################### # -# This function is used to set up a directory with switching headers +# This builder and wrapper method are used to set up a directory with +# switching headers. Those are headers which are in a generic location and +# that include more specific headers from a directory chosen at build time +# based on the current build settings. # ################################################### -main['ALL_ISA_LIST'] = all_isa_list -def make_switching_dir(dname, switch_headers, env): - # Generate the header. target[0] is the full path of the output - # header to generate. 'source' is a dummy variable, since we get the - # list of ISAs from env['ALL_ISA_LIST']. - def gen_switch_hdr(target, source, env): - fname = str(target[0]) - f = open(fname, 'w') - isa = env['TARGET_ISA'].lower() - print >>f, '#include "%s/%s/%s"' % (dname, isa, basename(fname)) - f.close() - - # Build SCons Action object. 'varlist' specifies env vars that this - # action depends on; when env['ALL_ISA_LIST'] changes these actions - # should get re-executed. - switch_hdr_action = MakeAction(gen_switch_hdr, - Transform("GENERATE"), varlist=['ALL_ISA_LIST']) - - # Instantiate actions for each header - for hdr in switch_headers: - env.Command(hdr, [], switch_hdr_action) -Export('make_switching_dir') +def build_switching_header(target, source, env): + path = str(target[0]) + subdir = str(source[0]) + dp, fp = os.path.split(path) + dp = os.path.relpath(os.path.realpath(dp), + os.path.realpath(env['BUILDDIR'])) + with open(path, 'w') as hdr: + print >>hdr, '#include "%s/%s/%s"' % (dp, subdir, fp) + +switching_header_action = MakeAction(build_switching_header, + Transform('GENERATE')) + +switching_header_builder = Builder(action=switching_header_action, + source_factory=Value, + single_source=True) + +main.Append(BUILDERS = { 'SwitchingHeader': switching_header_builder }) + +def switching_headers(self, headers, source): + for header in headers: + self.SwitchingHeader(header, source) + +main.AddMethod(switching_headers, 'SwitchingHeaders') + +# all-isas -> all-deps -> all-environs -> all_targets +main.Alias('#all-isas', []) +main.Alias('#all-deps', '#all-isas') + +# Dummy target to ensure all environments are created before telling +# SCons what to actually make (the command line arguments). We attach +# them to the dependence graph after the environments are complete. +ORIG_BUILD_TARGETS = list(BUILD_TARGETS) # force a copy; gets closure to work. +def environsComplete(target, source, env): + for t in ORIG_BUILD_TARGETS: + main.Depends('#all-targets', t) + +# Each build/* switching_dir attaches its *-environs target to #all-environs. +main.Append(BUILDERS = {'CompleteEnvirons' : + Builder(action=MakeAction(environsComplete, None))}) +main.CompleteEnvirons('#all-environs', []) + +def doNothing(**ignored): pass +main.Append(BUILDERS = {'Dummy': Builder(action=MakeAction(doNothing, None))}) + +# The final target to which all the original targets ultimately get attached. +main.Dummy('#all-targets', '#all-environs') +BUILD_TARGETS[:] = ['#all-targets'] ################################################### # @@ -957,8 +1435,14 @@ Export('make_switching_dir') # ################################################### +def variant_name(path): + return os.path.basename(path).lower().replace('_', '-') +main['variant_name'] = variant_name +main['VARIANT_NAME'] = '${variant_name(BUILDDIR)}' + for variant_path in variant_paths: - print "Building in", variant_path + if not GetOption('silent'): + print "Building in", variant_path # Make a copy of the build-root environment to use for this config. env = main.Clone() @@ -976,7 +1460,11 @@ for variant_path in variant_paths: current_vars_file = joinpath(build_root, 'variables', variant_dir) if isfile(current_vars_file): sticky_vars.files.append(current_vars_file) - print "Using saved variables file %s" % current_vars_file + if not GetOption('silent'): + print "Using saved variables file %s" % current_vars_file + elif variant_dir in ext_build_dirs: + # Things in ext are built without a variant directory. + continue else: # Build dir-specific variables file doesn't exist. @@ -1028,26 +1516,60 @@ for variant_path in variant_paths: if env['EFENCE']: env.Append(LIBS=['efence']) + if env['USE_KVM']: + if not have_kvm: + print "Warning: Can not enable KVM, host seems to lack KVM support" + env['USE_KVM'] = False + elif not is_isa_kvm_compatible(env['TARGET_ISA']): + print "Info: KVM support disabled due to unsupported host and " \ + "target ISA combination" + env['USE_KVM'] = False + + if env['USE_TUNTAP']: + if not have_tuntap: + print "Warning: Can't connect EtherTap with a tap device." + env['USE_TUNTAP'] = False + + if env['BUILD_GPU']: + env.Append(CPPDEFINES=['BUILD_GPU']) + + # Warn about missing optional functionality + if env['USE_KVM']: + if not main['HAVE_PERF_ATTR_EXCLUDE_HOST']: + print "Warning: perf_event headers lack support for the " \ + "exclude_host attribute. KVM instruction counts will " \ + "be inaccurate." + # Save sticky variable settings back to current variables file sticky_vars.Save(current_vars_file, env) if env['USE_SSE2']: env.Append(CCFLAGS=['-msse2']) - if have_tcmalloc: - env.Append(LIBS=['tcmalloc_minimal']) - # The src/SConscript file sets up the build rules in 'env' according # to the configured variables. It returns a list of environments, # one for each variant build (debug, opt, etc.) - envList = SConscript('src/SConscript', variant_dir = variant_path, - exports = 'env') - - # Set up the regression tests for each build. - for e in envList: - SConscript('tests/SConscript', - variant_dir = joinpath(variant_path, 'tests', e.Label), - exports = { 'env' : e }, duplicate = False) + SConscript('src/SConscript', variant_dir = variant_path, exports = 'env') + +def pairwise(iterable): + "s -> (s0,s1), (s1,s2), (s2, s3), ..." + a, b = itertools.tee(iterable) + b.next() + return itertools.izip(a, b) + +variant_names = [variant_name(path) for path in variant_paths] + +# Create false dependencies so SCons will parse ISAs, establish +# dependencies, and setup the build Environments serially. Either +# SCons (likely) and/or our SConscripts (possibly) cannot cope with -j +# greater than 1. It appears to be standard race condition stuff; it +# doesn't always fail, but usually, and the behaviors are different. +# Every time I tried to remove this, builds would fail in some +# creative new way. So, don't do that. You'll want to, though, because +# tests/SConscript takes a long time to make its Environments. +for t1, t2 in pairwise(sorted(variant_names)): + main.Depends('#%s-deps' % t2, '#%s-deps' % t1) + main.Depends('#%s-environs' % t2, '#%s-environs' % t1) # base help text Help('''