# -*- mode:python -*-
+# Copyright (c) 2013, 2015, 2016 ARM Limited
+# All rights reserved.
+#
+# The license below extends only to copyright in the software and shall
+# not be construed as granting a license to any other intellectual
+# property including but not limited to intellectual property relating
+# to a hardware implementation of the functionality of the software
+# licensed hereunder. You may use the software subject to the license
+# terms below provided that you ensure that this notice is replicated
+# unmodified and in its entirety in all distributions of the software,
+# modified or unmodified, in source code or in binary form.
+#
+# Copyright (c) 2011 Advanced Micro Devices, Inc.
+# Copyright (c) 2009 The Hewlett-Packard Development Company
# Copyright (c) 2004-2005 The Regents of The University of Michigan
# All rights reserved.
#
# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
#
# Authors: Steve Reinhardt
+# Nathan Binkert
###################################################
#
# SCons top-level build description (SConstruct) file.
#
-# While in this directory ('m5'), just type 'scons' to build the default
+# While in this directory ('gem5'), just type 'scons' to build the default
# configuration (see below), or type 'scons build/<CONFIG>/<binary>'
-# to build some other configuration (e.g., 'build/ALPHA_FS/m5.opt' for
+# to build some other configuration (e.g., 'build/ALPHA/gem5.opt' for
# the optimized full-system version).
#
-# You can build M5 in a different directory as long as there is a
+# You can build gem5 in a different directory as long as there is a
# 'build/<CONFIG>' somewhere along the target path. The build system
# expects that all configs under the same build directory are being
# built for the same host system.
#
# The following two commands are equivalent. The '-u' option tells
# scons to search up the directory tree for this SConstruct file.
-# % cd <path-to-src>/m5 ; scons build/ALPHA_FS/m5.debug
-# % cd <path-to-src>/m5/build/ALPHA_FS; scons -u m5.debug
+# % cd <path-to-src>/gem5 ; scons build/ALPHA/gem5.debug
+# % cd <path-to-src>/gem5/build/ALPHA; scons -u gem5.debug
#
# The following two commands are equivalent and demonstrate building
# in a directory outside of the source tree. The '-C' option tells
# scons to chdir to the specified directory to find this SConstruct
# file.
-# % cd <path-to-src>/m5 ; scons /local/foo/build/ALPHA_FS/m5.debug
-# % cd /local/foo/build/ALPHA_FS; scons -C <path-to-src>/m5 m5.debug
+# % cd <path-to-src>/gem5 ; scons /local/foo/build/ALPHA/gem5.debug
+# % cd /local/foo/build/ALPHA; scons -C <path-to-src>/gem5 gem5.debug
#
# You can use 'scons -H' to print scons options. If you're in this
-# 'm5' directory (or use -u or -C to tell scons where to find this
-# file), you can use 'scons -h' to print all the M5-specific build
+# 'gem5' directory (or use -u or -C to tell scons where to find this
+# file), you can use 'scons -h' to print all the gem5-specific build
# options as well.
#
###################################################
-# Python library imports
-import sys
+# Check for recent-enough Python and SCons versions.
+try:
+ # Really old versions of scons only take two options for the
+ # function, so check once without the revision and once with the
+ # revision, the first instance will fail for stuff other than
+ # 0.98, and the second will fail for 0.98.0
+ EnsureSConsVersion(0, 98)
+ EnsureSConsVersion(0, 98, 1)
+except SystemExit, e:
+ print """
+For more details, see:
+ http://gem5.org/Dependencies
+"""
+ raise
+
+# We ensure the python version early because because python-config
+# requires python 2.5
+try:
+ EnsurePythonVersion(2, 5)
+except SystemExit, e:
+ print """
+You can use a non-default installation of the Python interpreter by
+rearranging your PATH so that scons finds the non-default 'python' and
+'python-config' first.
+
+For more details, see:
+ http://gem5.org/wiki/index.php/Using_a_non-default_Python_installation
+"""
+ raise
+
+# Global Python includes
+import itertools
import os
+import re
+import shutil
+import subprocess
+import sys
-# Check for recent-enough Python and SCons versions. If your system's
-# default installation of Python is not recent enough, you can use a
-# non-default installation of the Python interpreter by either (1)
-# rearranging your PATH so that scons finds the non-default 'python'
-# first or (2) explicitly invoking an alternative interpreter on the
-# scons script, e.g., "/usr/local/bin/python2.4 `which scons` [args]".
-EnsurePythonVersion(2,4)
-
-# Ironically, SCons 0.96 dies if you give EnsureSconsVersion a
-# 3-element version number.
-min_scons_version = (0,96,91)
-try:
- EnsureSConsVersion(*min_scons_version)
-except:
- print "Error checking current SCons version."
- print "SCons", ".".join(map(str,min_scons_version)), "or greater required."
- Exit(2)
-
+from os import mkdir, environ
+from os.path import abspath, basename, dirname, expanduser, normpath
+from os.path import exists, isdir, isfile
+from os.path import join as joinpath, split as splitpath
+
+# SCons includes
+import SCons
+import SCons.Node
+
+extra_python_paths = [
+ Dir('src/python').srcnode().abspath, # gem5 includes
+ Dir('ext/ply').srcnode().abspath, # ply is used by several files
+ ]
+
+sys.path[1:1] = extra_python_paths
+
+from m5.util import compareVersions, readCommand
+from m5.util.terminal import get_termcap
+
+help_texts = {
+ "options" : "",
+ "global_vars" : "",
+ "local_vars" : ""
+}
+
+Export("help_texts")
+
+
+# There's a bug in scons in that (1) by default, the help texts from
+# AddOption() are supposed to be displayed when you type 'scons -h'
+# and (2) you can override the help displayed by 'scons -h' using the
+# Help() function, but these two features are incompatible: once
+# you've overridden the help text using Help(), there's no way to get
+# at the help texts from AddOptions. See:
+# http://scons.tigris.org/issues/show_bug.cgi?id=2356
+# http://scons.tigris.org/issues/show_bug.cgi?id=2611
+# This hack lets us extract the help text from AddOptions and
+# re-inject it via Help(). Ideally someday this bug will be fixed and
+# we can just use AddOption directly.
+def AddLocalOption(*args, **kwargs):
+ col_width = 30
+
+ help = " " + ", ".join(args)
+ if "help" in kwargs:
+ length = len(help)
+ if length >= col_width:
+ help += "\n" + " " * col_width
+ else:
+ help += " " * (col_width - length)
+ help += kwargs["help"]
+ help_texts["options"] += help + "\n"
+
+ AddOption(*args, **kwargs)
+
+AddLocalOption('--colors', dest='use_colors', action='store_true',
+ help="Add color to abbreviated scons output")
+AddLocalOption('--no-colors', dest='use_colors', action='store_false',
+ help="Don't add color to abbreviated scons output")
+AddLocalOption('--with-cxx-config', dest='with_cxx_config',
+ action='store_true',
+ help="Build with support for C++-based configuration")
+AddLocalOption('--default', dest='default', type='string', action='store',
+ help='Override which build_opts file to use for defaults')
+AddLocalOption('--ignore-style', dest='ignore_style', action='store_true',
+ help='Disable style checking hooks')
+AddLocalOption('--no-lto', dest='no_lto', action='store_true',
+ help='Disable Link-Time Optimization for fast')
+AddLocalOption('--update-ref', dest='update_ref', action='store_true',
+ help='Update test reference outputs')
+AddLocalOption('--verbose', dest='verbose', action='store_true',
+ help='Print full tool command lines')
+AddLocalOption('--without-python', dest='without_python',
+ action='store_true',
+ help='Build without Python configuration support')
+AddLocalOption('--without-tcmalloc', dest='without_tcmalloc',
+ action='store_true',
+ help='Disable linking against tcmalloc')
+AddLocalOption('--with-ubsan', dest='with_ubsan', action='store_true',
+ help='Build with Undefined Behavior Sanitizer if available')
+AddLocalOption('--with-asan', dest='with_asan', action='store_true',
+ help='Build with Address Sanitizer if available')
+
+termcap = get_termcap(GetOption('use_colors'))
+
+########################################################################
+#
+# Set up the main build environment.
+#
+########################################################################
+
+# export TERM so that clang reports errors in color
+use_vars = set([ 'AS', 'AR', 'CC', 'CXX', 'HOME', 'LD_LIBRARY_PATH',
+ 'LIBRARY_PATH', 'PATH', 'PKG_CONFIG_PATH', 'PROTOC',
+ 'PYTHONPATH', 'RANLIB', 'SWIG', 'TERM' ])
+
+use_prefixes = [
+ "ASAN_", # address sanitizer symbolizer path and settings
+ "CCACHE_", # ccache (caching compiler wrapper) configuration
+ "CCC_", # clang static analyzer configuration
+ "DISTCC_", # distcc (distributed compiler wrapper) configuration
+ "INCLUDE_SERVER_", # distcc pump server settings
+ "M5", # M5 configuration (e.g., path to kernels)
+ ]
+
+use_env = {}
+for key,val in sorted(os.environ.iteritems()):
+ if key in use_vars or \
+ any([key.startswith(prefix) for prefix in use_prefixes]):
+ use_env[key] = val
+
+# Tell scons to avoid implicit command dependencies to avoid issues
+# with the param wrappes being compiled twice (see
+# http://scons.tigris.org/issues/show_bug.cgi?id=2811)
+main = Environment(ENV=use_env, IMPLICIT_COMMAND_DEPENDENCIES=0)
+main.Decider('MD5-timestamp')
+main.root = Dir(".") # The current directory (where this file lives).
+main.srcdir = Dir("src") # The source directory
+
+main_dict_keys = main.Dictionary().keys()
+
+# Check that we have a C/C++ compiler
+if not ('CC' in main_dict_keys and 'CXX' in main_dict_keys):
+ print "No C++ compiler installed (package g++ on Ubuntu and RedHat)"
+ Exit(1)
+
+# Check that swig is present
+if not 'SWIG' in main_dict_keys:
+ print "swig is not installed (package swig on Ubuntu and RedHat)"
+ Exit(1)
+
+# add useful python code PYTHONPATH so it can be used by subprocesses
+# as well
+main.AppendENVPath('PYTHONPATH', extra_python_paths)
+
+########################################################################
+#
+# Mercurial Stuff.
+#
+# If the gem5 directory is a mercurial repository, we should do some
+# extra things.
+#
+########################################################################
+
+hgdir = main.root.Dir(".hg")
+
+
+style_message = """
+You're missing the gem5 style hook, which automatically checks your code
+against the gem5 style rules on %s.
+This script will now install the hook in your %s.
+Press enter to continue, or ctrl-c to abort: """
+
+mercurial_style_message = style_message % ("hg commit and qrefresh commands",
+ ".hg/hgrc file")
+git_style_message = style_message % ("'git commit'",
+ ".git/hooks/ directory")
+
+mercurial_style_upgrade_message = """
+Your Mercurial style hooks are not up-to-date. This script will now
+try to automatically update them. A backup of your hgrc will be saved
+in .hg/hgrc.old.
+Press enter to continue, or ctrl-c to abort: """
+
+mercurial_style_hook = """
+# The following lines were automatically added by gem5/SConstruct
+# to provide the gem5 style-checking hooks
+[extensions]
+hgstyle = %s/util/hgstyle.py
+
+[hooks]
+pretxncommit.style = python:hgstyle.check_style
+pre-qrefresh.style = python:hgstyle.check_style
+# End of SConstruct additions
+
+""" % (main.root.abspath)
+
+mercurial_lib_not_found = """
+Mercurial libraries cannot be found, ignoring style hook. If
+you are a gem5 developer, please fix this and run the style
+hook. It is important.
+"""
+
+# Check for style hook and prompt for installation if it's not there.
+# Skip this if --ignore-style was specified, there's no interactive
+# terminal to prompt, or no recognized revision control system can be
+# found.
+ignore_style = GetOption('ignore_style') or not sys.stdin.isatty()
+
+# Try wire up Mercurial to the style hooks
+if not ignore_style and hgdir.exists():
+ style_hook = True
+ style_hooks = tuple()
+ hgrc = hgdir.File('hgrc')
+ hgrc_old = hgdir.File('hgrc.old')
+ try:
+ from mercurial import ui
+ ui = ui.ui()
+ ui.readconfig(hgrc.abspath)
+ style_hooks = (ui.config('hooks', 'pretxncommit.style', None),
+ ui.config('hooks', 'pre-qrefresh.style', None))
+ style_hook = all(style_hooks)
+ style_extension = ui.config('extensions', 'style', None)
+ except ImportError:
+ print mercurial_lib_not_found
+
+ if "python:style.check_style" in style_hooks:
+ # Try to upgrade the style hooks
+ print mercurial_style_upgrade_message
+ # continue unless user does ctrl-c/ctrl-d etc.
+ try:
+ raw_input()
+ except:
+ print "Input exception, exiting scons.\n"
+ sys.exit(1)
+ shutil.copyfile(hgrc.abspath, hgrc_old.abspath)
+ re_style_hook = re.compile(r"^([^=#]+)\.style\s*=\s*([^#\s]+).*")
+ re_style_extension = re.compile("style\s*=\s*([^#\s]+).*")
+ old, new = open(hgrc_old.abspath, 'r'), open(hgrc.abspath, 'w')
+ for l in old:
+ m_hook = re_style_hook.match(l)
+ m_ext = re_style_extension.match(l)
+ if m_hook:
+ hook, check = m_hook.groups()
+ if check != "python:style.check_style":
+ print "Warning: %s.style is using a non-default " \
+ "checker: %s" % (hook, check)
+ if hook not in ("pretxncommit", "pre-qrefresh"):
+ print "Warning: Updating unknown style hook: %s" % hook
+
+ l = "%s.style = python:hgstyle.check_style\n" % hook
+ elif m_ext and m_ext.group(1) == style_extension:
+ l = "hgstyle = %s/util/hgstyle.py\n" % main.root.abspath
+
+ new.write(l)
+ elif not style_hook:
+ print mercurial_style_message,
+ # continue unless user does ctrl-c/ctrl-d etc.
+ try:
+ raw_input()
+ except:
+ print "Input exception, exiting scons.\n"
+ sys.exit(1)
+ hgrc_path = '%s/.hg/hgrc' % main.root.abspath
+ print "Adding style hook to", hgrc_path, "\n"
+ try:
+ with open(hgrc_path, 'a') as f:
+ f.write(mercurial_style_hook)
+ except:
+ print "Error updating", hgrc_path
+ sys.exit(1)
+
+def install_git_style_hooks():
+ try:
+ gitdir = Dir(readCommand(
+ ["git", "rev-parse", "--git-dir"]).strip("\n"))
+ except Exception, e:
+ print "Warning: Failed to find git repo directory: %s" % e
+ return
+
+ git_hooks = gitdir.Dir("hooks")
+ git_pre_commit_hook = git_hooks.File("pre-commit")
+ git_style_script = File("util/git-pre-commit.py")
+
+ if git_pre_commit_hook.exists():
+ return
-# The absolute path to the current directory (where this file lives).
-ROOT = Dir('.').abspath
+ print git_style_message,
+ try:
+ raw_input()
+ except:
+ print "Input exception, exiting scons.\n"
+ sys.exit(1)
-# Paths to the M5 and external source trees.
-SRCDIR = os.path.join(ROOT, 'src')
+ if not git_hooks.exists():
+ mkdir(git_hooks.get_abspath())
-# tell python where to find m5 python code
-sys.path.append(os.path.join(ROOT, 'src/python'))
+ # Use a relative symlink if the hooks live in the source directory
+ if git_pre_commit_hook.is_under(main.root):
+ script_path = os.path.relpath(
+ git_style_script.get_abspath(),
+ git_pre_commit_hook.Dir(".").get_abspath())
+ else:
+ script_path = git_style_script.get_abspath()
+
+ try:
+ os.symlink(script_path, git_pre_commit_hook.get_abspath())
+ except:
+ print "Error updating git pre-commit hook"
+ raise
+
+# Try to wire up git to the style hooks
+if not ignore_style and main.root.Entry(".git").exists():
+ install_git_style_hooks()
###################################################
#
###################################################
# Find default configuration & binary.
-Default(os.environ.get('M5_DEFAULT_BINARY', 'build/ALPHA_SE/m5.debug'))
-
-# Ask SCons which directory it was invoked from.
-launch_dir = GetLaunchDir()
-
-# Make targets relative to invocation directory
-abs_targets = map(lambda x: os.path.normpath(os.path.join(launch_dir, str(x))),
- BUILD_TARGETS)
+Default(environ.get('M5_DEFAULT_BINARY', 'build/ALPHA/gem5.debug'))
# helper function: find last occurrence of element in list
def rfind(l, elt, offs = -1):
return i
raise ValueError, "element not found"
+# Take a list of paths (or SCons Nodes) and return a list with all
+# paths made absolute and ~-expanded. Paths will be interpreted
+# relative to the launch directory unless a different root is provided
+def makePathListAbsolute(path_list, root=GetLaunchDir()):
+ return [abspath(joinpath(root, expanduser(str(p))))
+ for p in path_list]
+
# Each target must have 'build' in the interior of the path; the
# directory below this will determine the build parameters. For
# example, for target 'foo/bar/build/ALPHA_SE/arch/alpha/blah.do' we
# recognize that ALPHA_SE specifies the configuration because it
-# follow 'build' in the bulid path.
+# follow 'build' in the build path.
+
+# The funky assignment to "[:]" is needed to replace the list contents
+# in place rather than reassign the symbol to a new list, which
+# doesn't work (obviously!).
+BUILD_TARGETS[:] = makePathListAbsolute(BUILD_TARGETS)
# Generate a list of the unique build roots and configs that the
# collected targets reference.
-build_paths = []
+variant_paths = []
build_root = None
-for t in abs_targets:
+for t in BUILD_TARGETS:
path_dirs = t.split('/')
try:
build_top = rfind(path_dirs, 'build', -2)
except:
print "Error: no non-leaf 'build' dir found on target path", t
Exit(1)
- this_build_root = os.path.join('/',*path_dirs[:build_top+1])
+ this_build_root = joinpath('/',*path_dirs[:build_top+1])
if not build_root:
build_root = this_build_root
else:
print "Error: build targets not under same build root\n"\
" %s\n %s" % (build_root, this_build_root)
Exit(1)
- build_path = os.path.join('/',*path_dirs[:build_top+2])
- if build_path not in build_paths:
- build_paths.append(build_path)
+ variant_path = joinpath('/',*path_dirs[:build_top+2])
+ if variant_path not in variant_paths:
+ variant_paths.append(variant_path)
-###################################################
-#
-# Set up the default build environment. This environment is copied
-# and modified according to each selected configuration.
-#
-###################################################
+# Make sure build_root exists (might not if this is the first build there)
+if not isdir(build_root):
+ mkdir(build_root)
+main['BUILDROOT'] = build_root
-env = Environment(ENV = os.environ, # inherit user's environment vars
- ROOT = ROOT,
- SRCDIR = SRCDIR)
+Export('main')
-env.SConsignFile(os.path.join(build_root,"sconsign"))
+main.SConsignFile(joinpath(build_root, "sconsign"))
# Default duplicate option is to use hard links, but this messes up
# when you use emacs to edit a file in the target dir, as emacs moves
# file to file~ then copies to file, breaking the link. Symbolic
# (soft) links work better.
-env.SetOption('duplicate', 'soft-copy')
+main.SetOption('duplicate', 'soft-copy')
-# I waffle on this setting... it does avoid a few painful but
-# unnecessary builds, but it also seems to make trivial builds take
-# noticeably longer.
-if False:
- env.TargetSignatures('content')
+#
+# Set up global sticky variables... these are common to an entire build
+# tree (not specific to a particular build like ALPHA_SE)
+#
-# M5_PLY is used by isa_parser.py to find the PLY package.
-env.Append(ENV = { 'M5_PLY' : Dir('ext/ply') })
+global_vars_file = joinpath(build_root, 'variables.global')
+
+global_vars = Variables(global_vars_file, args=ARGUMENTS)
+
+global_vars.AddVariables(
+ ('CC', 'C compiler', environ.get('CC', main['CC'])),
+ ('CXX', 'C++ compiler', environ.get('CXX', main['CXX'])),
+ ('SWIG', 'SWIG tool', environ.get('SWIG', main['SWIG'])),
+ ('PROTOC', 'protoc tool', environ.get('PROTOC', 'protoc')),
+ ('BATCH', 'Use batch pool for build and tests', False),
+ ('BATCH_CMD', 'Batch pool submission command name', 'qdo'),
+ ('M5_BUILD_CACHE', 'Cache built objects in this directory', False),
+ ('EXTRAS', 'Add extra directories to the compilation', '')
+ )
+
+# Update main environment with values from ARGUMENTS & global_vars_file
+global_vars.Update(main)
+help_texts["global_vars"] += global_vars.GenerateHelpText(main)
+
+# Save sticky variable settings back to current variables file
+global_vars.Save(global_vars_file, main)
+
+# Parse EXTRAS variable to build list of all directories where we're
+# look for sources etc. This list is exported as extras_dir_list.
+base_dir = main.srcdir.abspath
+if main['EXTRAS']:
+ extras_dir_list = makePathListAbsolute(main['EXTRAS'].split(':'))
+else:
+ extras_dir_list = []
+
+Export('base_dir')
+Export('extras_dir_list')
+
+# the ext directory should be on the #includes path
+main.Append(CPPPATH=[Dir('ext')])
+
+def strip_build_path(path, env):
+ path = str(path)
+ variant_base = env['BUILDROOT'] + os.path.sep
+ if path.startswith(variant_base):
+ path = path[len(variant_base):]
+ elif path.startswith('build/'):
+ path = path[6:]
+ return path
+
+# Generate a string of the form:
+# common/path/prefix/src1, src2 -> tgt1, tgt2
+# to print while building.
+class Transform(object):
+ # all specific color settings should be here and nowhere else
+ tool_color = termcap.Normal
+ pfx_color = termcap.Yellow
+ srcs_color = termcap.Yellow + termcap.Bold
+ arrow_color = termcap.Blue + termcap.Bold
+ tgts_color = termcap.Yellow + termcap.Bold
+
+ def __init__(self, tool, max_sources=99):
+ self.format = self.tool_color + (" [%8s] " % tool) \
+ + self.pfx_color + "%s" \
+ + self.srcs_color + "%s" \
+ + self.arrow_color + " -> " \
+ + self.tgts_color + "%s" \
+ + termcap.Normal
+ self.max_sources = max_sources
+
+ def __call__(self, target, source, env, for_signature=None):
+ # truncate source list according to max_sources param
+ source = source[0:self.max_sources]
+ def strip(f):
+ return strip_build_path(str(f), env)
+ if len(source) > 0:
+ srcs = map(strip, source)
+ else:
+ srcs = ['']
+ tgts = map(strip, target)
+ # surprisingly, os.path.commonprefix is a dumb char-by-char string
+ # operation that has nothing to do with paths.
+ com_pfx = os.path.commonprefix(srcs + tgts)
+ com_pfx_len = len(com_pfx)
+ if com_pfx:
+ # do some cleanup and sanity checking on common prefix
+ if com_pfx[-1] == ".":
+ # prefix matches all but file extension: ok
+ # back up one to change 'foo.cc -> o' to 'foo.cc -> .o'
+ com_pfx = com_pfx[0:-1]
+ elif com_pfx[-1] == "/":
+ # common prefix is directory path: OK
+ pass
+ else:
+ src0_len = len(srcs[0])
+ tgt0_len = len(tgts[0])
+ if src0_len == com_pfx_len:
+ # source is a substring of target, OK
+ pass
+ elif tgt0_len == com_pfx_len:
+ # target is a substring of source, need to back up to
+ # avoid empty string on RHS of arrow
+ sep_idx = com_pfx.rfind(".")
+ if sep_idx != -1:
+ com_pfx = com_pfx[0:sep_idx]
+ else:
+ com_pfx = ''
+ elif src0_len > com_pfx_len and srcs[0][com_pfx_len] == ".":
+ # still splitting at file extension: ok
+ pass
+ else:
+ # probably a fluke; ignore it
+ com_pfx = ''
+ # recalculate length in case com_pfx was modified
+ com_pfx_len = len(com_pfx)
+ def fmt(files):
+ f = map(lambda s: s[com_pfx_len:], files)
+ return ', '.join(f)
+ return self.format % (com_pfx, fmt(srcs), fmt(tgts))
+
+Export('Transform')
+
+# enable the regression script to use the termcap
+main['TERMCAP'] = termcap
+
+if GetOption('verbose'):
+ def MakeAction(action, string, *args, **kwargs):
+ return Action(action, *args, **kwargs)
+else:
+ MakeAction = Action
+ main['CCCOMSTR'] = Transform("CC")
+ main['CXXCOMSTR'] = Transform("CXX")
+ main['ASCOMSTR'] = Transform("AS")
+ main['SWIGCOMSTR'] = Transform("SWIG")
+ main['ARCOMSTR'] = Transform("AR", 0)
+ main['LINKCOMSTR'] = Transform("LINK", 0)
+ main['RANLIBCOMSTR'] = Transform("RANLIB", 0)
+ main['M4COMSTR'] = Transform("M4")
+ main['SHCCCOMSTR'] = Transform("SHCC")
+ main['SHCXXCOMSTR'] = Transform("SHCXX")
+Export('MakeAction')
+
+# Initialize the Link-Time Optimization (LTO) flags
+main['LTO_CCFLAGS'] = []
+main['LTO_LDFLAGS'] = []
+
+# According to the readme, tcmalloc works best if the compiler doesn't
+# assume that we're using the builtin malloc and friends. These flags
+# are compiler-specific, so we need to set them after we detect which
+# compiler we're using.
+main['TCMALLOC_CCFLAGS'] = []
+
+CXX_version = readCommand([main['CXX'],'--version'], exception=False)
+CXX_V = readCommand([main['CXX'],'-V'], exception=False)
+
+main['GCC'] = CXX_version and CXX_version.find('g++') >= 0
+main['CLANG'] = CXX_version and CXX_version.find('clang') >= 0
+if main['GCC'] + main['CLANG'] > 1:
+ print 'Error: How can we have two at the same time?'
+ Exit(1)
# Set up default C++ compiler flags
-env.Append(CCFLAGS='-pipe')
-env.Append(CCFLAGS='-fno-strict-aliasing')
-env.Append(CCFLAGS=Split('-Wall -Wno-sign-compare -Werror -Wundef'))
+if main['GCC'] or main['CLANG']:
+ # As gcc and clang share many flags, do the common parts here
+ main.Append(CCFLAGS=['-pipe'])
+ main.Append(CCFLAGS=['-fno-strict-aliasing'])
+ # Enable -Wall and -Wextra and then disable the few warnings that
+ # we consistently violate
+ main.Append(CCFLAGS=['-Wall', '-Wundef', '-Wextra',
+ '-Wno-sign-compare', '-Wno-unused-parameter'])
+ # We always compile using C++11
+ main.Append(CXXFLAGS=['-std=c++11'])
+else:
+ print termcap.Yellow + termcap.Bold + 'Error' + termcap.Normal,
+ print "Don't know what compiler options to use for your compiler."
+ print termcap.Yellow + ' compiler:' + termcap.Normal, main['CXX']
+ print termcap.Yellow + ' version:' + termcap.Normal,
+ if not CXX_version:
+ print termcap.Yellow + termcap.Bold + "COMMAND NOT FOUND!" +\
+ termcap.Normal
+ else:
+ print CXX_version.replace('\n', '<nl>')
+ print " If you're trying to use a compiler other than GCC"
+ print " or clang, there appears to be something wrong with your"
+ print " environment."
+ print " "
+ print " If you are trying to use a compiler other than those listed"
+ print " above you will need to ease fix SConstruct and "
+ print " src/SConscript to support that compiler."
+ Exit(1)
+
+if main['GCC']:
+ # Check for a supported version of gcc. >= 4.8 is chosen for its
+ # level of c++11 support. See
+ # http://gcc.gnu.org/projects/cxx0x.html for details.
+ gcc_version = readCommand([main['CXX'], '-dumpversion'], exception=False)
+ if compareVersions(gcc_version, "4.8") < 0:
+ print 'Error: gcc version 4.8 or newer required.'
+ print ' Installed version:', gcc_version
+ Exit(1)
+
+ main['GCC_VERSION'] = gcc_version
+
+ # gcc from version 4.8 and above generates "rep; ret" instructions
+ # to avoid performance penalties on certain AMD chips. Older
+ # assemblers detect this as an error, "Error: expecting string
+ # instruction after `rep'"
+ as_version_raw = readCommand([main['AS'], '-v', '/dev/null'],
+ exception=False).split()
+
+ # version strings may contain extra distro-specific
+ # qualifiers, so play it safe and keep only what comes before
+ # the first hyphen
+ as_version = as_version_raw[-1].split('-')[0] if as_version_raw else None
+
+ if not as_version or compareVersions(as_version, "2.23") < 0:
+ print termcap.Yellow + termcap.Bold + \
+ 'Warning: This combination of gcc and binutils have' + \
+ ' known incompatibilities.\n' + \
+ ' If you encounter build problems, please update ' + \
+ 'binutils to 2.23.' + \
+ termcap.Normal
+
+ # Make sure we warn if the user has requested to compile with the
+ # Undefined Benahvior Sanitizer and this version of gcc does not
+ # support it.
+ if GetOption('with_ubsan') and \
+ compareVersions(gcc_version, '4.9') < 0:
+ print termcap.Yellow + termcap.Bold + \
+ 'Warning: UBSan is only supported using gcc 4.9 and later.' + \
+ termcap.Normal
+
+ # Add the appropriate Link-Time Optimization (LTO) flags
+ # unless LTO is explicitly turned off. Note that these flags
+ # are only used by the fast target.
+ if not GetOption('no_lto'):
+ # Pass the LTO flag when compiling to produce GIMPLE
+ # output, we merely create the flags here and only append
+ # them later
+ main['LTO_CCFLAGS'] = ['-flto=%d' % GetOption('num_jobs')]
+
+ # Use the same amount of jobs for LTO as we are running
+ # scons with
+ main['LTO_LDFLAGS'] = ['-flto=%d' % GetOption('num_jobs')]
+
+ main.Append(TCMALLOC_CCFLAGS=['-fno-builtin-malloc', '-fno-builtin-calloc',
+ '-fno-builtin-realloc', '-fno-builtin-free'])
+
+ # add option to check for undeclared overrides
+ if compareVersions(gcc_version, "5.0") > 0:
+ main.Append(CCFLAGS=['-Wno-error=suggest-override'])
+
+elif main['CLANG']:
+ # Check for a supported version of clang, >= 3.1 is needed to
+ # support similar features as gcc 4.8. See
+ # http://clang.llvm.org/cxx_status.html for details
+ clang_version_re = re.compile(".* version (\d+\.\d+)")
+ clang_version_match = clang_version_re.search(CXX_version)
+ if (clang_version_match):
+ clang_version = clang_version_match.groups()[0]
+ if compareVersions(clang_version, "3.1") < 0:
+ print 'Error: clang version 3.1 or newer required.'
+ print ' Installed version:', clang_version
+ Exit(1)
+ else:
+ print 'Error: Unable to determine clang version.'
+ Exit(1)
+
+ # clang has a few additional warnings that we disable, extraneous
+ # parantheses are allowed due to Ruby's printing of the AST,
+ # finally self assignments are allowed as the generated CPU code
+ # is relying on this
+ main.Append(CCFLAGS=['-Wno-parentheses',
+ '-Wno-self-assign',
+ # Some versions of libstdc++ (4.8?) seem to
+ # use struct hash and class hash
+ # interchangeably.
+ '-Wno-mismatched-tags',
+ ])
+
+ main.Append(TCMALLOC_CCFLAGS=['-fno-builtin'])
+
+ # On Mac OS X/Darwin we need to also use libc++ (part of XCode) as
+ # opposed to libstdc++, as the later is dated.
+ if sys.platform == "darwin":
+ main.Append(CXXFLAGS=['-stdlib=libc++'])
+ main.Append(LIBS=['c++'])
+
+else:
+ print termcap.Yellow + termcap.Bold + 'Error' + termcap.Normal,
+ print "Don't know what compiler options to use for your compiler."
+ print termcap.Yellow + ' compiler:' + termcap.Normal, main['CXX']
+ print termcap.Yellow + ' version:' + termcap.Normal,
+ if not CXX_version:
+ print termcap.Yellow + termcap.Bold + "COMMAND NOT FOUND!" +\
+ termcap.Normal
+ else:
+ print CXX_version.replace('\n', '<nl>')
+ print " If you're trying to use a compiler other than GCC"
+ print " or clang, there appears to be something wrong with your"
+ print " environment."
+ print " "
+ print " If you are trying to use a compiler other than those listed"
+ print " above you will need to ease fix SConstruct and "
+ print " src/SConscript to support that compiler."
+ Exit(1)
+
+# Set up common yacc/bison flags (needed for Ruby)
+main['YACCFLAGS'] = '-d'
+main['YACCHXXFILESUFFIX'] = '.hh'
+
+# Do this after we save setting back, or else we'll tack on an
+# extra 'qdo' every time we run scons.
+if main['BATCH']:
+ main['CC'] = main['BATCH_CMD'] + ' ' + main['CC']
+ main['CXX'] = main['BATCH_CMD'] + ' ' + main['CXX']
+ main['AS'] = main['BATCH_CMD'] + ' ' + main['AS']
+ main['AR'] = main['BATCH_CMD'] + ' ' + main['AR']
+ main['RANLIB'] = main['BATCH_CMD'] + ' ' + main['RANLIB']
+
if sys.platform == 'cygwin':
# cygwin has some header file issues...
- env.Append(CCFLAGS=Split("-Wno-uninitialized"))
-env.Append(CPPPATH=[Dir('ext/dnet')])
-
-# Find Python include and library directories for embedding the
-# interpreter. For consistency, we will use the same Python
-# installation used to run scons (and thus this script). If you want
-# to link in an alternate version, see above for instructions on how
-# to invoke scons with a different copy of the Python interpreter.
-
-# Get brief Python version name (e.g., "python2.4") for locating
-# include & library files
-py_version_name = 'python' + sys.version[:3]
-
-# include path, e.g. /usr/local/include/python2.4
-env.Append(CPPPATH = os.path.join(sys.exec_prefix, 'include', py_version_name))
-env.Append(LIBS = py_version_name)
-# add library path too if it's not in the default place
-if sys.exec_prefix != '/usr':
- env.Append(LIBPATH = os.path.join(sys.exec_prefix, 'lib'))
+ main.Append(CCFLAGS=["-Wno-uninitialized"])
+
+# Check for the protobuf compiler
+protoc_version = readCommand([main['PROTOC'], '--version'],
+ exception='').split()
+
+# First two words should be "libprotoc x.y.z"
+if len(protoc_version) < 2 or protoc_version[0] != 'libprotoc':
+ print termcap.Yellow + termcap.Bold + \
+ 'Warning: Protocol buffer compiler (protoc) not found.\n' + \
+ ' Please install protobuf-compiler for tracing support.' + \
+ termcap.Normal
+ main['PROTOC'] = False
+else:
+ # Based on the availability of the compress stream wrappers,
+ # require 2.1.0
+ min_protoc_version = '2.1.0'
+ if compareVersions(protoc_version[1], min_protoc_version) < 0:
+ print termcap.Yellow + termcap.Bold + \
+ 'Warning: protoc version', min_protoc_version, \
+ 'or newer required.\n' + \
+ ' Installed version:', protoc_version[1], \
+ termcap.Normal
+ main['PROTOC'] = False
+ else:
+ # Attempt to determine the appropriate include path and
+ # library path using pkg-config, that means we also need to
+ # check for pkg-config. Note that it is possible to use
+ # protobuf without the involvement of pkg-config. Later on we
+ # check go a library config check and at that point the test
+ # will fail if libprotobuf cannot be found.
+ if readCommand(['pkg-config', '--version'], exception=''):
+ try:
+ # Attempt to establish what linking flags to add for protobuf
+ # using pkg-config
+ main.ParseConfig('pkg-config --cflags --libs-only-L protobuf')
+ except:
+ print termcap.Yellow + termcap.Bold + \
+ 'Warning: pkg-config could not get protobuf flags.' + \
+ termcap.Normal
+
+# Check for SWIG
+if not main.has_key('SWIG'):
+ print 'Error: SWIG utility not found.'
+ print ' Please install (see http://www.swig.org) and retry.'
+ Exit(1)
+
+# Check for appropriate SWIG version
+swig_version = readCommand([main['SWIG'], '-version'], exception='').split()
+# First 3 words should be "SWIG Version x.y.z"
+if len(swig_version) < 3 or \
+ swig_version[0] != 'SWIG' or swig_version[1] != 'Version':
+ print 'Error determining SWIG version.'
+ Exit(1)
+
+min_swig_version = '2.0.4'
+if compareVersions(swig_version[2], min_swig_version) < 0:
+ print 'Error: SWIG version', min_swig_version, 'or newer required.'
+ print ' Installed version:', swig_version[2]
+ Exit(1)
+
+# Check for known incompatibilities. The standard library shipped with
+# gcc >= 4.9 does not play well with swig versions prior to 3.0
+if main['GCC'] and compareVersions(gcc_version, '4.9') >= 0 and \
+ compareVersions(swig_version[2], '3.0') < 0:
+ print termcap.Yellow + termcap.Bold + \
+ 'Warning: This combination of gcc and swig have' + \
+ ' known incompatibilities.\n' + \
+ ' If you encounter build problems, please update ' + \
+ 'swig to 3.0 or later.' + \
+ termcap.Normal
# Set up SWIG flags & scanner
+swig_flags=Split('-c++ -python -modern -templatereduce $_CPPINCFLAGS')
+main.Append(SWIGFLAGS=swig_flags)
+
+# Check for 'timeout' from GNU coreutils. If present, regressions will
+# be run with a time limit. We require version 8.13 since we rely on
+# support for the '--foreground' option.
+timeout_lines = readCommand(['timeout', '--version'],
+ exception='').splitlines()
+# Get the first line and tokenize it
+timeout_version = timeout_lines[0].split() if timeout_lines else []
+main['TIMEOUT'] = timeout_version and \
+ compareVersions(timeout_version[-1], '8.13') >= 0
+
+# filter out all existing swig scanners, they mess up the dependency
+# stuff for some reason
+scanners = []
+for scanner in main['SCANNERS']:
+ skeys = scanner.skeys
+ if skeys == '.i':
+ continue
+
+ if isinstance(skeys, (list, tuple)) and '.i' in skeys:
+ continue
+
+ scanners.append(scanner)
+
+# add the new swig scanner that we like better
+from SCons.Scanner import ClassicCPP as CPPScanner
+swig_inc_re = '^[ \t]*[%,#][ \t]*(?:include|import)[ \t]*(<|")([^>"]+)(>|")'
+scanners.append(CPPScanner("SwigScan", [ ".i" ], "CPPPATH", swig_inc_re))
+
+# replace the scanners list that has what we want
+main['SCANNERS'] = scanners
+
+# Add a custom Check function to test for structure members.
+def CheckMember(context, include, decl, member, include_quotes="<>"):
+ context.Message("Checking for member %s in %s..." %
+ (member, decl))
+ text = """
+#include %(header)s
+int main(){
+ %(decl)s test;
+ (void)test.%(member)s;
+ return 0;
+};
+""" % { "header" : include_quotes[0] + include + include_quotes[1],
+ "decl" : decl,
+ "member" : member,
+ }
+
+ ret = context.TryCompile(text, extension=".cc")
+ context.Result(ret)
+ return ret
-env.Append(SWIGFLAGS=Split('-c++ -python -modern $_CPPINCFLAGS'))
+# Platform-specific configuration. Note again that we assume that all
+# builds under a given build root run on the same host platform.
+conf = Configure(main,
+ conf_dir = joinpath(build_root, '.scons_config'),
+ log_file = joinpath(build_root, 'scons_config.log'),
+ custom_tests = {
+ 'CheckMember' : CheckMember,
+ })
+
+# Check if we should compile a 64 bit binary on Mac OS X/Darwin
+try:
+ import platform
+ uname = platform.uname()
+ if uname[0] == 'Darwin' and compareVersions(uname[2], '9.0.0') >= 0:
+ if int(readCommand('sysctl -n hw.cpu64bit_capable')[0]):
+ main.Append(CCFLAGS=['-arch', 'x86_64'])
+ main.Append(CFLAGS=['-arch', 'x86_64'])
+ main.Append(LINKFLAGS=['-arch', 'x86_64'])
+ main.Append(ASFLAGS=['-arch', 'x86_64'])
+except:
+ pass
+
+# Recent versions of scons substitute a "Null" object for Configure()
+# when configuration isn't necessary, e.g., if the "--help" option is
+# present. Unfortuantely this Null object always returns false,
+# breaking all our configuration checks. We replace it with our own
+# more optimistic null object that returns True instead.
+if not conf:
+ def NullCheck(*args, **kwargs):
+ return True
+
+ class NullConf:
+ def __init__(self, env):
+ self.env = env
+ def Finish(self):
+ return self.env
+ def __getattr__(self, mname):
+ return NullCheck
+
+ conf = NullConf(main)
+
+# Cache build files in the supplied directory.
+if main['M5_BUILD_CACHE']:
+ print 'Using build cache located at', main['M5_BUILD_CACHE']
+ CacheDir(main['M5_BUILD_CACHE'])
+
+if not GetOption('without_python'):
+ # Find Python include and library directories for embedding the
+ # interpreter. We rely on python-config to resolve the appropriate
+ # includes and linker flags. ParseConfig does not seem to understand
+ # the more exotic linker flags such as -Xlinker and -export-dynamic so
+ # we add them explicitly below. If you want to link in an alternate
+ # version of python, see above for instructions on how to invoke
+ # scons with the appropriate PATH set.
+ #
+ # First we check if python2-config exists, else we use python-config
+ python_config = readCommand(['which', 'python2-config'],
+ exception='').strip()
+ if not os.path.exists(python_config):
+ python_config = readCommand(['which', 'python-config'],
+ exception='').strip()
+ py_includes = readCommand([python_config, '--includes'],
+ exception='').split()
+ # Strip the -I from the include folders before adding them to the
+ # CPPPATH
+ main.Append(CPPPATH=map(lambda inc: inc[2:], py_includes))
+
+ # Read the linker flags and split them into libraries and other link
+ # flags. The libraries are added later through the call the CheckLib.
+ py_ld_flags = readCommand([python_config, '--ldflags'],
+ exception='').split()
+ py_libs = []
+ for lib in py_ld_flags:
+ if not lib.startswith('-l'):
+ main.Append(LINKFLAGS=[lib])
+ else:
+ lib = lib[2:]
+ if lib not in py_libs:
+ py_libs.append(lib)
+
+ # verify that this stuff works
+ if not conf.CheckHeader('Python.h', '<>'):
+ print "Error: can't find Python.h header in", py_includes
+ print "Install Python headers (package python-dev on Ubuntu and RedHat)"
+ Exit(1)
-import SCons.Scanner
+ for lib in py_libs:
+ if not conf.CheckLib(lib):
+ print "Error: can't find library %s required by python" % lib
+ Exit(1)
-swig_inc_re = '^[ \t]*[%,#][ \t]*(?:include|import)[ \t]*(<|")([^>"]+)(>|")'
+# On Solaris you need to use libsocket for socket ops
+if not conf.CheckLibWithHeader(None, 'sys/socket.h', 'C++', 'accept(0,0,0);'):
+ if not conf.CheckLibWithHeader('socket', 'sys/socket.h', 'C++', 'accept(0,0,0);'):
+ print "Can't find library with socket calls (e.g. accept())"
+ Exit(1)
+
+# Check for zlib. If the check passes, libz will be automatically
+# added to the LIBS environment variable.
+if not conf.CheckLibWithHeader('z', 'zlib.h', 'C++','zlibVersion();'):
+ print 'Error: did not find needed zlib compression library '\
+ 'and/or zlib.h header file.'
+ print ' Please install zlib and try again.'
+ Exit(1)
+
+# If we have the protobuf compiler, also make sure we have the
+# development libraries. If the check passes, libprotobuf will be
+# automatically added to the LIBS environment variable. After
+# this, we can use the HAVE_PROTOBUF flag to determine if we have
+# got both protoc and libprotobuf available.
+main['HAVE_PROTOBUF'] = main['PROTOC'] and \
+ conf.CheckLibWithHeader('protobuf', 'google/protobuf/message.h',
+ 'C++', 'GOOGLE_PROTOBUF_VERIFY_VERSION;')
+
+# If we have the compiler but not the library, print another warning.
+if main['PROTOC'] and not main['HAVE_PROTOBUF']:
+ print termcap.Yellow + termcap.Bold + \
+ 'Warning: did not find protocol buffer library and/or headers.\n' + \
+ ' Please install libprotobuf-dev for tracing support.' + \
+ termcap.Normal
+
+# Check for librt.
+have_posix_clock = \
+ conf.CheckLibWithHeader(None, 'time.h', 'C',
+ 'clock_nanosleep(0,0,NULL,NULL);') or \
+ conf.CheckLibWithHeader('rt', 'time.h', 'C',
+ 'clock_nanosleep(0,0,NULL,NULL);')
+
+have_posix_timers = \
+ conf.CheckLibWithHeader([None, 'rt'], [ 'time.h', 'signal.h' ], 'C',
+ 'timer_create(CLOCK_MONOTONIC, NULL, NULL);')
+
+if not GetOption('without_tcmalloc'):
+ if conf.CheckLib('tcmalloc'):
+ main.Append(CCFLAGS=main['TCMALLOC_CCFLAGS'])
+ elif conf.CheckLib('tcmalloc_minimal'):
+ main.Append(CCFLAGS=main['TCMALLOC_CCFLAGS'])
+ else:
+ print termcap.Yellow + termcap.Bold + \
+ "You can get a 12% performance improvement by "\
+ "installing tcmalloc (libgoogle-perftools-dev package "\
+ "on Ubuntu or RedHat)." + termcap.Normal
-swig_scanner = SCons.Scanner.ClassicCPP("SwigScan", ".i", "CPPPATH",
- swig_inc_re)
-env.Append(SCANNERS = swig_scanner)
+# Detect back trace implementations. The last implementation in the
+# list will be used by default.
+backtrace_impls = [ "none" ]
-# Other default libraries
-env.Append(LIBS=['z'])
+if conf.CheckLibWithHeader(None, 'execinfo.h', 'C',
+ 'backtrace_symbols_fd((void*)0, 0, 0);'):
+ backtrace_impls.append("glibc")
-# Platform-specific configuration. Note again that we assume that all
-# builds under a given build root run on the same host platform.
-conf = Configure(env,
- conf_dir = os.path.join(build_root, '.scons_config'),
- log_file = os.path.join(build_root, 'scons_config.log'))
+if backtrace_impls[-1] == "none":
+ default_backtrace_impl = "none"
+ print termcap.Yellow + termcap.Bold + \
+ "No suitable back trace implementation found." + \
+ termcap.Normal
+
+if not have_posix_clock:
+ print "Can't find library for POSIX clocks."
# Check for <fenv.h> (C99 FP environment control)
have_fenv = conf.CheckHeader('fenv.h', '<>')
print "Warning: Header file <fenv.h> not found."
print " This host has no IEEE FP rounding mode control."
-# Check for mysql.
-mysql_config = WhereIs('mysql_config')
-have_mysql = mysql_config != None
-
-# Check MySQL version.
-if have_mysql:
- mysql_version = os.popen(mysql_config + ' --version').read()
- mysql_version = mysql_version.split('.')
- mysql_major = int(mysql_version[0])
- mysql_minor = int(mysql_version[1])
- # This version check is probably overly conservative, but it deals
- # with the versions we have installed.
- if mysql_major < 4 or (mysql_major == 4 and mysql_minor < 1):
- print "Warning: MySQL v4.1 or newer required."
- have_mysql = False
-
-# Set up mysql_config commands.
-if have_mysql:
- mysql_config_include = mysql_config + ' --include'
- if os.system(mysql_config_include + ' > /dev/null') != 0:
- # older mysql_config versions don't support --include, use
- # --cflags instead
- mysql_config_include = mysql_config + ' --cflags | sed s/\\\'//g'
- # This seems to work in all versions
- mysql_config_libs = mysql_config + ' --libs'
-
-env = conf.Finish()
+# Check if we should enable KVM-based hardware virtualization. The API
+# we rely on exists since version 2.6.36 of the kernel, but somehow
+# the KVM_API_VERSION does not reflect the change. We test for one of
+# the types as a fall back.
+have_kvm = conf.CheckHeader('linux/kvm.h', '<>')
+if not have_kvm:
+ print "Info: Compatible header file <linux/kvm.h> not found, " \
+ "disabling KVM support."
+
+# x86 needs support for xsave. We test for the structure here since we
+# won't be able to run new tests by the time we know which ISA we're
+# targeting.
+have_kvm_xsave = conf.CheckTypeSize('struct kvm_xsave',
+ '#include <linux/kvm.h>') != 0
+
+# Check if the requested target ISA is compatible with the host
+def is_isa_kvm_compatible(isa):
+ try:
+ import platform
+ host_isa = platform.machine()
+ except:
+ print "Warning: Failed to determine host ISA."
+ return False
+
+ if not have_posix_timers:
+ print "Warning: Can not enable KVM, host seems to lack support " \
+ "for POSIX timers"
+ return False
+
+ if isa == "arm":
+ return host_isa in ( "armv7l", "aarch64" )
+ elif isa == "x86":
+ if host_isa != "x86_64":
+ return False
+
+ if not have_kvm_xsave:
+ print "KVM on x86 requires xsave support in kernel headers."
+ return False
+
+ return True
+ else:
+ return False
+
+
+# Check if the exclude_host attribute is available. We want this to
+# get accurate instruction counts in KVM.
+main['HAVE_PERF_ATTR_EXCLUDE_HOST'] = conf.CheckMember(
+ 'linux/perf_event.h', 'struct perf_event_attr', 'exclude_host')
+
+
+######################################################################
+#
+# Finish the configuration
+#
+main = conf.Finish()
+
+######################################################################
+#
+# Collect all non-global variables
+#
# Define the universe of supported ISAs
-env['ALL_ISA_LIST'] = ['alpha', 'sparc', 'mips']
+all_isa_list = [ ]
+all_gpu_isa_list = [ ]
+Export('all_isa_list')
+Export('all_gpu_isa_list')
-# Define the universe of supported CPU models
-env['ALL_CPU_LIST'] = ['AtomicSimpleCPU', 'TimingSimpleCPU',
- 'FullCPU', 'O3CPU',
- 'OzoneCPU']
+class CpuModel(object):
+ '''The CpuModel class encapsulates everything the ISA parser needs to
+ know about a particular CPU model.'''
-# Sticky options get saved in the options file so they persist from
-# one invocation to the next (unless overridden, in which case the new
-# value becomes sticky).
-sticky_opts = Options(args=ARGUMENTS)
-sticky_opts.AddOptions(
- EnumOption('TARGET_ISA', 'Target ISA', 'alpha', env['ALL_ISA_LIST']),
- BoolOption('FULL_SYSTEM', 'Full-system support', False),
- # There's a bug in scons 0.96.1 that causes ListOptions with list
- # values (more than one value) not to be able to be restored from
- # a saved option file. If this causes trouble then upgrade to
- # scons 0.96.90 or later.
- ListOption('CPU_MODELS', 'CPU models', 'AtomicSimpleCPU,TimingSimpleCPU',
- env['ALL_CPU_LIST']),
- BoolOption('ALPHA_TLASER',
- 'Model Alpha TurboLaser platform (vs. Tsunami)', False),
- BoolOption('NO_FAST_ALLOC', 'Disable fast object allocator', False),
- BoolOption('EFENCE', 'Link with Electric Fence malloc debugger',
- False),
- BoolOption('SS_COMPATIBLE_FP',
- 'Make floating-point results compatible with SimpleScalar',
- False),
- BoolOption('USE_SSE2',
- 'Compile for SSE2 (-msse2) to get IEEE FP on x86 hosts',
- False),
- BoolOption('USE_MYSQL', 'Use MySQL for stats output', have_mysql),
- BoolOption('USE_FENV', 'Use <fenv.h> IEEE mode control', have_fenv),
- BoolOption('USE_CHECKER', 'Use checker for detailed CPU models', False),
- ('CC', 'C compiler', os.environ.get('CC', env['CC'])),
- ('CXX', 'C++ compiler', os.environ.get('CXX', env['CXX'])),
- BoolOption('BATCH', 'Use batch pool for build and tests', False),
- ('BATCH_CMD', 'Batch pool submission command name', 'qdo')
- )
+ # Dict of available CPU model objects. Accessible as CpuModel.dict.
+ dict = {}
-# Non-sticky options only apply to the current build.
-nonsticky_opts = Options(args=ARGUMENTS)
-nonsticky_opts.AddOptions(
- BoolOption('update_ref', 'Update test reference outputs', False)
- )
+ # Constructor. Automatically adds models to CpuModel.dict.
+ def __init__(self, name, default=False):
+ self.name = name # name of model
-# These options get exported to #defines in config/*.hh (see src/SConscript).
-env.ExportOptions = ['FULL_SYSTEM', 'ALPHA_TLASER', 'USE_FENV', \
- 'USE_MYSQL', 'NO_FAST_ALLOC', 'SS_COMPATIBLE_FP', \
- 'USE_CHECKER']
+ # This cpu is enabled by default
+ self.default = default
-# Define a handy 'no-op' action
-def no_action(target, source, env):
- return 0
+ # Add self to dict
+ if name in CpuModel.dict:
+ raise AttributeError, "CpuModel '%s' already registered" % name
+ CpuModel.dict[name] = self
-env.NoAction = Action(no_action, None)
+Export('CpuModel')
+
+# Sticky variables get saved in the variables file so they persist from
+# one invocation to the next (unless overridden, in which case the new
+# value becomes sticky).
+sticky_vars = Variables(args=ARGUMENTS)
+Export('sticky_vars')
+
+# Sticky variables that should be exported
+export_vars = []
+Export('export_vars')
+
+# For Ruby
+all_protocols = []
+Export('all_protocols')
+protocol_dirs = []
+Export('protocol_dirs')
+slicc_includes = []
+Export('slicc_includes')
+
+# Walk the tree and execute all SConsopts scripts that wil add to the
+# above variables
+if GetOption('verbose'):
+ print "Reading SConsopts"
+for bdir in [ base_dir ] + extras_dir_list:
+ if not isdir(bdir):
+ print "Error: directory '%s' does not exist" % bdir
+ Exit(1)
+ for root, dirs, files in os.walk(bdir):
+ if 'SConsopts' in files:
+ if GetOption('verbose'):
+ print "Reading", joinpath(root, 'SConsopts')
+ SConscript(joinpath(root, 'SConsopts'))
+
+all_isa_list.sort()
+all_gpu_isa_list.sort()
+
+sticky_vars.AddVariables(
+ EnumVariable('TARGET_ISA', 'Target ISA', 'alpha', all_isa_list),
+ EnumVariable('TARGET_GPU_ISA', 'Target GPU ISA', 'hsail', all_gpu_isa_list),
+ ListVariable('CPU_MODELS', 'CPU models',
+ sorted(n for n,m in CpuModel.dict.iteritems() if m.default),
+ sorted(CpuModel.dict.keys())),
+ BoolVariable('EFENCE', 'Link with Electric Fence malloc debugger',
+ False),
+ BoolVariable('SS_COMPATIBLE_FP',
+ 'Make floating-point results compatible with SimpleScalar',
+ False),
+ BoolVariable('USE_SSE2',
+ 'Compile for SSE2 (-msse2) to get IEEE FP on x86 hosts',
+ False),
+ BoolVariable('USE_POSIX_CLOCK', 'Use POSIX Clocks', have_posix_clock),
+ BoolVariable('USE_FENV', 'Use <fenv.h> IEEE mode control', have_fenv),
+ BoolVariable('CP_ANNOTATE', 'Enable critical path annotation capability', False),
+ BoolVariable('USE_KVM', 'Enable hardware virtualized (KVM) CPU models', have_kvm),
+ BoolVariable('BUILD_GPU', 'Build the compute-GPU model', False),
+ EnumVariable('PROTOCOL', 'Coherence protocol for Ruby', 'None',
+ all_protocols),
+ EnumVariable('BACKTRACE_IMPL', 'Post-mortem dump implementation',
+ backtrace_impls[-1], backtrace_impls)
+ )
+
+# These variables get exported to #defines in config/*.hh (see src/SConscript).
+export_vars += ['USE_FENV', 'SS_COMPATIBLE_FP', 'TARGET_ISA', 'TARGET_GPU_ISA',
+ 'CP_ANNOTATE', 'USE_POSIX_CLOCK', 'USE_KVM', 'PROTOCOL',
+ 'HAVE_PROTOBUF', 'HAVE_PERF_ATTR_EXCLUDE_HOST']
###################################################
#
###################################################
# This function generates a config header file that #defines the
-# option symbol to the current option setting (0 or 1). The source
-# operands are the name of the option and a Value node containing the
-# value of the option.
+# variable symbol to the current variable setting (0 or 1). The source
+# operands are the name of the variable and a Value node containing the
+# value of the variable.
def build_config_file(target, source, env):
- (option, value) = [s.get_contents() for s in source]
+ (variable, value) = [s.get_contents() for s in source]
f = file(str(target[0]), 'w')
- print >> f, '#define', option, value
+ print >> f, '#define', variable, value
f.close()
return None
-# Generate the message to be printed when building the config file.
-def build_config_file_string(target, source, env):
- (option, value) = [s.get_contents() for s in source]
- return "Defining %s as %s in %s." % (option, value, target[0])
-
# Combine the two functions into a scons Action object.
-config_action = Action(build_config_file, build_config_file_string)
+config_action = MakeAction(build_config_file, Transform("CONFIG H", 2))
# The emitter munges the source & target node lists to reflect what
# we're really doing.
def config_emitter(target, source, env):
- # extract option name from Builder arg
- option = str(target[0])
+ # extract variable name from Builder arg
+ variable = str(target[0])
# True target is config header file
- target = os.path.join('config', option.lower() + '.hh')
- # Force value to 0/1 even if it's a Python bool
- val = int(eval(str(env[option])))
- # Sources are option name & value (packaged in SCons Value nodes)
- return ([target], [Value(option), Value(val)])
+ target = joinpath('config', variable.lower() + '.hh')
+ val = env[variable]
+ if isinstance(val, bool):
+ # Force value to 0/1
+ val = int(val)
+ elif isinstance(val, str):
+ val = '"' + val + '"'
+
+ # Sources are variable name & value (packaged in SCons Value nodes)
+ return ([target], [Value(variable), Value(val)])
config_builder = Builder(emitter = config_emitter, action = config_action)
-env.Append(BUILDERS = { 'ConfigFile' : config_builder })
+main.Append(BUILDERS = { 'ConfigFile' : config_builder })
-###################################################
-#
-# Define a SCons builder for copying files. This is used by the
-# Python zipfile code in src/python/SConscript, but is placed up here
-# since it's potentially more generally applicable.
-#
-###################################################
+# libelf build is shared across all configs in the build root.
+main.SConscript('ext/libelf/SConscript',
+ variant_dir = joinpath(build_root, 'libelf'))
-copy_builder = Builder(action = Copy("$TARGET", "$SOURCE"))
+# iostream3 build is shared across all configs in the build root.
+main.SConscript('ext/iostream3/SConscript',
+ variant_dir = joinpath(build_root, 'iostream3'))
-env.Append(BUILDERS = { 'CopyFile' : copy_builder })
+# libfdt build is shared across all configs in the build root.
+main.SConscript('ext/libfdt/SConscript',
+ variant_dir = joinpath(build_root, 'libfdt'))
-###################################################
-#
-# Define a simple SCons builder to concatenate files.
-#
-# Used to append the Python zip archive to the executable.
-#
-###################################################
+# fputils build is shared across all configs in the build root.
+main.SConscript('ext/fputils/SConscript',
+ variant_dir = joinpath(build_root, 'fputils'))
-concat_builder = Builder(action = Action(['cat $SOURCES > $TARGET',
- 'chmod +x $TARGET']))
+# DRAMSim2 build is shared across all configs in the build root.
+main.SConscript('ext/dramsim2/SConscript',
+ variant_dir = joinpath(build_root, 'dramsim2'))
-env.Append(BUILDERS = { 'Concat' : concat_builder })
+# DRAMPower build is shared across all configs in the build root.
+main.SConscript('ext/drampower/SConscript',
+ variant_dir = joinpath(build_root, 'drampower'))
+# nomali build is shared across all configs in the build root.
+main.SConscript('ext/nomali/SConscript',
+ variant_dir = joinpath(build_root, 'nomali'))
-# base help text
-help_text = '''
-Usage: scons [scons options] [build options] [target(s)]
-
-'''
+###################################################
+#
+# This function is used to set up a directory with switching headers
+#
+###################################################
-# libelf build is shared across all configs in the build root.
-env.SConscript('ext/libelf/SConscript',
- build_dir = os.path.join(build_root, 'libelf'),
- exports = 'env')
+main['ALL_ISA_LIST'] = all_isa_list
+main['ALL_GPU_ISA_LIST'] = all_gpu_isa_list
+all_isa_deps = {}
+def make_switching_dir(dname, switch_headers, env):
+ # Generate the header. target[0] is the full path of the output
+ # header to generate. 'source' is a dummy variable, since we get the
+ # list of ISAs from env['ALL_ISA_LIST'].
+ def gen_switch_hdr(target, source, env):
+ fname = str(target[0])
+ isa = env['TARGET_ISA'].lower()
+ try:
+ f = open(fname, 'w')
+ print >>f, '#include "%s/%s/%s"' % (dname, isa, basename(fname))
+ f.close()
+ except IOError:
+ print "Failed to create %s" % fname
+ raise
+
+ # Build SCons Action object. 'varlist' specifies env vars that this
+ # action depends on; when env['ALL_ISA_LIST'] changes these actions
+ # should get re-executed.
+ switch_hdr_action = MakeAction(gen_switch_hdr,
+ Transform("GENERATE"), varlist=['ALL_ISA_LIST'])
+
+ # Instantiate actions for each header
+ for hdr in switch_headers:
+ env.Command(hdr, [], switch_hdr_action)
+
+ isa_target = Dir('.').up().name.lower().replace('_', '-')
+ env['PHONY_BASE'] = '#'+isa_target
+ all_isa_deps[isa_target] = None
+
+Export('make_switching_dir')
+
+def make_gpu_switching_dir(dname, switch_headers, env):
+ # Generate the header. target[0] is the full path of the output
+ # header to generate. 'source' is a dummy variable, since we get the
+ # list of ISAs from env['ALL_ISA_LIST'].
+ def gen_switch_hdr(target, source, env):
+ fname = str(target[0])
+
+ isa = env['TARGET_GPU_ISA'].lower()
+
+ try:
+ f = open(fname, 'w')
+ print >>f, '#include "%s/%s/%s"' % (dname, isa, basename(fname))
+ f.close()
+ except IOError:
+ print "Failed to create %s" % fname
+ raise
+
+ # Build SCons Action object. 'varlist' specifies env vars that this
+ # action depends on; when env['ALL_ISA_LIST'] changes these actions
+ # should get re-executed.
+ switch_hdr_action = MakeAction(gen_switch_hdr,
+ Transform("GENERATE"), varlist=['ALL_ISA_GPU_LIST'])
+
+ # Instantiate actions for each header
+ for hdr in switch_headers:
+ env.Command(hdr, [], switch_hdr_action)
+
+Export('make_gpu_switching_dir')
+
+# all-isas -> all-deps -> all-environs -> all_targets
+main.Alias('#all-isas', [])
+main.Alias('#all-deps', '#all-isas')
+
+# Dummy target to ensure all environments are created before telling
+# SCons what to actually make (the command line arguments). We attach
+# them to the dependence graph after the environments are complete.
+ORIG_BUILD_TARGETS = list(BUILD_TARGETS) # force a copy; gets closure to work.
+def environsComplete(target, source, env):
+ for t in ORIG_BUILD_TARGETS:
+ main.Depends('#all-targets', t)
+
+# Each build/* switching_dir attaches its *-environs target to #all-environs.
+main.Append(BUILDERS = {'CompleteEnvirons' :
+ Builder(action=MakeAction(environsComplete, None))})
+main.CompleteEnvirons('#all-environs', [])
+
+def doNothing(**ignored): pass
+main.Append(BUILDERS = {'Dummy': Builder(action=MakeAction(doNothing, None))})
+
+# The final target to which all the original targets ultimately get attached.
+main.Dummy('#all-targets', '#all-environs')
+BUILD_TARGETS[:] = ['#all-targets']
###################################################
#
#
###################################################
-# rename base env
-base_env = env
+for variant_path in variant_paths:
+ if not GetOption('silent'):
+ print "Building in", variant_path
-for build_path in build_paths:
- print "Building in", build_path
- # build_dir is the tail component of build path, and is used to
- # determine the build parameters (e.g., 'ALPHA_SE')
- (build_root, build_dir) = os.path.split(build_path)
# Make a copy of the build-root environment to use for this config.
- env = base_env.Copy()
-
- # Set env options according to the build directory config.
- sticky_opts.files = []
- # Options for $BUILD_ROOT/$BUILD_DIR are stored in
- # $BUILD_ROOT/options/$BUILD_DIR so you can nuke
- # $BUILD_ROOT/$BUILD_DIR without losing your options settings.
- current_opts_file = os.path.join(build_root, 'options', build_dir)
- if os.path.isfile(current_opts_file):
- sticky_opts.files.append(current_opts_file)
- print "Using saved options file %s" % current_opts_file
+ env = main.Clone()
+ env['BUILDDIR'] = variant_path
+
+ # variant_dir is the tail component of build path, and is used to
+ # determine the build parameters (e.g., 'ALPHA_SE')
+ (build_root, variant_dir) = splitpath(variant_path)
+
+ # Set env variables according to the build directory config.
+ sticky_vars.files = []
+ # Variables for $BUILD_ROOT/$VARIANT_DIR are stored in
+ # $BUILD_ROOT/variables/$VARIANT_DIR so you can nuke
+ # $BUILD_ROOT/$VARIANT_DIR without losing your variables settings.
+ current_vars_file = joinpath(build_root, 'variables', variant_dir)
+ if isfile(current_vars_file):
+ sticky_vars.files.append(current_vars_file)
+ if not GetOption('silent'):
+ print "Using saved variables file %s" % current_vars_file
else:
- # Build dir-specific options file doesn't exist.
+ # Build dir-specific variables file doesn't exist.
# Make sure the directory is there so we can create it later
- opt_dir = os.path.dirname(current_opts_file)
- if not os.path.isdir(opt_dir):
- os.mkdir(opt_dir)
-
- # Get default build options from source tree. Options are
- # normally determined by name of $BUILD_DIR, but can be
- # overriden by 'default=' arg on command line.
- default_opts_file = os.path.join('build_opts',
- ARGUMENTS.get('default', build_dir))
- if os.path.isfile(default_opts_file):
- sticky_opts.files.append(default_opts_file)
- print "Options file %s not found,\n using defaults in %s" \
- % (current_opts_file, default_opts_file)
+ opt_dir = dirname(current_vars_file)
+ if not isdir(opt_dir):
+ mkdir(opt_dir)
+
+ # Get default build variables from source tree. Variables are
+ # normally determined by name of $VARIANT_DIR, but can be
+ # overridden by '--default=' arg on command line.
+ default = GetOption('default')
+ opts_dir = joinpath(main.root.abspath, 'build_opts')
+ if default:
+ default_vars_files = [joinpath(build_root, 'variables', default),
+ joinpath(opts_dir, default)]
else:
- print "Error: cannot find options file %s or %s" \
- % (current_opts_file, default_opts_file)
+ default_vars_files = [joinpath(opts_dir, variant_dir)]
+ existing_files = filter(isfile, default_vars_files)
+ if existing_files:
+ default_vars_file = existing_files[0]
+ sticky_vars.files.append(default_vars_file)
+ print "Variables file %s not found,\n using defaults in %s" \
+ % (current_vars_file, default_vars_file)
+ else:
+ print "Error: cannot find variables file %s or " \
+ "default file(s) %s" \
+ % (current_vars_file, ' or '.join(default_vars_files))
Exit(1)
- # Apply current option settings to env
- sticky_opts.Update(env)
- nonsticky_opts.Update(env)
+ # Apply current variable settings to env
+ sticky_vars.Update(env)
- help_text += "Sticky options for %s:\n" % build_dir \
- + sticky_opts.GenerateHelpText(env) \
- + "\nNon-sticky options for %s:\n" % build_dir \
- + nonsticky_opts.GenerateHelpText(env)
+ help_texts["local_vars"] += \
+ "Build variables for %s:\n" % variant_dir \
+ + sticky_vars.GenerateHelpText(env)
- # Process option settings.
+ # Process variable settings.
if not have_fenv and env['USE_FENV']:
print "Warning: <fenv.h> not available; " \
- "forcing USE_FENV to False in", build_dir + "."
+ "forcing USE_FENV to False in", variant_dir + "."
env['USE_FENV'] = False
if not env['USE_FENV']:
- print "Warning: No IEEE FP rounding mode control in", build_dir + "."
+ print "Warning: No IEEE FP rounding mode control in", variant_dir + "."
print " FP results may deviate slightly from other platforms."
if env['EFENCE']:
env.Append(LIBS=['efence'])
- if env['USE_MYSQL']:
- if not have_mysql:
- print "Warning: MySQL not available; " \
- "forcing USE_MYSQL to False in", build_dir + "."
- env['USE_MYSQL'] = False
- else:
- print "Compiling in", build_dir, "with MySQL support."
- env.ParseConfig(mysql_config_libs)
- env.ParseConfig(mysql_config_include)
-
- # Save sticky option settings back to current options file
- sticky_opts.Save(current_opts_file, env)
-
- # Do this after we save setting back, or else we'll tack on an
- # extra 'qdo' every time we run scons.
- if env['BATCH']:
- env['CC'] = env['BATCH_CMD'] + ' ' + env['CC']
- env['CXX'] = env['BATCH_CMD'] + ' ' + env['CXX']
+ if env['USE_KVM']:
+ if not have_kvm:
+ print "Warning: Can not enable KVM, host seems to lack KVM support"
+ env['USE_KVM'] = False
+ elif not is_isa_kvm_compatible(env['TARGET_ISA']):
+ print "Info: KVM support disabled due to unsupported host and " \
+ "target ISA combination"
+ env['USE_KVM'] = False
+
+ # Warn about missing optional functionality
+ if env['USE_KVM']:
+ if not main['HAVE_PERF_ATTR_EXCLUDE_HOST']:
+ print "Warning: perf_event headers lack support for the " \
+ "exclude_host attribute. KVM instruction counts will " \
+ "be inaccurate."
+
+ # Save sticky variable settings back to current variables file
+ sticky_vars.Save(current_vars_file, env)
if env['USE_SSE2']:
- env.Append(CCFLAGS='-msse2')
+ env.Append(CCFLAGS=['-msse2'])
# The src/SConscript file sets up the build rules in 'env' according
- # to the configured options. It returns a list of environments,
+ # to the configured variables. It returns a list of environments,
# one for each variant build (debug, opt, etc.)
- envList = SConscript('src/SConscript', build_dir = build_path,
- exports = 'env')
+ SConscript('src/SConscript', variant_dir = variant_path, exports = 'env')
+
+def pairwise(iterable):
+ "s -> (s0,s1), (s1,s2), (s2, s3), ..."
+ a, b = itertools.tee(iterable)
+ b.next()
+ return itertools.izip(a, b)
+
+# Create false dependencies so SCons will parse ISAs, establish
+# dependencies, and setup the build Environments serially. Either
+# SCons (likely) and/or our SConscripts (possibly) cannot cope with -j
+# greater than 1. It appears to be standard race condition stuff; it
+# doesn't always fail, but usually, and the behaviors are different.
+# Every time I tried to remove this, builds would fail in some
+# creative new way. So, don't do that. You'll want to, though, because
+# tests/SConscript takes a long time to make its Environments.
+for t1, t2 in pairwise(sorted(all_isa_deps.iterkeys())):
+ main.Depends('#%s-deps' % t2, '#%s-deps' % t1)
+ main.Depends('#%s-environs' % t2, '#%s-environs' % t1)
- # Set up the regression tests for each build.
-# for e in envList:
-# SConscript('m5-test/SConscript',
-# build_dir = os.path.join(build_dir, 'test', e.Label),
-# exports = { 'env' : e }, duplicate = False)
+# base help text
+Help('''
+Usage: scons [scons options] [build variables] [target(s)]
-Help(help_text)
+Extra scons options:
+%(options)s
-###################################################
-#
-# Let SCons do its thing. At this point SCons will use the defined
-# build environments to build the requested targets.
-#
-###################################################
+Global build variables:
+%(global_vars)s
+%(local_vars)s
+''' % help_texts)