# -*- mode:python -*-
+# Copyright (c) 2013, 2015-2017 ARM Limited
+# All rights reserved.
+#
+# The license below extends only to copyright in the software and shall
+# not be construed as granting a license to any other intellectual
+# property including but not limited to intellectual property relating
+# to a hardware implementation of the functionality of the software
+# licensed hereunder. You may use the software subject to the license
+# terms below provided that you ensure that this notice is replicated
+# unmodified and in its entirety in all distributions of the software,
+# modified or unmodified, in source code or in binary form.
+#
+# Copyright (c) 2011 Advanced Micro Devices, Inc.
# Copyright (c) 2009 The Hewlett-Packard Development Company
# Copyright (c) 2004-2005 The Regents of The University of Michigan
# All rights reserved.
#
# SCons top-level build description (SConstruct) file.
#
-# While in this directory ('m5'), just type 'scons' to build the default
+# While in this directory ('gem5'), just type 'scons' to build the default
# configuration (see below), or type 'scons build/<CONFIG>/<binary>'
-# to build some other configuration (e.g., 'build/ALPHA_FS/m5.opt' for
+# to build some other configuration (e.g., 'build/ALPHA/gem5.opt' for
# the optimized full-system version).
#
-# You can build M5 in a different directory as long as there is a
+# You can build gem5 in a different directory as long as there is a
# 'build/<CONFIG>' somewhere along the target path. The build system
# expects that all configs under the same build directory are being
# built for the same host system.
#
# The following two commands are equivalent. The '-u' option tells
# scons to search up the directory tree for this SConstruct file.
-# % cd <path-to-src>/m5 ; scons build/ALPHA_FS/m5.debug
-# % cd <path-to-src>/m5/build/ALPHA_FS; scons -u m5.debug
+# % cd <path-to-src>/gem5 ; scons build/ALPHA/gem5.debug
+# % cd <path-to-src>/gem5/build/ALPHA; scons -u gem5.debug
#
# The following two commands are equivalent and demonstrate building
# in a directory outside of the source tree. The '-C' option tells
# scons to chdir to the specified directory to find this SConstruct
# file.
-# % cd <path-to-src>/m5 ; scons /local/foo/build/ALPHA_FS/m5.debug
-# % cd /local/foo/build/ALPHA_FS; scons -C <path-to-src>/m5 m5.debug
+# % cd <path-to-src>/gem5 ; scons /local/foo/build/ALPHA/gem5.debug
+# % cd /local/foo/build/ALPHA; scons -C <path-to-src>/gem5 gem5.debug
#
# You can use 'scons -H' to print scons options. If you're in this
-# 'm5' directory (or use -u or -C to tell scons where to find this
-# file), you can use 'scons -h' to print all the M5-specific build
+# 'gem5' directory (or use -u or -C to tell scons where to find this
+# file), you can use 'scons -h' to print all the gem5-specific build
# options as well.
#
###################################################
-# Check for recent-enough Python and SCons versions.
-try:
- # Really old versions of scons only take two options for the
- # function, so check once without the revision and once with the
- # revision, the first instance will fail for stuff other than
- # 0.98, and the second will fail for 0.98.0
- EnsureSConsVersion(0, 98)
- EnsureSConsVersion(0, 98, 1)
-except SystemExit, e:
- print """
-For more details, see:
- http://m5sim.org/wiki/index.php/Compiling_M5
-"""
- raise
-
-# We ensure the python version early because we have stuff that
-# requires python 2.4
-try:
- EnsurePythonVersion(2, 4)
-except SystemExit, e:
- print """
-You can use a non-default installation of the Python interpreter by
-either (1) rearranging your PATH so that scons finds the non-default
-'python' first or (2) explicitly invoking an alternative interpreter
-on the scons script.
-
-For more details, see:
- http://m5sim.org/wiki/index.php/Using_a_non-default_Python_installation
-"""
- raise
+from __future__ import print_function
# Global Python includes
+import itertools
import os
import re
+import shutil
import subprocess
import sys
import SCons
import SCons.Node
-extra_python_paths = [
- Dir('src/python').srcnode().abspath, # M5 includes
- Dir('ext/ply').srcnode().abspath, # ply is used by several files
- ]
-
-sys.path[1:1] = extra_python_paths
-
from m5.util import compareVersions, readCommand
-########################################################################
-#
-# Set up the main build environment.
-#
-########################################################################
-use_vars = set([ 'AS', 'AR', 'CC', 'CXX', 'HOME', 'LD_LIBRARY_PATH', 'PATH',
- 'PYTHONPATH', 'RANLIB' ])
-
-use_env = {}
-for key,val in os.environ.iteritems():
- if key in use_vars or key.startswith("M5"):
- use_env[key] = val
-
-main = Environment(ENV=use_env)
-main.root = Dir(".") # The current directory (where this file lives).
-main.srcdir = Dir("src") # The source directory
-
-# add useful python code PYTHONPATH so it can be used by subprocesses
-# as well
-main.AppendENVPath('PYTHONPATH', extra_python_paths)
+help_texts = {
+ "options" : "",
+ "global_vars" : "",
+ "local_vars" : ""
+}
+
+Export("help_texts")
+
+
+# There's a bug in scons in that (1) by default, the help texts from
+# AddOption() are supposed to be displayed when you type 'scons -h'
+# and (2) you can override the help displayed by 'scons -h' using the
+# Help() function, but these two features are incompatible: once
+# you've overridden the help text using Help(), there's no way to get
+# at the help texts from AddOptions. See:
+# http://scons.tigris.org/issues/show_bug.cgi?id=2356
+# http://scons.tigris.org/issues/show_bug.cgi?id=2611
+# This hack lets us extract the help text from AddOptions and
+# re-inject it via Help(). Ideally someday this bug will be fixed and
+# we can just use AddOption directly.
+def AddLocalOption(*args, **kwargs):
+ col_width = 30
+
+ help = " " + ", ".join(args)
+ if "help" in kwargs:
+ length = len(help)
+ if length >= col_width:
+ help += "\n" + " " * col_width
+ else:
+ help += " " * (col_width - length)
+ help += kwargs["help"]
+ help_texts["options"] += help + "\n"
+
+ AddOption(*args, **kwargs)
+
+AddLocalOption('--colors', dest='use_colors', action='store_true',
+ help="Add color to abbreviated scons output")
+AddLocalOption('--no-colors', dest='use_colors', action='store_false',
+ help="Don't add color to abbreviated scons output")
+AddLocalOption('--with-cxx-config', dest='with_cxx_config',
+ action='store_true',
+ help="Build with support for C++-based configuration")
+AddLocalOption('--default', dest='default', type='string', action='store',
+ help='Override which build_opts file to use for defaults')
+AddLocalOption('--ignore-style', dest='ignore_style', action='store_true',
+ help='Disable style checking hooks')
+AddLocalOption('--no-lto', dest='no_lto', action='store_true',
+ help='Disable Link-Time Optimization for fast')
+AddLocalOption('--force-lto', dest='force_lto', action='store_true',
+ help='Use Link-Time Optimization instead of partial linking' +
+ ' when the compiler doesn\'t support using them together.')
+AddLocalOption('--update-ref', dest='update_ref', action='store_true',
+ help='Update test reference outputs')
+AddLocalOption('--verbose', dest='verbose', action='store_true',
+ help='Print full tool command lines')
+AddLocalOption('--without-python', dest='without_python',
+ action='store_true',
+ help='Build without Python configuration support')
+AddLocalOption('--without-tcmalloc', dest='without_tcmalloc',
+ action='store_true',
+ help='Disable linking against tcmalloc')
+AddLocalOption('--with-ubsan', dest='with_ubsan', action='store_true',
+ help='Build with Undefined Behavior Sanitizer if available')
+AddLocalOption('--with-asan', dest='with_asan', action='store_true',
+ help='Build with Address Sanitizer if available')
+
+if GetOption('no_lto') and GetOption('force_lto'):
+ print('--no-lto and --force-lto are mutually exclusive')
+ Exit(1)
########################################################################
#
-# Mercurial Stuff.
-#
-# If the M5 directory is a mercurial repository, we should do some
-# extra things.
+# Set up the main build environment.
#
########################################################################
-hgdir = main.root.Dir(".hg")
-
-mercurial_style_message = """
-You're missing the M5 style hook.
-Please install the hook so we can ensure that all code fits a common style.
-
-All you'd need to do is add the following lines to your repository .hg/hgrc
-or your personal .hgrc
-----------------
+main = Environment()
-[extensions]
-style = %s/util/style.py
+from gem5_scons import Transform
+from gem5_scons.util import get_termcap
+termcap = get_termcap()
-[hooks]
-pretxncommit.style = python:style.check_whitespace
-""" % (main.root)
-
-mercurial_bin_not_found = """
-Mercurial binary cannot be found, unfortunately this means that we
-cannot easily determine the version of M5 that you are running and
-this makes error messages more difficult to collect. Please consider
-installing mercurial if you choose to post an error message
-"""
-
-mercurial_lib_not_found = """
-Mercurial libraries cannot be found, ignoring style hook
-If you are actually a M5 developer, please fix this and
-run the style hook. It is important.
-"""
-
-hg_info = "Unknown"
-if hgdir.exists():
- # 1) Grab repository revision if we know it.
- cmd = "hg id -n -i -t -b"
- try:
- hg_info = readCommand(cmd, cwd=main.root.abspath).strip()
- except OSError:
- print mercurial_bin_not_found
+main_dict_keys = main.Dictionary().keys()
- # 2) Ensure that the style hook is in place.
- try:
- ui = None
- if ARGUMENTS.get('IGNORE_STYLE') != 'True':
- from mercurial import ui
- ui = ui.ui()
- except ImportError:
- print mercurial_lib_not_found
-
- if ui is not None:
- ui.readconfig(hgdir.File('hgrc').abspath)
- style_hook = ui.config('hooks', 'pretxncommit.style', None)
-
- if not style_hook:
- print mercurial_style_message
- sys.exit(1)
-else:
- print ".hg directory not found"
-
-main['HG_INFO'] = hg_info
+# Check that we have a C/C++ compiler
+if not ('CC' in main_dict_keys and 'CXX' in main_dict_keys):
+ print("No C++ compiler installed (package g++ on Ubuntu and RedHat)")
+ Exit(1)
###################################################
#
###################################################
# Find default configuration & binary.
-Default(environ.get('M5_DEFAULT_BINARY', 'build/ALPHA_SE/m5.debug'))
+Default(environ.get('M5_DEFAULT_BINARY', 'build/ALPHA/gem5.debug'))
# helper function: find last occurrence of element in list
def rfind(l, elt, offs = -1):
return i
raise ValueError, "element not found"
+# Take a list of paths (or SCons Nodes) and return a list with all
+# paths made absolute and ~-expanded. Paths will be interpreted
+# relative to the launch directory unless a different root is provided
+def makePathListAbsolute(path_list, root=GetLaunchDir()):
+ return [abspath(joinpath(root, expanduser(str(p))))
+ for p in path_list]
+
# Each target must have 'build' in the interior of the path; the
# directory below this will determine the build parameters. For
# example, for target 'foo/bar/build/ALPHA_SE/arch/alpha/blah.do' we
# recognize that ALPHA_SE specifies the configuration because it
-# follow 'build' in the bulid path.
-
-# Generate absolute paths to targets so we can see where the build dir is
-if COMMAND_LINE_TARGETS:
- # Ask SCons which directory it was invoked from
- launch_dir = GetLaunchDir()
- # Make targets relative to invocation directory
- abs_targets = [ normpath(joinpath(launch_dir, str(x))) for x in \
- COMMAND_LINE_TARGETS]
-else:
- # Default targets are relative to root of tree
- abs_targets = [ normpath(joinpath(main.root.abspath, str(x))) for x in \
- DEFAULT_TARGETS]
+# follow 'build' in the build path.
+# The funky assignment to "[:]" is needed to replace the list contents
+# in place rather than reassign the symbol to a new list, which
+# doesn't work (obviously!).
+BUILD_TARGETS[:] = makePathListAbsolute(BUILD_TARGETS)
# Generate a list of the unique build roots and configs that the
# collected targets reference.
variant_paths = []
build_root = None
-for t in abs_targets:
+for t in BUILD_TARGETS:
path_dirs = t.split('/')
try:
build_top = rfind(path_dirs, 'build', -2)
except:
- print "Error: no non-leaf 'build' dir found on target path", t
+ print("Error: no non-leaf 'build' dir found on target path", t)
Exit(1)
this_build_root = joinpath('/',*path_dirs[:build_top+1])
if not build_root:
build_root = this_build_root
else:
if this_build_root != build_root:
- print "Error: build targets not under same build root\n"\
- " %s\n %s" % (build_root, this_build_root)
+ print("Error: build targets not under same build root\n"
+ " %s\n %s" % (build_root, this_build_root))
Exit(1)
variant_path = joinpath('/',*path_dirs[:build_top+2])
if variant_path not in variant_paths:
# Make sure build_root exists (might not if this is the first build there)
if not isdir(build_root):
mkdir(build_root)
+main['BUILDROOT'] = build_root
Export('main')
# tree (not specific to a particular build like ALPHA_SE)
#
-# Variable validators & converters for global sticky variables
-def PathListMakeAbsolute(val):
- if not val:
- return val
- f = lambda p: abspath(expanduser(p))
- return ':'.join(map(f, val.split(':')))
-
-def PathListAllExist(key, val, env):
- if not val:
- return
- paths = val.split(':')
- for path in paths:
- if not isdir(path):
- raise SCons.Errors.UserError("Path does not exist: '%s'" % path)
+global_vars_file = joinpath(build_root, 'variables.global')
-global_sticky_vars_file = joinpath(build_root, 'variables.global')
+global_vars = Variables(global_vars_file, args=ARGUMENTS)
-global_sticky_vars = Variables(global_sticky_vars_file, args=ARGUMENTS)
-
-global_sticky_vars.AddVariables(
+global_vars.AddVariables(
('CC', 'C compiler', environ.get('CC', main['CC'])),
('CXX', 'C++ compiler', environ.get('CXX', main['CXX'])),
+ ('PROTOC', 'protoc tool', environ.get('PROTOC', 'protoc')),
('BATCH', 'Use batch pool for build and tests', False),
('BATCH_CMD', 'Batch pool submission command name', 'qdo'),
- ('EXTRAS', 'Add Extra directories to the compilation', '',
- PathListAllExist, PathListMakeAbsolute),
- BoolVariable('RUBY', 'Build with Ruby', False),
+ ('M5_BUILD_CACHE', 'Cache built objects in this directory', False),
+ ('EXTRAS', 'Add extra directories to the compilation', '')
)
-# base help text
-help_text = '''
-Usage: scons [scons options] [build options] [target(s)]
-
-Global sticky options:
-'''
-
-help_text += global_sticky_vars.GenerateHelpText(main)
-
-# Update main environment with values from ARGUMENTS & global_sticky_vars_file
-global_sticky_vars.Update(main)
+# Update main environment with values from ARGUMENTS & global_vars_file
+global_vars.Update(main)
+help_texts["global_vars"] += global_vars.GenerateHelpText(main)
# Save sticky variable settings back to current variables file
-global_sticky_vars.Save(global_sticky_vars_file, main)
+global_vars.Save(global_vars_file, main)
# Parse EXTRAS variable to build list of all directories where we're
-# look for sources etc. This list is exported as base_dir_list.
+# look for sources etc. This list is exported as extras_dir_list.
base_dir = main.srcdir.abspath
if main['EXTRAS']:
- extras_dir_list = main['EXTRAS'].split(':')
+ extras_dir_list = makePathListAbsolute(main['EXTRAS'].split(':'))
else:
extras_dir_list = []
# the ext directory should be on the #includes path
main.Append(CPPPATH=[Dir('ext')])
+# Add shared top-level headers
+main.Prepend(CPPPATH=Dir('include'))
+
+if GetOption('verbose'):
+ def MakeAction(action, string, *args, **kwargs):
+ return Action(action, *args, **kwargs)
+else:
+ MakeAction = Action
+ main['CCCOMSTR'] = Transform("CC")
+ main['CXXCOMSTR'] = Transform("CXX")
+ main['ASCOMSTR'] = Transform("AS")
+ main['ARCOMSTR'] = Transform("AR", 0)
+ main['LINKCOMSTR'] = Transform("LINK", 0)
+ main['SHLINKCOMSTR'] = Transform("SHLINK", 0)
+ main['RANLIBCOMSTR'] = Transform("RANLIB", 0)
+ main['M4COMSTR'] = Transform("M4")
+ main['SHCCCOMSTR'] = Transform("SHCC")
+ main['SHCXXCOMSTR'] = Transform("SHCXX")
+Export('MakeAction')
+
+# Initialize the Link-Time Optimization (LTO) flags
+main['LTO_CCFLAGS'] = []
+main['LTO_LDFLAGS'] = []
+
+# According to the readme, tcmalloc works best if the compiler doesn't
+# assume that we're using the builtin malloc and friends. These flags
+# are compiler-specific, so we need to set them after we detect which
+# compiler we're using.
+main['TCMALLOC_CCFLAGS'] = []
+
CXX_version = readCommand([main['CXX'],'--version'], exception=False)
CXX_V = readCommand([main['CXX'],'-V'], exception=False)
main['GCC'] = CXX_version and CXX_version.find('g++') >= 0
-main['SUNCC'] = CXX_V and CXX_V.find('Sun C++') >= 0
-main['ICC'] = CXX_V and CXX_V.find('Intel') >= 0
-if main['GCC'] + main['SUNCC'] + main['ICC'] > 1:
- print 'Error: How can we have two at the same time?'
+main['CLANG'] = CXX_version and CXX_version.find('clang') >= 0
+if main['GCC'] + main['CLANG'] > 1:
+ print('Error: How can we have two at the same time?')
Exit(1)
# Set up default C++ compiler flags
+if main['GCC'] or main['CLANG']:
+ # As gcc and clang share many flags, do the common parts here
+ main.Append(CCFLAGS=['-pipe'])
+ main.Append(CCFLAGS=['-fno-strict-aliasing'])
+ # Enable -Wall and -Wextra and then disable the few warnings that
+ # we consistently violate
+ main.Append(CCFLAGS=['-Wall', '-Wundef', '-Wextra',
+ '-Wno-sign-compare', '-Wno-unused-parameter'])
+ # We always compile using C++11
+ main.Append(CXXFLAGS=['-std=c++11'])
+ if sys.platform.startswith('freebsd'):
+ main.Append(CCFLAGS=['-I/usr/local/include'])
+ main.Append(CXXFLAGS=['-I/usr/local/include'])
+
+ main['FILTER_PSHLINKFLAGS'] = lambda x: str(x).replace(' -shared', '')
+ main['PSHLINKFLAGS'] = main.subst('${FILTER_PSHLINKFLAGS(SHLINKFLAGS)}')
+ main['PLINKFLAGS'] = main.subst('${LINKFLAGS}')
+ shared_partial_flags = ['-r', '-nostdlib']
+ main.Append(PSHLINKFLAGS=shared_partial_flags)
+ main.Append(PLINKFLAGS=shared_partial_flags)
+
+ # Treat warnings as errors but white list some warnings that we
+ # want to allow (e.g., deprecation warnings).
+ main.Append(CCFLAGS=['-Werror',
+ '-Wno-error=deprecated-declarations',
+ '-Wno-error=deprecated',
+ ])
+else:
+ print(termcap.Yellow + termcap.Bold + 'Error' + termcap.Normal, end=' ')
+ print("Don't know what compiler options to use for your compiler.")
+ print(termcap.Yellow + ' compiler:' + termcap.Normal, main['CXX'])
+ print(termcap.Yellow + ' version:' + termcap.Normal, end = ' ')
+ if not CXX_version:
+ print(termcap.Yellow + termcap.Bold + "COMMAND NOT FOUND!" +
+ termcap.Normal)
+ else:
+ print(CXX_version.replace('\n', '<nl>'))
+ print(" If you're trying to use a compiler other than GCC")
+ print(" or clang, there appears to be something wrong with your")
+ print(" environment.")
+ print(" ")
+ print(" If you are trying to use a compiler other than those listed")
+ print(" above you will need to ease fix SConstruct and ")
+ print(" src/SConscript to support that compiler.")
+ Exit(1)
+
if main['GCC']:
- main.Append(CCFLAGS='-pipe')
- main.Append(CCFLAGS='-fno-strict-aliasing')
- main.Append(CCFLAGS=['-Wall', '-Wno-sign-compare', '-Wundef'])
- main.Append(CXXFLAGS='-Wno-deprecated')
-elif main['ICC']:
- pass #Fix me... add warning flags once we clean up icc warnings
-elif main['SUNCC']:
- main.Append(CCFLAGS='-Qoption ccfe')
- main.Append(CCFLAGS='-features=gcc')
- main.Append(CCFLAGS='-features=extensions')
- main.Append(CCFLAGS='-library=stlport4')
- main.Append(CCFLAGS='-xar')
- #main.Append(CCFLAGS='-instances=semiexplicit')
+ # Check for a supported version of gcc. >= 4.8 is chosen for its
+ # level of c++11 support. See
+ # http://gcc.gnu.org/projects/cxx0x.html for details.
+ gcc_version = readCommand([main['CXX'], '-dumpversion'], exception=False)
+ if compareVersions(gcc_version, "4.8") < 0:
+ print('Error: gcc version 4.8 or newer required.')
+ print(' Installed version: ', gcc_version)
+ Exit(1)
+
+ main['GCC_VERSION'] = gcc_version
+
+ if compareVersions(gcc_version, '4.9') >= 0:
+ # Incremental linking with LTO is currently broken in gcc versions
+ # 4.9 and above. A version where everything works completely hasn't
+ # yet been identified.
+ #
+ # https://gcc.gnu.org/bugzilla/show_bug.cgi?id=67548
+ main['BROKEN_INCREMENTAL_LTO'] = True
+ if compareVersions(gcc_version, '6.0') >= 0:
+ # gcc versions 6.0 and greater accept an -flinker-output flag which
+ # selects what type of output the linker should generate. This is
+ # necessary for incremental lto to work, but is also broken in
+ # current versions of gcc. It may not be necessary in future
+ # versions. We add it here since it might be, and as a reminder that
+ # it exists. It's excluded if lto is being forced.
+ #
+ # https://gcc.gnu.org/gcc-6/changes.html
+ # https://gcc.gnu.org/ml/gcc-patches/2015-11/msg03161.html
+ # https://gcc.gnu.org/bugzilla/show_bug.cgi?id=69866
+ if not GetOption('force_lto'):
+ main.Append(PSHLINKFLAGS='-flinker-output=rel')
+ main.Append(PLINKFLAGS='-flinker-output=rel')
+
+ # Make sure we warn if the user has requested to compile with the
+ # Undefined Benahvior Sanitizer and this version of gcc does not
+ # support it.
+ if GetOption('with_ubsan') and \
+ compareVersions(gcc_version, '4.9') < 0:
+ print(termcap.Yellow + termcap.Bold +
+ 'Warning: UBSan is only supported using gcc 4.9 and later.' +
+ termcap.Normal)
+
+ disable_lto = GetOption('no_lto')
+ if not disable_lto and main.get('BROKEN_INCREMENTAL_LTO', False) and \
+ not GetOption('force_lto'):
+ print(termcap.Yellow + termcap.Bold +
+ 'Warning: Your compiler doesn\'t support incremental linking' +
+ ' and lto at the same time, so lto is being disabled. To force' +
+ ' lto on anyway, use the --force-lto option. That will disable' +
+ ' partial linking.' +
+ termcap.Normal)
+ disable_lto = True
+
+ # Add the appropriate Link-Time Optimization (LTO) flags
+ # unless LTO is explicitly turned off. Note that these flags
+ # are only used by the fast target.
+ if not disable_lto:
+ # Pass the LTO flag when compiling to produce GIMPLE
+ # output, we merely create the flags here and only append
+ # them later
+ main['LTO_CCFLAGS'] = ['-flto=%d' % GetOption('num_jobs')]
+
+ # Use the same amount of jobs for LTO as we are running
+ # scons with
+ main['LTO_LDFLAGS'] = ['-flto=%d' % GetOption('num_jobs')]
+
+ main.Append(TCMALLOC_CCFLAGS=['-fno-builtin-malloc', '-fno-builtin-calloc',
+ '-fno-builtin-realloc', '-fno-builtin-free'])
+
+ # The address sanitizer is available for gcc >= 4.8
+ if GetOption('with_asan'):
+ if GetOption('with_ubsan') and \
+ compareVersions(main['GCC_VERSION'], '4.9') >= 0:
+ main.Append(CCFLAGS=['-fsanitize=address,undefined',
+ '-fno-omit-frame-pointer'],
+ LINKFLAGS='-fsanitize=address,undefined')
+ else:
+ main.Append(CCFLAGS=['-fsanitize=address',
+ '-fno-omit-frame-pointer'],
+ LINKFLAGS='-fsanitize=address')
+ # Only gcc >= 4.9 supports UBSan, so check both the version
+ # and the command-line option before adding the compiler and
+ # linker flags.
+ elif GetOption('with_ubsan') and \
+ compareVersions(main['GCC_VERSION'], '4.9') >= 0:
+ main.Append(CCFLAGS='-fsanitize=undefined')
+ main.Append(LINKFLAGS='-fsanitize=undefined')
+
+elif main['CLANG']:
+ # Check for a supported version of clang, >= 3.1 is needed to
+ # support similar features as gcc 4.8. See
+ # http://clang.llvm.org/cxx_status.html for details
+ clang_version_re = re.compile(".* version (\d+\.\d+)")
+ clang_version_match = clang_version_re.search(CXX_version)
+ if (clang_version_match):
+ clang_version = clang_version_match.groups()[0]
+ if compareVersions(clang_version, "3.1") < 0:
+ print('Error: clang version 3.1 or newer required.')
+ print(' Installed version:', clang_version)
+ Exit(1)
+ else:
+ print('Error: Unable to determine clang version.')
+ Exit(1)
+
+ # clang has a few additional warnings that we disable, extraneous
+ # parantheses are allowed due to Ruby's printing of the AST,
+ # finally self assignments are allowed as the generated CPU code
+ # is relying on this
+ main.Append(CCFLAGS=['-Wno-parentheses',
+ '-Wno-self-assign',
+ # Some versions of libstdc++ (4.8?) seem to
+ # use struct hash and class hash
+ # interchangeably.
+ '-Wno-mismatched-tags',
+ ])
+
+ main.Append(TCMALLOC_CCFLAGS=['-fno-builtin'])
+
+ # On Mac OS X/Darwin we need to also use libc++ (part of XCode) as
+ # opposed to libstdc++, as the later is dated.
+ if sys.platform == "darwin":
+ main.Append(CXXFLAGS=['-stdlib=libc++'])
+ main.Append(LIBS=['c++'])
+
+ # On FreeBSD we need libthr.
+ if sys.platform.startswith('freebsd'):
+ main.Append(LIBS=['thr'])
+
+ # We require clang >= 3.1, so there is no need to check any
+ # versions here.
+ if GetOption('with_ubsan'):
+ if GetOption('with_asan'):
+ main.Append(CCFLAGS=['-fsanitize=address,undefined',
+ '-fno-omit-frame-pointer'],
+ LINKFLAGS='-fsanitize=address,undefined')
+ else:
+ main.Append(CCFLAGS='-fsanitize=undefined',
+ LINKFLAGS='-fsanitize=undefined')
+
+ elif GetOption('with_asan'):
+ main.Append(CCFLAGS=['-fsanitize=address',
+ '-fno-omit-frame-pointer'],
+ LINKFLAGS='-fsanitize=address')
+
else:
- print 'Error: Don\'t know what compiler options to use for your compiler.'
- print ' Please fix SConstruct and src/SConscript and try again.'
+ print(termcap.Yellow + termcap.Bold + 'Error' + termcap.Normal, end=' ')
+ print("Don't know what compiler options to use for your compiler.")
+ print(termcap.Yellow + ' compiler:' + termcap.Normal, main['CXX'])
+ print(termcap.Yellow + ' version:' + termcap.Normal, end=' ')
+ if not CXX_version:
+ print(termcap.Yellow + termcap.Bold + "COMMAND NOT FOUND!" +
+ termcap.Normal)
+ else:
+ print(CXX_version.replace('\n', '<nl>'))
+ print(" If you're trying to use a compiler other than GCC")
+ print(" or clang, there appears to be something wrong with your")
+ print(" environment.")
+ print(" ")
+ print(" If you are trying to use a compiler other than those listed")
+ print(" above you will need to ease fix SConstruct and ")
+ print(" src/SConscript to support that compiler.")
Exit(1)
# Set up common yacc/bison flags (needed for Ruby)
if sys.platform == 'cygwin':
# cygwin has some header file issues...
- main.Append(CCFLAGS="-Wno-uninitialized")
-
-# Check for SWIG
-if not main.has_key('SWIG'):
- print 'Error: SWIG utility not found.'
- print ' Please install (see http://www.swig.org) and retry.'
- Exit(1)
-
-# Check for appropriate SWIG version
-swig_version = readCommand(('swig', '-version'), exception='').split()
-# First 3 words should be "SWIG Version x.y.z"
-if len(swig_version) < 3 or \
- swig_version[0] != 'SWIG' or swig_version[1] != 'Version':
- print 'Error determining SWIG version.'
- Exit(1)
-
-min_swig_version = '1.3.28'
-if compareVersions(swig_version[2], min_swig_version) < 0:
- print 'Error: SWIG version', min_swig_version, 'or newer required.'
- print ' Installed version:', swig_version[2]
- Exit(1)
-
-# Set up SWIG flags & scanner
-swig_flags=Split('-c++ -python -modern -templatereduce $_CPPINCFLAGS')
-main.Append(SWIGFLAGS=swig_flags)
-
-# filter out all existing swig scanners, they mess up the dependency
-# stuff for some reason
-scanners = []
-for scanner in main['SCANNERS']:
- skeys = scanner.skeys
- if skeys == '.i':
- continue
-
- if isinstance(skeys, (list, tuple)) and '.i' in skeys:
- continue
-
- scanners.append(scanner)
-
-# add the new swig scanner that we like better
-from SCons.Scanner import ClassicCPP as CPPScanner
-swig_inc_re = '^[ \t]*[%,#][ \t]*(?:include|import)[ \t]*(<|")([^>"]+)(>|")'
-scanners.append(CPPScanner("SwigScan", [ ".i" ], "CPPPATH", swig_inc_re))
-
-# replace the scanners list that has what we want
-main['SCANNERS'] = scanners
-
-# Add a custom Check function to the Configure context so that we can
-# figure out if the compiler adds leading underscores to global
-# variables. This is needed for the autogenerated asm files that we
-# use for embedding the python code.
-def CheckLeading(context):
- context.Message("Checking for leading underscore in global variables...")
- # 1) Define a global variable called x from asm so the C compiler
- # won't change the symbol at all.
- # 2) Declare that variable.
- # 3) Use the variable
- #
- # If the compiler prepends an underscore, this will successfully
- # link because the external symbol 'x' will be called '_x' which
- # was defined by the asm statement. If the compiler does not
- # prepend an underscore, this will not successfully link because
- # '_x' will have been defined by assembly, while the C portion of
- # the code will be trying to use 'x'
- ret = context.TryLink('''
- asm(".globl _x; _x: .byte 0");
- extern int x;
- int main() { return x; }
- ''', extension=".c")
- context.env.Append(LEADING_UNDERSCORE=ret)
+ main.Append(CCFLAGS=["-Wno-uninitialized"])
+
+# Check for the protobuf compiler
+protoc_version = readCommand([main['PROTOC'], '--version'],
+ exception='').split()
+
+# First two words should be "libprotoc x.y.z"
+if len(protoc_version) < 2 or protoc_version[0] != 'libprotoc':
+ print(termcap.Yellow + termcap.Bold +
+ 'Warning: Protocol buffer compiler (protoc) not found.\n' +
+ ' Please install protobuf-compiler for tracing support.' +
+ termcap.Normal)
+ main['PROTOC'] = False
+else:
+ # Based on the availability of the compress stream wrappers,
+ # require 2.1.0
+ min_protoc_version = '2.1.0'
+ if compareVersions(protoc_version[1], min_protoc_version) < 0:
+ print(termcap.Yellow + termcap.Bold +
+ 'Warning: protoc version', min_protoc_version,
+ 'or newer required.\n' +
+ ' Installed version:', protoc_version[1],
+ termcap.Normal)
+ main['PROTOC'] = False
+ else:
+ # Attempt to determine the appropriate include path and
+ # library path using pkg-config, that means we also need to
+ # check for pkg-config. Note that it is possible to use
+ # protobuf without the involvement of pkg-config. Later on we
+ # check go a library config check and at that point the test
+ # will fail if libprotobuf cannot be found.
+ if readCommand(['pkg-config', '--version'], exception=''):
+ try:
+ # Attempt to establish what linking flags to add for protobuf
+ # using pkg-config
+ main.ParseConfig('pkg-config --cflags --libs-only-L protobuf')
+ except:
+ print(termcap.Yellow + termcap.Bold +
+ 'Warning: pkg-config could not get protobuf flags.' +
+ termcap.Normal)
+
+
+# Check for 'timeout' from GNU coreutils. If present, regressions will
+# be run with a time limit. We require version 8.13 since we rely on
+# support for the '--foreground' option.
+if sys.platform.startswith('freebsd'):
+ timeout_lines = readCommand(['gtimeout', '--version'],
+ exception='').splitlines()
+else:
+ timeout_lines = readCommand(['timeout', '--version'],
+ exception='').splitlines()
+# Get the first line and tokenize it
+timeout_version = timeout_lines[0].split() if timeout_lines else []
+main['TIMEOUT'] = timeout_version and \
+ compareVersions(timeout_version[-1], '8.13') >= 0
+
+# Add a custom Check function to test for structure members.
+def CheckMember(context, include, decl, member, include_quotes="<>"):
+ context.Message("Checking for member %s in %s..." %
+ (member, decl))
+ text = """
+#include %(header)s
+int main(){
+ %(decl)s test;
+ (void)test.%(member)s;
+ return 0;
+};
+""" % { "header" : include_quotes[0] + include + include_quotes[1],
+ "decl" : decl,
+ "member" : member,
+ }
+
+ ret = context.TryCompile(text, extension=".cc")
context.Result(ret)
return ret
conf = Configure(main,
conf_dir = joinpath(build_root, '.scons_config'),
log_file = joinpath(build_root, 'scons_config.log'),
- custom_tests = { 'CheckLeading' : CheckLeading })
-
-# Check for leading underscores. Don't really need to worry either
-# way so don't need to check the return code.
-conf.CheckLeading()
+ custom_tests = {
+ 'CheckMember' : CheckMember,
+ })
# Check if we should compile a 64 bit binary on Mac OS X/Darwin
try:
uname = platform.uname()
if uname[0] == 'Darwin' and compareVersions(uname[2], '9.0.0') >= 0:
if int(readCommand('sysctl -n hw.cpu64bit_capable')[0]):
- main.Append(CCFLAGS='-arch x86_64')
- main.Append(CFLAGS='-arch x86_64')
- main.Append(LINKFLAGS='-arch x86_64')
- main.Append(ASFLAGS='-arch x86_64')
+ main.Append(CCFLAGS=['-arch', 'x86_64'])
+ main.Append(CFLAGS=['-arch', 'x86_64'])
+ main.Append(LINKFLAGS=['-arch', 'x86_64'])
+ main.Append(ASFLAGS=['-arch', 'x86_64'])
except:
pass
conf = NullConf(main)
-# Find Python include and library directories for embedding the
-# interpreter. For consistency, we will use the same Python
-# installation used to run scons (and thus this script). If you want
-# to link in an alternate version, see above for instructions on how
-# to invoke scons with a different copy of the Python interpreter.
-from distutils import sysconfig
-
-py_getvar = sysconfig.get_config_var
-
-py_version = 'python' + py_getvar('VERSION')
-
-py_general_include = sysconfig.get_python_inc()
-py_platform_include = sysconfig.get_python_inc(plat_specific=True)
-py_includes = [ py_general_include ]
-if py_platform_include != py_general_include:
- py_includes.append(py_platform_include)
-
-py_lib_path = [ py_getvar('LIBDIR') ]
-# add the prefix/lib/pythonX.Y/config dir, but only if there is no
-# shared library in prefix/lib/.
-if not py_getvar('Py_ENABLE_SHARED'):
- py_lib_path.append(py_getvar('LIBPL'))
-
-py_libs = []
-for lib in py_getvar('LIBS').split() + py_getvar('SYSLIBS').split():
- assert lib.startswith('-l')
- lib = lib[2:]
- if lib not in py_libs:
- py_libs.append(lib)
-py_libs.append(py_version)
-
-main.Append(CPPPATH=py_includes)
-main.Append(LIBPATH=py_lib_path)
-
-# verify that this stuff works
-if not conf.CheckHeader('Python.h', '<>'):
- print "Error: can't find Python.h header in", py_includes
- Exit(1)
-
-for lib in py_libs:
- if not conf.CheckLib(lib):
- print "Error: can't find library %s required by python" % lib
+# Cache build files in the supplied directory.
+if main['M5_BUILD_CACHE']:
+ print('Using build cache located at', main['M5_BUILD_CACHE'])
+ CacheDir(main['M5_BUILD_CACHE'])
+
+main['USE_PYTHON'] = not GetOption('without_python')
+if main['USE_PYTHON']:
+ # Find Python include and library directories for embedding the
+ # interpreter. We rely on python-config to resolve the appropriate
+ # includes and linker flags. ParseConfig does not seem to understand
+ # the more exotic linker flags such as -Xlinker and -export-dynamic so
+ # we add them explicitly below. If you want to link in an alternate
+ # version of python, see above for instructions on how to invoke
+ # scons with the appropriate PATH set.
+ #
+ # First we check if python2-config exists, else we use python-config
+ python_config = readCommand(['which', 'python2-config'],
+ exception='').strip()
+ if not os.path.exists(python_config):
+ python_config = readCommand(['which', 'python-config'],
+ exception='').strip()
+ py_includes = readCommand([python_config, '--includes'],
+ exception='').split()
+ # Strip the -I from the include folders before adding them to the
+ # CPPPATH
+ main.Append(CPPPATH=map(lambda inc: inc[2:], py_includes))
+
+ # Read the linker flags and split them into libraries and other link
+ # flags. The libraries are added later through the call the CheckLib.
+ py_ld_flags = readCommand([python_config, '--ldflags'],
+ exception='').split()
+ py_libs = []
+ for lib in py_ld_flags:
+ if not lib.startswith('-l'):
+ main.Append(LINKFLAGS=[lib])
+ else:
+ lib = lib[2:]
+ if lib not in py_libs:
+ py_libs.append(lib)
+
+ # verify that this stuff works
+ if not conf.CheckHeader('Python.h', '<>'):
+ print("Error: Check failed for Python.h header in", py_includes)
+ print("Two possible reasons:")
+ print("1. Python headers are not installed (You can install the "
+ "package python-dev on Ubuntu and RedHat)")
+ print("2. SCons is using a wrong C compiler. This can happen if "
+ "CC has the wrong value.")
+ print("CC = %s" % main['CC'])
Exit(1)
+ for lib in py_libs:
+ if not conf.CheckLib(lib):
+ print("Error: can't find library %s required by python" % lib)
+ Exit(1)
+
# On Solaris you need to use libsocket for socket ops
if not conf.CheckLibWithHeader(None, 'sys/socket.h', 'C++', 'accept(0,0,0);'):
if not conf.CheckLibWithHeader('socket', 'sys/socket.h', 'C++', 'accept(0,0,0);'):
- print "Can't find library with socket calls (e.g. accept())"
+ print("Can't find library with socket calls (e.g. accept())")
Exit(1)
# Check for zlib. If the check passes, libz will be automatically
# added to the LIBS environment variable.
if not conf.CheckLibWithHeader('z', 'zlib.h', 'C++','zlibVersion();'):
- print 'Error: did not find needed zlib compression library '\
- 'and/or zlib.h header file.'
- print ' Please install zlib and try again.'
+ print('Error: did not find needed zlib compression library '
+ 'and/or zlib.h header file.')
+ print(' Please install zlib and try again.')
Exit(1)
+# If we have the protobuf compiler, also make sure we have the
+# development libraries. If the check passes, libprotobuf will be
+# automatically added to the LIBS environment variable. After
+# this, we can use the HAVE_PROTOBUF flag to determine if we have
+# got both protoc and libprotobuf available.
+main['HAVE_PROTOBUF'] = main['PROTOC'] and \
+ conf.CheckLibWithHeader('protobuf', 'google/protobuf/message.h',
+ 'C++', 'GOOGLE_PROTOBUF_VERIFY_VERSION;')
+
+# Valgrind gets much less confused if you tell it when you're using
+# alternative stacks.
+main['HAVE_VALGRIND'] = conf.CheckCHeader('valgrind/valgrind.h')
+
+# If we have the compiler but not the library, print another warning.
+if main['PROTOC'] and not main['HAVE_PROTOBUF']:
+ print(termcap.Yellow + termcap.Bold +
+ 'Warning: did not find protocol buffer library and/or headers.\n' +
+ ' Please install libprotobuf-dev for tracing support.' +
+ termcap.Normal)
+
+# Check for librt.
+have_posix_clock = \
+ conf.CheckLibWithHeader(None, 'time.h', 'C',
+ 'clock_nanosleep(0,0,NULL,NULL);') or \
+ conf.CheckLibWithHeader('rt', 'time.h', 'C',
+ 'clock_nanosleep(0,0,NULL,NULL);')
+
+have_posix_timers = \
+ conf.CheckLibWithHeader([None, 'rt'], [ 'time.h', 'signal.h' ], 'C',
+ 'timer_create(CLOCK_MONOTONIC, NULL, NULL);')
+
+if not GetOption('without_tcmalloc'):
+ if conf.CheckLib('tcmalloc'):
+ main.Append(CCFLAGS=main['TCMALLOC_CCFLAGS'])
+ elif conf.CheckLib('tcmalloc_minimal'):
+ main.Append(CCFLAGS=main['TCMALLOC_CCFLAGS'])
+ else:
+ print(termcap.Yellow + termcap.Bold +
+ "You can get a 12% performance improvement by "
+ "installing tcmalloc (libgoogle-perftools-dev package "
+ "on Ubuntu or RedHat)." + termcap.Normal)
+
+
+# Detect back trace implementations. The last implementation in the
+# list will be used by default.
+backtrace_impls = [ "none" ]
+
+backtrace_checker = 'char temp;' + \
+ ' backtrace_symbols_fd((void*)&temp, 0, 0);'
+if conf.CheckLibWithHeader(None, 'execinfo.h', 'C', backtrace_checker):
+ backtrace_impls.append("glibc")
+elif conf.CheckLibWithHeader('execinfo', 'execinfo.h', 'C',
+ backtrace_checker):
+ # NetBSD and FreeBSD need libexecinfo.
+ backtrace_impls.append("glibc")
+ main.Append(LIBS=['execinfo'])
+
+if backtrace_impls[-1] == "none":
+ default_backtrace_impl = "none"
+ print(termcap.Yellow + termcap.Bold +
+ "No suitable back trace implementation found." +
+ termcap.Normal)
+
+if not have_posix_clock:
+ print("Can't find library for POSIX clocks.")
+
# Check for <fenv.h> (C99 FP environment control)
have_fenv = conf.CheckHeader('fenv.h', '<>')
if not have_fenv:
- print "Warning: Header file <fenv.h> not found."
- print " This host has no IEEE FP rounding mode control."
+ print("Warning: Header file <fenv.h> not found.")
+ print(" This host has no IEEE FP rounding mode control.")
+
+# Check for <png.h> (libpng library needed if wanting to dump
+# frame buffer image in png format)
+have_png = conf.CheckHeader('png.h', '<>')
+if not have_png:
+ print("Warning: Header file <png.h> not found.")
+ print(" This host has no libpng library.")
+ print(" Disabling support for PNG framebuffers.")
+
+# Check if we should enable KVM-based hardware virtualization. The API
+# we rely on exists since version 2.6.36 of the kernel, but somehow
+# the KVM_API_VERSION does not reflect the change. We test for one of
+# the types as a fall back.
+have_kvm = conf.CheckHeader('linux/kvm.h', '<>')
+if not have_kvm:
+ print("Info: Compatible header file <linux/kvm.h> not found, "
+ "disabling KVM support.")
+
+# Check if the TUN/TAP driver is available.
+have_tuntap = conf.CheckHeader('linux/if_tun.h', '<>')
+if not have_tuntap:
+ print("Info: Compatible header file <linux/if_tun.h> not found.")
+
+# x86 needs support for xsave. We test for the structure here since we
+# won't be able to run new tests by the time we know which ISA we're
+# targeting.
+have_kvm_xsave = conf.CheckTypeSize('struct kvm_xsave',
+ '#include <linux/kvm.h>') != 0
+
+# Check if the requested target ISA is compatible with the host
+def is_isa_kvm_compatible(isa):
+ try:
+ import platform
+ host_isa = platform.machine()
+ except:
+ print("Warning: Failed to determine host ISA.")
+ return False
+
+ if not have_posix_timers:
+ print("Warning: Can not enable KVM, host seems to lack support "
+ "for POSIX timers")
+ return False
+
+ if isa == "arm":
+ return host_isa in ( "armv7l", "aarch64" )
+ elif isa == "x86":
+ if host_isa != "x86_64":
+ return False
+
+ if not have_kvm_xsave:
+ print("KVM on x86 requires xsave support in kernel headers.")
+ return False
+
+ return True
+ else:
+ return False
+
+
+# Check if the exclude_host attribute is available. We want this to
+# get accurate instruction counts in KVM.
+main['HAVE_PERF_ATTR_EXCLUDE_HOST'] = conf.CheckMember(
+ 'linux/perf_event.h', 'struct perf_event_attr', 'exclude_host')
-######################################################################
-#
-# Check for mysql.
-#
-mysql_config = WhereIs('mysql_config')
-have_mysql = bool(mysql_config)
-
-# Check MySQL version.
-if have_mysql:
- mysql_version = readCommand(mysql_config + ' --version')
- min_mysql_version = '4.1'
- if compareVersions(mysql_version, min_mysql_version) < 0:
- print 'Warning: MySQL', min_mysql_version, 'or newer required.'
- print ' Version', mysql_version, 'detected.'
- have_mysql = False
-
-# Set up mysql_config commands.
-if have_mysql:
- mysql_config_include = mysql_config + ' --include'
- if os.system(mysql_config_include + ' > /dev/null') != 0:
- # older mysql_config versions don't support --include, use
- # --cflags instead
- mysql_config_include = mysql_config + ' --cflags | sed s/\\\'//g'
- # This seems to work in all versions
- mysql_config_libs = mysql_config + ' --libs'
######################################################################
#
# Define the universe of supported ISAs
all_isa_list = [ ]
+all_gpu_isa_list = [ ]
Export('all_isa_list')
+Export('all_gpu_isa_list')
-# Define the universe of supported CPU models
-all_cpu_list = [ ]
-default_cpus = [ ]
-Export('all_cpu_list', 'default_cpus')
+class CpuModel(object):
+ '''The CpuModel class encapsulates everything the ISA parser needs to
+ know about a particular CPU model.'''
+
+ # Dict of available CPU model objects. Accessible as CpuModel.dict.
+ dict = {}
+
+ # Constructor. Automatically adds models to CpuModel.dict.
+ def __init__(self, name, default=False):
+ self.name = name # name of model
+
+ # This cpu is enabled by default
+ self.default = default
+
+ # Add self to dict
+ if name in CpuModel.dict:
+ raise AttributeError, "CpuModel '%s' already registered" % name
+ CpuModel.dict[name] = self
+
+Export('CpuModel')
# Sticky variables get saved in the variables file so they persist from
# one invocation to the next (unless overridden, in which case the new
export_vars = []
Export('export_vars')
-# Non-sticky variables only apply to the current build.
-nonsticky_vars = Variables(args=ARGUMENTS)
-Export('nonsticky_vars')
+# For Ruby
+all_protocols = []
+Export('all_protocols')
+protocol_dirs = []
+Export('protocol_dirs')
+slicc_includes = []
+Export('slicc_includes')
# Walk the tree and execute all SConsopts scripts that wil add to the
# above variables
+if GetOption('verbose'):
+ print("Reading SConsopts")
for bdir in [ base_dir ] + extras_dir_list:
+ if not isdir(bdir):
+ print("Error: directory '%s' does not exist" % bdir)
+ Exit(1)
for root, dirs, files in os.walk(bdir):
if 'SConsopts' in files:
- print "Reading", joinpath(root, 'SConsopts')
+ if GetOption('verbose'):
+ print("Reading", joinpath(root, 'SConsopts'))
SConscript(joinpath(root, 'SConsopts'))
all_isa_list.sort()
-all_cpu_list.sort()
-default_cpus.sort()
+all_gpu_isa_list.sort()
sticky_vars.AddVariables(
EnumVariable('TARGET_ISA', 'Target ISA', 'alpha', all_isa_list),
- BoolVariable('FULL_SYSTEM', 'Full-system support', False),
- ListVariable('CPU_MODELS', 'CPU models', default_cpus, all_cpu_list),
- BoolVariable('NO_FAST_ALLOC', 'Disable fast object allocator', False),
- BoolVariable('FAST_ALLOC_DEBUG', 'Enable fast object allocator debugging',
- False),
- BoolVariable('FAST_ALLOC_STATS', 'Enable fast object allocator statistics',
- False),
+ EnumVariable('TARGET_GPU_ISA', 'Target GPU ISA', 'hsail', all_gpu_isa_list),
+ ListVariable('CPU_MODELS', 'CPU models',
+ sorted(n for n,m in CpuModel.dict.iteritems() if m.default),
+ sorted(CpuModel.dict.keys())),
BoolVariable('EFENCE', 'Link with Electric Fence malloc debugger',
False),
BoolVariable('SS_COMPATIBLE_FP',
BoolVariable('USE_SSE2',
'Compile for SSE2 (-msse2) to get IEEE FP on x86 hosts',
False),
- BoolVariable('USE_MYSQL', 'Use MySQL for stats output', have_mysql),
+ BoolVariable('USE_POSIX_CLOCK', 'Use POSIX Clocks', have_posix_clock),
BoolVariable('USE_FENV', 'Use <fenv.h> IEEE mode control', have_fenv),
- BoolVariable('USE_CHECKER', 'Use checker for detailed CPU models', False),
- BoolVariable('CP_ANNOTATE', 'Enable critical path annotation capability', False),
- )
-
-nonsticky_vars.AddVariables(
- BoolVariable('update_ref', 'Update test reference outputs', False)
+ BoolVariable('USE_PNG', 'Enable support for PNG images', have_png),
+ BoolVariable('CP_ANNOTATE', 'Enable critical path annotation capability',
+ False),
+ BoolVariable('USE_KVM', 'Enable hardware virtualized (KVM) CPU models',
+ have_kvm),
+ BoolVariable('USE_TUNTAP',
+ 'Enable using a tap device to bridge to the host network',
+ have_tuntap),
+ BoolVariable('BUILD_GPU', 'Build the compute-GPU model', False),
+ EnumVariable('PROTOCOL', 'Coherence protocol for Ruby', 'None',
+ all_protocols),
+ EnumVariable('BACKTRACE_IMPL', 'Post-mortem dump implementation',
+ backtrace_impls[-1], backtrace_impls)
)
# These variables get exported to #defines in config/*.hh (see src/SConscript).
-export_vars += ['FULL_SYSTEM', 'USE_FENV', 'USE_MYSQL',
- 'NO_FAST_ALLOC', 'FAST_ALLOC_DEBUG', 'FAST_ALLOC_STATS',
- 'SS_COMPATIBLE_FP', 'USE_CHECKER', 'TARGET_ISA', 'CP_ANNOTATE']
+export_vars += ['USE_FENV', 'SS_COMPATIBLE_FP', 'TARGET_ISA', 'TARGET_GPU_ISA',
+ 'CP_ANNOTATE', 'USE_POSIX_CLOCK', 'USE_KVM', 'USE_TUNTAP',
+ 'PROTOCOL', 'HAVE_PROTOBUF', 'HAVE_VALGRIND',
+ 'HAVE_PERF_ATTR_EXCLUDE_HOST', 'USE_PNG']
###################################################
#
def build_config_file(target, source, env):
(variable, value) = [s.get_contents() for s in source]
f = file(str(target[0]), 'w')
- print >> f, '#define', variable, value
+ print('#define', variable, value, file=f)
f.close()
return None
-# Generate the message to be printed when building the config file.
-def build_config_file_string(target, source, env):
- (variable, value) = [s.get_contents() for s in source]
- return "Defining %s as %s in %s." % (variable, value, target[0])
-
# Combine the two functions into a scons Action object.
-config_action = Action(build_config_file, build_config_file_string)
+config_action = MakeAction(build_config_file, Transform("CONFIG H", 2))
# The emitter munges the source & target node lists to reflect what
# we're really doing.
main.Append(BUILDERS = { 'ConfigFile' : config_builder })
-# libelf build is shared across all configs in the build root.
-main.SConscript('ext/libelf/SConscript',
- variant_dir = joinpath(build_root, 'libelf'))
+###################################################
+#
+# Builders for static and shared partially linked object files.
+#
+###################################################
-# gzstream build is shared across all configs in the build root.
-main.SConscript('ext/gzstream/SConscript',
- variant_dir = joinpath(build_root, 'gzstream'))
+partial_static_builder = Builder(action=SCons.Defaults.LinkAction,
+ src_suffix='$OBJSUFFIX',
+ src_builder=['StaticObject', 'Object'],
+ LINKFLAGS='$PLINKFLAGS',
+ LIBS='')
+
+def partial_shared_emitter(target, source, env):
+ for tgt in target:
+ tgt.attributes.shared = 1
+ return (target, source)
+partial_shared_builder = Builder(action=SCons.Defaults.ShLinkAction,
+ emitter=partial_shared_emitter,
+ src_suffix='$SHOBJSUFFIX',
+ src_builder='SharedObject',
+ SHLINKFLAGS='$PSHLINKFLAGS',
+ LIBS='')
+
+main.Append(BUILDERS = { 'PartialShared' : partial_shared_builder,
+ 'PartialStatic' : partial_static_builder })
+
+# builds in ext are shared across all configs in the build root.
+ext_dir = abspath(joinpath(str(main.root), 'ext'))
+ext_build_dirs = []
+for root, dirs, files in os.walk(ext_dir):
+ if 'SConscript' in files:
+ build_dir = os.path.relpath(root, ext_dir)
+ ext_build_dirs.append(build_dir)
+ main.SConscript(joinpath(root, 'SConscript'),
+ variant_dir=joinpath(build_root, build_dir))
+
+main.Prepend(CPPPATH=Dir('ext/pybind11/include/'))
###################################################
#
-# This function is used to set up a directory with switching headers
+# This builder and wrapper method are used to set up a directory with
+# switching headers. Those are headers which are in a generic location and
+# that include more specific headers from a directory chosen at build time
+# based on the current build settings.
#
###################################################
-main['ALL_ISA_LIST'] = all_isa_list
-def make_switching_dir(dname, switch_headers, env):
- # Generate the header. target[0] is the full path of the output
- # header to generate. 'source' is a dummy variable, since we get the
- # list of ISAs from env['ALL_ISA_LIST'].
- def gen_switch_hdr(target, source, env):
- fname = str(target[0])
- bname = basename(fname)
- f = open(fname, 'w')
- f.write('#include "arch/isa_specific.hh"\n')
- cond = '#if'
- for isa in all_isa_list:
- f.write('%s THE_ISA == %s_ISA\n#include "%s/%s/%s"\n'
- % (cond, isa.upper(), dname, isa, bname))
- cond = '#elif'
- f.write('#else\n#error "THE_ISA not set"\n#endif\n')
- f.close()
- return 0
-
- # String to print when generating header
- def gen_switch_hdr_string(target, source, env):
- return "Generating switch header " + str(target[0])
-
- # Build SCons Action object. 'varlist' specifies env vars that this
- # action depends on; when env['ALL_ISA_LIST'] changes these actions
- # should get re-executed.
- switch_hdr_action = Action(gen_switch_hdr, gen_switch_hdr_string,
- varlist=['ALL_ISA_LIST'])
-
- # Instantiate actions for each header
- for hdr in switch_headers:
- env.Command(hdr, [], switch_hdr_action)
-Export('make_switching_dir')
+def build_switching_header(target, source, env):
+ path = str(target[0])
+ subdir = str(source[0])
+ dp, fp = os.path.split(path)
+ dp = os.path.relpath(os.path.realpath(dp),
+ os.path.realpath(env['BUILDDIR']))
+ with open(path, 'w') as hdr:
+ print('#include "%s/%s/%s"' % (dp, subdir, fp), file=hdr)
+
+switching_header_action = MakeAction(build_switching_header,
+ Transform('GENERATE'))
+
+switching_header_builder = Builder(action=switching_header_action,
+ source_factory=Value,
+ single_source=True)
+
+main.Append(BUILDERS = { 'SwitchingHeader': switching_header_builder })
+
+def switching_headers(self, headers, source):
+ for header in headers:
+ self.SwitchingHeader(header, source)
+
+main.AddMethod(switching_headers, 'SwitchingHeaders')
###################################################
#
###################################################
for variant_path in variant_paths:
- print "Building in", variant_path
+ if not GetOption('silent'):
+ print("Building in", variant_path)
# Make a copy of the build-root environment to use for this config.
env = main.Clone()
current_vars_file = joinpath(build_root, 'variables', variant_dir)
if isfile(current_vars_file):
sticky_vars.files.append(current_vars_file)
- print "Using saved variables file %s" % current_vars_file
+ if not GetOption('silent'):
+ print("Using saved variables file %s" % current_vars_file)
+ elif variant_dir in ext_build_dirs:
+ # Things in ext are built without a variant directory.
+ continue
else:
# Build dir-specific variables file doesn't exist.
# Get default build variables from source tree. Variables are
# normally determined by name of $VARIANT_DIR, but can be
- # overriden by 'default=' arg on command line.
- default_vars_file = joinpath('build_opts',
- ARGUMENTS.get('default', variant_dir))
- if isfile(default_vars_file):
+ # overridden by '--default=' arg on command line.
+ default = GetOption('default')
+ opts_dir = joinpath(main.root.abspath, 'build_opts')
+ if default:
+ default_vars_files = [joinpath(build_root, 'variables', default),
+ joinpath(opts_dir, default)]
+ else:
+ default_vars_files = [joinpath(opts_dir, variant_dir)]
+ existing_files = filter(isfile, default_vars_files)
+ if existing_files:
+ default_vars_file = existing_files[0]
sticky_vars.files.append(default_vars_file)
- print "Variables file %s not found,\n using defaults in %s" \
- % (current_vars_file, default_vars_file)
+ print("Variables file %s not found,\n using defaults in %s"
+ % (current_vars_file, default_vars_file))
else:
- print "Error: cannot find variables file %s or %s" \
- % (current_vars_file, default_vars_file)
+ print("Error: cannot find variables file %s or "
+ "default file(s) %s"
+ % (current_vars_file, ' or '.join(default_vars_files)))
Exit(1)
# Apply current variable settings to env
sticky_vars.Update(env)
- nonsticky_vars.Update(env)
- help_text += "\nSticky variables for %s:\n" % variant_dir \
- + sticky_vars.GenerateHelpText(env) \
- + "\nNon-sticky variables for %s:\n" % variant_dir \
- + nonsticky_vars.GenerateHelpText(env)
+ help_texts["local_vars"] += \
+ "Build variables for %s:\n" % variant_dir \
+ + sticky_vars.GenerateHelpText(env)
# Process variable settings.
if not have_fenv and env['USE_FENV']:
- print "Warning: <fenv.h> not available; " \
- "forcing USE_FENV to False in", variant_dir + "."
+ print("Warning: <fenv.h> not available; "
+ "forcing USE_FENV to False in", variant_dir + ".")
env['USE_FENV'] = False
if not env['USE_FENV']:
- print "Warning: No IEEE FP rounding mode control in", variant_dir + "."
- print " FP results may deviate slightly from other platforms."
+ print("Warning: No IEEE FP rounding mode control in",
+ variant_dir + ".")
+ print(" FP results may deviate slightly from other platforms.")
+
+ if not have_png and env['USE_PNG']:
+ print("Warning: <png.h> not available; "
+ "forcing USE_PNG to False in", variant_dir + ".")
+ env['USE_PNG'] = False
+
+ if env['USE_PNG']:
+ env.Append(LIBS=['png'])
if env['EFENCE']:
env.Append(LIBS=['efence'])
- if env['USE_MYSQL']:
- if not have_mysql:
- print "Warning: MySQL not available; " \
- "forcing USE_MYSQL to False in", variant_dir + "."
- env['USE_MYSQL'] = False
- else:
- print "Compiling in", variant_dir, "with MySQL support."
- env.ParseConfig(mysql_config_libs)
- env.ParseConfig(mysql_config_include)
+ if env['USE_KVM']:
+ if not have_kvm:
+ print("Warning: Can not enable KVM, host seems to "
+ "lack KVM support")
+ env['USE_KVM'] = False
+ elif not is_isa_kvm_compatible(env['TARGET_ISA']):
+ print("Info: KVM support disabled due to unsupported host and "
+ "target ISA combination")
+ env['USE_KVM'] = False
+
+ if env['USE_TUNTAP']:
+ if not have_tuntap:
+ print("Warning: Can't connect EtherTap with a tap device.")
+ env['USE_TUNTAP'] = False
+
+ if env['BUILD_GPU']:
+ env.Append(CPPDEFINES=['BUILD_GPU'])
+
+ # Warn about missing optional functionality
+ if env['USE_KVM']:
+ if not main['HAVE_PERF_ATTR_EXCLUDE_HOST']:
+ print("Warning: perf_event headers lack support for the "
+ "exclude_host attribute. KVM instruction counts will "
+ "be inaccurate.")
# Save sticky variable settings back to current variables file
sticky_vars.Save(current_vars_file, env)
if env['USE_SSE2']:
- env.Append(CCFLAGS='-msse2')
+ env.Append(CCFLAGS=['-msse2'])
# The src/SConscript file sets up the build rules in 'env' according
# to the configured variables. It returns a list of environments,
# one for each variant build (debug, opt, etc.)
- envList = SConscript('src/SConscript', variant_dir = variant_path,
- exports = 'env')
+ SConscript('src/SConscript', variant_dir = variant_path, exports = 'env')
+
+# base help text
+Help('''
+Usage: scons [scons options] [build variables] [target(s)]
+
+Extra scons options:
+%(options)s
- # Set up the regression tests for each build.
- for e in envList:
- SConscript('tests/SConscript',
- variant_dir = joinpath(variant_path, 'tests', e.Label),
- exports = { 'env' : e }, duplicate = False)
+Global build variables:
+%(global_vars)s
-Help(help_text)
+%(local_vars)s
+''' % help_texts)