# -*- mode:python -*-
+# Copyright (c) 2011 Advanced Micro Devices, Inc.
# Copyright (c) 2009 The Hewlett-Packard Development Company
# Copyright (c) 2004-2005 The Regents of The University of Michigan
# All rights reserved.
"""
raise
+# Global Python includes
import os
import re
import subprocess
from os.path import exists, isdir, isfile
from os.path import join as joinpath, split as splitpath
+# SCons includes
import SCons
import SCons.Node
-def read_command(cmd, **kwargs):
- """run the command cmd, read the results and return them
- this is sorta like `cmd` in shell"""
- from subprocess import Popen, PIPE, STDOUT
+extra_python_paths = [
+ Dir('src/python').srcnode().abspath, # M5 includes
+ Dir('ext/ply').srcnode().abspath, # ply is used by several files
+ ]
+
+sys.path[1:1] = extra_python_paths
- if isinstance(cmd, str):
- cmd = cmd.split()
+from m5.util import compareVersions, readCommand
- no_exception = 'exception' in kwargs
- exception = kwargs.pop('exception', None)
-
- kwargs.setdefault('shell', False)
- kwargs.setdefault('stdout', PIPE)
- kwargs.setdefault('stderr', STDOUT)
- kwargs.setdefault('close_fds', True)
- try:
- subp = Popen(cmd, **kwargs)
- except Exception, e:
- if no_exception:
- return exception
- raise
-
- return subp.communicate()[0]
-
-# helper function: compare arrays or strings of version numbers.
-# E.g., compare_version((1,3,25), (1,4,1)')
-# returns -1, 0, 1 if v1 is <, ==, > v2
-def compare_versions(v1, v2):
- def make_version_list(v):
- if isinstance(v, (list,tuple)):
- return v
- elif isinstance(v, str):
- return map(lambda x: int(re.match('\d+', x).group()), v.split('.'))
+help_texts = {
+ "options" : "",
+ "global_vars" : "",
+ "local_vars" : ""
+}
+
+Export("help_texts")
+
+def AddM5Option(*args, **kwargs):
+ col_width = 30
+
+ help = " " + ", ".join(args)
+ if "help" in kwargs:
+ length = len(help)
+ if length >= col_width:
+ help += "\n" + " " * col_width
else:
- raise TypeError
-
- v1 = make_version_list(v1)
- v2 = make_version_list(v2)
- # Compare corresponding elements of lists
- for n1,n2 in zip(v1, v2):
- if n1 < n2: return -1
- if n1 > n2: return 1
- # all corresponding values are equal... see if one has extra values
- if len(v1) < len(v2): return -1
- if len(v1) > len(v2): return 1
- return 0
+ help += " " * (col_width - length)
+ help += kwargs["help"]
+ help_texts["options"] += help + "\n"
+
+ AddOption(*args, **kwargs)
+
+AddM5Option('--colors', dest='use_colors', action='store_true',
+ help="Add color to abbreviated scons output")
+AddM5Option('--no-colors', dest='use_colors', action='store_false',
+ help="Don't add color to abbreviated scons output")
+AddM5Option('--default', dest='default', type='string', action='store',
+ help='Override which build_opts file to use for defaults')
+AddM5Option('--ignore-style', dest='ignore_style', action='store_true',
+ help='Disable style checking hooks')
+AddM5Option('--update-ref', dest='update_ref', action='store_true',
+ help='Update test reference outputs')
+AddM5Option('--verbose', dest='verbose', action='store_true',
+ help='Print full tool command lines')
+
+use_colors = GetOption('use_colors')
+if use_colors:
+ from m5.util.terminal import termcap
+elif use_colors is None:
+ # option unspecified; default behavior is to use colors iff isatty
+ from m5.util.terminal import tty_termcap as termcap
+else:
+ from m5.util.terminal import no_termcap as termcap
########################################################################
#
-# Set up the base build environment.
+# Set up the main build environment.
#
########################################################################
-use_vars = set([ 'AS', 'AR', 'CC', 'CXX', 'HOME', 'PATH', 'RANLIB' ])
+use_vars = set([ 'AS', 'AR', 'CC', 'CXX', 'HOME', 'LD_LIBRARY_PATH', 'PATH',
+ 'PYTHONPATH', 'RANLIB' ])
use_env = {}
for key,val in os.environ.iteritems():
if key in use_vars or key.startswith("M5"):
use_env[key] = val
-env = Environment(ENV=use_env)
-env.root = Dir(".") # The current directory (where this file lives).
-env.srcdir = Dir("src") # The source directory
+main = Environment(ENV=use_env)
+main.root = Dir(".") # The current directory (where this file lives).
+main.srcdir = Dir("src") # The source directory
+
+# add useful python code PYTHONPATH so it can be used by subprocesses
+# as well
+main.AppendENVPath('PYTHONPATH', extra_python_paths)
########################################################################
#
#
########################################################################
-hgdir = env.root.Dir(".hg")
+hgdir = main.root.Dir(".hg")
mercurial_style_message = """
-You're missing the M5 style hook.
-Please install the hook so we can ensure that all code fits a common style.
-
-All you'd need to do is add the following lines to your repository .hg/hgrc
-or your personal .hgrc
-----------------
-
+You're missing the gem5 style hook, which automatically checks your code
+against the gem5 style rules on hg commit and qrefresh commands. This
+script will now install the hook in your .hg/hgrc file.
+Press enter to continue, or ctrl-c to abort: """
+
+mercurial_style_hook = """
+# The following lines were automatically added by gem5/SConstruct
+# to provide the gem5 style-checking hooks
[extensions]
style = %s/util/style.py
[hooks]
-pretxncommit.style = python:style.check_whitespace
-""" % (env.root)
-
-mercurial_bin_not_found = """
-Mercurial binary cannot be found, unfortunately this means that we
-cannot easily determine the version of M5 that you are running and
-this makes error messages more difficult to collect. Please consider
-installing mercurial if you choose to post an error message
-"""
+pretxncommit.style = python:style.check_style
+pre-qrefresh.style = python:style.check_style
+# End of SConstruct additions
+
+""" % (main.root.abspath)
mercurial_lib_not_found = """
-Mercurial libraries cannot be found, ignoring style hook
-If you are actually a M5 developer, please fix this and
-run the style hook. It is important.
+Mercurial libraries cannot be found, ignoring style hook. If
+you are a gem5 developer, please fix this and run the style
+hook. It is important.
"""
-if hgdir.exists():
- # 1) Grab repository revision if we know it.
- cmd = "hg id -n -i -t -b"
- try:
- hg_info = read_command(cmd, cwd=env.root.abspath).strip()
- except OSError:
- hg_info = "Unknown"
- print mercurial_bin_not_found
-
- env['HG_INFO'] = hg_info
-
- # 2) Ensure that the style hook is in place.
+# Check for style hook and prompt for installation if it's not there.
+# Skip this if --ignore-style was specified, there's no .hg dir to
+# install a hook in, or there's no interactive terminal to prompt.
+if not GetOption('ignore_style') and hgdir.exists() and sys.stdin.isatty():
+ style_hook = True
try:
- ui = None
- if ARGUMENTS.get('IGNORE_STYLE') != 'True':
- from mercurial import ui
- ui = ui.ui()
+ from mercurial import ui
+ ui = ui.ui()
+ ui.readconfig(hgdir.File('hgrc').abspath)
+ style_hook = ui.config('hooks', 'pretxncommit.style', None) and \
+ ui.config('hooks', 'pre-qrefresh.style', None)
except ImportError:
print mercurial_lib_not_found
- if ui is not None:
- ui.readconfig(hgdir.File('hgrc').abspath)
- style_hook = ui.config('hooks', 'pretxncommit.style', None)
-
- if not style_hook:
- print mercurial_style_message
+ if not style_hook:
+ print mercurial_style_message,
+ # continue unless user does ctrl-c/ctrl-d etc.
+ try:
+ raw_input()
+ except:
+ print "Input exception, exiting scons.\n"
+ sys.exit(1)
+ hgrc_path = '%s/.hg/hgrc' % main.root.abspath
+ print "Adding style hook to", hgrc_path, "\n"
+ try:
+ hgrc = open(hgrc_path, 'a')
+ hgrc.write(mercurial_style_hook)
+ hgrc.close()
+ except:
+ print "Error updating", hgrc_path
sys.exit(1)
-else:
- print ".hg directory not found"
+
###################################################
#
return i
raise ValueError, "element not found"
+# Take a list of paths (or SCons Nodes) and return a list with all
+# paths made absolute and ~-expanded. Paths will be interpreted
+# relative to the launch directory unless a different root is provided
+def makePathListAbsolute(path_list, root=GetLaunchDir()):
+ return [abspath(joinpath(root, expanduser(str(p))))
+ for p in path_list]
+
# Each target must have 'build' in the interior of the path; the
# directory below this will determine the build parameters. For
# example, for target 'foo/bar/build/ALPHA_SE/arch/alpha/blah.do' we
# recognize that ALPHA_SE specifies the configuration because it
-# follow 'build' in the bulid path.
-
-# Generate absolute paths to targets so we can see where the build dir is
-if COMMAND_LINE_TARGETS:
- # Ask SCons which directory it was invoked from
- launch_dir = GetLaunchDir()
- # Make targets relative to invocation directory
- abs_targets = [ normpath(joinpath(launch_dir, str(x))) for x in \
- COMMAND_LINE_TARGETS]
-else:
- # Default targets are relative to root of tree
- abs_targets = [ normpath(joinpath(ROOT, str(x))) for x in \
- DEFAULT_TARGETS]
+# follow 'build' in the build path.
+# The funky assignment to "[:]" is needed to replace the list contents
+# in place rather than reassign the symbol to a new list, which
+# doesn't work (obviously!).
+BUILD_TARGETS[:] = makePathListAbsolute(BUILD_TARGETS)
# Generate a list of the unique build roots and configs that the
# collected targets reference.
variant_paths = []
build_root = None
-for t in abs_targets:
+for t in BUILD_TARGETS:
path_dirs = t.split('/')
try:
build_top = rfind(path_dirs, 'build', -2)
# Make sure build_root exists (might not if this is the first build there)
if not isdir(build_root):
mkdir(build_root)
+main['BUILDROOT'] = build_root
-Export('env')
+Export('main')
-env.SConsignFile(joinpath(build_root, "sconsign"))
+main.SConsignFile(joinpath(build_root, "sconsign"))
# Default duplicate option is to use hard links, but this messes up
# when you use emacs to edit a file in the target dir, as emacs moves
# file to file~ then copies to file, breaking the link. Symbolic
# (soft) links work better.
-env.SetOption('duplicate', 'soft-copy')
+main.SetOption('duplicate', 'soft-copy')
#
# Set up global sticky variables... these are common to an entire build
# tree (not specific to a particular build like ALPHA_SE)
#
-# Variable validators & converters for global sticky variables
-def PathListMakeAbsolute(val):
- if not val:
- return val
- f = lambda p: abspath(expanduser(p))
- return ':'.join(map(f, val.split(':')))
-
-def PathListAllExist(key, val, env):
- if not val:
- return
- paths = val.split(':')
- for path in paths:
- if not isdir(path):
- raise SCons.Errors.UserError("Path does not exist: '%s'" % path)
+global_vars_file = joinpath(build_root, 'variables.global')
-global_sticky_vars_file = joinpath(build_root, 'variables.global')
+global_vars = Variables(global_vars_file, args=ARGUMENTS)
-global_sticky_vars = Variables(global_sticky_vars_file, args=ARGUMENTS)
-
-global_sticky_vars.AddVariables(
- ('CC', 'C compiler', environ.get('CC', env['CC'])),
- ('CXX', 'C++ compiler', environ.get('CXX', env['CXX'])),
+global_vars.AddVariables(
+ ('CC', 'C compiler', environ.get('CC', main['CC'])),
+ ('CXX', 'C++ compiler', environ.get('CXX', main['CXX'])),
('BATCH', 'Use batch pool for build and tests', False),
('BATCH_CMD', 'Batch pool submission command name', 'qdo'),
- ('EXTRAS', 'Add Extra directories to the compilation', '',
- PathListAllExist, PathListMakeAbsolute)
- )
-
-# base help text
-help_text = '''
-Usage: scons [scons options] [build options] [target(s)]
-
-Global sticky options:
-'''
-
-help_text += global_sticky_vars.GenerateHelpText(env)
+ ('M5_BUILD_CACHE', 'Cache built objects in this directory', False),
+ ('EXTRAS', 'Add extra directories to the compilation', '')
+ )
-# Update env with values from ARGUMENTS & file global_sticky_vars_file
-global_sticky_vars.Update(env)
+# Update main environment with values from ARGUMENTS & global_vars_file
+global_vars.Update(main)
+help_texts["global_vars"] += global_vars.GenerateHelpText(main)
# Save sticky variable settings back to current variables file
-global_sticky_vars.Save(global_sticky_vars_file, env)
+global_vars.Save(global_vars_file, main)
# Parse EXTRAS variable to build list of all directories where we're
-# look for sources etc. This list is exported as base_dir_list.
-base_dir = env.srcdir.abspath
-if env['EXTRAS']:
- extras_dir_list = env['EXTRAS'].split(':')
+# look for sources etc. This list is exported as extras_dir_list.
+base_dir = main.srcdir.abspath
+if main['EXTRAS']:
+ extras_dir_list = makePathListAbsolute(main['EXTRAS'].split(':'))
else:
extras_dir_list = []
Export('base_dir')
Export('extras_dir_list')
-# M5_PLY is used by isa_parser.py to find the PLY package.
-env.Append(ENV = { 'M5_PLY' : str(Dir('ext/ply')) })
-
-CXX_version = read_command([env['CXX'],'--version'], exception=False)
-CXX_V = read_command([env['CXX'],'-V'], exception=False)
-
-env['GCC'] = CXX_version and CXX_version.find('g++') >= 0
-env['SUNCC'] = CXX_V and CXX_V.find('Sun C++') >= 0
-env['ICC'] = CXX_V and CXX_V.find('Intel') >= 0
-if env['GCC'] + env['SUNCC'] + env['ICC'] > 1:
+# the ext directory should be on the #includes path
+main.Append(CPPPATH=[Dir('ext')])
+
+def strip_build_path(path, env):
+ path = str(path)
+ variant_base = env['BUILDROOT'] + os.path.sep
+ if path.startswith(variant_base):
+ path = path[len(variant_base):]
+ elif path.startswith('build/'):
+ path = path[6:]
+ return path
+
+# Generate a string of the form:
+# common/path/prefix/src1, src2 -> tgt1, tgt2
+# to print while building.
+class Transform(object):
+ # all specific color settings should be here and nowhere else
+ tool_color = termcap.Normal
+ pfx_color = termcap.Yellow
+ srcs_color = termcap.Yellow + termcap.Bold
+ arrow_color = termcap.Blue + termcap.Bold
+ tgts_color = termcap.Yellow + termcap.Bold
+
+ def __init__(self, tool, max_sources=99):
+ self.format = self.tool_color + (" [%8s] " % tool) \
+ + self.pfx_color + "%s" \
+ + self.srcs_color + "%s" \
+ + self.arrow_color + " -> " \
+ + self.tgts_color + "%s" \
+ + termcap.Normal
+ self.max_sources = max_sources
+
+ def __call__(self, target, source, env, for_signature=None):
+ # truncate source list according to max_sources param
+ source = source[0:self.max_sources]
+ def strip(f):
+ return strip_build_path(str(f), env)
+ if len(source) > 0:
+ srcs = map(strip, source)
+ else:
+ srcs = ['']
+ tgts = map(strip, target)
+ # surprisingly, os.path.commonprefix is a dumb char-by-char string
+ # operation that has nothing to do with paths.
+ com_pfx = os.path.commonprefix(srcs + tgts)
+ com_pfx_len = len(com_pfx)
+ if com_pfx:
+ # do some cleanup and sanity checking on common prefix
+ if com_pfx[-1] == ".":
+ # prefix matches all but file extension: ok
+ # back up one to change 'foo.cc -> o' to 'foo.cc -> .o'
+ com_pfx = com_pfx[0:-1]
+ elif com_pfx[-1] == "/":
+ # common prefix is directory path: OK
+ pass
+ else:
+ src0_len = len(srcs[0])
+ tgt0_len = len(tgts[0])
+ if src0_len == com_pfx_len:
+ # source is a substring of target, OK
+ pass
+ elif tgt0_len == com_pfx_len:
+ # target is a substring of source, need to back up to
+ # avoid empty string on RHS of arrow
+ sep_idx = com_pfx.rfind(".")
+ if sep_idx != -1:
+ com_pfx = com_pfx[0:sep_idx]
+ else:
+ com_pfx = ''
+ elif src0_len > com_pfx_len and srcs[0][com_pfx_len] == ".":
+ # still splitting at file extension: ok
+ pass
+ else:
+ # probably a fluke; ignore it
+ com_pfx = ''
+ # recalculate length in case com_pfx was modified
+ com_pfx_len = len(com_pfx)
+ def fmt(files):
+ f = map(lambda s: s[com_pfx_len:], files)
+ return ', '.join(f)
+ return self.format % (com_pfx, fmt(srcs), fmt(tgts))
+
+Export('Transform')
+
+
+if GetOption('verbose'):
+ def MakeAction(action, string, *args, **kwargs):
+ return Action(action, *args, **kwargs)
+else:
+ MakeAction = Action
+ main['CCCOMSTR'] = Transform("CC")
+ main['CXXCOMSTR'] = Transform("CXX")
+ main['ASCOMSTR'] = Transform("AS")
+ main['SWIGCOMSTR'] = Transform("SWIG")
+ main['ARCOMSTR'] = Transform("AR", 0)
+ main['LINKCOMSTR'] = Transform("LINK", 0)
+ main['RANLIBCOMSTR'] = Transform("RANLIB", 0)
+ main['M4COMSTR'] = Transform("M4")
+ main['SHCCCOMSTR'] = Transform("SHCC")
+ main['SHCXXCOMSTR'] = Transform("SHCXX")
+Export('MakeAction')
+
+CXX_version = readCommand([main['CXX'],'--version'], exception=False)
+CXX_V = readCommand([main['CXX'],'-V'], exception=False)
+
+main['GCC'] = CXX_version and CXX_version.find('g++') >= 0
+main['SUNCC'] = CXX_V and CXX_V.find('Sun C++') >= 0
+main['ICC'] = CXX_V and CXX_V.find('Intel') >= 0
+if main['GCC'] + main['SUNCC'] + main['ICC'] > 1:
print 'Error: How can we have two at the same time?'
Exit(1)
# Set up default C++ compiler flags
-if env['GCC']:
- env.Append(CCFLAGS='-pipe')
- env.Append(CCFLAGS='-fno-strict-aliasing')
- env.Append(CCFLAGS=Split('-Wall -Wno-sign-compare -Werror -Wundef'))
- env.Append(CXXFLAGS='-Wno-deprecated')
-elif env['ICC']:
+if main['GCC']:
+ main.Append(CCFLAGS=['-pipe'])
+ main.Append(CCFLAGS=['-fno-strict-aliasing'])
+ main.Append(CCFLAGS=['-Wall', '-Wno-sign-compare', '-Wundef'])
+ main.Append(CXXFLAGS=['-Wno-deprecated'])
+ # Read the GCC version to check for versions with bugs
+ # Note CCVERSION doesn't work here because it is run with the CC
+ # before we override it from the command line
+ gcc_version = readCommand([main['CXX'], '-dumpversion'], exception=False)
+ main['GCC_VERSION'] = gcc_version
+ if not compareVersions(gcc_version, '4.4.1') or \
+ not compareVersions(gcc_version, '4.4.2'):
+ print 'Info: Tree vectorizer in GCC 4.4.1 & 4.4.2 is buggy, disabling.'
+ main.Append(CCFLAGS=['-fno-tree-vectorize'])
+elif main['ICC']:
pass #Fix me... add warning flags once we clean up icc warnings
-elif env['SUNCC']:
- env.Append(CCFLAGS='-Qoption ccfe')
- env.Append(CCFLAGS='-features=gcc')
- env.Append(CCFLAGS='-features=extensions')
- env.Append(CCFLAGS='-library=stlport4')
- env.Append(CCFLAGS='-xar')
- #env.Append(CCFLAGS='-instances=semiexplicit')
+elif main['SUNCC']:
+ main.Append(CCFLAGS=['-Qoption ccfe'])
+ main.Append(CCFLAGS=['-features=gcc'])
+ main.Append(CCFLAGS=['-features=extensions'])
+ main.Append(CCFLAGS=['-library=stlport4'])
+ main.Append(CCFLAGS=['-xar'])
+ #main.Append(CCFLAGS=['-instances=semiexplicit'])
else:
print 'Error: Don\'t know what compiler options to use for your compiler.'
print ' Please fix SConstruct and src/SConscript and try again.'
Exit(1)
+# Set up common yacc/bison flags (needed for Ruby)
+main['YACCFLAGS'] = '-d'
+main['YACCHXXFILESUFFIX'] = '.hh'
+
# Do this after we save setting back, or else we'll tack on an
# extra 'qdo' every time we run scons.
-if env['BATCH']:
- env['CC'] = env['BATCH_CMD'] + ' ' + env['CC']
- env['CXX'] = env['BATCH_CMD'] + ' ' + env['CXX']
- env['AS'] = env['BATCH_CMD'] + ' ' + env['AS']
- env['AR'] = env['BATCH_CMD'] + ' ' + env['AR']
- env['RANLIB'] = env['BATCH_CMD'] + ' ' + env['RANLIB']
+if main['BATCH']:
+ main['CC'] = main['BATCH_CMD'] + ' ' + main['CC']
+ main['CXX'] = main['BATCH_CMD'] + ' ' + main['CXX']
+ main['AS'] = main['BATCH_CMD'] + ' ' + main['AS']
+ main['AR'] = main['BATCH_CMD'] + ' ' + main['AR']
+ main['RANLIB'] = main['BATCH_CMD'] + ' ' + main['RANLIB']
if sys.platform == 'cygwin':
# cygwin has some header file issues...
- env.Append(CCFLAGS=Split("-Wno-uninitialized"))
-env.Append(CPPPATH=[Dir('ext/dnet')])
+ main.Append(CCFLAGS=["-Wno-uninitialized"])
# Check for SWIG
-if not env.has_key('SWIG'):
+if not main.has_key('SWIG'):
print 'Error: SWIG utility not found.'
print ' Please install (see http://www.swig.org) and retry.'
Exit(1)
# Check for appropriate SWIG version
-swig_version = read_command(('swig', '-version'), exception='').split()
+swig_version = readCommand(('swig', '-version'), exception='').split()
# First 3 words should be "SWIG Version x.y.z"
if len(swig_version) < 3 or \
swig_version[0] != 'SWIG' or swig_version[1] != 'Version':
Exit(1)
min_swig_version = '1.3.28'
-if compare_versions(swig_version[2], min_swig_version) < 0:
+if compareVersions(swig_version[2], min_swig_version) < 0:
print 'Error: SWIG version', min_swig_version, 'or newer required.'
print ' Installed version:', swig_version[2]
Exit(1)
# Set up SWIG flags & scanner
swig_flags=Split('-c++ -python -modern -templatereduce $_CPPINCFLAGS')
-env.Append(SWIGFLAGS=swig_flags)
+main.Append(SWIGFLAGS=swig_flags)
# filter out all existing swig scanners, they mess up the dependency
# stuff for some reason
scanners = []
-for scanner in env['SCANNERS']:
+for scanner in main['SCANNERS']:
skeys = scanner.skeys
if skeys == '.i':
continue
scanners.append(CPPScanner("SwigScan", [ ".i" ], "CPPPATH", swig_inc_re))
# replace the scanners list that has what we want
-env['SCANNERS'] = scanners
+main['SCANNERS'] = scanners
# Add a custom Check function to the Configure context so that we can
# figure out if the compiler adds leading underscores to global
# Platform-specific configuration. Note again that we assume that all
# builds under a given build root run on the same host platform.
-conf = Configure(env,
+conf = Configure(main,
conf_dir = joinpath(build_root, '.scons_config'),
log_file = joinpath(build_root, 'scons_config.log'),
custom_tests = { 'CheckLeading' : CheckLeading })
try:
import platform
uname = platform.uname()
- if uname[0] == 'Darwin' and compare_versions(uname[2], '9.0.0') >= 0:
- if int(read_command('sysctl -n hw.cpu64bit_capable')[0]):
- env.Append(CCFLAGS='-arch x86_64')
- env.Append(CFLAGS='-arch x86_64')
- env.Append(LINKFLAGS='-arch x86_64')
- env.Append(ASFLAGS='-arch x86_64')
+ if uname[0] == 'Darwin' and compareVersions(uname[2], '9.0.0') >= 0:
+ if int(readCommand('sysctl -n hw.cpu64bit_capable')[0]):
+ main.Append(CCFLAGS=['-arch', 'x86_64'])
+ main.Append(CFLAGS=['-arch', 'x86_64'])
+ main.Append(LINKFLAGS=['-arch', 'x86_64'])
+ main.Append(ASFLAGS=['-arch', 'x86_64'])
except:
pass
def __getattr__(self, mname):
return NullCheck
- conf = NullConf(env)
+ conf = NullConf(main)
# Find Python include and library directories for embedding the
# interpreter. For consistency, we will use the same Python
py_getvar = sysconfig.get_config_var
-py_version = 'python' + py_getvar('VERSION')
+py_debug = getattr(sys, 'pydebug', False)
+py_version = 'python' + py_getvar('VERSION') + (py_debug and "_d" or "")
py_general_include = sysconfig.get_python_inc()
py_platform_include = sysconfig.get_python_inc(plat_specific=True)
if py_platform_include != py_general_include:
py_includes.append(py_platform_include)
-py_lib_path = []
+py_lib_path = [ py_getvar('LIBDIR') ]
# add the prefix/lib/pythonX.Y/config dir, but only if there is no
# shared library in prefix/lib/.
if not py_getvar('Py_ENABLE_SHARED'):
- py_lib_path.append('-L' + py_getvar('LIBPL'))
+ py_lib_path.append(py_getvar('LIBPL'))
py_libs = []
for lib in py_getvar('LIBS').split() + py_getvar('SYSLIBS').split():
+ assert lib.startswith('-l')
+ lib = lib[2:]
if lib not in py_libs:
py_libs.append(lib)
-py_libs.append('-l' + py_version)
+py_libs.append(py_version)
+
+main.Append(CPPPATH=py_includes)
+main.Append(LIBPATH=py_lib_path)
+
+# Cache build files in the supplied directory.
+if main['M5_BUILD_CACHE']:
+ print 'Using build cache located at', main['M5_BUILD_CACHE']
+ CacheDir(main['M5_BUILD_CACHE'])
-env.Append(CPPPATH=py_includes)
-env.Append(LIBPATH=py_lib_path)
-#env.Append(LIBS=py_libs)
# verify that this stuff works
if not conf.CheckHeader('Python.h', '<>'):
Exit(1)
for lib in py_libs:
- assert lib.startswith('-l')
- lib = lib[2:]
if not conf.CheckLib(lib):
print "Error: can't find library %s required by python" % lib
Exit(1)
print ' Please install zlib and try again.'
Exit(1)
+# Check for librt.
+have_posix_clock = \
+ conf.CheckLibWithHeader(None, 'time.h', 'C',
+ 'clock_nanosleep(0,0,NULL,NULL);') or \
+ conf.CheckLibWithHeader('rt', 'time.h', 'C',
+ 'clock_nanosleep(0,0,NULL,NULL);')
+
+if not have_posix_clock:
+ print "Can't find library for POSIX clocks."
+
# Check for <fenv.h> (C99 FP environment control)
have_fenv = conf.CheckHeader('fenv.h', '<>')
if not have_fenv:
print "Warning: Header file <fenv.h> not found."
print " This host has no IEEE FP rounding mode control."
-######################################################################
-#
-# Check for mysql.
-#
-mysql_config = WhereIs('mysql_config')
-have_mysql = bool(mysql_config)
-
-# Check MySQL version.
-if have_mysql:
- mysql_version = read_command(mysql_config + ' --version')
- min_mysql_version = '4.1'
- if compare_versions(mysql_version, min_mysql_version) < 0:
- print 'Warning: MySQL', min_mysql_version, 'or newer required.'
- print ' Version', mysql_version, 'detected.'
- have_mysql = False
-
-# Set up mysql_config commands.
-if have_mysql:
- mysql_config_include = mysql_config + ' --include'
- if os.system(mysql_config_include + ' > /dev/null') != 0:
- # older mysql_config versions don't support --include, use
- # --cflags instead
- mysql_config_include = mysql_config + ' --cflags | sed s/\\\'//g'
- # This seems to work in all versions
- mysql_config_libs = mysql_config + ' --libs'
-
######################################################################
#
# Finish the configuration
#
-env = conf.Finish()
+main = conf.Finish()
######################################################################
#
# Collect all non-global variables
#
-Export('env')
-
# Define the universe of supported ISAs
all_isa_list = [ ]
Export('all_isa_list')
-# Define the universe of supported CPU models
-all_cpu_list = [ ]
-default_cpus = [ ]
-Export('all_cpu_list', 'default_cpus')
+class CpuModel(object):
+ '''The CpuModel class encapsulates everything the ISA parser needs to
+ know about a particular CPU model.'''
+
+ # Dict of available CPU model objects. Accessible as CpuModel.dict.
+ dict = {}
+ list = []
+ defaults = []
+
+ # Constructor. Automatically adds models to CpuModel.dict.
+ def __init__(self, name, filename, includes, strings, default=False):
+ self.name = name # name of model
+ self.filename = filename # filename for output exec code
+ self.includes = includes # include files needed in exec file
+ # The 'strings' dict holds all the per-CPU symbols we can
+ # substitute into templates etc.
+ self.strings = strings
+
+ # This cpu is enabled by default
+ self.default = default
+
+ # Add self to dict
+ if name in CpuModel.dict:
+ raise AttributeError, "CpuModel '%s' already registered" % name
+ CpuModel.dict[name] = self
+ CpuModel.list.append(name)
+
+Export('CpuModel')
# Sticky variables get saved in the variables file so they persist from
# one invocation to the next (unless overridden, in which case the new
sticky_vars = Variables(args=ARGUMENTS)
Export('sticky_vars')
-# Non-sticky variables only apply to the current build.
-nonsticky_vars = Variables(args=ARGUMENTS)
-Export('nonsticky_vars')
+# Sticky variables that should be exported
+export_vars = []
+Export('export_vars')
# Walk the tree and execute all SConsopts scripts that wil add to the
# above variables
+if not GetOption('verbose'):
+ print "Reading SConsopts"
for bdir in [ base_dir ] + extras_dir_list:
+ if not isdir(bdir):
+ print "Error: directory '%s' does not exist" % bdir
+ Exit(1)
for root, dirs, files in os.walk(bdir):
if 'SConsopts' in files:
- print "Reading", joinpath(root, 'SConsopts')
+ if GetOption('verbose'):
+ print "Reading", joinpath(root, 'SConsopts')
SConscript(joinpath(root, 'SConsopts'))
all_isa_list.sort()
-all_cpu_list.sort()
-default_cpus.sort()
sticky_vars.AddVariables(
EnumVariable('TARGET_ISA', 'Target ISA', 'alpha', all_isa_list),
BoolVariable('FULL_SYSTEM', 'Full-system support', False),
- ListVariable('CPU_MODELS', 'CPU models', default_cpus, all_cpu_list),
+ ListVariable('CPU_MODELS', 'CPU models',
+ sorted(n for n,m in CpuModel.dict.iteritems() if m.default),
+ sorted(CpuModel.list)),
BoolVariable('NO_FAST_ALLOC', 'Disable fast object allocator', False),
- BoolVariable('FAST_ALLOC_DEBUG', 'Enable fast object allocator debugging',
- False),
+ BoolVariable('FORCE_FAST_ALLOC',
+ 'Enable fast object allocator, even for m5.debug', False),
BoolVariable('FAST_ALLOC_STATS', 'Enable fast object allocator statistics',
False),
BoolVariable('EFENCE', 'Link with Electric Fence malloc debugger',
BoolVariable('USE_SSE2',
'Compile for SSE2 (-msse2) to get IEEE FP on x86 hosts',
False),
- BoolVariable('USE_MYSQL', 'Use MySQL for stats output', have_mysql),
+ BoolVariable('USE_POSIX_CLOCK', 'Use POSIX Clocks', have_posix_clock),
BoolVariable('USE_FENV', 'Use <fenv.h> IEEE mode control', have_fenv),
BoolVariable('USE_CHECKER', 'Use checker for detailed CPU models', False),
BoolVariable('CP_ANNOTATE', 'Enable critical path annotation capability', False),
)
-nonsticky_vars.AddVariables(
- BoolVariable('update_ref', 'Update test reference outputs', False)
- )
-
# These variables get exported to #defines in config/*.hh (see src/SConscript).
-env.ExportVariables = ['FULL_SYSTEM', 'ALPHA_TLASER', 'USE_FENV', \
- 'USE_MYSQL', 'NO_FAST_ALLOC', 'FAST_ALLOC_DEBUG', \
- 'FAST_ALLOC_STATS', 'SS_COMPATIBLE_FP', \
- 'USE_CHECKER', 'TARGET_ISA', 'CP_ANNOTATE']
+export_vars += ['FULL_SYSTEM', 'USE_FENV',
+ 'NO_FAST_ALLOC', 'FORCE_FAST_ALLOC', 'FAST_ALLOC_STATS',
+ 'SS_COMPATIBLE_FP', 'USE_CHECKER', 'TARGET_ISA', 'CP_ANNOTATE',
+ 'USE_POSIX_CLOCK' ]
###################################################
#
f.close()
return None
-# Generate the message to be printed when building the config file.
-def build_config_file_string(target, source, env):
- (variable, value) = [s.get_contents() for s in source]
- return "Defining %s as %s in %s." % (variable, value, target[0])
-
# Combine the two functions into a scons Action object.
-config_action = Action(build_config_file, build_config_file_string)
+config_action = MakeAction(build_config_file, Transform("CONFIG H", 2))
# The emitter munges the source & target node lists to reflect what
# we're really doing.
config_builder = Builder(emitter = config_emitter, action = config_action)
-env.Append(BUILDERS = { 'ConfigFile' : config_builder })
+main.Append(BUILDERS = { 'ConfigFile' : config_builder })
# libelf build is shared across all configs in the build root.
-env.SConscript('ext/libelf/SConscript',
- variant_dir = joinpath(build_root, 'libelf'))
+main.SConscript('ext/libelf/SConscript',
+ variant_dir = joinpath(build_root, 'libelf'))
# gzstream build is shared across all configs in the build root.
-env.SConscript('ext/gzstream/SConscript',
- variant_dir = joinpath(build_root, 'gzstream'))
+main.SConscript('ext/gzstream/SConscript',
+ variant_dir = joinpath(build_root, 'gzstream'))
###################################################
#
#
###################################################
-env['ALL_ISA_LIST'] = all_isa_list
+main['ALL_ISA_LIST'] = all_isa_list
def make_switching_dir(dname, switch_headers, env):
# Generate the header. target[0] is the full path of the output
# header to generate. 'source' is a dummy variable, since we get the
# list of ISAs from env['ALL_ISA_LIST'].
def gen_switch_hdr(target, source, env):
fname = str(target[0])
- bname = basename(fname)
f = open(fname, 'w')
- f.write('#include "arch/isa_specific.hh"\n')
- cond = '#if'
- for isa in all_isa_list:
- f.write('%s THE_ISA == %s_ISA\n#include "%s/%s/%s"\n'
- % (cond, isa.upper(), dname, isa, bname))
- cond = '#elif'
- f.write('#else\n#error "THE_ISA not set"\n#endif\n')
+ isa = env['TARGET_ISA'].lower()
+ print >>f, '#include "%s/%s/%s"' % (dname, isa, basename(fname))
f.close()
- return 0
-
- # String to print when generating header
- def gen_switch_hdr_string(target, source, env):
- return "Generating switch header " + str(target[0])
# Build SCons Action object. 'varlist' specifies env vars that this
# action depends on; when env['ALL_ISA_LIST'] changes these actions
# should get re-executed.
- switch_hdr_action = Action(gen_switch_hdr, gen_switch_hdr_string,
- varlist=['ALL_ISA_LIST'])
+ switch_hdr_action = MakeAction(gen_switch_hdr,
+ Transform("GENERATE"), varlist=['ALL_ISA_LIST'])
# Instantiate actions for each header
for hdr in switch_headers:
#
###################################################
-# rename base env
-base_env = env
-
for variant_path in variant_paths:
print "Building in", variant_path
# Make a copy of the build-root environment to use for this config.
- env = base_env.Clone()
+ env = main.Clone()
env['BUILDDIR'] = variant_path
# variant_dir is the tail component of build path, and is used to
# Get default build variables from source tree. Variables are
# normally determined by name of $VARIANT_DIR, but can be
- # overriden by 'default=' arg on command line.
- default_vars_file = joinpath('build_opts',
- ARGUMENTS.get('default', variant_dir))
- if isfile(default_vars_file):
+ # overridden by '--default=' arg on command line.
+ default = GetOption('default')
+ opts_dir = joinpath(main.root.abspath, 'build_opts')
+ if default:
+ default_vars_files = [joinpath(build_root, 'variables', default),
+ joinpath(opts_dir, default)]
+ else:
+ default_vars_files = [joinpath(opts_dir, variant_dir)]
+ existing_files = filter(isfile, default_vars_files)
+ if existing_files:
+ default_vars_file = existing_files[0]
sticky_vars.files.append(default_vars_file)
print "Variables file %s not found,\n using defaults in %s" \
% (current_vars_file, default_vars_file)
else:
- print "Error: cannot find variables file %s or %s" \
- % (current_vars_file, default_vars_file)
+ print "Error: cannot find variables file %s or " \
+ "default file(s) %s" \
+ % (current_vars_file, ' or '.join(default_vars_files))
Exit(1)
# Apply current variable settings to env
sticky_vars.Update(env)
- nonsticky_vars.Update(env)
- help_text += "\nSticky variables for %s:\n" % variant_dir \
- + sticky_vars.GenerateHelpText(env) \
- + "\nNon-sticky variables for %s:\n" % variant_dir \
- + nonsticky_vars.GenerateHelpText(env)
+ help_texts["local_vars"] += \
+ "Build variables for %s:\n" % variant_dir \
+ + sticky_vars.GenerateHelpText(env)
# Process variable settings.
if env['EFENCE']:
env.Append(LIBS=['efence'])
- if env['USE_MYSQL']:
- if not have_mysql:
- print "Warning: MySQL not available; " \
- "forcing USE_MYSQL to False in", variant_dir + "."
- env['USE_MYSQL'] = False
- else:
- print "Compiling in", variant_dir, "with MySQL support."
- env.ParseConfig(mysql_config_libs)
- env.ParseConfig(mysql_config_include)
-
# Save sticky variable settings back to current variables file
sticky_vars.Save(current_vars_file, env)
if env['USE_SSE2']:
- env.Append(CCFLAGS='-msse2')
+ env.Append(CCFLAGS=['-msse2'])
# The src/SConscript file sets up the build rules in 'env' according
# to the configured variables. It returns a list of environments,
variant_dir = joinpath(variant_path, 'tests', e.Label),
exports = { 'env' : e }, duplicate = False)
-Help(help_text)
+# base help text
+Help('''
+Usage: scons [scons options] [build variables] [target(s)]
+
+Extra scons options:
+%(options)s
+
+Global build variables:
+%(global_vars)s
+
+%(local_vars)s
+''' % help_texts)