1 |
# Copyright 2006 by ACcESS MNRF |
# Copyright 2006 by ACcESS MNRF |
|
# |
|
|
# http://www.access.edu.au |
|
|
# Primary Business: Queensland, Australia |
|
|
# Licensed under the Open Software License version 3.0 |
|
|
# http://www.opensource.org/licenses/osl-3.0.php |
|
|
# |
|
|
# |
|
2 |
# |
# |
3 |
|
# http://www.access.edu.au |
4 |
|
# Primary Business: Queensland, Australia |
5 |
|
# Licensed under the Open Software License version 3.0 |
6 |
|
# http://www.opensource.org/licenses/osl-3.0.php |
7 |
|
|
|
# top-level Scons configuration file for all esys13 modules |
|
|
# Begin initialisation Section |
|
|
# all of this section just intialises default environments and helper |
|
|
# scripts. You shouldn't need to modify this section. |
|
8 |
EnsureSConsVersion(0,96,91) |
EnsureSConsVersion(0,96,91) |
9 |
EnsurePythonVersion(2,3) |
EnsurePythonVersion(2,3) |
10 |
|
|
11 |
# import tools: |
import sys, os, re, socket |
12 |
import glob |
|
|
import sys, os |
|
13 |
# Add our extensions |
# Add our extensions |
14 |
if sys.path.count('scons')==0: sys.path.append('scons') |
if os.path.isdir('scons'): sys.path.append('scons') |
15 |
import scons_extensions |
import scons_extensions |
16 |
|
|
17 |
# Default options and options help text |
# Use /usr/lib64 if available, else /usr/lib |
18 |
# These are defaults and can be overridden using command line arguments or an options file. |
usr_lib = '/usr/lib' |
19 |
# if the options_file or ARGUMENTS do not exist then the ones listed as default here are used |
if os.path.isfile('/usr/lib64/libc.so'): usr_lib = '/usr/lib64' |
20 |
# DO NOT CHANGE THEM HERE |
|
21 |
if ARGUMENTS.get('options_file',0): |
# The string python2.4 or python2.5 |
22 |
options_file = ARGUMENTS.get('options_file',0) |
python_version = 'python%s.%s' % (sys.version_info[0], sys.version_info[1]) |
23 |
else: |
|
24 |
import socket |
# MS Windows support, many thanks to PH |
25 |
from string import ascii_letters,digits |
IS_WINDOWS_PLATFORM = (os.name== "nt") |
26 |
hostname="" |
|
27 |
for s in socket.gethostname().split('.')[0]: |
prefix = ARGUMENTS.get('prefix', Dir('#.').abspath) |
28 |
if s in ascii_letters+digits: |
|
29 |
hostname+=s |
# Read configuration options from file scons/<hostname>_options.py |
30 |
else: |
hostname = re.sub("[^0-9a-zA-Z]", "_", socket.gethostname().split('.')[0]) |
31 |
hostname+="_" |
tmp = os.path.join("scons",hostname+"_options.py") |
32 |
options_file = "scons/"+hostname+"_options.py" |
options_file = ARGUMENTS.get('options_file', tmp) |
33 |
|
if not os.path.isfile(options_file): options_file = False |
34 |
|
else: print "Options file is", options_file |
35 |
|
|
36 |
|
# Load options file and command-line arguments |
37 |
opts = Options(options_file, ARGUMENTS) |
opts = Options(options_file, ARGUMENTS) |
38 |
|
|
39 |
|
############ Load build options ################################ |
40 |
|
|
41 |
opts.AddOptions( |
opts.AddOptions( |
42 |
# Where to install esys stuff |
# Where to install esys stuff |
43 |
('incinstall', 'where the esys headers will be installed', Dir('#.').abspath+'/include'), |
('prefix', 'where everything will be installed', Dir('#.').abspath), |
44 |
('libinstall', 'where the esys libraries will be installed', Dir('#.').abspath+'/lib'), |
('incinstall', 'where the esys headers will be installed', os.path.join(Dir('#.').abspath,'include')), |
45 |
('pyinstall', 'where the esys python modules will be installed', Dir('#.').abspath), |
('libinstall', 'where the esys libraries will be installed', os.path.join(prefix,'lib')), |
46 |
('src_zipfile', 'the source zip file will be installed.', Dir('#.').abspath+"/release/escript_src.zip"), |
('pyinstall', 'where the esys python modules will be installed', os.path.join(prefix,'esys')), |
|
('test_zipfile', 'the test zip file will be installed.', Dir('#.').abspath+"/release/escript_tests.zip"), |
|
|
('src_tarfile', 'the source tar file will be installed.', Dir('#.').abspath+"/release/escript_src.tar.gz"), |
|
|
('test_tarfile', 'the test tar file will be installed.', Dir('#.').abspath+"/release/escript_tests.tar.gz"), |
|
|
('examples_tarfile', 'the examples tar file will be installed.', Dir('#.').abspath+"/release/doc/escript_examples.tar.gz"), |
|
|
('examples_zipfile', 'the examples zip file will be installed.', Dir('#.').abspath+"/release/doc/escript_examples.zip"), |
|
|
('guide_pdf', 'name of the user guide in pdf format', Dir('#.').abspath+"/release/doc/user/guide.pdf"), |
|
|
('api_epydoc', 'name of the epydoc api docs directory',Dir('#.').abspath+"/release/doc/epydoc"), |
|
|
('guide_html', 'name of the directory for user guide in html format', Dir('#.').abspath+"/release/doc/user/html"), |
|
47 |
# Compilation options |
# Compilation options |
48 |
BoolOption('dodebug', 'Do you want a debug build?', 'no'), |
BoolOption('dodebug', 'For backwards compatibility', 'no'), |
49 |
('options_file', "Optional file containing preferred options. Ignored if it doesn't exist (default: scons/hostname_options.py)", options_file), |
BoolOption('usedebug', 'Do you want a debug build?', 'no'), |
50 |
('cc_defines','C/C++ defines to use', None), |
BoolOption('usevtk', 'Do you want to use VTK?', 'yes'), |
51 |
('cc_flags','C compiler flags to use (Release build)', '-O3 -std=c99 -ffast-math -fpic -Wno-unknown-pragmas'), |
('options_file', 'File of paths/options. Default: scons/<hostname>_options.py', options_file), |
52 |
('cc_flags_debug', 'C compiler flags to use (Debug build)', '-g -O0 -ffast-math -std=c99 -fpic -Wno-unknown-pragmas'), |
# The strings -DDEFAULT_ get replaced by scons/<hostname>_options.py or by defaults below |
53 |
('cxx_flags', 'C++ compiler flags to use (Release build)', '--no-warn -ansi'), |
('cc_flags', 'C compiler flags to use', '-DEFAULT_1'), |
54 |
('cxx_flags_debug', 'C++ compiler flags to use (Debug build)', '--no-warn -ansi -DDOASSERT -DDOPROF'), |
('cc_optim', 'C compiler optimization flags to use', '-DEFAULT_2'), |
55 |
('ar_flags', 'Static library archiver flags to use', None), |
('cc_debug', 'C compiler debug flags to use', '-DEFAULT_3'), |
56 |
('sys_libs', 'System libraries to link with', None), |
('omp_optim', 'OpenMP compiler flags to use (Release build)', '-DEFAULT_4'), |
57 |
('tar_flags','flags for zip files','-c -z'), |
('omp_debug', 'OpenMP compiler flags to use (Debug build)', '-DEFAULT_5'), |
58 |
# MKL |
('omp_libs', 'OpenMP compiler libraries to link with', '-DEFAULT_6'), |
59 |
PathOption('mkl_path', 'Path to MKL includes', None), |
('cc_extra', 'Extra C/C++ flags', ''), |
60 |
PathOption('mkl_lib_path', 'Path to MKL libs', None), |
('sys_libs', 'System libraries to link with', []), |
61 |
('mkl_libs', 'MKL libraries to link with', None), |
('ar_flags', 'Static library archiver flags to use', ''), |
62 |
# SCSL |
BoolOption('useopenmp', 'Compile parallel version using OpenMP', 'yes'), |
63 |
PathOption('scsl_path', 'Path to SCSL includes', None), |
BoolOption('usepedantic', 'Compile with -pedantic if using gcc', 'yes'), |
|
PathOption('scsl_lib_path', 'Path to SCSL libs', None), |
|
|
('scsl_libs', 'SCSL libraries to link with', None), |
|
|
# UMFPACK |
|
|
PathOption('umf_path', 'Path to UMF includes', None), |
|
|
PathOption('umf_lib_path', 'Path to UMF libs', None), |
|
|
('umf_libs', 'UMF libraries to link with', None), |
|
64 |
# Python |
# Python |
65 |
# locations of include files for python |
('python_path', 'Path to Python includes', '/usr/include/'+python_version), |
66 |
PathOption('python_path', 'Path to Python includes', '/usr/include/python%s.%s'%(sys.version_info[0],sys.version_info[1])), |
('python_lib_path', 'Path to Python libs', usr_lib), |
67 |
PathOption('python_lib_path', 'Path to Python libs', '/usr/lib'), |
('python_libs', 'Python libraries to link with', [python_version]), |
68 |
('python_lib', 'Python libraries to link with', ["python%s.%s"%(sys.version_info[0],sys.version_info[1]),]), |
('python_cmd', 'Python command', 'python'), |
69 |
# Boost |
# Boost |
70 |
PathOption('boost_path', 'Path to Boost includes', '/usr/include'), |
('boost_path', 'Path to Boost includes', '/usr/include'), |
71 |
PathOption('boost_lib_path', 'Path to Boost libs', '/usr/lib'), |
('boost_lib_path', 'Path to Boost libs', usr_lib), |
72 |
('boost_lib', 'Boost libraries to link with', ['boost_python',]), |
('boost_libs', 'Boost libraries to link with', ['boost_python']), |
73 |
# Doc building |
# NetCDF |
74 |
PathOption('doxygen_path', 'Path to Doxygen executable', None), |
BoolOption('usenetcdf', 'switch on/off the usage of netCDF', 'yes'), |
75 |
PathOption('epydoc_path', 'Path to Epydoc executable', None), |
('netCDF_path', 'Path to netCDF includes', '/usr/include'), |
76 |
PathOption('epydoc_pythonpath', 'Path to Epydoc python files', None), |
('netCDF_lib_path', 'Path to netCDF libs', usr_lib), |
77 |
|
('netCDF_libs', 'netCDF C++ libraries to link with', ['netcdf_c++', 'netcdf']), |
78 |
|
# MPI |
79 |
|
BoolOption('useMPI', 'For backwards compatibility', 'no'), |
80 |
|
BoolOption('usempi', 'Compile parallel version using MPI', 'no'), |
81 |
|
('MPICH_IGNORE_CXX_SEEK', 'name of macro to ignore MPI settings of C++ SEEK macro (for MPICH)' , 'MPICH_IGNORE_CXX_SEEK'), |
82 |
|
('mpi_path', 'Path to MPI includes', '/usr/include'), |
83 |
|
('mpi_run', 'mpirun name' , 'mpiexec -np 1'), |
84 |
|
('mpi_lib_path', 'Path to MPI libs (needs to be added to the LD_LIBRARY_PATH)', usr_lib), |
85 |
|
('mpi_libs', 'MPI libraries to link with (needs to be shared!)', ['mpich' , 'pthread', 'rt']), |
86 |
|
# ParMETIS |
87 |
|
BoolOption('useparmetis', 'Compile parallel version using ParMETIS', 'yes'), |
88 |
|
('parmetis_path', 'Path to ParMETIS includes', '/usr/include'), |
89 |
|
('parmetis_lib_path', 'Path to ParMETIS library', usr_lib), |
90 |
|
('parmetis_libs', 'ParMETIS library to link with', ['parmetis', 'metis']), |
91 |
# PAPI |
# PAPI |
92 |
PathOption('papi_path', 'Path to PAPI includes', None), |
BoolOption('usepapi', 'switch on/off the usage of PAPI', 'no'), |
93 |
PathOption('papi_lib_path', 'Path to PAPI libs', None), |
('papi_path', 'Path to PAPI includes', '/usr/include'), |
94 |
('papi_libs', 'PAPI libraries to link with', None), |
('papi_lib_path', 'Path to PAPI libs', usr_lib), |
95 |
|
('papi_libs', 'PAPI libraries to link with', ['papi']), |
96 |
|
BoolOption('papi_instrument_solver', 'use PAPI in Solver.c to instrument each iteration of the solver', False), |
97 |
|
# MKL |
98 |
|
BoolOption('usemkl', 'switch on/off the usage of MKL', 'no'), |
99 |
|
('mkl_path', 'Path to MKL includes', '/sw/sdev/cmkl/10.0.2.18/include'), |
100 |
|
('mkl_lib_path', 'Path to MKL libs', '/sw/sdev/cmkl/10.0.2.18/lib/em64t'), |
101 |
|
('mkl_libs', 'MKL libraries to link with', ['mkl_solver', 'mkl_em64t', 'guide', 'pthread']), |
102 |
|
# UMFPACK |
103 |
|
BoolOption('useumfpack', 'switch on/off the usage of UMFPACK', 'no'), |
104 |
|
('ufc_path', 'Path to UFconfig includes', '/usr/include/suitesparse'), |
105 |
|
('umf_path', 'Path to UMFPACK includes', '/usr/include/suitesparse'), |
106 |
|
('umf_lib_path', 'Path to UMFPACK libs', usr_lib), |
107 |
|
('umf_libs', 'UMFPACK libraries to link with', ['umfpack']), |
108 |
|
# AMD (used by UMFPACK) |
109 |
|
('amd_path', 'Path to AMD includes', '/usr/include/suitesparse'), |
110 |
|
('amd_lib_path', 'Path to AMD libs', usr_lib), |
111 |
|
('amd_libs', 'AMD libraries to link with', ['amd']), |
112 |
|
# BLAS (used by UMFPACK) |
113 |
|
('blas_path', 'Path to BLAS includes', '/usr/include/suitesparse'), |
114 |
|
('blas_lib_path', 'Path to BLAS libs', usr_lib), |
115 |
|
('blas_libs', 'BLAS libraries to link with', ['blas']) |
116 |
) |
) |
117 |
|
|
118 |
# Initialise Scons Build Environment |
############ Specify which compilers to use #################### |
|
# check for user environment variables we are interested in |
|
|
try: |
|
|
python_path = os.environ['PYTHONPATH'] |
|
|
except KeyError: |
|
|
python_path = '' |
|
|
try: |
|
|
path = os.environ['PATH'] |
|
|
except KeyError: |
|
|
path = '' |
|
|
try: |
|
|
ld_library_path = os.environ['LD_LIBRARY_PATH'] |
|
|
except KeyError: |
|
|
ld_library_path = '' |
|
|
|
|
|
# Note: On the Altix the intel compilers are not automatically |
|
|
# detected by scons intelc.py script. The Altix has a different directory |
|
|
# path and in some locations the "modules" facility is used to support |
|
|
# multiple compiler versions. This forces the need to import the users PATH |
|
|
# environment which isn't the "scons way" |
|
|
# This doesn't impact linux and windows which will use the default compiler (g++ or msvc, or the intel compiler if it is installed on both platforms) |
|
|
# FIXME: Perhaps a modification to intelc.py will allow better support for ia64 on altix |
|
|
|
|
|
if os.name != "nt" and os.uname()[4]=='ia64': |
|
|
env = Environment(ENV = {'PATH':path}, tools = ['default', 'intelc'], options = opts) |
|
|
env['ENV']['PATH'] = path |
|
|
env['ENV']['LD_LIBRARY_PATH'] = ld_library_path |
|
|
env['ENV']['PYTHONPATH'] = python_path |
|
|
if env['CXX'] == 'icpc': |
|
|
env['LINK'] = env['CXX'] # version >=9 of intel c++ compiler requires use of icpc to link in C++ runtimes (icc does not). FIXME: this behaviour could be directly incorporated into scons intelc.py |
|
|
elif os.name == "nt": |
|
|
# FIXME: Need to implement equivalent of ld library path for windoze |
|
|
env = Environment(tools = ['default', 'intelc'], options = opts) |
|
|
env['ENV']['PYTHONPATH'] = python_path |
|
|
else: |
|
|
env = Environment(tools = ['default'], options = opts) |
|
|
env['ENV']['PATH'] = path |
|
|
env['ENV']['LD_LIBRARY_PATH'] = ld_library_path |
|
|
env['ENV']['PYTHONPATH'] = python_path |
|
119 |
|
|
120 |
# Setup help for options |
# intelc uses regular expressions improperly and emits a warning about |
121 |
|
# failing to find the compilers. This warning can be safely ignored. |
122 |
|
|
123 |
|
if IS_WINDOWS_PLATFORM: |
124 |
|
env = Environment(tools = ['default', 'msvc'], options = opts) |
125 |
|
else: |
126 |
|
if socket.gethostname().split('.')[0] == 'service0': |
127 |
|
env = Environment(tools = ['default', 'intelc'], options = opts) |
128 |
|
elif os.uname()[4]=='ia64': |
129 |
|
env = Environment(tools = ['default', 'intelc'], options = opts) |
130 |
|
if env['CXX'] == 'icpc': |
131 |
|
env['LINK'] = env['CXX'] # version >=9 of intel c++ compiler requires use of icpc to link in C++ runtimes (icc does not) |
132 |
|
else: |
133 |
|
env = Environment(tools = ['default'], options = opts) |
134 |
Help(opts.GenerateHelpText(env)) |
Help(opts.GenerateHelpText(env)) |
135 |
|
|
136 |
# Add some customer builders |
############ Fill in compiler options if not set above ######### |
|
py_builder = Builder(action = scons_extensions.build_py, suffix = '.pyc', src_suffix = '.py', single_source=True) |
|
|
env.Append(BUILDERS = {'PyCompile' : py_builder}); |
|
137 |
|
|
138 |
if env['PLATFORM'] == "win32": |
# Backwards compatibility: allow dodebug=yes and useMPI=yes |
139 |
runUnitTest_builder = Builder(action = scons_extensions.runUnitTest, suffix = '.passed', src_suffix='.exe', single_source=True) |
if env['dodebug']: env['usedebug'] = 1 |
140 |
|
if env['useMPI']: env['usempi'] = 1 |
141 |
|
|
142 |
|
# Default compiler options (override allowed in hostname_options.py, but should not be necessary) |
143 |
|
# For both C and C++ you get: cc_flags and either the optim flags or debug flags |
144 |
|
|
145 |
|
if env["CC"] == "icc": |
146 |
|
# Intel compilers |
147 |
|
cc_flags = "-fPIC -ansi -wd161 -w1 -vec-report0 -DBLOCKTIMER -DCORE_ID1" |
148 |
|
cc_optim = "-O3 -ftz -IPF_ftlacc- -IPF_fma -fno-alias" |
149 |
|
cc_debug = "-g -O0 -UDOASSERT -DDOPROF -DBOUNDS_CHECK" |
150 |
|
omp_optim = "-openmp -openmp_report0" |
151 |
|
omp_debug = "-openmp -openmp_report0" |
152 |
|
omp_libs = ['guide', 'pthread'] |
153 |
|
pedantic = "" |
154 |
|
elif env["CC"] == "gcc": |
155 |
|
# GNU C on any system |
156 |
|
cc_flags = "-fPIC -ansi -ffast-math -Wno-unknown-pragmas -DBLOCKTIMER" |
157 |
|
cc_optim = "-O3" |
158 |
|
cc_debug = "-g -O0 -UDOASSERT -DDOPROF -DBOUNDS_CHECK" |
159 |
|
omp_optim = "" |
160 |
|
omp_debug = "" |
161 |
|
omp_libs = [] |
162 |
|
pedantic = "-pedantic-errors -Wno-long-long" |
163 |
|
elif env["CC"] == "cl": |
164 |
|
# Microsoft Visual C on Windows |
165 |
|
cc_flags = "/FD /EHsc /GR /wd4068 -D_USE_MATH_DEFINES -DDLL_NETCDF" |
166 |
|
cc_optim = "/O2 /Op /MT /W3" |
167 |
|
cc_debug = "/Od /RTC1 /MTd /ZI -DBOUNDS_CHECK" |
168 |
|
omp_optim = "" |
169 |
|
omp_debug = "" |
170 |
|
omp_libs = [] |
171 |
|
pedantic = "" |
172 |
|
|
173 |
|
# If not specified in hostname_options.py then set them here |
174 |
|
if env["cc_flags"] == "-DEFAULT_1": env['cc_flags'] = cc_flags |
175 |
|
if env["cc_optim"] == "-DEFAULT_2": env['cc_optim'] = cc_optim |
176 |
|
if env["cc_debug"] == "-DEFAULT_3": env['cc_debug'] = cc_debug |
177 |
|
if env["omp_optim"] == "-DEFAULT_4": env['omp_optim'] = omp_optim |
178 |
|
if env["omp_debug"] == "-DEFAULT_5": env['omp_debug'] = omp_debug |
179 |
|
if env["omp_libs"] == "-DEFAULT_6": env['omp_libs'] = omp_libs |
180 |
|
|
181 |
|
# OpenMP is disabled if useopenmp=no or both variables omp_optim and omp_debug are empty |
182 |
|
if not env["useopenmp"]: |
183 |
|
env['omp_optim'] = "" |
184 |
|
env['omp_debug'] = "" |
185 |
|
env['omp_libs'] = [] |
186 |
|
|
187 |
|
if env['omp_optim'] == "" and env['omp_debug'] == "": env["useopenmp"] = 0 |
188 |
|
|
189 |
|
############ Copy environment variables into scons env ######### |
190 |
|
|
191 |
|
try: env['ENV']['OMP_NUM_THREADS'] = os.environ['OMP_NUM_THREADS'] |
192 |
|
except KeyError: env['ENV']['OMP_NUM_THREADS'] = 1 |
193 |
|
|
194 |
|
try: env['ENV']['PATH'] = os.environ['PATH'] |
195 |
|
except KeyError: pass |
196 |
|
|
197 |
|
try: env['ENV']['PYTHONPATH'] = os.environ['PYTHONPATH'] |
198 |
|
except KeyError: pass |
199 |
|
|
200 |
|
try: env['ENV']['C_INCLUDE_PATH'] = os.environ['C_INCLUDE_PATH'] |
201 |
|
except KeyError: pass |
202 |
|
|
203 |
|
try: env['ENV']['CPLUS_INCLUDE_PATH'] = os.environ['CPLUS_INCLUDE_PATH'] |
204 |
|
except KeyError: pass |
205 |
|
|
206 |
|
try: env['ENV']['LD_LIBRARY_PATH'] = os.environ['LD_LIBRARY_PATH'] |
207 |
|
except KeyError: pass |
208 |
|
|
209 |
|
try: env['ENV']['LIBRARY_PATH'] = os.environ['LIBRARY_PATH'] |
210 |
|
except KeyError: pass |
211 |
|
|
212 |
|
try: env['ENV']['DISPLAY'] = os.environ['DISPLAY'] |
213 |
|
except KeyError: pass |
214 |
|
|
215 |
|
try: env['ENV']['XAUTHORITY'] = os.environ['XAUTHORITY'] |
216 |
|
except KeyError: pass |
217 |
|
|
218 |
|
try: env['ENV']['HOME'] = os.environ['HOME'] |
219 |
|
except KeyError: pass |
220 |
|
|
221 |
|
# Configure for test suite |
222 |
|
env.PrependENVPath('PYTHONPATH', prefix) |
223 |
|
env.PrependENVPath('LD_LIBRARY_PATH', env['libinstall']) |
224 |
|
|
225 |
|
############ Set up paths for Configure() ###################### |
226 |
|
|
227 |
|
# Make a copy of an environment |
228 |
|
# Use env.Clone if available, but fall back on env.Copy for older version of scons |
229 |
|
def clone_env(env): |
230 |
|
if 'Clone' in dir(env): return env.Clone() # scons-0.98 |
231 |
|
else: return env.Copy() # scons-0.96 |
232 |
|
|
233 |
|
# Add cc option -I<Escript>/trunk/include |
234 |
|
env.Append(CPPPATH = [Dir('include')]) |
235 |
|
|
236 |
|
# Add cc option -L<Escript>/trunk/lib |
237 |
|
env.Append(LIBPATH = [Dir('lib')]) |
238 |
|
|
239 |
|
env.Append(CPPDEFINES = ['ESCRIPT_EXPORTS', 'FINLEY_EXPORTS']) |
240 |
|
|
241 |
|
if env['cc_extra'] != '': env.Append(CCFLAGS = env['cc_extra']) |
242 |
|
|
243 |
|
if env['usepedantic']: env.Append(CCFLAGS = pedantic) |
244 |
|
|
245 |
|
# MS Windows |
246 |
|
if IS_WINDOWS_PLATFORM: |
247 |
|
env.PrependENVPath('PATH', [env['boost_lib_path']]) |
248 |
|
env.PrependENVPath('PATH', [env['libinstall']]) |
249 |
|
if env['usenetcdf']: |
250 |
|
env.PrependENVPath('PATH', [env['netCDF_lib_path']]) |
251 |
|
|
252 |
|
env.Append(ARFLAGS = env['ar_flags']) |
253 |
|
|
254 |
|
# Get the global Subversion revision number for getVersion() method |
255 |
|
try: |
256 |
|
global_revision = os.popen("svnversion -n .").read() |
257 |
|
global_revision = re.sub(":.*", "", global_revision) |
258 |
|
global_revision = re.sub("[^0-9]", "", global_revision) |
259 |
|
except: |
260 |
|
global_revision="-1" |
261 |
|
if global_revision == "": global_revision="-2" |
262 |
|
env.Append(CPPDEFINES = ["SVN_VERSION="+global_revision]) |
263 |
|
|
264 |
|
############ numarray (required) ############################### |
265 |
|
|
266 |
|
try: |
267 |
|
from numarray import identity |
268 |
|
except ImportError: |
269 |
|
print "Cannot import numarray, you need to set your PYTHONPATH" |
270 |
|
sys.exit(1) |
271 |
|
|
272 |
|
############ C compiler (required) ############################# |
273 |
|
|
274 |
|
# Create a Configure() environment for checking existence of required libraries and headers |
275 |
|
conf = Configure(clone_env(env)) |
276 |
|
|
277 |
|
# Test that the compiler is working |
278 |
|
if not conf.CheckFunc('printf'): |
279 |
|
print "Cannot run C compiler '%s' (or libc is missing)" % (env['CC']) |
280 |
|
sys.exit(1) |
281 |
|
|
282 |
|
if not conf.CheckFunc('gethostname'): |
283 |
|
env.Append(CPPDEFINES = ['HAVE_GETHOSTNAME']) |
284 |
|
|
285 |
|
############ python libraries (required) ####################### |
286 |
|
|
287 |
|
conf.env.AppendUnique(CPPPATH = [env['python_path']]) |
288 |
|
conf.env.AppendUnique(LIBPATH = [env['python_lib_path']]) |
289 |
|
conf.env.AppendUnique(LIBS = [env['python_libs']]) |
290 |
|
|
291 |
|
if not conf.CheckCHeader('Python.h'): |
292 |
|
print "Cannot find python include files (tried 'Python.h' in directory %s)" % (env['python_path']) |
293 |
|
sys.exit(1) |
294 |
|
if not conf.CheckFunc('Py_Main'): |
295 |
|
print "Cannot find python library method Py_Main (tried lib %s in directory %s)" % (env['python_libs'], env['python_lib_path']) |
296 |
|
sys.exit(1) |
297 |
|
|
298 |
|
############ boost (required) ################################## |
299 |
|
|
300 |
|
conf.env.AppendUnique(CPPPATH = [env['boost_path']]) |
301 |
|
conf.env.AppendUnique(LIBPATH = [env['boost_lib_path']]) |
302 |
|
conf.env.AppendUnique(LIBS = [env['boost_libs']]) |
303 |
|
|
304 |
|
if not conf.CheckCXXHeader('boost/python.hpp'): |
305 |
|
print "Cannot find boost include files (tried boost/python.hpp in directory %s)" % (env['boost_path']) |
306 |
|
sys.exit(1) |
307 |
|
if not conf.CheckFunc('PyObject_SetAttr'): |
308 |
|
print "Cannot find boost library method PyObject_SetAttr (tried method PyObject_SetAttr in library %s in directory %s)" % (env['boost_libs'], env['boost_lib_path']) |
309 |
|
sys.exit(1) |
310 |
|
|
311 |
|
# Commit changes to environment |
312 |
|
env = conf.Finish() |
313 |
|
|
314 |
|
############ VTK (optional) #################################### |
315 |
|
|
316 |
|
if env['usevtk']: |
317 |
|
try: |
318 |
|
import vtk |
319 |
|
env['usevtk'] = 1 |
320 |
|
except ImportError: |
321 |
|
env['usevtk'] = 0 |
322 |
|
|
323 |
|
# Add VTK to environment env if it was found |
324 |
|
if env['usevtk']: |
325 |
|
env.Append(CPPDEFINES = ['USE_VTK']) |
326 |
|
|
327 |
|
############ NetCDF (optional) ################################# |
328 |
|
|
329 |
|
conf = Configure(clone_env(env)) |
330 |
|
|
331 |
|
if env['usenetcdf']: |
332 |
|
conf.env.AppendUnique(CPPPATH = [env['netCDF_path']]) |
333 |
|
conf.env.AppendUnique(LIBPATH = [env['netCDF_lib_path']]) |
334 |
|
conf.env.AppendUnique(LIBS = [env['netCDF_libs']]) |
335 |
|
|
336 |
|
if env['usenetcdf'] and not conf.CheckCHeader('netcdf.h'): env['usenetcdf'] = 0 |
337 |
|
if env['usenetcdf'] and not conf.CheckFunc('nc_open'): env['usenetcdf'] = 0 |
338 |
|
|
339 |
|
# Add NetCDF to environment env if it was found |
340 |
|
if env['usenetcdf']: |
341 |
|
env = conf.Finish() |
342 |
|
env.Append(CPPDEFINES = ['USE_NETCDF']) |
343 |
else: |
else: |
344 |
runUnitTest_builder = Builder(action = scons_extensions.runUnitTest, suffix = '.passed', single_source=True) |
conf.Finish() |
|
env.Append(BUILDERS = {'RunUnitTest' : runUnitTest_builder}); |
|
345 |
|
|
346 |
runPyUnitTest_builder = Builder(action = scons_extensions.runPyUnitTest, suffix = '.passed', src_suffic='.py', single_source=True) |
############ PAPI (optional) ################################### |
|
env.Append(BUILDERS = {'RunPyUnitTest' : runPyUnitTest_builder}); |
|
347 |
|
|
348 |
# Convert the options which are held in environment variable into python variables for ease of handling and configure compilation options |
# Start a new configure environment that reflects what we've already found |
349 |
try: |
conf = Configure(clone_env(env)) |
|
incinstall = env['incinstall'] |
|
|
env.Append(CPPPATH = [incinstall,]) |
|
|
except KeyError: |
|
|
incinstall = None |
|
|
try: |
|
|
libinstall = env['libinstall'] |
|
|
env.Append(LIBPATH = [libinstall,]) |
|
|
env.PrependENVPath('LD_LIBRARY_PATH', libinstall) |
|
|
except KeyError: |
|
|
libinstall = None |
|
|
try: |
|
|
pyinstall = env['pyinstall']+'/esys' # all targets will install into pyinstall/esys but PYTHONPATH points at straight pyinstall so you go import esys.escript etc |
|
|
env.PrependENVPath('PYTHONPATH', env['pyinstall']) |
|
|
except KeyError: |
|
|
pyinstall = None |
|
|
try: |
|
|
dodebug = env['dodebug'] |
|
|
except KeyError: |
|
|
dodebug = None |
|
|
try: |
|
|
cc_defines = env['cc_defines'] |
|
|
env.Append(CPPDEFINES = [cc_defines,]) |
|
|
except KeyError: |
|
|
pass |
|
|
if dodebug: |
|
|
try: |
|
|
flags = env['cc_flags_debug'] |
|
|
env.Append(CCFLAGS = flags) |
|
|
except KeyError: |
|
|
pass |
|
|
else: |
|
|
try: |
|
|
flags = env['cc_flags'] |
|
|
env.Append(CCFLAGS = flags) |
|
|
except KeyError: |
|
|
pass |
|
|
|
|
|
if dodebug: |
|
|
try: |
|
|
flags = env['cxx_flags_debug'] |
|
|
env.Append(CXXFLAGS = flags) |
|
|
except KeyError: |
|
|
pass |
|
|
else: |
|
|
try: |
|
|
flags = env['cxx_flags'] |
|
|
env.Append(CXXFLAGS = flags) |
|
|
except KeyError: |
|
|
pass |
|
350 |
|
|
351 |
try: |
if env['usepapi']: |
352 |
flags = env['ar_flags'] |
conf.env.AppendUnique(CPPPATH = [env['papi_path']]) |
353 |
env.Append(ARFLAGS = flags) |
conf.env.AppendUnique(LIBPATH = [env['papi_lib_path']]) |
354 |
except KeyError: |
conf.env.AppendUnique(LIBS = [env['papi_libs']]) |
|
ar_flags = None |
|
|
try: |
|
|
sys_libs = env['sys_libs'] |
|
|
except KeyError: |
|
|
sys_libs = '' |
|
355 |
|
|
356 |
try: |
if env['usepapi'] and not conf.CheckCHeader('papi.h'): env['usepapi'] = 0 |
357 |
tar_flags = env['tar_flags'] |
if env['usepapi'] and not conf.CheckFunc('PAPI_start_counters'): env['usepapi'] = 0 |
|
env.Replace(TARFLAGS = tar_flags) |
|
|
except KeyError: |
|
|
pass |
|
358 |
|
|
359 |
try: |
# Add PAPI to environment env if it was found |
360 |
includes = env['mkl_path'] |
if env['usepapi']: |
361 |
env.Append(CPPPATH = [includes,]) |
env = conf.Finish() |
362 |
except KeyError: |
env.Append(CPPDEFINES = ['BLOCKPAPI']) |
363 |
pass |
else: |
364 |
|
conf.Finish() |
365 |
|
|
366 |
try: |
############ MKL (optional) #################################### |
|
lib_path = env['mkl_lib_path'] |
|
|
env.Append(LIBPATH = [lib_path,]) |
|
|
except KeyError: |
|
|
pass |
|
367 |
|
|
368 |
try: |
# Start a new configure environment that reflects what we've already found |
369 |
mkl_libs = env['mkl_libs'] |
conf = Configure(clone_env(env)) |
|
except KeyError: |
|
|
mkl_libs = '' |
|
|
try: |
|
|
includes = env['scsl_path'] |
|
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
lib_path = env['scsl_lib_path'] |
|
|
env.Append(LIBPATH = [lib_path,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
scsl_libs = env['scsl_libs'] |
|
|
except KeyError: |
|
|
scsl_libs = '' |
|
|
try: |
|
|
includes = env['umf_path'] |
|
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
lib_path = env['umf_lib_path'] |
|
|
env.Append(LIBPATH = [lib_path,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
umf_libs = env['umf_libs'] |
|
|
except KeyError: |
|
|
umf_libs = '' |
|
|
try: |
|
|
includes = env['boost_path'] |
|
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
lib_path = env['boost_lib_path'] |
|
|
env.Append(LIBPATH = [lib_path,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
boost_lib = env['boost_lib'] |
|
|
except KeyError: |
|
|
boost_lib = None |
|
|
try: |
|
|
includes = env['python_path'] |
|
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
lib_path = env['python_lib_path'] |
|
|
env.Append(LIBPATH = [lib_path,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
python_lib = env['python_lib'] |
|
|
except KeyError: |
|
|
python_lib = None |
|
|
try: |
|
|
doxygen_path = env['doxygen_path'] |
|
|
except KeyError: |
|
|
doxygen_path = None |
|
|
try: |
|
|
epydoc_path = env['epydoc_path'] |
|
|
except KeyError: |
|
|
epydoc_path = None |
|
|
try: |
|
|
epydoc_pythonpath = env['epydoc_pythonpath'] |
|
|
except KeyError: |
|
|
epydoc_pythonpath = None |
|
|
try: |
|
|
includes = env['papi_path'] |
|
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
lib_path = env['papi_lib_path'] |
|
|
env.Append(LIBPATH = [lib_path,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
papi_libs = env['papi_libs'] |
|
|
except KeyError: |
|
|
papi_libs = None |
|
370 |
|
|
371 |
|
if env['usemkl']: |
372 |
|
conf.env.AppendUnique(CPPPATH = [env['mkl_path']]) |
373 |
|
conf.env.AppendUnique(LIBPATH = [env['mkl_lib_path']]) |
374 |
|
conf.env.AppendUnique(LIBS = [env['mkl_libs']]) |
375 |
|
|
376 |
try: |
if env['usemkl'] and not conf.CheckCHeader('mkl_solver.h'): env['usemkl'] = 0 |
377 |
src_zipfile = env.File(env['src_zipfile']) |
if env['usemkl'] and not conf.CheckFunc('pardiso_'): env['usemkl'] = 0 |
|
except KeyError: |
|
|
src_zipfile = None |
|
|
try: |
|
|
test_zipfile = env.File(env['test_zipfile']) |
|
|
except KeyError: |
|
|
test_zipfile = None |
|
|
try: |
|
|
examples_zipfile = env.File(env['examples_zipfile']) |
|
|
except KeyError: |
|
|
examples_zipfile = None |
|
378 |
|
|
379 |
try: |
# Add MKL to environment env if it was found |
380 |
src_tarfile = env.File(env['src_tarfile']) |
if env['usemkl']: |
381 |
except KeyError: |
env = conf.Finish() |
382 |
src_tarfile = None |
env.Append(CPPDEFINES = ['MKL']) |
383 |
try: |
else: |
384 |
test_tarfile = env.File(env['test_tarfile']) |
conf.Finish() |
|
except KeyError: |
|
|
test_tarfile = None |
|
|
try: |
|
|
examples_tarfile = env.File(env['examples_tarfile']) |
|
|
except KeyError: |
|
|
examples_tarfile = None |
|
385 |
|
|
386 |
try: |
############ UMFPACK (optional) ################################ |
|
guide_pdf = env.File(env['guide_pdf']) |
|
|
except KeyError: |
|
|
guide_pdf = None |
|
387 |
|
|
388 |
try: |
# Start a new configure environment that reflects what we've already found |
389 |
guide_html_index = env.File('index.htm',env['guide_html']) |
conf = Configure(clone_env(env)) |
390 |
except KeyError: |
|
391 |
guide_html_index = None |
if env['useumfpack']: |
392 |
|
conf.env.AppendUnique(CPPPATH = [env['ufc_path']]) |
393 |
|
conf.env.AppendUnique(CPPPATH = [env['umf_path']]) |
394 |
|
conf.env.AppendUnique(LIBPATH = [env['umf_lib_path']]) |
395 |
|
conf.env.AppendUnique(LIBS = [env['umf_libs']]) |
396 |
|
conf.env.AppendUnique(CPPPATH = [env['amd_path']]) |
397 |
|
conf.env.AppendUnique(LIBPATH = [env['amd_lib_path']]) |
398 |
|
conf.env.AppendUnique(LIBS = [env['amd_libs']]) |
399 |
|
conf.env.AppendUnique(CPPPATH = [env['blas_path']]) |
400 |
|
conf.env.AppendUnique(LIBPATH = [env['blas_lib_path']]) |
401 |
|
conf.env.AppendUnique(LIBS = [env['blas_libs']]) |
402 |
|
|
403 |
|
if env['useumfpack'] and not conf.CheckCHeader('umfpack.h'): env['useumfpack'] = 0 |
404 |
|
if env['useumfpack'] and not conf.CheckFunc('umfpack_di_symbolic'): env['useumfpack'] = 0 |
405 |
|
|
406 |
|
# Add UMFPACK to environment env if it was found |
407 |
|
if env['useumfpack']: |
408 |
|
env = conf.Finish() |
409 |
|
env.Append(CPPDEFINES = ['UMFPACK']) |
410 |
|
else: |
411 |
|
conf.Finish() |
412 |
|
|
413 |
|
############ Add the compiler flags ############################ |
414 |
|
|
415 |
|
# Enable debug by choosing either cc_debug or cc_optim |
416 |
|
if env['usedebug']: |
417 |
|
env.Append(CCFLAGS = env['cc_debug']) |
418 |
|
env.Append(CCFLAGS = env['omp_debug']) |
419 |
|
else: |
420 |
|
env.Append(CCFLAGS = env['cc_optim']) |
421 |
|
env.Append(CCFLAGS = env['omp_optim']) |
422 |
|
|
423 |
|
# Always use cc_flags |
424 |
|
env.Append(CCFLAGS = env['cc_flags']) |
425 |
|
env.Append(LIBS = [env['omp_libs']]) |
426 |
|
|
427 |
|
############ MPI (optional) #################################### |
428 |
|
|
429 |
|
# Create a modified environment for MPI programs (identical to env if usempi=no) |
430 |
|
env_mpi = clone_env(env) |
431 |
|
|
432 |
|
# Start a new configure environment that reflects what we've already found |
433 |
|
conf = Configure(clone_env(env_mpi)) |
434 |
|
|
435 |
|
if env_mpi['usempi']: |
436 |
|
conf.env.AppendUnique(CPPPATH = [env_mpi['mpi_path']]) |
437 |
|
conf.env.AppendUnique(LIBPATH = [env_mpi['mpi_lib_path']]) |
438 |
|
conf.env.AppendUnique(LIBS = [env_mpi['mpi_libs']]) |
439 |
|
|
440 |
|
if env_mpi['usempi'] and not conf.CheckCHeader('mpi.h'): env_mpi['usempi'] = 0 |
441 |
|
if env_mpi['usempi'] and not conf.CheckFunc('MPI_Init'): env_mpi['usempi'] = 0 |
442 |
|
|
443 |
|
# Add MPI to environment env_mpi if it was found |
444 |
|
if env_mpi['usempi']: |
445 |
|
env_mpi = conf.Finish() |
446 |
|
env_mpi.Append(CPPDEFINES = ['PASO_MPI', 'MPI_NO_CPPBIND', env_mpi['MPICH_IGNORE_CXX_SEEK']]) |
447 |
|
else: |
448 |
|
conf.Finish() |
449 |
|
|
450 |
|
env['usempi'] = env_mpi['usempi'] |
451 |
|
|
452 |
|
############ ParMETIS (optional) ############################### |
453 |
|
|
454 |
|
# Start a new configure environment that reflects what we've already found |
455 |
|
conf = Configure(clone_env(env_mpi)) |
456 |
|
|
457 |
|
if not env_mpi['usempi']: env_mpi['useparmetis'] = 0 |
458 |
|
|
459 |
|
if env_mpi['useparmetis']: |
460 |
|
conf.env.AppendUnique(CPPPATH = [env_mpi['parmetis_path']]) |
461 |
|
conf.env.AppendUnique(LIBPATH = [env_mpi['parmetis_lib_path']]) |
462 |
|
conf.env.AppendUnique(LIBS = [env_mpi['parmetis_libs']]) |
463 |
|
|
464 |
|
if env_mpi['useparmetis'] and not conf.CheckCHeader('parmetis.h'): env_mpi['useparmetis'] = 0 |
465 |
|
if env_mpi['useparmetis'] and not conf.CheckFunc('ParMETIS_V3_PartGeomKway'): env_mpi['useparmetis'] = 0 |
466 |
|
|
467 |
|
# Add ParMETIS to environment env_mpi if it was found |
468 |
|
if env_mpi['useparmetis']: |
469 |
|
env_mpi = conf.Finish() |
470 |
|
env_mpi.Append(CPPDEFINES = ['USE_PARMETIS']) |
471 |
|
else: |
472 |
|
conf.Finish() |
473 |
|
|
474 |
|
env['useparmetis'] = env_mpi['useparmetis'] |
475 |
|
|
476 |
|
############ Summarize our environment ######################### |
477 |
|
|
478 |
|
print "" |
479 |
|
print "Summary of configuration (see ./config.log for information)" |
480 |
|
print " Using python libraries" |
481 |
|
print " Using numarray" |
482 |
|
print " Using boost" |
483 |
|
if env['usenetcdf']: print " Using NetCDF" |
484 |
|
else: print " Not using NetCDF" |
485 |
|
if env['usevtk']: print " Using VTK" |
486 |
|
else: print " Not using VTK" |
487 |
|
if env['usemkl']: print " Using MKL" |
488 |
|
else: print " Not using MKL" |
489 |
|
if env['useumfpack']: print " Using UMFPACK" |
490 |
|
else: print " Not using UMFPACK" |
491 |
|
if env['useopenmp']: print " Using OpenMP" |
492 |
|
else: print " Not using OpenMP" |
493 |
|
if env['usempi']: print " Using MPI" |
494 |
|
else: print " Not using MPI" |
495 |
|
if env['useparmetis']: print " Using ParMETIS" |
496 |
|
else: print " Not using ParMETIS (requires MPI)" |
497 |
|
if env['usepapi']: print " Using PAPI" |
498 |
|
else: print " Not using PAPI" |
499 |
|
if env['usedebug']: print " Compiling for debug" |
500 |
|
else: print " Not compiling for debug" |
501 |
|
print " Installing in", prefix |
502 |
|
print "" |
503 |
|
|
504 |
|
############ Add some custom builders ########################## |
505 |
|
|
506 |
|
py_builder = Builder(action = scons_extensions.build_py, suffix = '.pyc', src_suffix = '.py', single_source=True) |
507 |
|
env.Append(BUILDERS = {'PyCompile' : py_builder}); |
508 |
|
|
509 |
|
runUnitTest_builder = Builder(action = scons_extensions.runUnitTest, suffix = '.passed', src_suffix=env['PROGSUFFIX'], single_source=True) |
510 |
|
env.Append(BUILDERS = {'RunUnitTest' : runUnitTest_builder}); |
511 |
|
|
512 |
|
runPyUnitTest_builder = Builder(action = scons_extensions.runPyUnitTest, suffix = '.passed', src_suffic='.py', single_source=True) |
513 |
|
env.Append(BUILDERS = {'RunPyUnitTest' : runPyUnitTest_builder}); |
514 |
|
|
515 |
|
############ Build the desired subdirectories ################## |
516 |
|
|
517 |
|
Export(["env", "env_mpi", "clone_env"]) |
518 |
|
|
|
try: |
|
|
api_epydoc = env.Dir(env['api_epydoc']) |
|
|
except KeyError: |
|
|
api_epydoc = None |
|
|
|
|
|
# Zipgets |
|
|
|
|
|
env.Default(libinstall) |
|
|
env.Default(incinstall) |
|
|
env.Default(pyinstall) |
|
|
env.Alias('release_src',[ src_zipfile, src_tarfile ]) |
|
|
env.Alias('release_tests',[ test_zipfile, test_tarfile]) |
|
|
env.Alias('release_examples',[ examples_zipfile, examples_tarfile]) |
|
|
env.Alias('api_epydoc',api_epydoc) |
|
|
env.Alias('docs',[ 'release_examples', guide_pdf, guide_html_index, api_epydoc]) |
|
|
env.Alias('release', ['release_src', 'release_tests', 'docs']) |
|
|
env.Alias('build_tests') # target to build all C++ tests |
|
|
env.Alias('build_py_tests') # target to build all python tests |
|
|
env.Alias('build_all_tests', [ 'build_tests', 'build_py_tests' ] ) # target to build all python tests |
|
|
env.Alias('run_tests', 'build_tests') # target to run all C++ test |
|
|
env.Alias('py_tests', 'build_py_tests') # taget to run all released python tests |
|
|
env.Alias('all_tests', ['run_tests', 'py_tests']) # target to run all C++ and released python tests |
|
|
|
|
|
# Python install - esys __init__.py |
|
|
# This is just an empty file but stills need to be touched so add a special target and Command. Note you can't use the scons Touch() function as it will not |
|
|
# create the file if it doesn't exist |
|
|
env.Command(pyinstall+'/__init__.py', None, 'touch $TARGET') |
|
|
|
|
|
# Allow sconscripts to see the env |
|
|
Export(["env", "incinstall", "libinstall", "pyinstall", "dodebug", "mkl_libs", "scsl_libs", "umf_libs", |
|
|
"boost_lib", "python_lib", "doxygen_path", "epydoc_path", "epydoc_pythonpath", "papi_libs", |
|
|
"sys_libs", "test_zipfile", "src_zipfile", "test_tarfile", "src_tarfile", "examples_tarfile", "examples_zipfile", |
|
|
"guide_pdf", "guide_html_index", "api_epydoc"]) |
|
|
|
|
|
# End initialisation section |
|
|
# Begin configuration section |
|
|
# adds this file and the scons option directore to the source tar |
|
|
release_srcfiles=[env.File('SConstruct'),]+[ env.File(x) for x in glob.glob('scons/*.py') ] |
|
|
release_testfiles=[env.File('README_TESTS'),] |
|
|
env.Zip(src_zipfile, release_srcfiles) |
|
|
env.Zip(test_zipfile, release_testfiles) |
|
|
env.Tar(src_tarfile, release_srcfiles) |
|
|
env.Tar(test_tarfile, release_testfiles) |
|
|
|
|
|
# Insert new components to be build here |
|
|
# FIXME: might be nice to replace this verbosity with a list of targets and some |
|
|
# FIXME: nifty python to create the lengthy but very similar env.Sconscript lines |
|
|
# Third Party libraries |
|
519 |
env.SConscript(dirs = ['tools/CppUnitTest/src'], build_dir='build/$PLATFORM/tools/CppUnitTest', duplicate=0) |
env.SConscript(dirs = ['tools/CppUnitTest/src'], build_dir='build/$PLATFORM/tools/CppUnitTest', duplicate=0) |
|
# C/C++ Libraries |
|
520 |
env.SConscript(dirs = ['paso/src'], build_dir='build/$PLATFORM/paso', duplicate=0) |
env.SConscript(dirs = ['paso/src'], build_dir='build/$PLATFORM/paso', duplicate=0) |
|
env.SConscript(dirs = ['bruce/src'], build_dir='build/$PLATFORM/bruce', duplicate=0) |
|
521 |
env.SConscript(dirs = ['escript/src'], build_dir='build/$PLATFORM/escript', duplicate=0) |
env.SConscript(dirs = ['escript/src'], build_dir='build/$PLATFORM/escript', duplicate=0) |
522 |
env.SConscript(dirs = ['esysUtils/src'], build_dir='build/$PLATFORM/esysUtils', duplicate=0) |
env.SConscript(dirs = ['esysUtils/src'], build_dir='build/$PLATFORM/esysUtils', duplicate=0) |
523 |
env.SConscript(dirs = ['finley/src'], build_dir='build/$PLATFORM/finley', duplicate=0) |
env.SConscript(dirs = ['finley/src'], build_dir='build/$PLATFORM/finley', duplicate=0) |
524 |
env.SConscript(dirs = ['modellib/py_src'], build_dir='build/$PLATFORM/modellib', duplicate=0) |
env.SConscript(dirs = ['modellib/py_src'], build_dir='build/$PLATFORM/modellib', duplicate=0) |
525 |
env.SConscript(dirs = ['doc'], build_dir='build/$PLATFORM/doc', duplicate=0) |
env.SConscript(dirs = ['doc'], build_dir='build/$PLATFORM/doc', duplicate=0) |
526 |
env.SConscript(dirs = ['pyvisi/py_src'], build_dir='build/$PLATFORM/pyvisi', duplicate=0) |
env.SConscript(dirs = ['pyvisi/py_src'], build_dir='build/$PLATFORM/pyvisi', duplicate=0) |
527 |
|
env.SConscript(dirs = ['pycad/py_src'], build_dir='build/$PLATFORM/pycad', duplicate=0) |
528 |
|
env.SConscript(dirs = ['pythonMPI/src'], build_dir='build/$PLATFORM/pythonMPI', duplicate=0) |
529 |
|
|
530 |
|
############ Remember what optimizations we used ############### |
531 |
|
|
532 |
|
remember_list = [] |
533 |
|
|
534 |
|
if env['usedebug']: |
535 |
|
remember_list += env.Command(env['libinstall'] + "/Compiled.with.debug", None, Touch('$TARGET')) |
536 |
|
else: |
537 |
|
remember_list += env.Command(env['libinstall'] + "/Compiled.with.debug", None, Delete('$TARGET')) |
538 |
|
|
539 |
|
if env['usempi']: |
540 |
|
remember_list += env.Command(env['libinstall'] + "/Compiled.with.mpi", None, Touch('$TARGET')) |
541 |
|
else: |
542 |
|
remember_list += env.Command(env['libinstall'] + "/Compiled.with.mpi", None, Delete('$TARGET')) |
543 |
|
|
544 |
|
if env['omp_optim'] != '': |
545 |
|
remember_list += env.Command(env['libinstall'] + "/Compiled.with.openmp", None, Touch('$TARGET')) |
546 |
|
else: |
547 |
|
remember_list += env.Command(env['libinstall'] + "/Compiled.with.openmp", None, Delete('$TARGET')) |
548 |
|
|
549 |
|
env.Alias('remember_options', remember_list) |
550 |
|
|
551 |
|
############ Targets to build and install libraries ############ |
552 |
|
|
553 |
|
target_init = env.Command(env['pyinstall']+'/__init__.py', None, Touch('$TARGET')) |
554 |
|
env.Alias('target_init', [target_init]) |
555 |
|
|
556 |
|
# The headers have to be installed prior to build in order to satisfy #include <paso/Common.h> |
557 |
|
env.Alias('build_esysUtils', ['target_install_esysUtils_headers', 'target_esysUtils_a']) |
558 |
|
env.Alias('install_esysUtils', ['build_esysUtils', 'target_install_esysUtils_a']) |
559 |
|
|
560 |
|
env.Alias('build_paso', ['target_install_paso_headers', 'target_paso_a']) |
561 |
|
env.Alias('install_paso', ['build_paso', 'target_install_paso_a']) |
562 |
|
|
563 |
|
env.Alias('build_escript', ['target_install_escript_headers', 'target_escript_so', 'target_escriptcpp_so']) |
564 |
|
env.Alias('install_escript', ['build_escript', 'target_install_escript_so', 'target_install_escriptcpp_so', 'target_install_escript_py']) |
565 |
|
|
566 |
|
env.Alias('build_finley', ['target_install_finley_headers', 'target_finley_so', 'target_finleycpp_so']) |
567 |
|
env.Alias('install_finley', ['build_finley', 'target_install_finley_so', 'target_install_finleycpp_so', 'target_install_finley_py']) |
568 |
|
|
569 |
|
# Now gather all the above into a couple easy targets: build_all and install_all |
570 |
|
build_all_list = [] |
571 |
|
build_all_list += ['build_esysUtils'] |
572 |
|
build_all_list += ['build_paso'] |
573 |
|
build_all_list += ['build_escript'] |
574 |
|
build_all_list += ['build_finley'] |
575 |
|
if env['usempi']: build_all_list += ['target_pythonMPI_exe'] |
576 |
|
env.Alias('build_all', build_all_list) |
577 |
|
|
578 |
|
install_all_list = [] |
579 |
|
install_all_list += ['target_init'] |
580 |
|
install_all_list += ['install_esysUtils'] |
581 |
|
install_all_list += ['install_paso'] |
582 |
|
install_all_list += ['install_escript'] |
583 |
|
install_all_list += ['install_finley'] |
584 |
|
install_all_list += ['target_install_pyvisi_py'] |
585 |
|
install_all_list += ['target_install_modellib_py'] |
586 |
|
install_all_list += ['target_install_pycad_py'] |
587 |
|
if env['usempi']: install_all_list += ['target_install_pythonMPI_exe'] |
588 |
|
install_all_list += ['remember_options'] |
589 |
|
env.Alias('install_all', install_all_list) |
590 |
|
|
591 |
|
# Default target is install |
592 |
|
env.Default('install_all') |
593 |
|
|
594 |
|
############ Targets to build and run the test suite ########### |
595 |
|
|
596 |
|
env.Alias('build_cppunittest', ['target_install_cppunittest_headers', 'target_cppunittest_a']) |
597 |
|
env.Alias('install_cppunittest', ['build_cppunittest', 'target_install_cppunittest_a']) |
598 |
|
env.Alias('run_tests', ['install_all', 'target_install_cppunittest_a']) |
599 |
|
env.Alias('all_tests', ['install_all', 'target_install_cppunittest_a', 'run_tests', 'py_tests']) |
600 |
|
|
601 |
|
############ Targets to build the documentation ################ |
602 |
|
|
603 |
|
env.Alias('docs', ['examples_tarfile', 'examples_zipfile', 'api_epydoc', 'api_doxygen', 'guide_pdf', 'guide_html']) |
604 |
|
|