1 |
# Copyright 2006 by ACcESS MNRF |
|
2 |
|
######################################################## |
3 |
|
# |
4 |
|
# Copyright (c) 2003-2010 by University of Queensland |
5 |
|
# Earth Systems Science Computational Center (ESSCC) |
6 |
|
# http://www.uq.edu.au/esscc |
7 |
# |
# |
8 |
# http://www.access.edu.au |
# Primary Business: Queensland, Australia |
9 |
# Primary Business: Queensland, Australia |
# Licensed under the Open Software License version 3.0 |
10 |
# Licensed under the Open Software License version 3.0 |
# http://www.opensource.org/licenses/osl-3.0.php |
11 |
# http://www.opensource.org/licenses/osl-3.0.php |
# |
12 |
|
######################################################## |
13 |
# top-level Scons configuration file for all esys13 modules |
|
14 |
# Begin initialisation Section |
|
|
# all of this section just intialises default environments and helper |
|
|
# scripts. You shouldn't need to modify this section. |
|
15 |
EnsureSConsVersion(0,96,91) |
EnsureSConsVersion(0,96,91) |
16 |
EnsurePythonVersion(2,3) |
EnsurePythonVersion(2,3) |
17 |
|
|
18 |
#=============================================================== |
import sys, os, re, socket, platform, stat |
19 |
# import tools: |
# For copy() |
20 |
import glob |
import shutil |
21 |
import sys, os, re |
|
22 |
# Add our extensions |
# Add our extensions |
23 |
if sys.path.count('scons')==0: sys.path.append('scons') |
if os.path.isdir('scons'): sys.path.append('scons') |
24 |
import scons_extensions |
import scons_extensions |
25 |
|
|
26 |
# We may also need to know where python's site-packages subdirectory lives |
# Use /usr/lib64 if available, else /usr/lib |
27 |
python_version = 'python%s.%s'%(sys.version_info[0],sys.version_info[1]) |
usr_lib = '/usr/lib' |
28 |
|
if os.path.isfile('/usr/lib64/libc.so'): usr_lib = '/usr/lib64' |
|
#=============================================================== |
|
|
|
|
|
tools_prefix="/usr" |
|
|
|
|
|
#============================================================================================== |
|
|
# |
|
|
# get the installation prefix |
|
|
# |
|
|
prefix = ARGUMENTS.get('prefix', sys.prefix ) |
|
|
|
|
|
# We may also need to know where python's site-packages subdirectory lives |
|
|
python_version = 'python%s.%s'%(sys.version_info[0],sys.version_info[1]) |
|
|
# Install as a standard python package in /usr/lib64 if available, else in /usr/lib |
|
|
if os.path.isdir( prefix+"/lib64/"+python_version+"/site-packages"): |
|
|
sys_dir_packages = prefix+"/lib64/"+python_version+"/site-packages/esys" |
|
|
sys_dir_libraries = prefix+"/lib64" |
|
|
else: |
|
|
sys_dir_packages = prefix+"/lib/"+python_version+"/site-packages/esys" |
|
|
sys_dir_libraries = prefix+"/lib" |
|
|
|
|
|
sys_dir_examples = prefix+"/share/doc/esys" |
|
|
|
|
|
source_root = Dir('#.').abspath |
|
|
|
|
|
dir_packages = os.path.join(source_root,"esys") |
|
|
dir_examples = os.path.join(source_root,"examples") |
|
|
dir_libraries = os.path.join(source_root,"lib") |
|
|
|
|
|
print "Source root is : ",source_root |
|
|
print " Default packages local installation: ", dir_packages |
|
|
print " Default library local installation ", dir_libraries |
|
|
print " Default example local installation: ", dir_examples |
|
|
print "Install prefix is: ", prefix |
|
|
print " Default packages system installation: ", sys_dir_packages |
|
|
print " Default library system installation ", sys_dir_libraries |
|
|
print " Default example system installation: ", sys_dir_examples |
|
|
|
|
|
#============================================================================================== |
|
|
|
|
|
# Default options and options help text |
|
|
# These are defaults and can be overridden using command line arguments or an options file. |
|
|
# if the options_file or ARGUMENTS do not exist then the ones listed as default here are used |
|
|
# DO NOT CHANGE THEM HERE |
|
|
# Where to install? |
|
|
#============================================================================================== |
|
|
# |
|
|
# get the options file if present: |
|
|
# |
|
|
options_file = ARGUMENTS.get('options_file','') |
|
|
|
|
|
if not os.path.isfile(options_file) : |
|
|
options_file = False |
|
|
|
|
|
if not options_file : |
|
|
import socket |
|
|
hostname = re.sub("[^0-9a-zA-Z]", "_", socket.gethostname().split('.')[0]) |
|
|
tmp = os.path.join("scons",hostname+"_options.py") |
|
29 |
|
|
30 |
if os.path.isfile(tmp) : |
# The string python2.4 or python2.5 |
31 |
options_file = tmp |
python_version = 'python%s.%s' % (sys.version_info[0], sys.version_info[1]) |
32 |
|
|
33 |
|
# MS Windows support, many thanks to PH |
34 |
IS_WINDOWS_PLATFORM = (os.name== "nt") |
IS_WINDOWS_PLATFORM = (os.name== "nt") |
35 |
|
|
36 |
# If you're not going to tell me then...... |
prefix = ARGUMENTS.get('prefix', Dir('#.').abspath) |
|
# FIXME: add one for the altix too. |
|
|
if not options_file : |
|
|
if IS_WINDOWS_PLATFORM : |
|
|
options_file = "scons/windows_mscv71_options.py" |
|
|
else: |
|
|
options_file = "scons/linux_gcc_eg_options.py" |
|
37 |
|
|
38 |
# and load it |
#Holds names of variables from the calling environment which need to be passed |
39 |
opts = Options(options_file, ARGUMENTS) |
#to tools |
40 |
#================================================================ |
env_export=[] |
41 |
# |
|
42 |
# check if UMFPACK is installed on the system: |
#Determine where to read options from use: |
43 |
# |
#1. command line |
44 |
uf_root=None |
#2. scons/<hostname>_options.py |
45 |
for i in [ 'UMFPACK', 'umfpack', 'ufsparse', 'UFSPARSE']: |
#3. name as part of a cluster |
46 |
if os.path.isdir(os.path.join(tools_prefix,'include',i)): |
options_file=ARGUMENTS.get('options_file', None) |
47 |
uf_root=i |
effective_hostname=socket.gethostname().split('.')[0] |
48 |
print i," is used form ",tools_prefix |
if not options_file: |
49 |
break |
mangledhostname = re.sub("[^0-9a-zA-Z]", "_", effective_hostname) |
50 |
if not uf_root==None: |
options_file = os.path.join("scons",mangledhostname+"_options.py") |
51 |
umf_path_default=os.path.join(tools_prefix,'include',uf_root) |
#If there is no options file with that name see if there is a substitute |
52 |
umf_lib_path_default=os.path.join(tools_prefix,'lib') |
if not os.path.isfile(options_file): |
53 |
umf_libs_default=['umfpack'] |
effective_hostname = scons_extensions.effectiveName(effective_hostname) |
54 |
amd_path_default=os.path.join(tools_prefix,'include',uf_root) |
mangledhostname = re.sub("[^0-9a-zA-Z]", "_", effective_hostname) |
55 |
amd_lib_path_default=os.path.join(tools_prefix,'lib') |
options_file = os.path.join("scons",mangledhostname+"_options.py") |
56 |
amd_libs_default=['amd'] |
|
57 |
ufc_path_default=os.path.join(tools_prefix,'include',uf_root) |
if not os.path.isfile(options_file): |
58 |
else: |
print "Options file not found (expected '%s')" % options_file |
59 |
umf_path_default=None |
options_file = False |
60 |
umf_lib_path_default=None |
else: |
61 |
umf_libs_default=None |
print "Options file is", options_file |
|
amd_path_default=None |
|
|
amd_lib_path_default=None |
|
|
amd_libs_default=None |
|
|
ufc_path_default=None |
|
|
# |
|
|
#========================================================================== |
|
|
# |
|
|
# python installation: |
|
|
# |
|
|
python_path_default=os.path.join(tools_prefix,'include','python%s.%s'%(sys.version_info[0],sys.version_info[1])) |
|
|
python_lib_path_default=os.path.join(tools_prefix,'lib') |
|
|
python_lib_default="python%s.%s"%(sys.version_info[0],sys.version_info[1]) |
|
62 |
|
|
63 |
#========================================================================== |
#Does our scons support the newer Variables class or do we need to use Options? |
|
# |
|
|
# boost installation: |
|
|
# |
|
|
boost_path_default=os.path.join(tools_prefix,'include') |
|
|
boost_lib_path_default=os.path.join(tools_prefix,'lib') |
|
|
boost_lib_default=['boost_python'] |
|
64 |
|
|
65 |
#========================================================================== |
try: |
66 |
# |
dummyvar=Variables |
67 |
# check if netCDF is installed on the system: |
opts = Variables(options_file, ARGUMENTS) |
68 |
# |
adder = opts.AddVariables |
69 |
netCDF_path_default=os.path.join(tools_prefix,'include','netcdf-3') |
except: |
70 |
netCDF_lib_path_default=os.path.join(tools_prefix,'lib') |
opts = Options(options_file, ARGUMENTS) |
71 |
|
adder = opts.AddOptions |
72 |
|
BoolVariable = BoolOption |
73 |
|
|
74 |
if os.path.isdir(netCDF_path_default) and os.path.isdir(netCDF_lib_path_default): |
############ Load build options ################################ |
|
useNetCDF_default='yes' |
|
|
netCDF_libs_default=[ 'netcdf_c++', 'netcdf' ] |
|
|
else: |
|
|
useNetCDF_default='no' |
|
|
netCDF_path_default=None |
|
|
netCDF_lib_path_default=None |
|
|
netCDF_libs_default=None |
|
75 |
|
|
76 |
#========================================================================== |
adder( |
77 |
# |
#opts.AddOptions( |
|
# MPI: |
|
|
# |
|
|
if IS_WINDOWS_PLATFORM: |
|
|
useMPI_default='no' |
|
|
mpi_path_default=None |
|
|
mpi_lib_path_default=None |
|
|
mpi_libs_default=[] |
|
|
mpi_run_default=None |
|
|
else: |
|
|
useMPI_default='no' |
|
|
mpi_root='/usr/local' |
|
|
mpi_path_default=os.path.join(mpi_root,'include') |
|
|
mpi_lib_path_default=os.path.join(mpi_root,'lib') |
|
|
mpi_libs_default=[ 'mpich' , 'pthread', 'rt' ] |
|
|
mpi_run_default='mpiexec -np 1' |
|
|
# |
|
|
#========================================================================== |
|
|
# |
|
|
# compile: |
|
|
# |
|
|
cc_flags_default='-O3 -std=c99 -ffast-math -fpic -Wno-unknown-pragmas -ansi' |
|
|
cc_flags_debug_default='-g -O0 -ffast-math -std=c99 -fpic -Wno-unknown-pragmas -ansi' |
|
|
cxx_flags_default='--no-warn -ansi' |
|
|
cxx_flags_debug_default='--no-warn -ansi -DDOASSERT' |
|
|
|
|
|
#============================================================================================== |
|
|
# Default options and options help text |
|
|
# These are defaults and can be overridden using command line arguments or an options file. |
|
|
# if the options_file or ARGUMENTS do not exist then the ones listed as default here are used |
|
|
# DO NOT CHANGE THEM HERE |
|
|
opts.AddOptions( |
|
78 |
# Where to install esys stuff |
# Where to install esys stuff |
79 |
('incinstall', 'where the esys headers will be installed', Dir('#.').abspath+'/include'), |
('prefix', 'where everything will be installed', Dir('#.').abspath), |
80 |
('libinstall', 'where the esys libraries will be installed', dir_libraries), |
('incinstall', 'where the esys headers will be installed', os.path.join(Dir('#.').abspath,'include')), |
81 |
('pyinstall', 'where the esys python modules will be installed', dir_packages), |
('bininstall', 'where the esys binaries will be installed', os.path.join(prefix,'bin')), |
82 |
('exinstall', 'where the esys examples will be installed', dir_examples), |
('libinstall', 'where the esys libraries will be installed', os.path.join(prefix,'lib')), |
83 |
('sys_libinstall', 'where the system esys libraries will be installed', sys_dir_libraries), |
('pyinstall', 'where the esys python modules will be installed', os.path.join(prefix,'esys')), |
|
('sys_pyinstall', 'where the system esys python modules will be installed', sys_dir_packages), |
|
|
('sys_exinstall', 'where the system esys examples will be installed', sys_dir_examples), |
|
|
('src_zipfile', 'the source zip file will be installed.', Dir('#.').abspath+"/release/escript_src.zip"), |
|
|
('test_zipfile', 'the test zip file will be installed.', Dir('#.').abspath+"/release/escript_tests.zip"), |
|
|
('src_tarfile', 'the source tar file will be installed.', Dir('#.').abspath+"/release/escript_src.tar.gz"), |
|
|
('test_tarfile', 'the test tar file will be installed.', Dir('#.').abspath+"/release/escript_tests.tar.gz"), |
|
|
('examples_tarfile', 'the examples tar file will be installed.', Dir('#.').abspath+"/release/doc/escript_examples.tar.gz"), |
|
|
('examples_zipfile', 'the examples zip file will be installed.', Dir('#.').abspath+"/release/doc/escript_examples.zip"), |
|
|
('guide_pdf', 'name of the user guide in pdf format', Dir('#.').abspath+"/release/doc/user/guide.pdf"), |
|
|
('api_epydoc', 'name of the epydoc api docs directory', Dir('#.').abspath+"/release/doc/epydoc"), |
|
|
('guide_html', 'name of the directory for user guide in html format', Dir('#.').abspath+"/release/doc/user/html"), |
|
|
('api_doxygen', 'name of the doxygen api docs directory',prefix+"/release/doc/doxygen"), |
|
84 |
# Compilation options |
# Compilation options |
85 |
BoolOption('dodebug', 'Do you want a debug build?', 'no'), |
BoolVariable('dodebug', 'For backwards compatibility', 'no'), |
86 |
BoolOption('bounds_check', 'Do you want extra array bounds checking?', 'no'), |
BoolVariable('usedebug', 'Do you want a debug build?', 'no'), |
87 |
('options_file', "Optional file containing preferred options. Ignored if it doesn't exist (default: scons/<hostname>_options.py)", options_file), |
BoolVariable('usevtk', 'Do you want to use VTK?', 'yes'), |
88 |
('cc_defines','C/C++ defines to use', None), |
('options_file', 'File of paths/options. Default: scons/<hostname>_options.py', options_file), |
89 |
('cc_flags','C compiler flags to use (Release build)', cc_flags_default), |
('cc', 'path to C compiler', 'DEFAULT'), |
90 |
('cc_flags_debug', 'C compiler flags to use (Debug build)', cc_flags_debug_default), |
('cxx', 'path to C++ compiler', 'DEFAULT'), |
91 |
('cxx_flags', 'C++ compiler flags to use (Release build)', cxx_flags_default), |
('win_cc_name', 'windows C compiler name if needed', 'msvc'), |
92 |
('cxx_flags_debug', 'C++ compiler flags to use (Debug build)', cxx_flags_debug_default), |
# The strings -DDEFAULT_ get replaced by scons/<hostname>_options.py or by defaults below |
93 |
('omp_flags', 'OpenMP compiler flags to use (Release build)', ''), |
('cc_flags', 'C/C++ compiler flags to use', '-DEFAULT_1'), |
94 |
('omp_flags_debug', 'OpenMP compiler flags to use (Debug build)', ''), |
('cc_optim', 'C/C++ optimization flags to use', '-DEFAULT_2'), |
95 |
('ar_flags', 'Static library archiver flags to use', None), |
('cc_debug', 'C/C++ debug flags to use', '-DEFAULT_3'), |
96 |
('sys_libs', 'System libraries to link with', None), |
('omp_optim', 'OpenMP compiler flags to use (Release build)', '-DEFAULT_4'), |
97 |
('tar_flags','flags for zip files','-c -z'), |
('omp_debug', 'OpenMP compiler flags to use (Debug build)', '-DEFAULT_5'), |
98 |
# MKL |
('omp_libs', 'OpenMP compiler libraries to link with', '-DEFAULT_6'), |
99 |
PathOption('mkl_path', 'Path to MKL includes', None), |
('cc_extra', 'Extra C compiler flags', ''), |
100 |
PathOption('mkl_lib_path', 'Path to MKL libs', None), |
('cxx_extra', 'Extra C++ compiler flags', ''), |
101 |
('mkl_libs', 'MKL libraries to link with', None), |
('ld_extra', 'Extra linker flags', ''), |
102 |
# SCSL |
('sys_libs', 'System libraries to link with', []), |
103 |
PathOption('scsl_path', 'Path to SCSL includes', None), |
('ar_flags', 'Static library archiver flags to use', ''), |
104 |
PathOption('scsl_lib_path', 'Path to SCSL libs', None), |
BoolVariable('useopenmp', 'Compile parallel version using OpenMP', 'no'), |
105 |
('scsl_libs', 'SCSL libraries to link with', None), |
BoolVariable('usepedantic', 'Compile with -pedantic if using gcc', 'no'), |
106 |
('scsl_libs_MPI', 'SCSL libraries to link with for MPI build', None), |
BoolVariable('usewarnings','Compile with warnings as errors if using gcc','yes'), |
107 |
# UMFPACK |
('forcelazy','for testing use only - set the default value for autolazy','leave_alone'), |
108 |
PathOption('ufc_path', 'Path to UFconfig includes', ufc_path_default), |
('forcecollres','for testing use only - set the default value for force resolving collective ops','leave_alone'), |
|
PathOption('umf_path', 'Path to UMFPACK includes', umf_path_default), |
|
|
PathOption('umf_lib_path', 'Path to UMFPACK libs', umf_lib_path_default), |
|
|
('umf_libs', 'UMFPACK libraries to link with', umf_libs_default), |
|
|
# AMD (used by UMFPACK) |
|
|
PathOption('amd_path', 'Path to AMD includes', amd_path_default), |
|
|
PathOption('amd_lib_path', 'Path to AMD libs', amd_lib_path_default), |
|
|
('amd_libs', 'AMD libraries to link with', amd_libs_default), |
|
|
# ParMETIS |
|
|
('parmetis_path', 'Path to ParMETIS includes', ''), |
|
|
('parmetis_lib_path', 'Path to ParMETIS library', ''), |
|
|
('parmetis_lib', 'ParMETIS library to link with', []), |
|
|
# TRILINOS |
|
|
PathOption('trilinos_path', 'Path to TRILINOS includes', None), |
|
|
PathOption('trilinos_lib_path', 'Path to TRILINOS libs', None), |
|
|
('trilinos_libs', 'TRILINOS libraries to link with', None), |
|
|
# BLAS |
|
|
PathOption('blas_path', 'Path to BLAS includes', None), |
|
|
PathOption('blas_lib_path', 'Path to BLAS libs', None), |
|
|
('blas_libs', 'BLAS libraries to link with', None), |
|
|
# netCDF |
|
|
('useNetCDF', 'switch on/off the usage of netCDF', useNetCDF_default), |
|
|
PathOption('netCDF_path', 'Path to netCDF includes', netCDF_path_default), |
|
|
PathOption('netCDF_lib_path', 'Path to netCDF libs', netCDF_lib_path_default), |
|
|
('netCDF_libs', 'netCDF C++ libraries to link with', netCDF_libs_default), |
|
109 |
# Python |
# Python |
110 |
# locations of include files for python |
('python_path', 'Path to Python includes', '/usr/include/'+python_version), |
111 |
# FIXME: python_path should be python_inc_path and the same for boost etc. |
('python_lib_path', 'Path to Python libs', usr_lib), |
112 |
PathOption('python_path', 'Path to Python includes', python_path_default), |
('python_libs', 'Python libraries to link with', [python_version]), |
|
PathOption('python_lib_path', 'Path to Python libs', python_lib_path_default), |
|
|
('python_lib', 'Python libraries to link with', python_lib_default), |
|
113 |
('python_cmd', 'Python command', 'python'), |
('python_cmd', 'Python command', 'python'), |
114 |
# Boost |
# Boost |
115 |
PathOption('boost_path', 'Path to Boost includes', boost_path_default), |
('boost_path', 'Path to Boost includes', '/usr/include'), |
116 |
PathOption('boost_lib_path', 'Path to Boost libs', boost_lib_path_default), |
('boost_lib_path', 'Path to Boost libs', usr_lib), |
117 |
('boost_lib', 'Boost libraries to link with', boost_lib_default), |
('boost_libs', 'Boost libraries to link with', ['boost_python']), |
118 |
# Doc building |
# NetCDF |
119 |
# PathOption('doxygen_path', 'Path to Doxygen executable', None), |
BoolVariable('usenetcdf', 'switch on/off the usage of netCDF', 'yes'), |
120 |
# PathOption('epydoc_path', 'Path to Epydoc executable', None), |
('netCDF_path', 'Path to netCDF includes', '/usr/include'), |
121 |
# PAPI |
('netCDF_lib_path', 'Path to netCDF libs', usr_lib), |
122 |
PathOption('papi_path', 'Path to PAPI includes', None), |
('netCDF_libs', 'netCDF C++ libraries to link with', ['netcdf_c++', 'netcdf']), |
|
PathOption('papi_lib_path', 'Path to PAPI libs', None), |
|
|
('papi_libs', 'PAPI libraries to link with', None), |
|
|
('papi_instrument_solver', 'use PAPI in Solver.c to instrument each iteration of the solver', None), |
|
123 |
# MPI |
# MPI |
124 |
BoolOption('useMPI', 'Compile parallel version using MPI', useMPI_default), |
BoolVariable('useMPI', 'For backwards compatibility', 'no'), |
125 |
|
BoolVariable('usempi', 'Compile parallel version using MPI', 'no'), |
126 |
('MPICH_IGNORE_CXX_SEEK', 'name of macro to ignore MPI settings of C++ SEEK macro (for MPICH)' , 'MPICH_IGNORE_CXX_SEEK'), |
('MPICH_IGNORE_CXX_SEEK', 'name of macro to ignore MPI settings of C++ SEEK macro (for MPICH)' , 'MPICH_IGNORE_CXX_SEEK'), |
127 |
PathOption('mpi_path', 'Path to MPI includes', mpi_path_default), |
('mpi_path', 'Path to MPI includes', '/usr/include'), |
128 |
('mpi_run', 'mpirun name' , mpi_run_default), |
('mpi_run', 'mpirun name' , 'mpiexec -np 1'), |
129 |
PathOption('mpi_lib_path', 'Path to MPI libs (needs to be added to the LD_LIBRARY_PATH)',mpi_lib_path_default), |
('mpi_lib_path', 'Path to MPI libs (needs to be added to the LD_LIBRARY_PATH)', usr_lib), |
130 |
('mpi_libs', 'MPI libraries to link with (needs to be shared!)', mpi_libs_default) |
('mpi_libs', 'MPI libraries to link with (needs to be shared!)', []), |
131 |
|
('mpi_flavour','Type of MPI execution environment','none'), |
132 |
|
# ParMETIS |
133 |
|
BoolVariable('useparmetis', 'Compile parallel version using ParMETIS', 'yes'), |
134 |
|
('parmetis_path', 'Path to ParMETIS includes', '/usr/include'), |
135 |
|
('parmetis_lib_path', 'Path to ParMETIS library', usr_lib), |
136 |
|
('parmetis_libs', 'ParMETIS library to link with', ['parmetis', 'metis']), |
137 |
|
# PAPI |
138 |
|
BoolVariable('usepapi', 'switch on/off the usage of PAPI', 'no'), |
139 |
|
('papi_path', 'Path to PAPI includes', '/usr/include'), |
140 |
|
('papi_lib_path', 'Path to PAPI libs', usr_lib), |
141 |
|
('papi_libs', 'PAPI libraries to link with', ['papi']), |
142 |
|
BoolVariable('papi_instrument_solver', 'use PAPI in Solver.c to instrument each iteration of the solver', False), |
143 |
|
# MKL |
144 |
|
BoolVariable('usemkl', 'switch on/off the usage of MKL', 'no'), |
145 |
|
('mkl_path', 'Path to MKL includes', '/sw/sdev/cmkl/10.0.2.18/include'), |
146 |
|
('mkl_lib_path', 'Path to MKL libs', '/sw/sdev/cmkl/10.0.2.18/lib/em64t'), |
147 |
|
('mkl_libs', 'MKL libraries to link with', ['mkl_solver', 'mkl_em64t', 'guide', 'pthread']), |
148 |
|
# UMFPACK |
149 |
|
BoolVariable('useumfpack', 'switch on/off the usage of UMFPACK', 'no'), |
150 |
|
('ufc_path', 'Path to UFconfig includes', '/usr/include/suitesparse'), |
151 |
|
('umf_path', 'Path to UMFPACK includes', '/usr/include/suitesparse'), |
152 |
|
('umf_lib_path', 'Path to UMFPACK libs', usr_lib), |
153 |
|
('umf_libs', 'UMFPACK libraries to link with', ['umfpack']), |
154 |
|
# Silo |
155 |
|
BoolVariable('usesilo', 'switch on/off the usage of Silo', 'yes'), |
156 |
|
('silo_path', 'Path to Silo includes', '/usr/include'), |
157 |
|
('silo_lib_path', 'Path to Silo libs', usr_lib), |
158 |
|
('silo_libs', 'Silo libraries to link with', ['siloh5', 'hdf5']), |
159 |
|
# VisIt |
160 |
|
BoolVariable('usevisit', 'switch on/off the usage of the VisIt sim library', 'no'), |
161 |
|
('visit_path', 'Path to VisIt libsim includes', '/usr/include'), |
162 |
|
('visit_lib_path', 'Path to VisIt sim library', usr_lib), |
163 |
|
# AMD (used by UMFPACK) |
164 |
|
('amd_path', 'Path to AMD includes', '/usr/include/suitesparse'), |
165 |
|
('amd_lib_path', 'Path to AMD libs', usr_lib), |
166 |
|
('amd_libs', 'AMD libraries to link with', ['amd']), |
167 |
|
# BLAS (used by UMFPACK) |
168 |
|
('blas_path', 'Path to BLAS includes', '/usr/include/suitesparse'), |
169 |
|
('blas_lib_path', 'Path to BLAS libs', usr_lib), |
170 |
|
('blas_libs', 'BLAS libraries to link with', ['blas']), |
171 |
|
#Lapack options |
172 |
|
BoolVariable('uselapack','switch on/off use of Lapack','no'), |
173 |
|
('lapack_path', 'Path to Lapack includes','/usr/include'), |
174 |
|
('lapack_lib_path', 'Path to Lapack libs', usr_lib), |
175 |
|
('lapack_libs', 'Lapack libraries to link with', []), |
176 |
|
('lapack_type', '{clapack,mkl}','clapack'), |
177 |
|
# An option for specifying the compiler tools set (see windows branch). |
178 |
|
('tools_names', 'allow control over the tools in the env setup', ['default']), |
179 |
|
# finer control over library building, intel aggressive global optimisation |
180 |
|
# works with dynamic libraries on windows. |
181 |
|
('share_esysUtils', 'control static or dynamic esysUtils lib', False), |
182 |
|
('share_paso', 'control static or dynamic paso lib', False), |
183 |
|
('env_export','Environment variables to be passed to children',[]), |
184 |
|
#To enable passing function pointers through python |
185 |
|
BoolVariable('iknowwhatimdoing','allow nonstandard C',False) |
186 |
) |
) |
187 |
#================================================================================================= |
|
188 |
# |
|
189 |
# Note: On the Altix the intel compilers are not automatically |
################### |
190 |
# detected by scons intelc.py script. The Altix has a different directory |
|
191 |
# path and in some locations the "modules" facility is used to support |
# This is only to support old versions of scons which don't accept |
192 |
# multiple compiler versions. This forces the need to import the users PATH |
# the variant_dir parameter (older than 0.98 I think). |
193 |
# environment which isn't the "scons way" |
# Once these are no longer an issue we can go back to a direct call |
194 |
# This doesn't impact linux and windows which will use the default compiler (g++ or msvc, or the intel compiler if it is installed on both platforms) |
# to obj.SConscript |
195 |
# FIXME: Perhaps a modification to intelc.py will allow better support for ia64 on altix |
import SCons |
196 |
# |
vs=SCons.__version__.split('.') |
197 |
|
cantusevariantdir=float(vs[0]+'.'+vs[1])<0.98 |
198 |
|
|
199 |
|
|
200 |
|
def CallSConscript(obj, **kw): |
201 |
|
if cantusevariantdir: |
202 |
|
if 'variant_dir' in kw: |
203 |
|
kw['build_dir']=kw['variant_dir'] |
204 |
|
del kw['variant_dir'] |
205 |
|
obj.SConscript(**kw) |
206 |
|
|
207 |
|
|
208 |
|
############ Specify which compilers to use #################### |
209 |
|
|
210 |
|
# intelc uses regular expressions improperly and emits a warning about |
211 |
|
# failing to find the compilers. This warning can be safely ignored. |
212 |
|
|
213 |
if IS_WINDOWS_PLATFORM: |
if IS_WINDOWS_PLATFORM: |
214 |
env = Environment(tools = ['default', 'msvc'], options = opts) |
env = Environment(options = opts) |
215 |
#env = Environment(tools = ['default', 'intelc'], options = opts) |
env = Environment(tools = ['default'] + env['tools_names'], |
216 |
|
options = opts) |
217 |
else: |
else: |
218 |
if socket.gethostname().split('.')[0] == 'service0': |
if os.uname()[4]=='ia64': |
|
env = Environment(tools = ['default', 'intelc'], options = opts) |
|
|
elif os.uname()[4]=='ia64': |
|
219 |
env = Environment(tools = ['default', 'intelc'], options = opts) |
env = Environment(tools = ['default', 'intelc'], options = opts) |
220 |
if env['CXX'] == 'icpc': |
if env['CXX'] == 'icpc': |
221 |
env['LINK'] = env['CXX'] # version >=9 of intel c++ compiler requires use of icpc to link in C++ runtimes (icc does not). FIXME: this behaviour could be directly incorporated into scons intelc.py |
env['LINK'] = env['CXX'] # version >=9 of intel c++ compiler requires use of icpc to link in C++ runtimes (icc does not) |
222 |
else: |
else: |
223 |
env = Environment(tools = ['default'], options = opts) |
env = Environment(tools = ['default'], options = opts) |
224 |
|
if env['tools_names']!='default': |
225 |
|
env=Environment(tools = ['default'] +env['tools_names'], options=opts) |
226 |
|
|
227 |
|
# Override compiler choice if provided |
228 |
|
if env['cc'] != 'DEFAULT': env['CC']=env['cc'] |
229 |
|
if env['cxx'] != 'DEFAULT': env['CXX']=env['cxx'] |
230 |
|
|
231 |
Help(opts.GenerateHelpText(env)) |
Help(opts.GenerateHelpText(env)) |
232 |
|
|
233 |
if env['bounds_check']: |
############ Make sure target directories exist ################ |
234 |
env.Append(CPPDEFINES = [ 'BOUNDS_CHECK' ]) |
|
235 |
env.Append(CXXDEFINES = [ 'BOUNDS_CHECK' ]) |
if not os.path.isdir(env['bininstall']): |
236 |
bounds_check = env['bounds_check'] |
os.makedirs(env['bininstall']) |
237 |
|
if not os.path.isdir(env['libinstall']): |
238 |
|
os.makedirs(env['libinstall']) |
239 |
|
if not os.path.isdir(env['pyinstall']): |
240 |
|
os.makedirs(env['pyinstall']) |
241 |
|
|
242 |
|
########## Copy required environment vars ###################### |
243 |
|
|
244 |
|
for i in env['env_export']: |
245 |
|
env.Append(ENV = {i:os.environ[i]}) |
246 |
|
|
247 |
|
############ Fill in compiler options if not set above ######### |
248 |
|
|
249 |
|
# Backwards compatibility: allow dodebug=yes and useMPI=yes |
250 |
|
if env['dodebug']: env['usedebug'] = 1 |
251 |
|
if env['useMPI']: env['usempi'] = 1 |
252 |
|
|
253 |
|
# Default compiler options (override allowed in hostname_options.py, but should not be necessary) |
254 |
|
# For both C and C++ you get: cc_flags and either the optim flags or debug flags |
255 |
|
|
256 |
|
sysheaderopt = "" # how do we indicate that a header is a system header. Use "" for no action. |
257 |
|
|
258 |
|
cc_flags = "" |
259 |
|
cc_optim = "" |
260 |
|
cc_debug = "" |
261 |
|
omp_optim = "" |
262 |
|
omp_debug = "" |
263 |
|
omp_libs = [] |
264 |
|
|
265 |
|
if env["CC"] == "icc": |
266 |
|
# Intel compilers |
267 |
|
cc_flags = "-std=c99 -fPIC -wd161 -w1 -vec-report0 -DBLOCKTIMER -DCORE_ID1" |
268 |
|
cc_optim = "-O3 -ftz -IPF_ftlacc- -IPF_fma -fno-alias -ip" |
269 |
|
cc_debug = "-g -O0 -DDOASSERT -DDOPROF -DBOUNDS_CHECK" |
270 |
|
omp_optim = "-openmp -openmp_report0" |
271 |
|
omp_debug = "-openmp -openmp_report0" |
272 |
|
omp_libs = ['guide', 'pthread'] |
273 |
|
pedantic = "" |
274 |
|
fatalwarning = "" # Switch to turn warnings into errors |
275 |
|
sysheaderopt = "" |
276 |
|
elif env["CC"][:3] == "gcc": |
277 |
|
# GNU C on any system |
278 |
|
cc_flags = "-pedantic -Wall -fPIC -ffast-math -Wno-unknown-pragmas -DBLOCKTIMER -Wno-sign-compare -Wno-system-headers -Wno-long-long -Wno-strict-aliasing -finline-functions" |
279 |
|
#the long long warning occurs on the Mac |
280 |
|
cc_optim = "-O3" |
281 |
|
cc_debug = "-g -O0 -DDOASSERT -DDOPROF -DBOUNDS_CHECK" |
282 |
|
omp_optim = "-fopenmp" |
283 |
|
omp_debug = "-fopenmp" |
284 |
|
omp_libs = [] |
285 |
|
pedantic = "-pedantic-errors -Wno-long-long" |
286 |
|
fatalwarning = "-Werror" |
287 |
|
sysheaderopt = "-isystem " |
288 |
|
elif env["CC"] == "cl": |
289 |
|
# Microsoft Visual C on Windows |
290 |
|
cc_flags = "/FD /EHsc /GR /wd4068 -D_USE_MATH_DEFINES -DDLL_NETCDF" |
291 |
|
cc_optim = "/O2 /Op /MT /W3" |
292 |
|
cc_debug = "/Od /RTC1 /MTd /ZI -DBOUNDS_CHECK" |
293 |
|
omp_optim = "" |
294 |
|
omp_debug = "" |
295 |
|
omp_libs = [] |
296 |
|
pedantic = "" |
297 |
|
fatalwarning = "" |
298 |
|
sysheaderopt = "" |
299 |
|
elif env["CC"] == "icl": |
300 |
|
# intel C on Windows, see windows_intelc_options.py for a start |
301 |
|
pedantic = "" |
302 |
|
fatalwarning = "" |
303 |
|
sysheaderopt = "" |
304 |
|
|
305 |
|
|
306 |
|
# If not specified in hostname_options.py then set them here |
307 |
|
if env["cc_flags"] == "-DEFAULT_1": env['cc_flags'] = cc_flags |
308 |
|
if env["cc_optim"] == "-DEFAULT_2": env['cc_optim'] = cc_optim |
309 |
|
if env["cc_debug"] == "-DEFAULT_3": env['cc_debug'] = cc_debug |
310 |
|
if env["omp_optim"] == "-DEFAULT_4": env['omp_optim'] = omp_optim |
311 |
|
if env["omp_debug"] == "-DEFAULT_5": env['omp_debug'] = omp_debug |
312 |
|
if env["omp_libs"] == "-DEFAULT_6": env['omp_libs'] = omp_libs |
313 |
|
|
314 |
|
#set up the autolazy values |
315 |
|
if env['forcelazy'] != "leave_alone": |
316 |
|
if env['forcelazy'] == 'on': |
317 |
|
env.Append(CPPDEFINES=['FAUTOLAZYON']) |
318 |
|
else: |
319 |
|
if env['forcelazy'] == 'off': |
320 |
|
env.Append(CPPDEFINES=['FAUTOLAZYOFF']) |
321 |
|
|
322 |
|
#set up the colective resolve values |
323 |
|
if env['forcecollres'] != "leave_alone": |
324 |
|
print env['forcecollres'] |
325 |
|
if env['forcecollres'] == 'on': |
326 |
|
env.Append(CPPDEFINES=['FRESCOLLECTON']) |
327 |
|
else: |
328 |
|
if env['forcecollres'] == 'off': |
329 |
|
env.Append(CPPDEFINES=['FRESCOLLECTOFF']) |
330 |
|
|
331 |
|
|
332 |
|
if env['iknowwhatimdoing']: |
333 |
|
env.Append(CPPDEFINES=['IKNOWWHATIMDOING']) |
334 |
|
|
335 |
|
# OpenMP is disabled if useopenmp=no or both variables omp_optim and omp_debug are empty |
336 |
|
if not env["useopenmp"]: |
337 |
|
env['omp_optim'] = "" |
338 |
|
env['omp_debug'] = "" |
339 |
|
env['omp_libs'] = [] |
340 |
|
|
341 |
|
if env['omp_optim'] == "" and env['omp_debug'] == "": env["useopenmp"] = 0 |
342 |
|
|
343 |
|
# Windows doesn't use LD_LIBRARY_PATH but PATH instead |
344 |
|
if IS_WINDOWS_PLATFORM: |
345 |
|
LD_LIBRARY_PATH_KEY='PATH' |
346 |
|
env['ENV']['LD_LIBRARY_PATH']='' |
347 |
else: |
else: |
348 |
bounds_check = 0 |
LD_LIBRARY_PATH_KEY='LD_LIBRARY_PATH' |
349 |
|
############ Copy environment variables into scons env ######### |
350 |
|
|
351 |
#================================================================================================= |
try: env['ENV']['OMP_NUM_THREADS'] = os.environ['OMP_NUM_THREADS'] |
352 |
# |
except KeyError: env['ENV']['OMP_NUM_THREADS'] = 1 |
|
# Initialise Scons Build Environment |
|
|
# check for user environment variables we are interested in |
|
|
try: |
|
|
tmp = os.environ['PYTHONPATH'] |
|
|
env['ENV']['PYTHONPATH'] = tmp |
|
|
except KeyError: |
|
|
pass |
|
353 |
|
|
354 |
env.PrependENVPath('PYTHONPATH', source_root) |
try: env['ENV']['ESCRIPT_NUM_THREADS'] = os.environ['ESCRIPT_NUM_THREADS'] |
355 |
|
except KeyError: pass |
356 |
|
|
357 |
try: |
try: env['ENV']['ESCRIPT_NUM_PROCS'] = os.environ['ESCRIPT_NUM_PROCS'] |
358 |
omp_num_threads = os.environ['OMP_NUM_THREADS'] |
except KeyError: env['ENV']['ESCRIPT_NUM_PROCS']=1 |
|
except KeyError: |
|
|
omp_num_threads = 1 |
|
|
env['ENV']['OMP_NUM_THREADS'] = omp_num_threads |
|
359 |
|
|
360 |
try: |
try: env['ENV']['ESCRIPT_NUM_NODES'] = os.environ['ESCRIPT_NUM_NODES'] |
361 |
path = os.environ['PATH'] |
except KeyError: env['ENV']['ESCRIPT_NUM_NODES']=1 |
|
env['ENV']['PATH'] = path |
|
|
except KeyError: |
|
|
omp_num_threads = 1 |
|
362 |
|
|
363 |
env['ENV']['OMP_NUM_THREADS'] = omp_num_threads |
try: env['ENV']['ESCRIPT_HOSTFILE'] = os.environ['ESCRIPT_HOSTFILE'] |
364 |
|
except KeyError: pass |
365 |
|
|
366 |
|
try: env['ENV']['PATH'] = os.environ['PATH'] |
367 |
|
except KeyError: pass |
368 |
|
|
369 |
# Copy some variables from the system environment to the build environment |
try: env['ENV']['PYTHONPATH'] = os.environ['PYTHONPATH'] |
370 |
try: |
except KeyError: pass |
|
env['ENV']['DISPLAY'] = os.environ['DISPLAY'] |
|
|
env['ENV']['XAUTHORITY'] = os.environ['XAUTHORITY'] |
|
|
home_temp = os.environ['HOME'] # MPICH2's mpd needs $HOME to find $HOME/.mpd.conf |
|
|
env['ENV']['HOME'] = home_temp |
|
|
except KeyError: |
|
|
pass |
|
371 |
|
|
372 |
try: |
try: env['ENV']['C_INCLUDE_PATH'] = os.environ['C_INCLUDE_PATH'] |
373 |
tmp = os.environ['PATH'] |
except KeyError: pass |
|
env['ENV']['PATH'] = tmp |
|
|
except KeyError: |
|
|
pass |
|
374 |
|
|
375 |
try: |
try: env['ENV']['CPLUS_INCLUDE_PATH'] = os.environ['CPLUS_INCLUDE_PATH'] |
376 |
tmp = os.environ['LD_LIBRARY_PATH'] |
except KeyError: pass |
|
env['ENV']['LD_LIBRARY_PATH'] = tmp |
|
|
except KeyError: |
|
|
pass |
|
|
#========================================================================== |
|
|
# |
|
|
# Add some customer builders |
|
|
# |
|
|
py_builder = Builder(action = scons_extensions.build_py, suffix = '.pyc', src_suffix = '.py', single_source=True) |
|
|
env.Append(BUILDERS = {'PyCompile' : py_builder}); |
|
377 |
|
|
378 |
runUnitTest_builder = Builder(action = scons_extensions.runUnitTest, suffix = '.passed', |
try: env.PrependENVPath(LD_LIBRARY_PATH_KEY,os.environ['LD_LIBRARY_PATH']) |
379 |
src_suffix=env['PROGSUFFIX'], single_source=True) |
except KeyError: pass |
380 |
|
|
381 |
env.Append(BUILDERS = {'RunUnitTest' : runUnitTest_builder}); |
try: env['ENV']['LIBRARY_PATH'] = os.environ['LIBRARY_PATH'] |
382 |
|
except KeyError: pass |
383 |
|
|
384 |
runPyUnitTest_builder = Builder(action = scons_extensions.runPyUnitTest, suffix = '.passed', src_suffic='.py', single_source=True) |
try: env['ENV']['DISPLAY'] = os.environ['DISPLAY'] |
385 |
env.Append(BUILDERS = {'RunPyUnitTest' : runPyUnitTest_builder}); |
except KeyError: pass |
386 |
|
|
387 |
# Convert the options which are held in environment variable into python variables for ease of handling and configure compilation options |
try: env['ENV']['XAUTHORITY'] = os.environ['XAUTHORITY'] |
388 |
try: |
except KeyError: pass |
|
incinstall = env['incinstall'] |
|
|
env.Append(CPPPATH = [incinstall,]) |
|
|
except KeyError: |
|
|
incinstall = None |
|
|
try: |
|
|
libinstall = env['libinstall'] |
|
|
env.Append(LIBPATH = [libinstall,]) # Adds -L for building of libescript.so libfinley.so escriptcpp.so finleycpp.so |
|
|
env.PrependENVPath('LD_LIBRARY_PATH', libinstall) |
|
|
if IS_WINDOWS_PLATFORM : |
|
|
env.PrependENVPath('PATH', libinstall) |
|
|
env.PrependENVPath('PATH', env['boost_lib_path']) |
|
|
except KeyError: |
|
|
libinstall = None |
|
|
try: |
|
|
pyinstall = env['pyinstall'] # all targets will install into pyinstall/esys but PYTHONPATH points at straight pyinstall so you go import esys.escript etc |
|
|
except KeyError: |
|
|
pyinstall = None |
|
389 |
|
|
390 |
try: |
try: env['ENV']['HOME'] = os.environ['HOME'] |
391 |
cc_defines = env['cc_defines'] |
except KeyError: pass |
|
env.Append(CPPDEFINES = cc_defines) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
flags = env['ar_flags'] |
|
|
env.Append(ARFLAGS = flags) |
|
|
except KeyError: |
|
|
ar_flags = None |
|
|
try: |
|
|
sys_libs = env['sys_libs'] |
|
|
except KeyError: |
|
|
sys_libs = [] |
|
392 |
|
|
393 |
try: |
# Configure for test suite |
|
tar_flags = env['tar_flags'] |
|
|
env.Replace(TARFLAGS = tar_flags) |
|
|
except KeyError: |
|
|
pass |
|
394 |
|
|
|
try: |
|
|
exinstall = env['exinstall'] |
|
|
except KeyError: |
|
|
exinstall = None |
|
|
try: |
|
|
sys_libinstall = env['sys_libinstall'] |
|
|
except KeyError: |
|
|
sys_libinstall = None |
|
|
try: |
|
|
sys_pyinstall = env['sys_pyinstall'] |
|
|
except KeyError: |
|
|
sys_pyinstall = None |
|
|
try: |
|
|
sys_exinstall = env['sys_exinstall'] |
|
|
except KeyError: |
|
|
sys_exinstall = None |
|
395 |
|
|
396 |
# ====================== debugging =================================== |
env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['libinstall']) |
397 |
try: |
env.PrependENVPath('PYTHONPATH', prefix) |
398 |
dodebug = env['dodebug'] |
env['ENV']['ESCRIPT_ROOT'] = prefix |
|
except KeyError: |
|
|
dodebug = None |
|
399 |
|
|
400 |
# === switch on omp =================================================== |
############ Set up paths for Configure() ###################### |
401 |
try: |
|
402 |
omp_flags = env['omp_flags'] |
# Make a copy of an environment |
403 |
except KeyError: |
# Use env.Clone if available, but fall back on env.Copy for older version of scons |
404 |
omp_flags = '' |
def clone_env(env): |
405 |
|
if 'Clone' in dir(env): return env.Clone() # scons-0.98 |
406 |
|
else: return env.Copy() # scons-0.96 |
407 |
|
|
408 |
|
# Add cc option -I<Escript>/trunk/include |
409 |
|
env.Append(CPPPATH = [Dir('include')]) |
410 |
|
|
411 |
|
# Add cc option -L<Escript>/trunk/lib |
412 |
|
env.Append(LIBPATH = [Dir(env['libinstall'])]) |
413 |
|
|
414 |
|
if env['cc_extra'] != '': env.Append(CFLAGS = env['cc_extra']) |
415 |
|
if env['cxx_extra'] != '': env.Append(CXXFLAGS = env['cxx_extra']) |
416 |
|
if env['ld_extra'] != '': env.Append(LINKFLAGS = env['ld_extra']) |
417 |
|
|
418 |
|
if env['usepedantic']: env.Append(CCFLAGS = pedantic) |
419 |
|
|
420 |
|
# MS Windows |
421 |
|
if IS_WINDOWS_PLATFORM: |
422 |
|
env.AppendENVPath('PATH', [env['boost_lib_path']]) |
423 |
|
env.AppendENVPath('PATH', [env['libinstall']]) |
424 |
|
if not env['share_esysUtils'] : |
425 |
|
env.Append(CPPDEFINES = ['ESYSUTILS_STATIC_LIB']) |
426 |
|
if not env['share_paso'] : |
427 |
|
env.Append(CPPDEFINES = ['PASO_STATIC_LIB']) |
428 |
|
|
429 |
|
if env['usenetcdf']: |
430 |
|
env.AppendENVPath('PATH', [env['netCDF_lib_path']]) |
431 |
|
|
432 |
|
env.Append(ARFLAGS = env['ar_flags']) |
433 |
|
|
434 |
|
# Get the global Subversion revision number for getVersion() method |
435 |
try: |
try: |
436 |
omp_flags_debug = env['omp_flags_debug'] |
global_revision = os.popen("svnversion -n .").read() |
437 |
except KeyError: |
global_revision = re.sub(":.*", "", global_revision) |
438 |
omp_flags_debug = '' |
global_revision = re.sub("[^0-9]", "", global_revision) |
439 |
|
except: |
440 |
|
global_revision="-1" |
441 |
|
if global_revision == "": global_revision="-2" |
442 |
|
env.Append(CPPDEFINES = ["SVN_VERSION="+global_revision]) |
443 |
|
|
444 |
|
############ numpy (required) ############################### |
445 |
|
|
|
# ========= use mpi? ===================================================== |
|
446 |
try: |
try: |
447 |
useMPI = env['useMPI'] |
from numpy import identity |
448 |
except KeyError: |
except ImportError: |
449 |
useMPI = None |
print "Cannot import numpy, you need to set your PYTHONPATH" |
450 |
# ========= set compiler flags =========================================== |
sys.exit(1) |
451 |
|
|
452 |
if dodebug: |
############ C compiler (required) ############################# |
453 |
try: |
|
454 |
flags = env['cc_flags_debug'] + ' ' + omp_flags_debug |
# Create a Configure() environment for checking existence of required libraries and headers |
455 |
env.Append(CCFLAGS = flags) |
conf = Configure(clone_env(env)) |
456 |
except KeyError: |
|
457 |
pass |
# Test that the compiler is working |
458 |
|
if not conf.CheckFunc('printf'): |
459 |
|
print "Cannot run C compiler '%s' (or libc is missing)" % (env['CC']) |
460 |
|
sys.exit(1) |
461 |
|
|
462 |
|
if conf.CheckFunc('gethostname'): |
463 |
|
conf.env.Append(CPPDEFINES = ['HAVE_GETHOSTNAME']) |
464 |
|
|
465 |
|
############ python libraries (required) ####################### |
466 |
|
|
467 |
|
|
468 |
|
if not sysheaderopt =="": |
469 |
|
conf.env.Append(CCFLAGS=sysheaderopt+env['python_path']) |
470 |
else: |
else: |
471 |
try: |
conf.env.AppendUnique(CPPPATH = [env['python_path']]) |
472 |
flags = env['cc_flags'] + ' ' + omp_flags |
|
473 |
env.Append(CCFLAGS = flags) |
conf.env.AppendUnique(LIBPATH = [env['python_lib_path']]) |
474 |
except KeyError: |
conf.env.AppendUnique(LIBS = [env['python_libs']]) |
|
pass |
|
|
if dodebug: |
|
|
try: |
|
|
flags = env['cxx_flags_debug'] |
|
|
env.Append(CXXFLAGS = flags) |
|
|
except KeyError: |
|
|
pass |
|
|
else: |
|
|
try: |
|
|
flags = env['cxx_flags'] |
|
|
env.Append(CXXFLAGS = flags) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
if env['CC'] == 'gcc': env.Append(CCFLAGS = "-pedantic-errors -Wno-long-long") |
|
|
except: |
|
|
pass |
|
475 |
|
|
476 |
# ============= Remember what options were used in the compile ===================================== |
conf.env.PrependENVPath('PYTHONPATH', prefix) |
477 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['python_lib_path']) # The wrapper script needs to find these libs |
478 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['libinstall']) |
479 |
|
|
480 |
|
if not conf.CheckCHeader('Python.h'): |
481 |
|
print "Cannot find python include files (tried 'Python.h' in directory %s)" % (env['python_path']) |
482 |
|
sys.exit(1) |
483 |
|
if not conf.CheckFunc('Py_Exit'): |
484 |
|
print "Cannot find python library method Py_Main (tried lib %s in directory %s)" % (env['python_libs'], env['python_lib_path']) |
485 |
|
sys.exit(1) |
486 |
|
|
487 |
|
############ boost (required) ################################## |
488 |
|
|
489 |
|
if not sysheaderopt =="": |
490 |
|
# This is required because we can't -isystem /usr/system because it breaks std includes |
491 |
|
if os.path.normpath(env['boost_path']) =="/usr/include": |
492 |
|
conf.env.Append(CCFLAGS=sysheaderopt+os.path.join(env['boost_path'],'boost')) |
493 |
|
else: |
494 |
|
conf.env.Append(CCFLAGS=sysheaderopt+env['boost_path']) |
495 |
|
else: |
496 |
|
conf.env.AppendUnique(CPPPATH = [env['boost_path']]) |
497 |
|
|
498 |
|
conf.env.AppendUnique(LIBPATH = [env['boost_lib_path']]) |
499 |
|
conf.env.AppendUnique(LIBS = [env['boost_libs']]) |
500 |
|
|
501 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['boost_lib_path']) # The wrapper script needs to find these libs |
502 |
|
#ensure that our path entries remain at the front |
503 |
|
conf.env.PrependENVPath('PYTHONPATH', prefix) |
504 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['libinstall']) |
505 |
|
|
506 |
|
|
507 |
|
#Yep we still cant figure this one out. - working on it. |
508 |
if not IS_WINDOWS_PLATFORM: |
if not IS_WINDOWS_PLATFORM: |
509 |
env.Execute("/bin/rm -f " + libinstall + "/Compiled.with.*") |
if not conf.CheckCXXHeader('boost/python.hpp'): |
510 |
if dodebug: env.Execute("touch " + libinstall + "/Compiled.with.debug") |
print "Cannot find boost include files (tried boost/python.hpp in directory %s)" % (env['boost_path']) |
511 |
if useMPI: env.Execute("touch " + libinstall + "/Compiled.with.mpi") |
sys.exit(1) |
512 |
if omp_flags != '': env.Execute("touch " + libinstall + "/Compiled.with.OpenMP") |
|
513 |
if bounds_check: env.Execute("touch " + libinstall + "/Compiled.with.bounds_check") |
if not conf.CheckFunc('PyObject_SetAttr'): |
514 |
|
print "Cannot find boost library method PyObject_SetAttr (tried method PyObject_SetAttr in library %s in directory %s)" % (env['boost_libs'], env['boost_lib_path']) |
515 |
|
sys.exit(1) |
516 |
|
|
517 |
|
|
518 |
|
# Commit changes to environment |
519 |
|
env = conf.Finish() |
520 |
|
|
521 |
|
############ VTK (optional) #################################### |
522 |
|
|
523 |
|
if env['usevtk']: |
524 |
|
try: |
525 |
|
import vtk |
526 |
|
env['usevtk'] = 1 |
527 |
|
except ImportError: |
528 |
|
env['usevtk'] = 0 |
529 |
|
|
530 |
|
# Add VTK to environment env if it was found |
531 |
|
if env['usevtk']: |
532 |
|
env.Append(CPPDEFINES = ['USE_VTK']) |
533 |
|
|
534 |
|
############ NetCDF (optional) ################################# |
535 |
|
|
536 |
|
conf = Configure(clone_env(env)) |
537 |
|
|
538 |
|
if env['usenetcdf']: |
539 |
|
conf.env.AppendUnique(CPPPATH = [env['netCDF_path']]) |
540 |
|
conf.env.AppendUnique(LIBPATH = [env['netCDF_lib_path']]) |
541 |
|
conf.env.AppendUnique(LIBS = [env['netCDF_libs']]) |
542 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['netCDF_lib_path']) # The wrapper script needs to find these libs |
543 |
|
#ensure that our path entries remain at the front |
544 |
|
conf.env.PrependENVPath('PYTHONPATH', prefix) |
545 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['libinstall']) |
546 |
|
|
547 |
|
if env['usenetcdf'] and not conf.CheckCHeader('netcdf.h'): env['usenetcdf'] = 0 |
548 |
|
if env['usenetcdf'] and not conf.CheckFunc('nc_open'): env['usenetcdf'] = 0 |
549 |
|
|
550 |
|
# Add NetCDF to environment env if it was found |
551 |
|
if env['usenetcdf']: |
552 |
|
env = conf.Finish() |
553 |
|
env.Append(CPPDEFINES = ['USE_NETCDF']) |
554 |
|
else: |
555 |
|
conf.Finish() |
556 |
|
|
557 |
# ============= set mkl (but only of no MPI) ===================================== |
############ PAPI (optional) ################################### |
|
if not useMPI: |
|
|
try: |
|
|
includes = env['mkl_path'] |
|
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
558 |
|
|
559 |
try: |
# Start a new configure environment that reflects what we've already found |
560 |
lib_path = env['mkl_lib_path'] |
conf = Configure(clone_env(env)) |
|
env.Append(LIBPATH = [lib_path,]) |
|
|
env['ENV']['LD_LIBRARY_PATH']+=":"+lib_path |
|
|
except KeyError: |
|
|
pass |
|
561 |
|
|
562 |
try: |
if env['usepapi']: |
563 |
mkl_libs = env['mkl_libs'] |
conf.env.AppendUnique(CPPPATH = [env['papi_path']]) |
564 |
except KeyError: |
conf.env.AppendUnique(LIBPATH = [env['papi_lib_path']]) |
565 |
mkl_libs = [] |
conf.env.AppendUnique(LIBS = [env['papi_libs']]) |
566 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['papi_lib_path']) # The wrapper script needs to find these libs |
567 |
|
#ensure that our path entries remain at the front |
568 |
|
conf.env.PrependENVPath('PYTHONPATH', prefix) |
569 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['libinstall']) |
570 |
|
|
571 |
|
if env['usepapi'] and not conf.CheckCHeader('papi.h'): env['usepapi'] = 0 |
572 |
|
if env['usepapi'] and not conf.CheckFunc('PAPI_start_counters'): env['usepapi'] = 0 |
573 |
|
|
574 |
|
# Add PAPI to environment env if it was found |
575 |
|
if env['usepapi']: |
576 |
|
env = conf.Finish() |
577 |
|
env.Append(CPPDEFINES = ['BLOCKPAPI']) |
578 |
else: |
else: |
579 |
mkl_libs = [] |
conf.Finish() |
580 |
|
|
581 |
# ============= set scsl (but only of no MPI) ===================================== |
############ MKL (optional) #################################### |
|
if not useMPI: |
|
|
try: |
|
|
includes = env['scsl_path'] |
|
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
582 |
|
|
583 |
try: |
# Start a new configure environment that reflects what we've already found |
584 |
lib_path = env['scsl_lib_path'] |
conf = Configure(clone_env(env)) |
|
env.Append(LIBPATH = [lib_path,]) |
|
|
env['ENV']['LD_LIBRARY_PATH']+=":"+lib_path |
|
|
except KeyError: |
|
|
pass |
|
|
|
|
|
try: |
|
|
scsl_libs = env['scsl_libs'] |
|
|
except KeyError: |
|
|
scsl_libs = [ ] |
|
585 |
|
|
586 |
|
if env['usemkl']: |
587 |
|
conf.env.AppendUnique(CPPPATH = [env['mkl_path']]) |
588 |
|
conf.env.AppendUnique(LIBPATH = [env['mkl_lib_path']]) |
589 |
|
conf.env.AppendUnique(LIBS = [env['mkl_libs']]) |
590 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['mkl_lib_path']) # The wrapper script needs to find these libs |
591 |
|
#ensure that our path entries remain at the front |
592 |
|
conf.env.PrependENVPath('PYTHONPATH', prefix) |
593 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['libinstall']) |
594 |
|
|
595 |
|
if env['usemkl'] and not conf.CheckCHeader('mkl_solver.h'): env['usemkl'] = 0 |
596 |
|
if env['usemkl'] and not conf.CheckFunc('pardiso'): env['usemkl'] = 0 |
597 |
|
|
598 |
|
|
599 |
|
# Add MKL to environment env if it was found |
600 |
|
if env['usemkl']: |
601 |
|
env = conf.Finish() |
602 |
|
env.Append(CPPDEFINES = ['MKL']) |
603 |
else: |
else: |
604 |
scsl_libs = [] |
conf.Finish() |
605 |
|
|
606 |
# ============= set TRILINOS (but only with MPI) ===================================== |
############ UMFPACK (optional) ################################ |
|
if useMPI: |
|
|
try: |
|
|
includes = env['trilinos_path'] |
|
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
607 |
|
|
608 |
try: |
# Start a new configure environment that reflects what we've already found |
609 |
lib_path = env['trilinos_lib_path'] |
conf = Configure(clone_env(env)) |
|
env.Append(LIBPATH = [lib_path,]) |
|
|
except KeyError: |
|
|
pass |
|
610 |
|
|
611 |
try: |
if env['useumfpack']: |
612 |
trilinos_libs = env['trilinos_libs'] |
conf.env.AppendUnique(CPPPATH = [env['ufc_path']]) |
613 |
except KeyError: |
conf.env.AppendUnique(CPPPATH = [env['umf_path']]) |
614 |
trilinos_libs = [] |
conf.env.AppendUnique(LIBPATH = [env['umf_lib_path']]) |
615 |
|
conf.env.AppendUnique(LIBS = [env['umf_libs']]) |
616 |
|
conf.env.AppendUnique(CPPPATH = [env['amd_path']]) |
617 |
|
conf.env.AppendUnique(LIBPATH = [env['amd_lib_path']]) |
618 |
|
conf.env.AppendUnique(LIBS = [env['amd_libs']]) |
619 |
|
conf.env.AppendUnique(CPPPATH = [env['blas_path']]) |
620 |
|
conf.env.AppendUnique(LIBPATH = [env['blas_lib_path']]) |
621 |
|
conf.env.AppendUnique(LIBS = [env['blas_libs']]) |
622 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['umf_lib_path']) # The wrapper script needs to find these libs |
623 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['amd_lib_path']) # The wrapper script needs to find these libs |
624 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['blas_lib_path']) # The wrapper script needs to find these libs |
625 |
|
#ensure that our path entries remain at the front |
626 |
|
conf.env.PrependENVPath('PYTHONPATH', prefix) |
627 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['libinstall']) |
628 |
|
|
629 |
|
if env['useumfpack'] and not conf.CheckCHeader('umfpack.h'): env['useumfpack'] = 0 |
630 |
|
if env['useumfpack'] and not conf.CheckFunc('umfpack_di_symbolic'): env['useumfpack'] = 0 |
631 |
|
# if env['useumfpack'] and not conf.CheckFunc('daxpy'): env['useumfpack'] = 0 # this does not work on shake73? |
632 |
|
|
633 |
|
# Add UMFPACK to environment env if it was found |
634 |
|
if env['useumfpack']: |
635 |
|
env = conf.Finish() |
636 |
|
env.Append(CPPDEFINES = ['UMFPACK']) |
637 |
else: |
else: |
638 |
trilinos_libs = [] |
conf.Finish() |
639 |
|
|
640 |
|
############ Silo (optional) ################################### |
641 |
|
|
642 |
# ============= set umfpack (but only without MPI) ===================================== |
if env['usesilo']: |
643 |
umf_libs=[ ] |
conf = Configure(clone_env(env)) |
644 |
if not useMPI: |
conf.env.AppendUnique(CPPPATH = [env['silo_path']]) |
645 |
try: |
conf.env.AppendUnique(LIBPATH = [env['silo_lib_path']]) |
646 |
includes = env['umf_path'] |
conf.env.AppendUnique(LIBS = [env['silo_libs']]) |
647 |
env.Append(CPPPATH = [includes,]) |
if not conf.CheckCHeader('silo.h'): env['usesilo'] = 0 |
648 |
except KeyError: |
if not conf.CheckFunc('DBMkDir'): env['usesilo'] = 0 |
649 |
pass |
conf.Finish() |
650 |
|
|
651 |
|
# Add the path to Silo to environment env if it was found. |
652 |
|
# Note that we do not add the libs since they are only needed for the |
653 |
|
# weipa library and tools. |
654 |
|
if env['usesilo']: |
655 |
|
env.AppendUnique(CPPPATH = [env['silo_path']]) |
656 |
|
env.AppendUnique(LIBPATH = [env['silo_lib_path']]) |
657 |
|
|
658 |
|
############ VisIt (optional) ################################### |
659 |
|
|
660 |
|
if env['usevisit']: |
661 |
|
env.AppendUnique(CPPPATH = [env['visit_path']]) |
662 |
|
env.AppendUnique(LIBPATH = [env['visit_lib_path']]) |
663 |
|
|
664 |
|
########### Lapack (optional) ################################## |
665 |
|
|
666 |
|
if env['uselapack']: |
667 |
|
env.AppendUnique(CPPDEFINES='USE_LAPACK') |
668 |
|
env.AppendUnique(CPPPATH = [env['lapack_path']]) |
669 |
|
env.AppendUnique(LIBPATH =[env['lapack_lib_path']]) |
670 |
|
|
671 |
|
env.Append(LIBPATH = '/usr/lib/atlas') |
672 |
|
env.Append(LIBS = [env['lapack_libs']]) |
673 |
|
if env['lapack_type']=='mkl': |
674 |
|
if not env['usemkl']: |
675 |
|
env['uselapack']=0 |
676 |
|
print "mkl_lapack requires mkl" |
677 |
|
else: |
678 |
|
env.AppendUnique(CPPDEFINES='MKL_LAPACK') |
679 |
|
|
680 |
|
|
681 |
|
############ Add the compiler flags ############################ |
682 |
|
|
683 |
|
# Enable debug by choosing either cc_debug or cc_optim |
684 |
|
if env['usedebug']: |
685 |
|
env.Append(CCFLAGS = env['cc_debug']) |
686 |
|
env.Append(CCFLAGS = env['omp_debug']) |
687 |
|
else: |
688 |
|
env.Append(CCFLAGS = env['cc_optim']) |
689 |
|
env.Append(CCFLAGS = env['omp_optim']) |
690 |
|
|
691 |
try: |
# Always use cc_flags |
692 |
lib_path = env['umf_lib_path'] |
env.Append(CCFLAGS = env['cc_flags']) |
693 |
env.Append(LIBPATH = [lib_path,]) |
env.Append(LIBS = [env['omp_libs']]) |
|
except KeyError: |
|
|
pass |
|
694 |
|
|
695 |
try: |
############ Add some custom builders ########################## |
|
umf_libs = env['umf_libs'] |
|
|
umf_libs+=umf_libs |
|
|
except KeyError: |
|
|
pass |
|
696 |
|
|
697 |
try: |
py_builder = Builder(action = scons_extensions.build_py, suffix = '.pyc', src_suffix = '.py', single_source=True) |
698 |
includes = env['ufc_path'] |
env.Append(BUILDERS = {'PyCompile' : py_builder}); |
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
699 |
|
|
700 |
try: |
runUnitTest_builder = Builder(action = scons_extensions.runUnitTest, suffix = '.passed', src_suffix=env['PROGSUFFIX'], single_source=True) |
701 |
includes = env['amd_path'] |
env.Append(BUILDERS = {'RunUnitTest' : runUnitTest_builder}); |
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
702 |
|
|
703 |
try: |
runPyUnitTest_builder = Builder(action = scons_extensions.runPyUnitTest, suffix = '.passed', src_suffic='.py', single_source=True) |
704 |
lib_path = env['amd_lib_path'] |
env.Append(BUILDERS = {'RunPyUnitTest' : runPyUnitTest_builder}); |
|
env.Append(LIBPATH = [lib_path,]) |
|
|
except KeyError: |
|
|
pass |
|
705 |
|
|
706 |
try: |
epstopdfbuilder = Builder(action = scons_extensions.eps2pdf, suffix=".pdf", src_suffix=".eps", single_source=True) |
707 |
amd_libs = env['amd_libs'] |
env.Append(BUILDERS = {'EpsToPDF' : epstopdfbuilder}); |
|
umf_libs+=amd_libs |
|
|
except KeyError: |
|
|
pass |
|
708 |
|
|
709 |
# ============= set TRILINOS (but only with MPI) ===================================== |
############ MPI (optional) #################################### |
710 |
if useMPI: |
if not env['usempi']: env['mpi_flavour']='none' |
|
try: |
|
|
includes = env['trilinos_path'] |
|
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
711 |
|
|
712 |
try: |
# Create a modified environment for MPI programs (identical to env if usempi=no) |
713 |
lib_path = env['trilinos_lib_path'] |
env_mpi = clone_env(env) |
|
env.Append(LIBPATH = [lib_path,]) |
|
|
except KeyError: |
|
|
pass |
|
714 |
|
|
715 |
try: |
# Start a new configure environment that reflects what we've already found |
716 |
trilinos_libs = env['trilinos_libs'] |
conf = Configure(clone_env(env_mpi)) |
717 |
except KeyError: |
|
718 |
trilinos_libs = [] |
if env_mpi['usempi']: |
719 |
|
VALID_MPIs=[ "MPT", "MPICH", "MPICH2", "OPENMPI", "INTELMPI" ] |
720 |
|
if not env_mpi['mpi_flavour'] in VALID_MPIs: |
721 |
|
raise ValueError,"MPI is enabled but mpi_flavour = %s is not a valid key from %s."%( env_mpi['mpi_flavour'],VALID_MPIs) |
722 |
|
conf.env.AppendUnique(CPPPATH = [env_mpi['mpi_path']]) |
723 |
|
conf.env.AppendUnique(LIBPATH = [env_mpi['mpi_lib_path']]) |
724 |
|
conf.env.AppendUnique(LIBS = [env_mpi['mpi_libs']]) |
725 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['mpi_lib_path']) # The wrapper script needs to find these libs |
726 |
|
#ensure that our path entries remain at the front |
727 |
|
conf.env.PrependENVPath('PYTHONPATH', prefix) |
728 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['libinstall']) |
729 |
|
|
730 |
|
if env_mpi['usempi'] and not conf.CheckCHeader('mpi.h'): env_mpi['usempi'] = 0 |
731 |
|
# if env_mpi['usempi'] and not conf.CheckFunc('MPI_Init'): env_mpi['usempi'] = 0 |
732 |
|
|
733 |
|
# Add MPI to environment env_mpi if it was found |
734 |
|
if env_mpi['usempi']: |
735 |
|
env_mpi = conf.Finish() |
736 |
|
env_mpi.Append(CPPDEFINES = ['PASO_MPI', 'MPI_NO_CPPBIND', env_mpi['MPICH_IGNORE_CXX_SEEK']]) |
737 |
|
# NetCDF 4.1 defines MPI_Comm et al. if MPI_INCLUDED is not defined! |
738 |
|
# On the other hand MPT and OpenMPI don't define the latter so we have to |
739 |
|
# do that here |
740 |
|
if env['usenetcdf'] and env_mpi['mpi_flavour'] in ["MPT","OPENMPI"]: |
741 |
|
env_mpi.Append(CPPDEFINES = ['MPI_INCLUDED']) |
742 |
else: |
else: |
743 |
trilinos_libs = [] |
conf.Finish() |
744 |
|
|
745 |
# ============= set blas ===================================== |
env['usempi'] = env_mpi['usempi'] |
|
try: |
|
|
includes = env['blas_path'] |
|
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
746 |
|
|
747 |
try: |
############ ParMETIS (optional) ############################### |
|
lib_path = env['blas_lib_path'] |
|
|
env.Append(LIBPATH = [lib_path,]) |
|
|
except KeyError: |
|
|
pass |
|
748 |
|
|
749 |
try: |
# Start a new configure environment that reflects what we've already found |
750 |
blas_libs = env['blas_libs'] |
conf = Configure(clone_env(env_mpi)) |
|
except KeyError: |
|
|
blas_libs = [ ] |
|
751 |
|
|
752 |
# ========== netcdf ==================================== |
if not env_mpi['usempi']: env_mpi['useparmetis'] = 0 |
|
try: |
|
|
useNetCDF = env['useNetCDF'] |
|
|
except KeyError: |
|
|
useNetCDF = 'yes' |
|
|
pass |
|
|
|
|
|
if useNetCDF == 'yes': |
|
|
try: |
|
|
netCDF_libs = env['netCDF_libs'] |
|
|
except KeyError: |
|
|
pass |
|
753 |
|
|
754 |
env.Append(LIBS = netCDF_libs) |
if env_mpi['useparmetis']: |
755 |
env.Append(CPPDEFINES = [ 'USE_NETCDF' ]) |
conf.env.AppendUnique(CPPPATH = [env_mpi['parmetis_path']]) |
756 |
try: |
conf.env.AppendUnique(LIBPATH = [env_mpi['parmetis_lib_path']]) |
757 |
includes = env['netCDF_path'] |
conf.env.AppendUnique(LIBS = [env_mpi['parmetis_libs']]) |
758 |
env.Append(CPPPATH = [includes,]) |
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['parmetis_lib_path']) # The wrapper script needs to find these libs |
759 |
except KeyError: |
#ensure that our path entries remain at the front |
760 |
pass |
conf.env.PrependENVPath('PYTHONPATH', prefix) |
761 |
|
conf.env.PrependENVPath(LD_LIBRARY_PATH_KEY, env['libinstall']) |
762 |
|
|
763 |
try: |
if env_mpi['useparmetis'] and not conf.CheckCHeader('parmetis.h'): env_mpi['useparmetis'] = 0 |
764 |
lib_path = env['netCDF_lib_path'] |
if env_mpi['useparmetis'] and not conf.CheckFunc('ParMETIS_V3_PartGeomKway'): env_mpi['useparmetis'] = 0 |
765 |
env.Append(LIBPATH = [ lib_path, ]) |
|
766 |
env['ENV']['LD_LIBRARY_PATH']+=":"+lib_path |
# Add ParMETIS to environment env_mpi if it was found |
767 |
if IS_WINDOWS_PLATFORM : |
if env_mpi['useparmetis']: |
768 |
env.PrependENVPath('PATH', lib_path) |
env_mpi = conf.Finish() |
769 |
except KeyError: |
env_mpi.Append(CPPDEFINES = ['USE_PARMETIS']) |
|
pass |
|
770 |
else: |
else: |
771 |
print "Warning: Installation is not configured with netCDF. Some I/O function may not be available." |
conf.Finish() |
|
netCDF_libs=[ ] |
|
772 |
|
|
773 |
# ====================== boost ====================================== |
env['useparmetis'] = env_mpi['useparmetis'] |
|
try: |
|
|
includes = env['boost_path'] |
|
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
lib_path = env['boost_lib_path'] |
|
|
env.Append(LIBPATH = [lib_path,]) |
|
|
env['ENV']['LD_LIBRARY_PATH']+=":"+lib_path |
|
|
if IS_WINDOWS_PLATFORM : |
|
|
env.PrependENVPath('PATH', lib_path) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
boost_lib = env['boost_lib'] |
|
|
except KeyError: |
|
|
boost_lib = None |
|
|
# ====================== python ====================================== |
|
|
try: |
|
|
includes = env['python_path'] |
|
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
lib_path = env['python_lib_path'] |
|
|
env['ENV']['LD_LIBRARY_PATH']+=":"+lib_path |
|
|
env.Append(LIBPATH = [lib_path,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
python_lib = env['python_lib'] |
|
|
except KeyError: |
|
|
python_lib = None |
|
|
# =============== documentation ======================================= |
|
|
try: |
|
|
doxygen_path = env['doxygen_path'] |
|
|
except KeyError: |
|
|
doxygen_path = None |
|
|
try: |
|
|
epydoc_path = env['epydoc_path'] |
|
|
except KeyError: |
|
|
epydoc_path = None |
|
|
# =============== ParMETIS ======================================= |
|
|
try: |
|
|
parmetis_path = env['parmetis_path'] |
|
|
parmetis_lib_path = env['parmetis_lib_path'] |
|
|
parmetis_lib = env['parmetis_lib'] |
|
|
except KeyError: |
|
|
parmetis_path = '' |
|
|
parmetis_lib_path = '' |
|
|
parmetis_lib = '' |
|
|
|
|
|
if useMPI and os.path.isdir(parmetis_lib_path): |
|
|
env.Append(CPPDEFINES = [ 'PARMETIS' ]) |
|
|
env.Append(CXXDEFINES = [ 'PARMETIS' ]) |
|
|
env.Append(CPPPATH = [parmetis_path]) |
|
|
env.Append(LIBPATH = [parmetis_lib_path]) |
|
|
env.Append(LIBS = parmetis_lib) |
|
|
# =============== PAPI ======================================= |
|
|
try: |
|
|
includes = env['papi_path'] |
|
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
lib_path = env['papi_lib_path'] |
|
|
env.Append(LIBPATH = [lib_path,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
papi_libs = env['papi_libs'] |
|
|
except KeyError: |
|
|
papi_libs = None |
|
|
# ============= set mpi ===================================== |
|
|
if useMPI: |
|
|
env.Append(CPPDEFINES=['PASO_MPI',]) |
|
|
try: |
|
|
includes = env['mpi_path'] |
|
|
env.Append(CPPPATH = [includes,]) |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
lib_path = env['mpi_lib_path'] |
|
|
env.Append(LIBPATH = [lib_path,]) |
|
|
env['ENV']['LD_LIBRARY_PATH']+=":"+lib_path |
|
|
except KeyError: |
|
|
pass |
|
|
try: |
|
|
mpi_libs = env['mpi_libs'] |
|
|
except KeyError: |
|
|
mpi_libs = [] |
|
774 |
|
|
775 |
try: |
############ Summarize our environment ######################### |
|
mpi_run = env['mpi_run'] |
|
|
except KeyError: |
|
|
mpi_run = '' |
|
776 |
|
|
777 |
try: |
print "" |
778 |
mpich_ignore_cxx_seek=env['MPICH_IGNORE_CXX_SEEK'] |
print "Summary of configuration (see ./config.log for information)" |
779 |
env.Append(CPPDEFINES = [ mpich_ignore_cxx_seek ] ) |
print " Using python libraries" |
780 |
except KeyError: |
print " Using numpy" |
781 |
pass |
print " Using boost" |
782 |
else: |
if env['usenetcdf']: print " Using NetCDF" |
783 |
mpi_libs=[] |
else: print " Not using NetCDF" |
784 |
mpi_run = mpi_run_default |
if env['usevtk']: print " Using VTK" |
785 |
# =========== zip files =========================================== |
else: print " Not using VTK" |
786 |
try: |
if env['usevisit']: print " Using VisIt" |
787 |
includes = env['papi_path'] |
else: print " Not using VisIt" |
788 |
env.Append(CPPPATH = [includes,]) |
if env['usemkl']: print " Using MKL" |
789 |
except KeyError: |
else: print " Not using MKL" |
790 |
pass |
if env['useumfpack']: print " Using UMFPACK" |
791 |
try: |
else: print " Not using UMFPACK" |
792 |
lib_path = env['papi_lib_path'] |
if env['usesilo']: print " Using Silo" |
793 |
env.Append(LIBPATH = [lib_path,]) |
else: print " Not using Silo" |
794 |
except KeyError: |
if env['useopenmp']: print " Using OpenMP" |
795 |
pass |
else: print " Not using OpenMP" |
796 |
try: |
if env['usempi']: print " Using MPI (flavour = %s)"%env['mpi_flavour'] |
797 |
papi_libs = env['papi_libs'] |
else: print " Not using MPI" |
798 |
except KeyError: |
if env['useparmetis']: print " Using ParMETIS" |
799 |
papi_libs = None |
else: print " Not using ParMETIS (requires MPI)" |
800 |
try: |
if env['usepapi']: print " Using PAPI" |
801 |
papi_instrument_solver = env['papi_instrument_solver'] |
else: print " Not using PAPI" |
802 |
except KeyError: |
if env['uselapack']: print " Using Lapack" |
803 |
papi_instrument_solver = None |
else: print " Not using Lapack" |
804 |
|
if env['usedebug']: print " Compiling for debug" |
805 |
|
else: print " Not compiling for debug" |
806 |
|
print " Installing in", prefix |
807 |
|
if ((fatalwarning != "") and (env['usewarnings'])): print " Treating warnings as errors" |
808 |
|
else: print " Not treating warnings as errors" |
809 |
|
print "" |
810 |
|
|
811 |
|
############ Delete option-dependent files ##################### |
812 |
|
|
813 |
|
Execute(Delete(os.path.join(env['libinstall'],"Compiled.with.debug"))) |
814 |
|
Execute(Delete(os.path.join(env['libinstall'],"Compiled.with.mpi"))) |
815 |
|
Execute(Delete(os.path.join(env['libinstall'],"Compiled.with.openmp"))) |
816 |
|
Execute(Delete(os.path.join(env['libinstall'],"buildvars"))) |
817 |
|
if not env['usempi']: Execute(Delete(os.path.join(env['libinstall'],"pythonMPI"))) |
818 |
|
|
819 |
|
|
820 |
|
############ Build the subdirectories ########################## |
821 |
|
|
822 |
|
if env['usepedantic']: env_mpi.Append(CCFLAGS = pedantic) |
823 |
|
|
824 |
|
|
825 |
|
from grouptest import * |
826 |
|
|
827 |
|
TestGroups=[] |
828 |
|
|
829 |
|
dodgy_env=clone_env(env_mpi) # Environment without pedantic options |
830 |
|
|
831 |
|
############ Now we switch on Warnings as errors ############### |
832 |
|
|
833 |
|
#this needs to be done after configuration because the scons test files have warnings in them |
834 |
|
|
835 |
|
if ((fatalwarning != "") and (env['usewarnings'])): |
836 |
|
env.Append(CCFLAGS = fatalwarning) |
837 |
|
env_mpi.Append(CCFLAGS = fatalwarning) |
838 |
|
|
839 |
|
|
840 |
|
Export( |
841 |
|
["env", |
842 |
|
"env_mpi", |
843 |
|
"clone_env", |
844 |
|
"dodgy_env", |
845 |
|
"IS_WINDOWS_PLATFORM", |
846 |
|
"TestGroups", |
847 |
|
"CallSConscript", |
848 |
|
"cantusevariantdir" |
849 |
|
] |
850 |
|
) |
851 |
|
|
852 |
|
CallSConscript(env, dirs = ['tools/CppUnitTest/src'], variant_dir='build/$PLATFORM/tools/CppUnitTest', duplicate=0) |
853 |
|
CallSConscript(env, dirs = ['tools/escriptconvert'], variant_dir='build/$PLATFORM/tools/escriptconvert', duplicate=0) |
854 |
|
CallSConscript(env, dirs = ['paso/src'], variant_dir='build/$PLATFORM/paso', duplicate=0) |
855 |
|
CallSConscript(env, dirs = ['weipa/src'], variant_dir='build/$PLATFORM/weipa', duplicate=0) |
856 |
|
CallSConscript(env, dirs = ['escript/src'], variant_dir='build/$PLATFORM/escript', duplicate=0) |
857 |
|
CallSConscript(env, dirs = ['esysUtils/src'], variant_dir='build/$PLATFORM/esysUtils', duplicate=0) |
858 |
|
CallSConscript(env, dirs = ['finley/src'], variant_dir='build/$PLATFORM/finley', duplicate=0) |
859 |
|
CallSConscript(env, dirs = ['modellib/py_src'], variant_dir='build/$PLATFORM/modellib', duplicate=0) |
860 |
|
CallSConscript(env, dirs = ['doc'], variant_dir='build/$PLATFORM/doc', duplicate=0) |
861 |
|
CallSConscript(env, dirs = ['pyvisi/py_src'], variant_dir='build/$PLATFORM/pyvisi', duplicate=0) |
862 |
|
CallSConscript(env, dirs = ['pycad/py_src'], variant_dir='build/$PLATFORM/pycad', duplicate=0) |
863 |
|
CallSConscript(env, dirs = ['pythonMPI/src'], variant_dir='build/$PLATFORM/pythonMPI', duplicate=0) |
864 |
|
CallSConscript(env, dirs = ['scripts'], variant_dir='build/$PLATFORM/scripts', duplicate=0) |
865 |
|
CallSConscript(env, dirs = ['paso/profiling'], variant_dir='build/$PLATFORM/paso/profiling', duplicate=0) |
866 |
|
|
867 |
|
|
868 |
|
############ Remember what optimizations we used ############### |
869 |
|
|
870 |
|
remember_list = [] |
871 |
|
|
872 |
|
if env['usedebug']: |
873 |
|
remember_list += env.Command(os.path.join(env['libinstall'],"Compiled.with.debug"), None, Touch('$TARGET')) |
874 |
|
|
875 |
|
if env['usempi']: |
876 |
|
remember_list += env.Command(os.path.join(env['libinstall'],"Compiled.with.mpi"), None, Touch('$TARGET')) |
877 |
|
|
878 |
|
if env['useopenmp']: |
879 |
|
remember_list += env.Command(os.path.join(env['libinstall'],"Compiled.with.openmp"), None, Touch('$TARGET')) |
880 |
|
|
881 |
# ============= and some helpers ===================================== |
env.Alias('remember_options', remember_list) |
|
try: |
|
|
doxygen_path = env['doxygen_path'] |
|
|
except KeyError: |
|
|
doxygen_path = None |
|
|
try: |
|
|
epydoc_path = env['epydoc_path'] |
|
|
except KeyError: |
|
|
epydoc_path = None |
|
|
try: |
|
|
src_zipfile = env.File(env['src_zipfile']) |
|
|
except KeyError: |
|
|
src_zipfile = None |
|
|
try: |
|
|
test_zipfile = env.File(env['test_zipfile']) |
|
|
except KeyError: |
|
|
test_zipfile = None |
|
|
try: |
|
|
examples_zipfile = env.File(env['examples_zipfile']) |
|
|
except KeyError: |
|
|
examples_zipfile = None |
|
882 |
|
|
|
try: |
|
|
src_tarfile = env.File(env['src_tarfile']) |
|
|
except KeyError: |
|
|
src_tarfile = None |
|
|
try: |
|
|
test_tarfile = env.File(env['test_tarfile']) |
|
|
except KeyError: |
|
|
test_tarfile = None |
|
|
try: |
|
|
examples_tarfile = env.File(env['examples_tarfile']) |
|
|
except KeyError: |
|
|
examples_tarfile = None |
|
883 |
|
|
884 |
try: |
############### Record python interpreter version ############## |
|
guide_pdf = env.File(env['guide_pdf']) |
|
|
except KeyError: |
|
|
guide_pdf = None |
|
885 |
|
|
886 |
try: |
if not IS_WINDOWS_PLATFORM: |
|
guide_html_index = env.File('index.htm',env['guide_html']) |
|
|
except KeyError: |
|
|
guide_html_index = None |
|
887 |
|
|
888 |
try: |
versionstring="Python "+str(sys.version_info[0])+"."+str(sys.version_info[1])+"."+str(sys.version_info[2]) |
889 |
api_epydoc = env.Dir(env['api_epydoc']) |
# if sys.version_info[4] >0 : versionstring+="rc%s"%sys.version_info[4] |
|
except KeyError: |
|
|
api_epydoc = None |
|
890 |
|
|
891 |
try: |
############## Populate the buildvars file ##################### |
|
api_doxygen = env.Dir(env['api_doxygen']) |
|
|
except KeyError: |
|
|
api_doxygen = None |
|
892 |
|
|
893 |
try: |
buildvars=open(os.path.join(env['libinstall'],'buildvars'),'w') |
894 |
svn_pipe = os.popen("svnversion -n .") |
buildvars.write('python='+str(sys.version_info[0])+"."+str(sys.version_info[1])+"."+str(sys.version_info[2])+'\n') |
895 |
global_revision = svn_pipe.readlines() |
|
896 |
svn_pipe.close() |
# Find the boost version by extracting it from version.hpp |
897 |
global_revision = re.sub(":.*", "", global_revision[0]) |
boosthpp=open(os.path.join(env['boost_path'],'boost','version.hpp')) |
898 |
global_revision = re.sub("[^0-9]", "", global_revision) |
boostversion='unknown' |
899 |
except: |
try: |
900 |
global_revision="-1" |
for line in boosthpp: |
901 |
print "Warning: unable to recover global revsion number." |
ver=re.match(r'#define BOOST_VERSION (\d+)',line) |
902 |
if global_revision == "": global_revision="0" |
if ver: |
903 |
print "Revision number is %s."%global_revision |
boostversion=ver.group(1) |
904 |
env.Append(CPPDEFINES = "SVN_VERSION="+global_revision) |
except StopIteration: |
905 |
|
pass |
906 |
# Python install - esys __init__.py |
buildvars.write("boost="+boostversion+"\n") |
907 |
init_target = env.Command(pyinstall+'/__init__.py', None, Touch('$TARGET')) |
buildvars.write("svn_revision="+str(global_revision)+"\n") |
908 |
|
out="usedebug=" |
909 |
# FIXME: exinstall and friends related to examples are not working. |
if env['usedebug']: |
910 |
build_target = env.Alias('build',[libinstall,incinstall,pyinstall,init_target]) |
out+="y" |
911 |
|
else: |
912 |
env.Default(build_target) |
out+="n" |
913 |
|
out+="\nusempi=" |
914 |
# Zipgets |
if env['usempi']: |
915 |
env.Alias('release_src',[ src_zipfile, src_tarfile ]) |
out+="y" |
916 |
env.Alias('release_tests',[ test_zipfile, test_tarfile]) |
else: |
917 |
env.Alias('release_examples',[ examples_zipfile, examples_tarfile]) |
out+="n" |
918 |
env.Alias('examples_zipfile',examples_zipfile) |
out+="\nuseopenmp=" |
919 |
env.Alias('examples_tarfile',examples_tarfile) |
if env['useopenmp']: |
920 |
env.Alias('api_epydoc',api_epydoc) |
out+="y" |
921 |
env.Alias('api_doxygen',api_doxygen) |
else: |
922 |
env.Alias('guide_html_index',guide_html_index) |
out+="n" |
923 |
env.Alias('guide_pdf', guide_pdf) |
buildvars.write(out+"\n") |
924 |
env.Alias('docs',[ 'release_examples', 'guide_pdf', api_epydoc, api_doxygen, guide_html_index]) |
buildvars.write("mpi_flavour="+env['mpi_flavour']+'\n') |
925 |
env.Alias('release', ['release_src', 'release_tests', 'docs']) |
out="lapack=" |
926 |
|
if env['uselapack']: |
927 |
env.Alias('build_tests',build_target) # target to build all C++ tests |
out+="y" |
928 |
env.Alias('build_py_tests',build_target) # target to build all python tests |
else: |
929 |
env.Alias('build_all_tests', [ 'build_tests', 'build_py_tests' ] ) # target to build all python tests |
out+="n" |
930 |
env.Alias('run_tests', 'build_tests') # target to run all C++ test |
out+="\nsilo=" |
931 |
env.Alias('py_tests', 'build_py_tests') # taget to run all released python tests |
if env['usesilo']: |
932 |
env.Alias('all_tests', ['run_tests', 'py_tests']) # target to run all C++ and released python tests |
out+="y" |
933 |
|
else: |
934 |
|
out+="n" |
935 |
# Allow sconscripts to see the env |
out+="\nusevisit=" |
936 |
Export(["IS_WINDOWS_PLATFORM", "env", "incinstall", "libinstall", "pyinstall", "dodebug", "mkl_libs", "scsl_libs", "umf_libs", "blas_libs", "netCDF_libs", "useNetCDF", "mpi_run", |
if env['usevisit']: |
937 |
"boost_lib", "python_lib", "doxygen_path", "epydoc_path", "papi_libs", |
out+="y" |
938 |
"sys_libs", "test_zipfile", "src_zipfile", "test_tarfile", "src_tarfile", "examples_tarfile", "examples_zipfile", "trilinos_libs", "mpi_libs", "papi_instrument_solver", |
else: |
939 |
"guide_pdf", "guide_html_index", "api_epydoc", "api_doxygen", "useMPI" ]) |
out+="n" |
940 |
|
buildvars.write(out+"\n") |
941 |
# End initialisation section |
buildvars.close() |
942 |
# Begin configuration section |
|
943 |
# adds this file and the scons option directore to the source tar |
|
944 |
release_srcfiles=[env.File('SConstruct'),env.Dir('lib'),env.Dir('include'),]+[ env.File(x) for x in glob.glob('scons/*.py') ] |
############ Targets to build and install libraries ############ |
945 |
release_testfiles=[env.File('README_TESTS'),] |
|
946 |
env.Zip(src_zipfile, release_srcfiles) |
target_init = env.Command(env['pyinstall']+'/__init__.py', None, Touch('$TARGET')) |
947 |
env.Zip(test_zipfile, release_testfiles) |
env.Alias('target_init', [target_init]) |
948 |
try: |
|
949 |
env.Tar(src_tarfile, release_srcfiles) |
# The headers have to be installed prior to build in order to satisfy #include <paso/Common.h> |
950 |
env.Tar(test_tarfile, release_testfiles) |
env.Alias('build_esysUtils', ['target_install_esysUtils_headers', 'target_esysUtils_a']) |
951 |
except AttributeError: |
env.Alias('install_esysUtils', ['build_esysUtils', 'target_install_esysUtils_a']) |
952 |
pass |
|
953 |
# Insert new components to be build here |
env.Alias('build_paso', ['target_install_paso_headers', 'target_paso_a']) |
954 |
# FIXME: might be nice to replace this verbosity with a list of targets and some |
env.Alias('install_paso', ['build_paso', 'target_install_paso_a']) |
955 |
# FIXME: nifty python to create the lengthy but very similar env.Sconscript lines |
|
956 |
# Third Party libraries |
env.Alias('build_weipa', ['target_install_weipa_headers', 'target_weipa_so', 'target_weipacpp_so']) |
957 |
env.SConscript(dirs = ['tools/CppUnitTest/src'], build_dir='build/$PLATFORM/tools/CppUnitTest', duplicate=0) |
env.Alias('install_weipa', ['build_weipa', 'target_install_weipa_so', 'target_install_weipacpp_so', 'target_install_weipa_py']) |
958 |
# C/C++ Libraries |
|
959 |
env.SConscript(dirs = ['paso/src'], build_dir='build/$PLATFORM/paso', duplicate=0) |
|
960 |
# bruce is removed for now as it doesn't really do anything |
env.Alias('build_escriptreader', ['target_install_weipa_headers', 'target_escriptreader_a']) |
961 |
# env.SConscript(dirs = ['bruce/src'], build_dir='build/$PLATFORM/bruce', duplicate=0) |
env.Alias('install_escriptreader', ['build_escriptreader', 'target_install_escriptreader_a']) |
962 |
env.SConscript(dirs = ['escript/src'], build_dir='build/$PLATFORM/escript', duplicate=0) |
|
963 |
env.SConscript(dirs = ['esysUtils/src'], build_dir='build/$PLATFORM/esysUtils', duplicate=0) |
env.Alias('build_escript', ['target_install_escript_headers', 'target_escript_so', 'target_escriptcpp_so']) |
964 |
env.SConscript(dirs = ['finley/src'], build_dir='build/$PLATFORM/finley', duplicate=0) |
env.Alias('install_escript', ['build_escript', 'target_install_escript_so', 'target_install_escriptcpp_so', 'target_install_escript_py']) |
965 |
env.SConscript(dirs = ['modellib/py_src'], build_dir='build/$PLATFORM/modellib', duplicate=0) |
|
966 |
env.SConscript(dirs = ['doc'], build_dir='build/$PLATFORM/doc', duplicate=0) |
env.Alias('build_finley', ['target_install_finley_headers', 'target_finley_so', 'target_finleycpp_so']) |
967 |
env.SConscript(dirs = ['pyvisi/py_src'], build_dir='build/$PLATFORM/pyvisi', duplicate=0) |
env.Alias('install_finley', ['build_finley', 'target_install_finley_so', 'target_install_finleycpp_so', 'target_install_finley_py']) |
968 |
env.SConscript(dirs = ['pycad/py_src'], build_dir='build/$PLATFORM/pycad', duplicate=0) |
|
969 |
env.SConscript(dirs = ['pythonMPI/src'], build_dir='build/$PLATFORM/pythonMPI', duplicate=0) |
# Now gather all the above into a couple easy targets: build_all and install_all |
970 |
#env.SConscript(dirs = ['../test'], build_dir='../test/build', duplicate=0) |
build_all_list = [] |
971 |
|
build_all_list += ['build_esysUtils'] |
972 |
|
build_all_list += ['build_paso'] |
973 |
|
build_all_list += ['build_weipa'] |
974 |
|
build_all_list += ['build_escript'] |
975 |
|
build_all_list += ['build_finley'] |
976 |
|
if env['usempi']: build_all_list += ['target_pythonMPI_exe'] |
977 |
|
#if not IS_WINDOWS_PLATFORM: build_all_list += ['target_escript_wrapper'] |
978 |
|
build_all_list += ['target_escriptconvert'] |
979 |
|
env.Alias('build_all', build_all_list) |
980 |
|
|
981 |
|
install_all_list = [] |
982 |
|
install_all_list += ['target_init'] |
983 |
|
install_all_list += ['install_esysUtils'] |
984 |
|
install_all_list += ['install_paso'] |
985 |
|
install_all_list += ['install_weipa'] |
986 |
|
install_all_list += ['install_escript'] |
987 |
|
install_all_list += ['install_finley'] |
988 |
|
install_all_list += ['target_install_pyvisi_py'] |
989 |
|
install_all_list += ['target_install_modellib_py'] |
990 |
|
install_all_list += ['target_install_pycad_py'] |
991 |
|
if env['usempi']: install_all_list += ['target_install_pythonMPI_exe'] |
992 |
|
#if not IS_WINDOWS_PLATFORM: install_all_list += ['target_install_escript_wrapper'] |
993 |
|
if env['usesilo']: install_all_list += ['target_install_escriptconvert'] |
994 |
|
install_all_list += ['remember_options'] |
995 |
|
env.Alias('install_all', install_all_list) |
996 |
|
|
997 |
|
# Default target is install |
998 |
|
env.Default('install_all') |
999 |
|
|
1000 |
|
############ Targets to build and run the test suite ########### |
1001 |
|
|
1002 |
|
env.Alias('build_cppunittest', ['target_install_cppunittest_headers', 'target_cppunittest_a']) |
1003 |
|
env.Alias('install_cppunittest', ['build_cppunittest', 'target_install_cppunittest_a']) |
1004 |
|
env.Alias('run_tests', ['install_all', 'target_install_cppunittest_a']) |
1005 |
|
env.Alias('all_tests', ['install_all', 'target_install_cppunittest_a', 'run_tests', 'py_tests']) |
1006 |
|
env.Alias('build_full',['install_all','build_tests','build_py_tests']) |
1007 |
|
|
1008 |
|
|
1009 |
|
############ Targets to build the documentation ################ |
1010 |
|
|
1011 |
|
env.Alias('api_epydoc','install_all') |
1012 |
|
|
1013 |
|
env.Alias('docs', ['examples_tarfile', 'examples_zipfile', 'api_epydoc', 'api_doxygen', 'guide_pdf', 'guide_html','install_pdf', 'cookbook_pdf']) |
1014 |
|
|
1015 |
|
build_platform=os.name |
1016 |
|
|
1017 |
|
if not IS_WINDOWS_PLATFORM: |
1018 |
|
try: |
1019 |
|
utest=open("utest.sh","w") |
1020 |
|
#Sometimes Mac python says it is posix |
1021 |
|
if (build_platform=='posix') and platform.system()=="Darwin": |
1022 |
|
build_platform='darwin' |
1023 |
|
utest.write(GroupTest.makeHeader(build_platform)) |
1024 |
|
for tests in TestGroups: |
1025 |
|
utest.write(tests.makeString()) |
1026 |
|
utest.close() |
1027 |
|
os.chmod("utest.sh",stat.S_IRWXU|stat.S_IRGRP|stat.S_IXGRP|stat.S_IROTH|stat.S_IXOTH) |
1028 |
|
print "utest.sh written" |
1029 |
|
except IOError: |
1030 |
|
print "Error attempting to write unittests file." |
1031 |
|
sys.exit(1) |
1032 |
|
|
1033 |
|
#Make sure that the escript wrapper is in place |
1034 |
|
if not os.path.isfile(os.path.join(env['bininstall'],'escript')): |
1035 |
|
print "Copying escript wrapper" |
1036 |
|
shutil.copy("bin/escript",os.path.join(env['bininstall'],'escript')) |
1037 |
|
|
1038 |
|
############ Targets to build PasoTests suite ################ |
1039 |
|
|
1040 |
syslib_install_target = env.installDirectory(sys_libinstall,libinstall) |
env.Alias('build_PasoTests','build/'+build_platform+'/paso/profiling/PasoTests') |
|
syspy_install_target = env.installDirectory(sys_pyinstall,pyinstall,recursive=True) |
|
1041 |
|
|
1042 |
install_target = env.Alias("install", env.Flatten([syslib_install_target, syspy_install_target]) ) |
env.Alias('release_prep', ['docs', 'install_all']) |