diff --git a/etc/spack/sdsc/expanse/0.17.3/cpu/b/specs/intel@19.1.3.304/intel-tbb@2020.3.o33452916.exp-15-56 b/etc/spack/sdsc/expanse/0.17.3/cpu/b/specs/intel@19.1.3.304/intel-tbb@2020.3.o33452916.exp-15-56 new file mode 100644 index 00000000000000..439f03baf94c93 --- /dev/null +++ b/etc/spack/sdsc/expanse/0.17.3/cpu/b/specs/intel@19.1.3.304/intel-tbb@2020.3.o33452916.exp-15-56 @@ -0,0 +1,668 @@ +1724800658 33452916 3687437c65bd7cbbb304e656d9c2c57f /cm/shared/apps/spack/0.17.3/cpu/b/etc/spack/sdsc/expanse/0.17.3/cpu/b/specs/intel@19.1.3.304/intel-tbb@2020.3.sh + +#!/usr/bin/env bash + +#SBATCH --job-name=intel-tbb@2020.3 +#SBATCH --account=use300 +#SBATCH --reservation=root_73 +#SBATCH --partition=ind-shared +#SBATCH --nodes=1 +#SBATCH --ntasks-per-node=1 +#SBATCH --cpus-per-task=16 +#SBATCH --mem=32G +#SBATCH --time=00:30:00 +#SBATCH --output=%x.o%j.%N + +declare -xr LOCAL_TIME="$(date +'%Y%m%dT%H%M%S%z')" +declare -xir UNIX_TIME="$(date +'%s')" + +declare -xr LOCAL_SCRATCH_DIR="/scratch/${USER}/job_${SLURM_JOB_ID}" +declare -xr TMPDIR="${LOCAL_SCRATCH_DIR}" + +declare -xr SYSTEM_NAME='expanse' + +declare -xr SPACK_VERSION='0.17.3' +declare -xr SPACK_INSTANCE_NAME='cpu' +declare -xr SPACK_INSTANCE_VERSION='b' +declare -xr SPACK_INSTANCE_DIR="/cm/shared/apps/spack/${SPACK_VERSION}/${SPACK_INSTANCE_NAME}/${SPACK_INSTANCE_VERSION}" + +declare -xr SLURM_JOB_SCRIPT="$(scontrol show job ${SLURM_JOB_ID} | awk -F= '/Command=/{print $2}')" +declare -xr SLURM_JOB_MD5SUM="$(md5sum ${SLURM_JOB_SCRIPT})" + +declare -xr SCHEDULER_MODULE='slurm' + +echo "${UNIX_TIME} ${SLURM_JOB_ID} ${SLURM_JOB_MD5SUM} ${SLURM_JOB_DEPENDENCY}" +echo "" + +cat "${SLURM_JOB_SCRIPT}" + +module purge +module load "${SCHEDULER_MODULE}" +module list +. "${SPACK_INSTANCE_DIR}/share/spack/setup-env.sh" + +declare -xr INTEL_LICENSE_FILE='40000@elprado.sdsc.edu:40200@elprado.sdsc.edu' +declare -xr SPACK_PACKAGE='intel-tbb@2020.3' +declare -xr SPACK_COMPILER='intel@19.1.3.304' +declare -xr SPACK_VARIANTS='~ipo +shared +tm' +declare -xr SPACK_DEPENDENCIES='' +declare -xr SPACK_SPEC="${SPACK_PACKAGE} % ${SPACK_COMPILER} ${SPACK_VARIANTS}" + +printenv + +spack config get compilers +spack config get config +spack config get mirrors +spack config get modules +spack config get packages +spack config get repos +spack config get upstreams + +time -p spack spec --long --namespaces --types "${SPACK_SPEC}" +if [[ "${?}" -ne 0 ]]; then + echo 'ERROR: spack concretization failed.' + exit 1 +fi + +time -p spack install --jobs "${SLURM_CPUS_PER_TASK}" --fail-fast --yes-to-all "${SPACK_SPEC}" +if [[ "${?}" -ne 0 ]]; then + echo 'ERROR: spack install failed.' + exit 1 +fi + +#spack module lmod refresh --delete-tree -y + +#sbatch --dependency="afterok:${SLURM_JOB_ID}" 'gsl@2.7.sh' + +Currently Loaded Modules: + 1) slurm/expanse/23.02.7 + + + +SLURM_JOB_GROUP=spack +LD_LIBRARY_PATH=/cm/shared/apps/slurm/current/lib64/slurm:/cm/shared/apps/slurm/current/lib64 +LS_COLORS=rs=0:di=38;5;33:ln=38;5;51:mh=00:pi=40;38;5;11:so=38;5;13:do=38;5;5:bd=48;5;232;38;5;11:cd=48;5;232;38;5;3:or=48;5;232;38;5;9:mi=01;05;37;41:su=48;5;196;38;5;15:sg=48;5;11;38;5;16:ca=48;5;196;38;5;226:tw=48;5;10;38;5;16:ow=48;5;10;38;5;21:st=48;5;21;38;5;15:ex=38;5;40:*.tar=38;5;9:*.tgz=38;5;9:*.arc=38;5;9:*.arj=38;5;9:*.taz=38;5;9:*.lha=38;5;9:*.lz4=38;5;9:*.lzh=38;5;9:*.lzma=38;5;9:*.tlz=38;5;9:*.txz=38;5;9:*.tzo=38;5;9:*.t7z=38;5;9:*.zip=38;5;9:*.z=38;5;9:*.dz=38;5;9:*.gz=38;5;9:*.lrz=38;5;9:*.lz=38;5;9:*.lzo=38;5;9:*.xz=38;5;9:*.zst=38;5;9:*.tzst=38;5;9:*.bz2=38;5;9:*.bz=38;5;9:*.tbz=38;5;9:*.tbz2=38;5;9:*.tz=38;5;9:*.deb=38;5;9:*.rpm=38;5;9:*.jar=38;5;9:*.war=38;5;9:*.ear=38;5;9:*.sar=38;5;9:*.rar=38;5;9:*.alz=38;5;9:*.ace=38;5;9:*.zoo=38;5;9:*.cpio=38;5;9:*.7z=38;5;9:*.rz=38;5;9:*.cab=38;5;9:*.wim=38;5;9:*.swm=38;5;9:*.dwm=38;5;9:*.esd=38;5;9:*.jpg=38;5;13:*.jpeg=38;5;13:*.mjpg=38;5;13:*.mjpeg=38;5;13:*.gif=38;5;13:*.bmp=38;5;13:*.pbm=38;5;13:*.pgm=38;5;13:*.ppm=38;5;13:*.tga=38;5;13:*.xbm=38;5;13:*.xpm=38;5;13:*.tif=38;5;13:*.tiff=38;5;13:*.png=38;5;13:*.svg=38;5;13:*.svgz=38;5;13:*.mng=38;5;13:*.pcx=38;5;13:*.mov=38;5;13:*.mpg=38;5;13:*.mpeg=38;5;13:*.m2v=38;5;13:*.mkv=38;5;13:*.webm=38;5;13:*.ogm=38;5;13:*.mp4=38;5;13:*.m4v=38;5;13:*.mp4v=38;5;13:*.vob=38;5;13:*.qt=38;5;13:*.nuv=38;5;13:*.wmv=38;5;13:*.asf=38;5;13:*.rm=38;5;13:*.rmvb=38;5;13:*.flc=38;5;13:*.avi=38;5;13:*.fli=38;5;13:*.flv=38;5;13:*.gl=38;5;13:*.dl=38;5;13:*.xcf=38;5;13:*.xwd=38;5;13:*.yuv=38;5;13:*.cgm=38;5;13:*.emf=38;5;13:*.ogv=38;5;13:*.ogx=38;5;13:*.aac=38;5;45:*.au=38;5;45:*.flac=38;5;45:*.m4a=38;5;45:*.mid=38;5;45:*.midi=38;5;45:*.mka=38;5;45:*.mp3=38;5;45:*.mpc=38;5;45:*.ogg=38;5;45:*.ra=38;5;45:*.wav=38;5;45:*.oga=38;5;45:*.opus=38;5;45:*.spx=38;5;45:*.xspf=38;5;45: +SRUN_DEBUG=3 +SLURM_STEP_ID=0 +LOCAL_SCRATCH_DIR=/scratch/spack_cpu/job_33452916 +SLURM_NODEID=0 +SLURM_TASK_PID=3127422 +__LMOD_REF_COUNT_PATH=/cm/shared/apps/slurm/current/sbin:1;/cm/shared/apps/slurm/current/bin:1;/home/spack_cpu/.local/bin:2;/home/spack_cpu/bin:2;/usr/local/bin:1;/usr/bin:1;/usr/local/sbin:1;/usr/sbin:1 +_ModuleTable002_=L21vZHVsZWZpbGVzIiwiL3Vzci9zaGFyZS9Nb2R1bGVzL21vZHVsZWZpbGVzIiwiL2NtL3NoYXJlZC9tb2R1bGVmaWxlcyIsfSxbInN5c3RlbUJhc2VNUEFUSCJdPSIvY20vbG9jYWwvbW9kdWxlZmlsZXM6L2NtL3NoYXJlZC9hcHBzL2FjY2Vzcy9tb2R1bGVmaWxlczovY20vc2hhcmVkL21vZHVsZWZpbGVzOi9ldGMvbW9kdWxlZmlsZXM6L3Vzci9zaGFyZS9tb2R1bGVmaWxlczovdXNyL3NoYXJlL01vZHVsZXMvbW9kdWxlZmlsZXMiLH0= +SSH_CONNECTION=198.202.100.13 33740 198.202.100.13 22 +SPACK_PYTHON=/usr/bin/python3 +SLURM_PRIO_PROCESS=0 +SLURM_CPU_BIND_VERBOSE=quiet +LANG=en_US.UTF-8 +SLURM_SUBMIT_DIR=/cm/shared/apps/spack/0.17.3/cpu/b/etc/spack/sdsc/expanse/0.17.3/cpu/b/specs/intel@19.1.3.304 +HISTCONTROL=ignoredups +HOSTNAME=exp-15-56 +LMOD_SYSTEM_DEFAULT_MODULES=DefaultModules +OLDPWD=/cm/shared/apps/spack/0.17.3/cpu/b/etc/spack/sdsc/expanse/0.17.3/cpu/b/specs +__LMOD_REF_COUNT__LMFILES_=/cm/local/modulefiles/slurm/expanse/23.02.7:1 +SLURM_CPUS_PER_TASK=16 +SLURM_STEPID=0 +SLURM_SRUN_COMM_HOST=10.21.0.19 +SPACK_DEPENDENCIES= +SLURM_DISTRIBUTION=cyclic +ENVIRONMENT=BATCH +SPACK_VERSION=0.17.3 +SLURM_PROCID=0 +SPACK_INSTANCE_VERSION=b +SLURM_JOB_GID=11491 +SPACK_VARIANTS=~ipo +shared +tm +SPACK_INSTANCE_DIR=/cm/shared/apps/spack/0.17.3/cpu/b +__LMOD_REF_COUNT_LD_LIBRARY_PATH=/cm/shared/apps/slurm/current/lib64/slurm:1;/cm/shared/apps/slurm/current/lib64:1 +SLURM_CPU_BIND=quiet,mask_cpu:0x0000000000000000000000000000FFFF +SLURM_JOB_SCRIPT=/cm/shared/apps/spack/0.17.3/cpu/b/etc/spack/sdsc/expanse/0.17.3/cpu/b/specs/intel@19.1.3.304/intel-tbb@2020.3.sh +SLURMD_NODENAME=exp-15-56 +SLURM_JOB_END_TIME=1724802458 +SLURM_TASKS_PER_NODE=1 +S_COLORS=auto +which_declare=declare -f +SPACK_COMPILER=intel@19.1.3.304 +SLURM_JOB_RESERVATION=root_73 +USER=spack_cpu +SLURM_NNODES=1 +SLURM_LAUNCH_NODE_IPADDR=10.21.0.19 +SLURM_PMIX_TMPDIR=/scratch/spack_cpu/job_33452916 +SLURM_TMPDIR=/scratch/spack_cpu/job_33452916 +SLURM_STEP_TASKS_PER_NODE=1 +SLURM_JOB_START_TIME=1724800658 +SLURM_NTASKS_PER_NODE=1 +__LMOD_REF_COUNT_MODULEPATH=/cm/local/modulefiles:1;/cm/shared/apps/access/modulefiles:1;/etc/modulefiles:1;/usr/share/modulefiles:1;/usr/share/Modules/modulefiles:1;/cm/shared/modulefiles:1 +__LMOD_REF_COUNT_LOADEDMODULES=slurm/expanse/23.02.7:1 +PWD=/cm/shared/apps/spack/0.17.3/cpu/b/etc/spack/sdsc/expanse/0.17.3/cpu/b/specs/intel@19.1.3.304 +ENABLE_LMOD=1 +SLURM_JOB_NODELIST=exp-15-56 +HOME=/home/spack_cpu +SLURM_CLUSTER_NAME=expanse +LMOD_COLORIZE=yes +LOCAL_TIME=20240827T161738-0700 +LMOD_SYSHOST=expanse +SLURM_NODELIST=exp-15-56 +SSH_CLIENT=198.202.100.13 33740 22 +LMOD_VERSION=8.2.4 +CPATH=/cm/shared/apps/slurm/current/include +SLURM_NTASKS=1 +LMOD_SETTARG_CMD=: +SLURM_JOB_CPUS_PER_NODE=16 +BASH_ENV=/usr/share/lmod/lmod/init/bash +SLURM_TOPOLOGY_ADDR=exp-15-56 +SLURMD_DEBUG=2 +SLURM_WORKING_CLUSTER=expanse:mgr1:6817:9984:109 +SLURM_JOB_MD5SUM=3687437c65bd7cbbb304e656d9c2c57f /cm/shared/apps/spack/0.17.3/cpu/b/etc/spack/sdsc/expanse/0.17.3/cpu/b/specs/intel@19.1.3.304/intel-tbb@2020.3.sh +SPACK_PACKAGE=intel-tbb@2020.3 +__LMOD_REF_COUNT_LIBRARY_PATH=/cm/shared/apps/slurm/current/lib64/slurm:1;/cm/shared/apps/slurm/current/lib64:1 +SLURM_STEP_NODELIST=exp-15-56 +SLURM_JOB_NAME=intel-tbb@2020.3 +SLURM_SRUN_COMM_PORT=33991 +TMPDIR=/scratch/spack_cpu/job_33452916 +LIBRARY_PATH=/cm/shared/apps/slurm/current/lib64/slurm:/cm/shared/apps/slurm/current/lib64 +LMOD_sys=Linux +SLURM_JOBID=33452916 +SYSTEM_NAME=expanse +_ModuleTable001_=X01vZHVsZVRhYmxlXz17WyJNVHZlcnNpb24iXT0zLFsiY19yZWJ1aWxkVGltZSJdPWZhbHNlLFsiY19zaG9ydFRpbWUiXT1mYWxzZSxkZXB0aFQ9e30sZmFtaWx5PXt9LG1UPXtzbHVybT17WyJmbiJdPSIvY20vbG9jYWwvbW9kdWxlZmlsZXMvc2x1cm0vZXhwYW5zZS8yMy4wMi43IixbImZ1bGxOYW1lIl09InNsdXJtL2V4cGFuc2UvMjMuMDIuNyIsWyJsb2FkT3JkZXIiXT0xLHByb3BUPXt9LFsic3RhY2tEZXB0aCJdPTAsWyJzdGF0dXMiXT0iYWN0aXZlIixbInVzZXJOYW1lIl09InNsdXJtIix9LH0sbXBhdGhBPXsiL2NtL2xvY2FsL21vZHVsZWZpbGVzIiwiL2NtL3NoYXJlZC9hcHBzL2FjY2Vzcy9tb2R1bGVmaWxlcyIsIi9ldGMvbW9kdWxlZmlsZXMiLCIvdXNyL3NoYXJl +SLURM_CONF=/cm/shared/apps/slurm/var/etc/expanse/slurm.conf +LOADEDMODULES=slurm/expanse/23.02.7 +__LMOD_REF_COUNT_MANPATH=/cm/shared/apps/slurm/current/man:3;/usr/share/lmod/lmod/share/man:1;/usr/local/share/man:2;/usr/share/man:2;/cm/local/apps/environment-modules/current/share/man:2 +_ModuleTable003_=TmFtZSJdPSJzaGFyZWQiLH0sc2x1cm09e1siZm4iXT0iL2NtL2xvY2FsL21vZHVsZWZpbGVzL3NsdXJtL2V4cGFuc2UvMjMuMDIuNyIsWyJmdWxsTmFtZSJdPSJzbHVybS9leHBhbnNlLzIzLjAyLjciLFsibG9hZE9yZGVyIl09Myxwcm9wVD17fSxbInN0YWNrRGVwdGgiXT0yLFsic3RhdHVzIl09ImFjdGl2ZSIsWyJ1c2VyTmFtZSJdPSJzbHVybSIsfSx9LG1wYXRoQT17Ii9jbS9zaGFyZWQvYXBwcy9zcGFjay8wLjE3LjMvY3B1L2Ivc2hhcmUvc3BhY2svbG1vZC9saW51eC1yb2NreTgteDg2XzY0L0NvcmUiLCIvY20vbG9jYWwvbW9kdWxlZmlsZXMiLCIvY20vc2hhcmVkL2FwcHMvYWNjZXNzL21vZHVsZWZpbGVzIiwiL2V0Yy9tb2R1bGVmaWxlcyIsIi91c3Ivc2hhcmUvbW9k +INTEL_LICENSE_FILE=40000@elprado.sdsc.edu:40200@elprado.sdsc.edu +SLURM_NODE_ALIASES=(null) +LMOD_ROOT=/usr/share/lmod +SLURM_JOB_QOS=ind-shared-normal +SLURM_TOPOLOGY_ADDR_PATTERN=node +SSH_TTY=/dev/pts/136 +MAIL=/var/spool/mail/spack_cpu +SLURM_CPUS_ON_NODE=16 +LMOD_arch=x86_64 +SLURM_JOB_NUM_NODES=1 +__Init_Default_Modules=1 +CMD_WLM_CLUSTER_NAME=expanse +SLURM_MEM_PER_NODE=32768 +SPACK_ROOT=/cm/shared/apps/spack/0.17.3/cpu/b +TERM=xterm-256color +SHELL=/bin/bash +_ModuleTable_Sz_=2 +SLURM_JOB_UID=527834 +__LMOD_REF_COUNT_CPATH=/cm/shared/apps/slurm/current/include:1 +SLURM_JOB_PARTITION=ind-shared +SLURM_SCRIPT_CONTEXT=prolog_task +SLURM_PTY_WIN_ROW=24 +SLURM_CPU_BIND_LIST=0x0000000000000000000000000000FFFF +SPACK_INSTANCE_NAME=cpu +SLURM_JOB_USER=spack_cpu +SLURM_PTY_WIN_COL=80 +SLURM_NPROCS=1 +SHLVL=3 +SLURM_SUBMIT_HOST=exp-15-56 +UNIX_TIME=1724800658 +SLURM_JOB_ACCOUNT=use300 +MANPATH=/cm/shared/apps/slurm/current/man:/usr/share/lmod/lmod/share/man:/usr/local/share/man:/usr/share/man:/cm/local/apps/environment-modules/current/share/man +LMOD_PREPEND_BLOCK=normal +SLURM_STEP_LAUNCHER_PORT=33991 +MODULEPATH=/cm/local/modulefiles:/cm/shared/apps/access/modulefiles:/etc/modulefiles:/usr/share/modulefiles:/usr/share/Modules/modulefiles:/cm/shared/modulefiles +SLURM_PTY_PORT=39435 +SLURM_GTIDS=0 +LOGNAME=spack_cpu +MODULEPATH_ROOT=/usr/share/modulefiles +LMOD_PACKAGE_PATH=/usr/share/lmod/etc/site/lmod +PATH=/cm/shared/apps/spack/0.17.3/cpu/b/bin:/cm/shared/apps/slurm/current/sbin:/cm/shared/apps/slurm/current/bin:/home/spack_cpu/.local/bin:/home/spack_cpu/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin +SLURM_JOB_ID=33452916 +_LMFILES_=/cm/local/modulefiles/slurm/expanse/23.02.7 +SLURM_CPU_BIND_TYPE=mask_cpu: +DEBUGINFOD_URLS=https://debuginfod.centos.org/ +SLURM_STEP_NUM_TASKS=1 +SCHEDULER_MODULE=slurm +MODULESHOME=/usr/share/lmod/lmod +LMOD_SETTARG_FULL_SUPPORT=no +HISTSIZE=-1 +LMOD_PKG=/usr/share/lmod/lmod +SLURM_STEP_NUM_NODES=1 +SPACK_SPEC=intel-tbb@2020.3 % intel@19.1.3.304 ~ipo +shared +tm +LMOD_CMD=/usr/share/lmod/lmod/libexec/lmod +SLURM_LOCALID=0 +LESSOPEN=||/usr/bin/lesspipe.sh %s +LMOD_FULL_SETTARG_SUPPORT=no +LMOD_DIR=/usr/share/lmod/lmod/libexec +BASH_FUNC_which%%=() { ( alias; + eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@ +} +BASH_FUNC_module%%=() { eval $($LMOD_CMD bash "$@") && eval $(${LMOD_SETTARG_CMD:-:} -s sh) +} +BASH_FUNC_spack%%=() { : this is a shell function from: /cm/shared/apps/spack/0.17.3/cpu/b/share/spack/setup-env.sh; + : the real spack script is here: /cm/shared/apps/spack/0.17.3/cpu/b/bin/spack; + _spack_shell_wrapper "$@"; + return $? +} +BASH_FUNC__spack_shell_wrapper%%=() { for var in LD_LIBRARY_PATH DYLD_LIBRARY_PATH DYLD_FALLBACK_LIBRARY_PATH; + do + eval "if [ -n \"\${${var}-}\" ]; then export SPACK_$var=\${${var}}; fi"; + done; + if [ -n "${ZSH_VERSION:-}" ]; then + emulate -L sh; + fi; + _sp_flags=""; + while [ ! -z ${1+x} ] && [ "${1#-}" != "${1}" ]; do + _sp_flags="$_sp_flags $1"; + shift; + done; + if [ -n "$_sp_flags" ] && [ "${_sp_flags#*h}" != "${_sp_flags}" ] || [ "${_sp_flags#*V}" != "${_sp_flags}" ]; then + command spack $_sp_flags "$@"; + return; + fi; + _sp_subcommand=""; + if [ ! -z ${1+x} ]; then + _sp_subcommand="$1"; + shift; + fi; + case $_sp_subcommand in + "cd") + _sp_arg=""; + if [ -n "$1" ]; then + _sp_arg="$1"; + shift; + fi; + if [ "$_sp_arg" = "-h" ] || [ "$_sp_arg" = "--help" ]; then + command spack cd -h; + else + LOC="$(spack location $_sp_arg "$@")"; + if [ -d "$LOC" ]; then + cd "$LOC"; + else + return 1; + fi; + fi; + return + ;; + "env") + _sp_arg=""; + if [ -n "$1" ]; then + _sp_arg="$1"; + shift; + fi; + if [ "$_sp_arg" = "-h" ] || [ "$_sp_arg" = "--help" ]; then + command spack env -h; + else + case $_sp_arg in + activate) + _a=" $@"; + if [ -z ${1+x} ] || [ "${_a#* --sh}" != "$_a" ] || [ "${_a#* --csh}" != "$_a" ] || [ "${_a#* -h}" != "$_a" ] || [ "${_a#* --help}" != "$_a" ]; then + command spack env activate "$@"; + else + stdout="$(command spack $_sp_flags env activate --sh "$@")" || return; + eval "$stdout"; + fi + ;; + deactivate) + _a=" $@"; + if [ "${_a#* --sh}" != "$_a" ] || [ "${_a#* --csh}" != "$_a" ]; then + command spack env deactivate "$@"; + else + if [ -n "$*" ]; then + command spack env deactivate -h; + else + stdout="$(command spack $_sp_flags env deactivate --sh)" || return; + eval "$stdout"; + fi; + fi + ;; + *) + command spack env $_sp_arg "$@" + ;; + esac; + fi; + return + ;; + "load" | "unload") + _a=" $@"; + if [ "${_a#* --sh}" != "$_a" ] || [ "${_a#* --csh}" != "$_a" ] || [ "${_a#* -h}" != "$_a" ] || [ "${_a#* --list}" != "$_a" ] || [ "${_a#* --help}" != "$_a" ]; then + command spack $_sp_flags $_sp_subcommand "$@"; + else + stdout="$(command spack $_sp_flags $_sp_subcommand --sh "$@")" || return; + eval "$stdout"; + fi + ;; + *) + command spack $_sp_flags $_sp_subcommand "$@" + ;; + esac +} +BASH_FUNC_ml%%=() { eval $($LMOD_DIR/ml_cmd "$@") +} +_=/usr/bin/printenv +compilers: +- compiler: + spec: gcc@8.5.0 + paths: + cc: /usr/bin/gcc + cxx: /usr/bin/g++ + f77: /usr/bin/gfortran + fc: /usr/bin/gfortran + flags: + cflags: -O2 -march=native + cxxflags: -O2 -march=native + fflags: -O2 -march=native + operating_system: rocky8 + target: x86_64 + modules: [] + environment: {} + extra_rpaths: [] +- compiler: + spec: gcc@10.2.0 + paths: + cc: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/gcc-10.2.0-npcyll4gxjhf4tejksmdzlsl3d3usqpd/bin/gcc + cxx: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/gcc-10.2.0-npcyll4gxjhf4tejksmdzlsl3d3usqpd/bin/g++ + f77: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/gcc-10.2.0-npcyll4gxjhf4tejksmdzlsl3d3usqpd/bin/gfortran + fc: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/gcc-10.2.0-npcyll4gxjhf4tejksmdzlsl3d3usqpd/bin/gfortran + flags: + cflags: -O2 -march=native + cxxflags: -O2 -march=native + fflags: -O2 -march=native + operating_system: rocky8 + target: x86_64 + modules: [] + environment: {} + extra_rpaths: [] +- compiler: + spec: aocc@3.2.0 + paths: + cc: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/aocc-3.2.0-io3s466wsnnichqc2o2rikbuloev5bmq/bin/clang + cxx: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/aocc-3.2.0-io3s466wsnnichqc2o2rikbuloev5bmq/bin/clang++ + f77: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/aocc-3.2.0-io3s466wsnnichqc2o2rikbuloev5bmq/bin/flang + fc: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/aocc-3.2.0-io3s466wsnnichqc2o2rikbuloev5bmq/bin/flang + flags: + cflags: -O2 -march=native + cxxflags: -O2 -march=native + fflags: -O2 -march=native + operating_system: rocky8 + target: x86_64 + modules: [] + environment: {} + extra_rpaths: [] +- compiler: + spec: intel@19.1.3.304 + paths: + cc: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/intel-19.1.3.304-6pv46sozip5o35ezrswc4p77gsiawktn/bin/icc + cxx: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/intel-19.1.3.304-6pv46sozip5o35ezrswc4p77gsiawktn/bin/icpc + f77: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/intel-19.1.3.304-6pv46sozip5o35ezrswc4p77gsiawktn/bin/ifort + fc: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/intel-19.1.3.304-6pv46sozip5o35ezrswc4p77gsiawktn/bin/ifort + flags: + cflags: -O2 -march=native + cxxflags: -O2 -march=native + fflags: -O2 -march=native + operating_system: rocky8 + target: x86_64 + modules: [] + environment: {} + extra_rpaths: [] +- compiler: + spec: nvhpc@21.9 + paths: + cc: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/nvhpc-21.9-xxpthf52cloih3j5abdsx65rlg4yk2kk/Linux_x86_64/21.9/compilers/bin/nvc + cxx: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/nvhpc-21.9-xxpthf52cloih3j5abdsx65rlg4yk2kk/Linux_x86_64/21.9/compilers/bin/nvc++ + f77: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/nvhpc-21.9-xxpthf52cloih3j5abdsx65rlg4yk2kk/Linux_x86_64/21.9/compilers/bin/nvfortran + fc: /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen/gcc-8.5.0/nvhpc-21.9-xxpthf52cloih3j5abdsx65rlg4yk2kk/Linux_x86_64/21.9/compilers/bin/nvfortran + flags: + cflags: -fast + cxxflags: -fast + fflags: -fast + operating_system: rocky8 + target: x86_64 + modules: [] + environment: {} + extra_rpaths: [] +config: + verify_ssl: false + install_tree: + root: $spack/opt/spack + projections: + all: ${ARCHITECTURE}/${COMPILERNAME}-${COMPILERVER}/${PACKAGE}-${VERSION}-${HASH} + template_dirs: + - $spack/share/spack/templates + + # Temporary locations Spack can try to use for builds. + # + # Recommended options are given below. + # + # Builds can be faster in temporary directories on some (e.g., HPC) systems. + # Specifying `$tempdir` will ensure use of the default temporary directory + # (i.e., ``$TMP` or ``$TMPDIR``). + # + # Another option that prevents conflicts and potential permission issues is + # to specify `$user_cache_path/stage`, which ensures each user builds in their + # home directory. + # + # A more traditional path uses the value of `$spack/var/spack/stage`, which + # builds directly inside Spack's instance without staging them in a + # temporary space. Problems with specifying a path inside a Spack instance + # are that it precludes its use as a system package and its ability to be + # pip installable. + # + # In any case, if the username is not already in the path, Spack will append + # the value of `$user` in an attempt to avoid potential conflicts between + # users in shared temporary spaces. + # + # The build stage can be purged with `spack clean --stage` and + # `spack clean -a`, so it is important that the specified directory uniquely + # identifies Spack staging to avoid accidentally wiping out non-Spack work. + build_stage: + - $tempdir/$user/spack-stage + - $user_cache_path/stage + # - $spack/var/spack/stage + + # Directory in which to run tests and store test results. + # Tests will be stored in directories named by date/time and package + # name/hash. + test_stage: $user_cache_path/test + source_cache: $spack/var/spack/cache + misc_cache: $user_cache_path/cache + connect_timeout: 10 + suppress_gpg_warnings: false + install_missing_compilers: false + checksum: true + deprecated: false + dirty: false + build_language: C + locks: true + url_fetch_method: urllib + ccache: false + concretizer: clingo + db_lock_timeout: 3 + package_lock_timeout: null + shared_linking: rpath + allow_sgid: true + terminal_title: false + debug: false + build_jobs: 16 +mirrors: + spack-public: https://mirror.spack.io +modules: + default: + 'enable:': + - lmod + lmod: + core_compilers: + - gcc@8.5.0 + hierarchy: + - mpi + hash_length: 0 + blacklist_implicits: true + whitelist: [py-jupyter-core, hadoop] + naming_scheme: '{name}/{version}/{hash:7}' + projections: + all: '{name}/{version}/{hash:7}' + all: + suffixes: + +openmp: omp + threads=openmp: omp + +ipl64: i64 + environment: + set: + '{name}HOME': '{prefix}' + intel: + environment: + set: + INTEL_LICENSE_FILE: 40000@elprado.sdsc.edu:40200@elprado.sdsc.edu + petsc: + suffixes: + +complex: cmplx + slepc: + suffixes: + ^petsc +complex: cmplx + ^python: + autoload: direct + ^ucx: + autoload: direct + prefix_inspections: + lib: + - LD_LIBRARY_PATH + lib64: + - LD_LIBRARY_PATH + bin: + - PATH + man: + - MANPATH + share/man: + - MANPATH + share/aclocal: + - ACLOCAL_PATH + lib/pkgconfig: + - PKG_CONFIG_PATH + lib64/pkgconfig: + - PKG_CONFIG_PATH + share/pkgconfig: + - PKG_CONFIG_PATH + ? '' + : - CMAKE_PREFIX_PATH + + # These are configurations for the module set named "default" +packages: + lmod: + externals: + - spec: lmod@8.2.4 + prefix: /usr + buildable: false + lustre: + externals: + - spec: lustre@2.15.2 + prefix: /usr + buildable: false + openssl: + externals: + - spec: openssl@1.1.1k + prefix: /usr + buildable: false + rdma-core: + externals: + - spec: rdma-core@43.0 + prefix: /usr + buildable: false + slurm: + externals: + - spec: slurm@21.08.8 + prefix: /cm/shared/apps/slurm/21.08.8 + buildable: false + librsvg: + externals: + - spec: librsvg@2.42.7 + prefix: /usr + buildable: false + ghostscript: + externals: + - spec: ghostscript@9.27 + prefix: /usr + buildable: false + gaussian: + permissions: + read: group + group: gaussian + vasp6: + permissions: + read: group + group: vasp-6 + all: + compiler: [gcc, intel, pgi, clang, xl, nag, fj, aocc] + providers: + awk: [gawk] + blas: [openblas, amdblis] + D: [ldc] + daal: [intel-daal] + elf: [elfutils] + fftw-api: [fftw, amdfftw] + flame: [libflame, amdlibflame] + fuse: [libfuse] + gl: [mesa+opengl, mesa18, opengl] + glu: [mesa-glu, openglu] + glx: [mesa+glx, mesa18+glx, opengl] + golang: [gcc] + iconv: [libiconv] + ipp: [intel-ipp] + java: [openjdk, jdk, ibm-java] + jpeg: [libjpeg-turbo, libjpeg] + lapack: [openblas, amdlibflame] + lua-lang: [lua, lua-luajit] + mariadb-client: [mariadb-c-client, mariadb] + mkl: [intel-mkl] + mpe: [mpe2] + mpi: [openmpi, mpich] + mysql-client: [mysql, mariadb-c-client] + opencl: [pocl] + onedal: [intel-oneapi-dal] + osmesa: [mesa+osmesa, mesa18+osmesa] + pbs: [openpbs, torque] + pil: [py-pillow] + pkgconfig: [pkgconf, pkg-config] + rpc: [libtirpc] + scalapack: [netlib-scalapack, amdscalapack] + sycl: [hipsycl] + szip: [libaec, libszip] + tbb: [intel-tbb] + unwind: [libunwind] + uuid: [util-linux-uuid, libuuid] + xxd: [xxd-standalone, vim] + yacc: [bison, byacc] + ziglang: [zig] + permissions: + read: world + write: user +repos: +- $spack/var/spack/repos/sdsc +- $spack/var/spack/repos/builtin +upstreams: {} +==> Warning: Skipping package at /cm/shared/apps/spack/0.17.3/cpu/b/var/spack/repos/sdsc/packages/amber.configure. "amber.configure" is not a valid Spack module name. +Input spec +-------------------------------- +[ ] .intel-tbb@2020.3%intel@19.1.3.304~ipo+shared+tm + +Concretized +-------------------------------- +lfesfxm [ ] builtin.intel-tbb@2020.3%intel@19.1.3.304 cflags="-O2 -march=native" cxxflags="-O2 -march=native" fflags="-O2 -march=native" ~ipo+shared+tm build_type=RelWithDebInfo cxxstd=default patches=62ba015ebd1819c45bef47411540b789b493e31ca668c4ff4cb2afcbc306b476,ce1fb16fb932ce86a82ca87cf0431d1a8c83652af9f552b264213b2ff2945d73,d62cb666de4010998c339cde6f41c7623a07e9fc69e498f2e149821c0c2c6dd0 arch=linux-rocky8-zen2 +2ofeisf [b ] ^builtin.cmake@3.21.4%intel@19.1.3.304 cflags="-O2 -march=native" cxxflags="-O2 -march=native" fflags="-O2 -march=native" ~doc+ncurses~openssl+ownlibs~qt build_type=Release arch=linux-rocky8-zen2 +5lwrc3q [bl ] ^builtin.ncurses@6.2%intel@19.1.3.304 cflags="-O2 -march=native" cxxflags="-O2 -march=native" fflags="-O2 -march=native" ~symlinks+termlib abi=none arch=linux-rocky8-zen2 +xf5f3g4 [b ] ^builtin.pkgconf@1.8.0%intel@19.1.3.304 cflags="-O2 -march=native" cxxflags="-O2 -march=native" fflags="-O2 -march=native" arch=linux-rocky8-zen2 + +real 10.06 +user 8.24 +sys 0.25 +[+] /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen2/intel-19.1.3.304/pkgconf-1.8.0-xf5f3g4mxwxfxdtof2qrinltr6zpuhmj +[+] /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen2/intel-19.1.3.304/ncurses-6.2-5lwrc3qbdcwmox56wglshuf3lsvc35s7 +[+] /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen2/intel-19.1.3.304/cmake-3.21.4-2ofeisfhz6e7jb6ovev4s66ndhsqdce2 +==> Installing intel-tbb-2020.3-lfesfxmmanw5j6wbvboq5c7bbxblgaid +==> No binary for intel-tbb-2020.3-lfesfxmmanw5j6wbvboq5c7bbxblgaid found: installing from source +==> Warning: Skipping package at /cm/shared/apps/spack/0.17.3/cpu/b/var/spack/repos/sdsc/packages/amber.configure. "amber.configure" is not a valid Spack module name. +==> Warning: Expected user 527834 to own /scratch/spack_cpu, but it is owned by 0 +==> Fetching https://mirror.spack.io/_source-cache/archive/eb/ebc4f6aa47972daed1f7bf71d100ae5bf6931c2e3144cf299c8cc7d041dca2f3.tar.gz +==> Applied patch /cm/shared/apps/spack/0.17.3/cpu/b/var/spack/repos/builtin/packages/intel-tbb/tbb_cmakeConfig-2019.5.patch +==> Applied patch /cm/shared/apps/spack/0.17.3/cpu/b/var/spack/repos/builtin/packages/intel-tbb/makefile-debug.patch +==> Applied patch /cm/shared/apps/spack/0.17.3/cpu/b/var/spack/repos/builtin/packages/intel-tbb/macos-arm64.patch +==> Warning: Intel's compilers may or may not optimize to the same degree for non-Intel microprocessors for optimizations that are not unique to Intel microprocessors +==> intel-tbb: Executing phase: 'cmake' +==> intel-tbb: Executing phase: 'build' +==> intel-tbb: Executing phase: 'install' +==> intel-tbb: Successfully installed intel-tbb-2020.3-lfesfxmmanw5j6wbvboq5c7bbxblgaid + Fetch: 0.06s. Build: 11.84s. Total: 11.90s. +[+] /cm/shared/apps/spack/0.17.3/cpu/b/opt/spack/linux-rocky8-zen2/intel-19.1.3.304/intel-tbb-2020.3-lfesfxmmanw5j6wbvboq5c7bbxblgaid +real 28.68 +user 40.32 +sys 10.40 diff --git a/etc/spack/sdsc/expanse/0.17.3/cpu/b/specs/intel@19.1.3.304/intel-tbb@2020.3.sh b/etc/spack/sdsc/expanse/0.17.3/cpu/b/specs/intel@19.1.3.304/intel-tbb@2020.3.sh index e844819651fd92..7925835235dc47 100644 --- a/etc/spack/sdsc/expanse/0.17.3/cpu/b/specs/intel@19.1.3.304/intel-tbb@2020.3.sh +++ b/etc/spack/sdsc/expanse/0.17.3/cpu/b/specs/intel@19.1.3.304/intel-tbb@2020.3.sh @@ -2,7 +2,7 @@ #SBATCH --job-name=intel-tbb@2020.3 #SBATCH --account=use300 -#SBATCH --reservation=rocky8u7_testing +#SBATCH --reservation=root_73 #SBATCH --partition=ind-shared #SBATCH --nodes=1 #SBATCH --ntasks-per-node=1 @@ -70,4 +70,4 @@ fi #spack module lmod refresh --delete-tree -y -#sbatch --dependency="afterok:${SLURM_JOB_ID}" 'gsl@2.7.sh' +sbatch --dependency="afterok:${SLURM_JOB_ID}" 'gsl@2.7.sh'