From 306d8165ed85d8abc8b0f4d4d96297db984220a1 Mon Sep 17 00:00:00 2001 From: guadabsb15 Date: Wed, 29 May 2019 12:06:50 -0400 Subject: [PATCH] added submission scripts and inputs files --- .../summit_201905/CPU_runs/inputs.3d.sph | 58 +++++++++++++++++++ .../CPU_runs/inputs.3d.sph_1level | 58 +++++++++++++++++++ .../summit_201905/CPU_runs/probin.3d.sph | 29 ++++++++++ .../CPU_runs/summit_16node_nogpu.sh | 24 ++++++++ .../CPU_runs/summit_32node_nogpu.sh | 24 ++++++++ .../CPU_runs/summit_4node_nogpu.sh | 24 ++++++++ .../CPU_runs/summit_64node_nogpu.sh | 24 ++++++++ .../CPU_runs/summit_8node_nogpu.sh | 24 ++++++++ .../summit_201905/GPU_runs/inputs.3d.sph | 58 +++++++++++++++++++ .../GPU_runs/inputs.3d.sph_1level | 58 +++++++++++++++++++ .../summit_201905/GPU_runs/probin.3d.sph | 29 ++++++++++ .../summit_201905/GPU_runs/summit_16nodes.sh | 24 ++++++++ .../summit_201905/GPU_runs/summit_32nodes.sh | 24 ++++++++ .../summit_201905/GPU_runs/summit_4nodes.sh | 24 ++++++++ .../summit_201905/GPU_runs/summit_64nodes.sh | 24 ++++++++ .../summit_201905/GPU_runs/summit_8nodes.sh | 24 ++++++++ 16 files changed, 530 insertions(+) create mode 100644 Util/scaling/sedov/summit_201905/CPU_runs/inputs.3d.sph create mode 100644 Util/scaling/sedov/summit_201905/CPU_runs/inputs.3d.sph_1level create mode 100644 Util/scaling/sedov/summit_201905/CPU_runs/probin.3d.sph create mode 100644 Util/scaling/sedov/summit_201905/CPU_runs/summit_16node_nogpu.sh create mode 100644 Util/scaling/sedov/summit_201905/CPU_runs/summit_32node_nogpu.sh create mode 100644 Util/scaling/sedov/summit_201905/CPU_runs/summit_4node_nogpu.sh create mode 100644 Util/scaling/sedov/summit_201905/CPU_runs/summit_64node_nogpu.sh create mode 100644 Util/scaling/sedov/summit_201905/CPU_runs/summit_8node_nogpu.sh create mode 100644 Util/scaling/sedov/summit_201905/GPU_runs/inputs.3d.sph create mode 100644 Util/scaling/sedov/summit_201905/GPU_runs/inputs.3d.sph_1level create mode 100644 Util/scaling/sedov/summit_201905/GPU_runs/probin.3d.sph create mode 100644 Util/scaling/sedov/summit_201905/GPU_runs/summit_16nodes.sh create mode 100644 Util/scaling/sedov/summit_201905/GPU_runs/summit_32nodes.sh create mode 100644 Util/scaling/sedov/summit_201905/GPU_runs/summit_4nodes.sh create mode 100644 Util/scaling/sedov/summit_201905/GPU_runs/summit_64nodes.sh create mode 100644 Util/scaling/sedov/summit_201905/GPU_runs/summit_8nodes.sh diff --git a/Util/scaling/sedov/summit_201905/CPU_runs/inputs.3d.sph b/Util/scaling/sedov/summit_201905/CPU_runs/inputs.3d.sph new file mode 100644 index 0000000000..faea323b54 --- /dev/null +++ b/Util/scaling/sedov/summit_201905/CPU_runs/inputs.3d.sph @@ -0,0 +1,58 @@ +# ------------------ INPUTS TO MAIN PROGRAM ------------------- +max_step = 10 +stop_time = 0.01 + +# PROBLEM SIZE & GEOMETRY +geometry.is_periodic = 0 0 0 +geometry.coord_sys = 0 # 0 => cart +geometry.prob_lo = 0 0 0 +geometry.prob_hi = 1 1 1 +amr.n_cell = 256 256 256 + +# >>>>>>>>>>>>> BC FLAGS <<<<<<<<<<<<<<<< +# 0 = Interior 3 = Symmetry +# 1 = Inflow 4 = SlipWall +# 2 = Outflow 5 = NoSlipWall +# >>>>>>>>>>>>> BC FLAGS <<<<<<<<<<<<<<<< +castro.lo_bc = 2 2 2 +castro.hi_bc = 2 2 2 + +# WHICH PHYSICS +castro.do_hydro = 1 +castro.do_react = 0 +castro.ppm_type = 1 + +# TIME STEP CONTROL +castro.dt_cutoff = 5.e-20 # level 0 timestep below which we halt +castro.cfl = 0.5 # cfl number for hyperbolic system +castro.init_shrink = 0.01 # scale back initial timestep +castro.change_max = 1.1 # maximum increase in dt over successive steps + +# DIAGNOSTICS & VERBOSITY +castro.sum_interval = 1 # timesteps between computing mass +castro.v = 1 # verbosity in Castro.cpp +amr.v = 1 # verbosity in Amr.cpp +#amr.grid_log = grdlog # name of grid logging file + +# REFINEMENT / REGRIDDING +amr.max_level = 0 # maximum level number allowed +amr.ref_ratio = 4 2 2 2 # refinement ratio +amr.regrid_int = 2 # how often to regrid +amr.blocking_factor = 4 # block factor in grid generation +amr.max_grid_size = 32 + + +amr.checkpoint_files_output = 0 +amr.plot_files_output = 0 + +# CHECKPOINT FILES +#amr.check_file = sedov_3d_chk # root name of checkpoint file +#amr.check_int = 200 # number of timesteps between checkpoints + +# PLOTFILES +#amr.plot_file = sedov_3d_plt +#amr.plot_int = 50 +amr.derive_plot_vars=ALL + +# PROBIN FILENAME +amr.probin_file = probin.3d.sph diff --git a/Util/scaling/sedov/summit_201905/CPU_runs/inputs.3d.sph_1level b/Util/scaling/sedov/summit_201905/CPU_runs/inputs.3d.sph_1level new file mode 100644 index 0000000000..a991aeb3dc --- /dev/null +++ b/Util/scaling/sedov/summit_201905/CPU_runs/inputs.3d.sph_1level @@ -0,0 +1,58 @@ +# ------------------ INPUTS TO MAIN PROGRAM ------------------- +max_step = 10 +stop_time = 0.01 + +# PROBLEM SIZE & GEOMETRY +geometry.is_periodic = 0 0 0 +geometry.coord_sys = 0 # 0 => cart +geometry.prob_lo = 0 0 0 +geometry.prob_hi = 1 1 1 +amr.n_cell = 256 256 256 + +# >>>>>>>>>>>>> BC FLAGS <<<<<<<<<<<<<<<< +# 0 = Interior 3 = Symmetry +# 1 = Inflow 4 = SlipWall +# 2 = Outflow 5 = NoSlipWall +# >>>>>>>>>>>>> BC FLAGS <<<<<<<<<<<<<<<< +castro.lo_bc = 2 2 2 +castro.hi_bc = 2 2 2 + +# WHICH PHYSICS +castro.do_hydro = 1 +castro.do_react = 0 +castro.ppm_type = 1 + +# TIME STEP CONTROL +castro.dt_cutoff = 5.e-20 # level 0 timestep below which we halt +castro.cfl = 0.5 # cfl number for hyperbolic system +castro.init_shrink = 0.01 # scale back initial timestep +castro.change_max = 1.1 # maximum increase in dt over successive steps + +# DIAGNOSTICS & VERBOSITY +castro.sum_interval = 1 # timesteps between computing mass +castro.v = 1 # verbosity in Castro.cpp +amr.v = 1 # verbosity in Amr.cpp +#amr.grid_log = grdlog # name of grid logging file + +# REFINEMENT / REGRIDDING +amr.max_level = 1 # maximum level number allowed +amr.ref_ratio = 4 2 2 2 # refinement ratio +amr.regrid_int = 2 # how often to regrid +amr.blocking_factor = 4 # block factor in grid generation +amr.max_grid_size = 32 + + +amr.checkpoint_files_output = 0 +amr.plot_files_output = 0 + +# CHECKPOINT FILES +#amr.check_file = sedov_3d_chk # root name of checkpoint file +#amr.check_int = 200 # number of timesteps between checkpoints + +# PLOTFILES +#amr.plot_file = sedov_3d_plt +#amr.plot_int = 50 +amr.derive_plot_vars=ALL + +# PROBIN FILENAME +amr.probin_file = probin.3d.sph diff --git a/Util/scaling/sedov/summit_201905/CPU_runs/probin.3d.sph b/Util/scaling/sedov/summit_201905/CPU_runs/probin.3d.sph new file mode 100644 index 0000000000..28bf051d65 --- /dev/null +++ b/Util/scaling/sedov/summit_201905/CPU_runs/probin.3d.sph @@ -0,0 +1,29 @@ +&fortin + + r_init = 0.01 + p_ambient = 1.d-5 + exp_energy = 1.0 + dens_ambient = 1.0 + nsub = 10 + +/ + +&tagging + + denerr = 3 + dengrad = 0.01 + max_denerr_lev = 3 + max_dengrad_lev = 3 + + presserr = 3 + pressgrad = 0.01 + max_presserr_lev = 3 + max_pressgrad_lev = 3 + +/ + +&extern + + eos_assume_neutral = T + +/ diff --git a/Util/scaling/sedov/summit_201905/CPU_runs/summit_16node_nogpu.sh b/Util/scaling/sedov/summit_201905/CPU_runs/summit_16node_nogpu.sh new file mode 100644 index 0000000000..e2c605a3fc --- /dev/null +++ b/Util/scaling/sedov/summit_201905/CPU_runs/summit_16node_nogpu.sh @@ -0,0 +1,24 @@ +#!/bin/bash +#BSUB -P ast106 +#BSUB -W 10 +#BSUB -nnodes 16 +#BSUB -alloc_flags smt1 +#BSUB -J Sedov_cpu +#BSUB -o Sedov_cpu.%J +#BSUB -e Sedov_cpu.%J + +cd $LS_SUBCWD + +inputs_file=inputs.3d.sph + +n_mpi=96 # 16 nodes * 6 mpi per node +n_omp=7 +n_gpu=0 +n_cores=7 +n_rs_per_node=6 + +export OMP_NUM_THREADS=$n_omp + +Castro_ex=./Castro3d.pgi.MPI.OMP.ex + +jsrun -n $n_mpi -r $n_rs_per_node -c $n_cores -a 1 -g $n_gpu $Castro_ex $inputs_file diff --git a/Util/scaling/sedov/summit_201905/CPU_runs/summit_32node_nogpu.sh b/Util/scaling/sedov/summit_201905/CPU_runs/summit_32node_nogpu.sh new file mode 100644 index 0000000000..64fa3c8099 --- /dev/null +++ b/Util/scaling/sedov/summit_201905/CPU_runs/summit_32node_nogpu.sh @@ -0,0 +1,24 @@ +#!/bin/bash +#BSUB -P ast106 +#BSUB -W 10 +#BSUB -nnodes 32 +#BSUB -alloc_flags smt1 +#BSUB -J Sedov_cpu +#BSUB -o Sedov_cpu.%J +#BSUB -e Sedov_cpu.%J + +cd $LS_SUBCWD + +inputs_file=inputs.3d.sph_1level + +n_mpi=192 # 32 nodes * 6 mpi per node +n_omp=7 +n_gpu=0 +n_cores=7 +n_rs_per_node=6 + +export OMP_NUM_THREADS=$n_omp + +Castro_ex=./Castro3d.pgi.MPI.OMP.ex + +jsrun -n $n_mpi -r $n_rs_per_node -c $n_cores -a 1 -g $n_gpu $Castro_ex $inputs_file diff --git a/Util/scaling/sedov/summit_201905/CPU_runs/summit_4node_nogpu.sh b/Util/scaling/sedov/summit_201905/CPU_runs/summit_4node_nogpu.sh new file mode 100644 index 0000000000..3e4a986a30 --- /dev/null +++ b/Util/scaling/sedov/summit_201905/CPU_runs/summit_4node_nogpu.sh @@ -0,0 +1,24 @@ +#!/bin/bash +#BSUB -P ast106 +#BSUB -W 25 +#BSUB -nnodes 4 +#BSUB -alloc_flags smt1 +#BSUB -J Sedov_cpu +#BSUB -o Sedov_cpu.%J +#BSUB -e Sedov_cpu.%J + +cd $LS_SUBCWD + +inputs_file=inputs.3d.sph_1level + +n_mpi=24 # 4 nodes * 6 mpi per node +n_omp=7 +n_gpu=0 +n_cores=7 +n_rs_per_node=6 + +export OMP_NUM_THREADS=$n_omp + +Castro_ex=./Castro3d.pgi.MPI.OMP.ex + +jsrun -n $n_mpi -r $n_rs_per_node -c $n_cores -a 1 -g $n_gpu $Castro_ex $inputs_file diff --git a/Util/scaling/sedov/summit_201905/CPU_runs/summit_64node_nogpu.sh b/Util/scaling/sedov/summit_201905/CPU_runs/summit_64node_nogpu.sh new file mode 100644 index 0000000000..29ed0370a1 --- /dev/null +++ b/Util/scaling/sedov/summit_201905/CPU_runs/summit_64node_nogpu.sh @@ -0,0 +1,24 @@ +#!/bin/bash +#BSUB -P ast106 +#BSUB -W 25 +#BSUB -nnodes 64 +#BSUB -alloc_flags smt1 +#BSUB -J Sedov_cpu +#BSUB -o Sedov_cpu.%J +#BSUB -e Sedov_cpu.%J + +cd $LS_SUBCWD + +inputs_file=inputs.3d.sph_1level + +n_mpi=384 # 64 nodes * 6 mpi per node +n_omp=7 +n_gpu=0 +n_cores=7 +n_rs_per_node=6 + +export OMP_NUM_THREADS=$n_omp + +Castro_ex=./Castro3d.pgi.MPI.OMP.ex + +jsrun -n $n_mpi -r $n_rs_per_node -c $n_cores -a 1 -g $n_gpu $Castro_ex $inputs_file diff --git a/Util/scaling/sedov/summit_201905/CPU_runs/summit_8node_nogpu.sh b/Util/scaling/sedov/summit_201905/CPU_runs/summit_8node_nogpu.sh new file mode 100644 index 0000000000..1c0eb15369 --- /dev/null +++ b/Util/scaling/sedov/summit_201905/CPU_runs/summit_8node_nogpu.sh @@ -0,0 +1,24 @@ +#!/bin/bash +#BSUB -P ast106 +#BSUB -W 25 +#BSUB -nnodes 8 +#BSUB -alloc_flags smt1 +#BSUB -J Sedov_cpu +#BSUB -o Sedov_cpu.%J +#BSUB -e Sedov_cpu.%J + +cd $LS_SUBCWD + +inputs_file=inputs.3d.sph_1level + +n_mpi=48 # 8 nodes * 6 mpi per node +n_omp=7 +n_gpu=0 +n_cores=7 +n_rs_per_node=6 + +export OMP_NUM_THREADS=$n_omp + +Castro_ex=./Castro3d.pgi.MPI.OMP.ex + +jsrun -n $n_mpi -r $n_rs_per_node -c $n_cores -a 1 -g $n_gpu $Castro_ex $inputs_file diff --git a/Util/scaling/sedov/summit_201905/GPU_runs/inputs.3d.sph b/Util/scaling/sedov/summit_201905/GPU_runs/inputs.3d.sph new file mode 100644 index 0000000000..bc337cc37b --- /dev/null +++ b/Util/scaling/sedov/summit_201905/GPU_runs/inputs.3d.sph @@ -0,0 +1,58 @@ +# ------------------ INPUTS TO MAIN PROGRAM ------------------- +max_step = 10 +stop_time = 0.01 + +# PROBLEM SIZE & GEOMETRY +geometry.is_periodic = 0 0 0 +geometry.coord_sys = 0 # 0 => cart +geometry.prob_lo = 0 0 0 +geometry.prob_hi = 1 1 1 +amr.n_cell = 256 256 256 + +# >>>>>>>>>>>>> BC FLAGS <<<<<<<<<<<<<<<< +# 0 = Interior 3 = Symmetry +# 1 = Inflow 4 = SlipWall +# 2 = Outflow 5 = NoSlipWall +# >>>>>>>>>>>>> BC FLAGS <<<<<<<<<<<<<<<< +castro.lo_bc = 2 2 2 +castro.hi_bc = 2 2 2 + +# WHICH PHYSICS +castro.do_hydro = 1 +castro.do_react = 0 +castro.ppm_type = 1 + +# TIME STEP CONTROL +castro.dt_cutoff = 5.e-20 # level 0 timestep below which we halt +castro.cfl = 0.5 # cfl number for hyperbolic system +castro.init_shrink = 0.01 # scale back initial timestep +castro.change_max = 1.1 # maximum increase in dt over successive steps + +# DIAGNOSTICS & VERBOSITY +castro.sum_interval = 1 # timesteps between computing mass +castro.v = 1 # verbosity in Castro.cpp +amr.v = 1 # verbosity in Amr.cpp +#amr.grid_log = grdlog # name of grid logging file + +# REFINEMENT / REGRIDDING +amr.max_level = 0 # maximum level number allowed +amr.ref_ratio = 2 2 2 2 # refinement ratio +amr.regrid_int = 2 # how often to regrid +amr.blocking_factor = 8 # block factor in grid generation +amr.max_grid_size = 64 + + +amr.checkpoint_files_output = 0 +amr.plot_files_output = 0 + +# CHECKPOINT FILES +#amr.check_file = sedov_3d_chk # root name of checkpoint file +#amr.check_int = 200 # number of timesteps between checkpoints + +# PLOTFILES +#amr.plot_file = sedov_3d_plt +#amr.plot_int = 50 +amr.derive_plot_vars=ALL + +# PROBIN FILENAME +amr.probin_file = probin.3d.sph diff --git a/Util/scaling/sedov/summit_201905/GPU_runs/inputs.3d.sph_1level b/Util/scaling/sedov/summit_201905/GPU_runs/inputs.3d.sph_1level new file mode 100644 index 0000000000..39c4746348 --- /dev/null +++ b/Util/scaling/sedov/summit_201905/GPU_runs/inputs.3d.sph_1level @@ -0,0 +1,58 @@ +# ------------------ INPUTS TO MAIN PROGRAM ------------------- +max_step = 10 +stop_time = 0.01 + +# PROBLEM SIZE & GEOMETRY +geometry.is_periodic = 0 0 0 +geometry.coord_sys = 0 # 0 => cart +geometry.prob_lo = 0 0 0 +geometry.prob_hi = 1 1 1 +amr.n_cell = 256 256 256 + +# >>>>>>>>>>>>> BC FLAGS <<<<<<<<<<<<<<<< +# 0 = Interior 3 = Symmetry +# 1 = Inflow 4 = SlipWall +# 2 = Outflow 5 = NoSlipWall +# >>>>>>>>>>>>> BC FLAGS <<<<<<<<<<<<<<<< +castro.lo_bc = 2 2 2 +castro.hi_bc = 2 2 2 + +# WHICH PHYSICS +castro.do_hydro = 1 +castro.do_react = 0 +castro.ppm_type = 1 + +# TIME STEP CONTROL +castro.dt_cutoff = 5.e-20 # level 0 timestep below which we halt +castro.cfl = 0.5 # cfl number for hyperbolic system +castro.init_shrink = 0.01 # scale back initial timestep +castro.change_max = 1.1 # maximum increase in dt over successive steps + +# DIAGNOSTICS & VERBOSITY +castro.sum_interval = 1 # timesteps between computing mass +castro.v = 1 # verbosity in Castro.cpp +amr.v = 1 # verbosity in Amr.cpp +#amr.grid_log = grdlog # name of grid logging file + +# REFINEMENT / REGRIDDING +amr.max_level = 1 # maximum level number allowed +amr.ref_ratio = 4 2 2 2 # refinement ratio +amr.regrid_int = 2 # how often to regrid +amr.blocking_factor = 8 # block factor in grid generation +amr.max_grid_size = 64 + + +amr.checkpoint_files_output = 0 +amr.plot_files_output = 0 + +# CHECKPOINT FILES +#amr.check_file = sedov_3d_chk # root name of checkpoint file +#amr.check_int = 200 # number of timesteps between checkpoints + +# PLOTFILES +#amr.plot_file = sedov_3d_plt +#amr.plot_int = 50 +amr.derive_plot_vars=ALL + +# PROBIN FILENAME +amr.probin_file = probin.3d.sph diff --git a/Util/scaling/sedov/summit_201905/GPU_runs/probin.3d.sph b/Util/scaling/sedov/summit_201905/GPU_runs/probin.3d.sph new file mode 100644 index 0000000000..28bf051d65 --- /dev/null +++ b/Util/scaling/sedov/summit_201905/GPU_runs/probin.3d.sph @@ -0,0 +1,29 @@ +&fortin + + r_init = 0.01 + p_ambient = 1.d-5 + exp_energy = 1.0 + dens_ambient = 1.0 + nsub = 10 + +/ + +&tagging + + denerr = 3 + dengrad = 0.01 + max_denerr_lev = 3 + max_dengrad_lev = 3 + + presserr = 3 + pressgrad = 0.01 + max_presserr_lev = 3 + max_pressgrad_lev = 3 + +/ + +&extern + + eos_assume_neutral = T + +/ diff --git a/Util/scaling/sedov/summit_201905/GPU_runs/summit_16nodes.sh b/Util/scaling/sedov/summit_201905/GPU_runs/summit_16nodes.sh new file mode 100644 index 0000000000..70148b8c11 --- /dev/null +++ b/Util/scaling/sedov/summit_201905/GPU_runs/summit_16nodes.sh @@ -0,0 +1,24 @@ +#!/bin/bash +#BSUB -P ast106 +#BSUB -W 30 +#BSUB -nnodes 16 +#BSUB -alloc_flags smt1 +#BSUB -J Sedov_gpu +#BSUB -o Sedov_gpu.%J +#BSUB -e Sedov_gpu.%J + +cd $LS_SUBCWD + +inputs_file=inputs.3d.sph_1level + +n_mpi=96 # 16 nodes * 6 gpu per node +n_omp=1 +n_gpu=1 +n_cores=1 +n_rs_per_node=6 + +export OMP_NUM_THREADS=$n_omp + +Castro_ex=./Castro3d.pgi.MPI.CUDA.ex + +jsrun -n $n_mpi -r $n_rs_per_node -c $n_cores -a 1 -g $n_gpu $Castro_ex $inputs_file diff --git a/Util/scaling/sedov/summit_201905/GPU_runs/summit_32nodes.sh b/Util/scaling/sedov/summit_201905/GPU_runs/summit_32nodes.sh new file mode 100644 index 0000000000..1f3b7c8194 --- /dev/null +++ b/Util/scaling/sedov/summit_201905/GPU_runs/summit_32nodes.sh @@ -0,0 +1,24 @@ +#!/bin/bash +#BSUB -P ast106 +#BSUB -W 15 +#BSUB -nnodes 32 +#BSUB -alloc_flags smt1 +#BSUB -J Sedov_gpu +#BSUB -o Sedov_gpu.%J +#BSUB -e Sedov_gpu.%J + +cd $LS_SUBCWD + +inputs_file=inputs.3d.sph_1level + +n_mpi=192 # 16 nodes * 6 gpu per node +n_omp=1 +n_gpu=1 +n_cores=1 +n_rs_per_node=6 + +export OMP_NUM_THREADS=$n_omp + +Castro_ex=./Castro3d.pgi.MPI.CUDA.ex + +jsrun -n $n_mpi -r $n_rs_per_node -c $n_cores -a 1 -g $n_gpu $Castro_ex $inputs_file diff --git a/Util/scaling/sedov/summit_201905/GPU_runs/summit_4nodes.sh b/Util/scaling/sedov/summit_201905/GPU_runs/summit_4nodes.sh new file mode 100644 index 0000000000..bd1200d8ca --- /dev/null +++ b/Util/scaling/sedov/summit_201905/GPU_runs/summit_4nodes.sh @@ -0,0 +1,24 @@ +#!/bin/bash +#BSUB -P ast106 +#BSUB -W 15 +#BSUB -nnodes 4 +#BSUB -alloc_flags smt1 +#BSUB -J Sedov_gpu +#BSUB -o Sedov_gpu.%J +#BSUB -e Sedov_gpu.%J + +cd $LS_SUBCWD + +inputs_file=inputs.3d.sph_1level + +n_mpi=24 # 4 nodes * 6 gpu per node +n_omp=1 +n_gpu=1 +n_cores=1 +n_rs_per_node=6 + +export OMP_NUM_THREADS=$n_omp + +Castro_ex=./Castro3d.pgi.MPI.CUDA.ex + +jsrun -n $n_mpi -r $n_rs_per_node -c $n_cores -a 1 -g $n_gpu $Castro_ex $inputs_file diff --git a/Util/scaling/sedov/summit_201905/GPU_runs/summit_64nodes.sh b/Util/scaling/sedov/summit_201905/GPU_runs/summit_64nodes.sh new file mode 100644 index 0000000000..e3e4e748e0 --- /dev/null +++ b/Util/scaling/sedov/summit_201905/GPU_runs/summit_64nodes.sh @@ -0,0 +1,24 @@ +#!/bin/bash +#BSUB -P ast106 +#BSUB -W 15 +#BSUB -nnodes 64 +#BSUB -alloc_flags smt1 +#BSUB -J Sedov_gpu +#BSUB -o Sedov_gpu.%J +#BSUB -e Sedov_gpu.%J + +cd $LS_SUBCWD + +inputs_file=inputs.3d.sph_1level + +n_mpi=384 # 64 nodes * 6 gpu per node +n_omp=1 +n_gpu=1 +n_cores=1 +n_rs_per_node=6 + +export OMP_NUM_THREADS=$n_omp + +Castro_ex=./Castro3d.pgi.MPI.CUDA.ex + +jsrun -n $n_mpi -r $n_rs_per_node -c $n_cores -a 1 -g $n_gpu $Castro_ex $inputs_file diff --git a/Util/scaling/sedov/summit_201905/GPU_runs/summit_8nodes.sh b/Util/scaling/sedov/summit_201905/GPU_runs/summit_8nodes.sh new file mode 100644 index 0000000000..76f9be3313 --- /dev/null +++ b/Util/scaling/sedov/summit_201905/GPU_runs/summit_8nodes.sh @@ -0,0 +1,24 @@ +#!/bin/bash +#BSUB -P ast106 +#BSUB -W 30 +#BSUB -nnodes 8 +#BSUB -alloc_flags smt1 +#BSUB -J Sedov_gpu +#BSUB -o Sedov_gpu.%J +#BSUB -e Sedov_gpu.%J + +cd $LS_SUBCWD + +inputs_file=inputs.3d.sph_1level + +n_mpi=48 # 8 nodes * 6 gpu per node +n_omp=1 +n_gpu=1 +n_cores=1 +n_rs_per_node=6 + +export OMP_NUM_THREADS=$n_omp + +Castro_ex=./Castro3d.pgi.MPI.CUDA.ex + +jsrun -n $n_mpi -r $n_rs_per_node -c $n_cores -a 1 -g $n_gpu $Castro_ex $inputs_file