ml21cm / frontera_generate_dataset.sbatch
Xsmos's picture
0815-184556
aa037e1 verified
#!/bin/bash
#----------------------------------------------------
# Sample Slurm job script
# for TACC Frontera CLX nodes
#
# *** MPI Job in Normal Queue ***
#
# Last revised: 20 May 2019
#
# Notes:
#
# -- Launch this script by executing
# "sbatch clx.mpi.slurm" on a Frontera login node.
#
# -- Use ibrun to launch MPI codes on TACC systems.
# Do NOT use mpirun or mpiexec.
#
# -- Max recommended MPI ranks per CLX node: 56
# (start small, increase gradually).
#
# -- If you're running out of memory, try running
# fewer tasks per node to give each task more memory.
#
#----------------------------------------------------
#SBATCH -J datasets # Job name
#SBATCH -o Report-%j # Name of stdout output file
#SBATCH -p small # Queue (partition) name
#SBATCH -N 2 # 50 # Total # of nodes
#SBATCH -t 09:00:00 # Run time (hh:mm:ss)
#SBATCH --mail-type=all # Send email at begin and end of job
#SBATCH --mail-user=xiabin@gatech.edu
#SBATCH --ntasks-per-node=1
# Any other commands must follow all #SBATCH directives...
############# #SBATCH -c 56 # Total # of mpi tasks
#----------------------------------------------------
cat $0
date
pwd
module list
conda env list
srun python generate_dataset.py \
--save_direc $SCRATCH \
--num_images 800\
--BOX_LEN 128 \
--HII_DIM 64 \
--NON_CUBIC_FACTOR 16 \
--cpus_per_node 38 \
#----------------------------------------------------