Loading doc/jobs.md +3 −3 Original line number Original line Diff line number Diff line Loading @@ -29,7 +29,7 @@ Not many options are needed for a basic serial job: #SBATCH --output=log.%x.job_%j # file name for stdout/stderr (%x will be replaced with the job name, %j with the jobid) #SBATCH --output=log.%x.job_%j # file name for stdout/stderr (%x will be replaced with the job name, %j with the jobid) #SBATCH --time=0:01:00 # maximum wall time allocated for the job (D-H:MM:SS) #SBATCH --time=0:01:00 # maximum wall time allocated for the job (D-H:MM:SS) #SBATCH --partition=gpXY # partition/queue name for the job submission #SBATCH --partition=gp # partition/queue name for the job submission #SBATCH --ntasks=1 # number of tasks/processes #SBATCH --ntasks=1 # number of tasks/processes # start the job in the directory it was submitted from # start the job in the directory it was submitted from Loading Loading @@ -150,7 +150,7 @@ This can be achieved by exporting the `OMP_NUM_THREADS` according to the Slurm c #SBATCH --output=log.%x.job_%j # file name for stdout/stderr (%x will be replaced with the job name, %j with the jobid) #SBATCH --output=log.%x.job_%j # file name for stdout/stderr (%x will be replaced with the job name, %j with the jobid) #SBATCH --time=0:01:00 # maximum wall time allocated for the job (D-H:MM:SS) #SBATCH --time=0:01:00 # maximum wall time allocated for the job (D-H:MM:SS) #SBATCH --partition=gpXY # partition/queue name for the job submission #SBATCH --partition=gp # partition/queue name for the job submission #SBATCH --ntasks=1 # number of tasks/processes #SBATCH --ntasks=1 # number of tasks/processes #SBATCH --threads-per-core=1 # do not use hyperthreads (i.e. CPUs = physical cores below) #SBATCH --threads-per-core=1 # do not use hyperthreads (i.e. CPUs = physical cores below) Loading Loading @@ -184,7 +184,7 @@ For example: #SBATCH --output=log.%x.job_%j # file name for stdout/stderr (%x will be replaced with the job name, %j with the jobid) #SBATCH --output=log.%x.job_%j # file name for stdout/stderr (%x will be replaced with the job name, %j with the jobid) #SBATCH --time=0:01:00 # maximum wall time allocated for the job (D-H:MM:SS) #SBATCH --time=0:01:00 # maximum wall time allocated for the job (D-H:MM:SS) #SBATCH --partition=gpXY # partition/queue name for the job submission #SBATCH --partition=gp # partition/queue name for the job submission #SBATCH --nodes=2 # number of nodes #SBATCH --nodes=2 # number of nodes #SBATCH --ntasks-per-node=2 # MPI processes per node #SBATCH --ntasks-per-node=2 # MPI processes per node Loading Loading
doc/jobs.md +3 −3 Original line number Original line Diff line number Diff line Loading @@ -29,7 +29,7 @@ Not many options are needed for a basic serial job: #SBATCH --output=log.%x.job_%j # file name for stdout/stderr (%x will be replaced with the job name, %j with the jobid) #SBATCH --output=log.%x.job_%j # file name for stdout/stderr (%x will be replaced with the job name, %j with the jobid) #SBATCH --time=0:01:00 # maximum wall time allocated for the job (D-H:MM:SS) #SBATCH --time=0:01:00 # maximum wall time allocated for the job (D-H:MM:SS) #SBATCH --partition=gpXY # partition/queue name for the job submission #SBATCH --partition=gp # partition/queue name for the job submission #SBATCH --ntasks=1 # number of tasks/processes #SBATCH --ntasks=1 # number of tasks/processes # start the job in the directory it was submitted from # start the job in the directory it was submitted from Loading Loading @@ -150,7 +150,7 @@ This can be achieved by exporting the `OMP_NUM_THREADS` according to the Slurm c #SBATCH --output=log.%x.job_%j # file name for stdout/stderr (%x will be replaced with the job name, %j with the jobid) #SBATCH --output=log.%x.job_%j # file name for stdout/stderr (%x will be replaced with the job name, %j with the jobid) #SBATCH --time=0:01:00 # maximum wall time allocated for the job (D-H:MM:SS) #SBATCH --time=0:01:00 # maximum wall time allocated for the job (D-H:MM:SS) #SBATCH --partition=gpXY # partition/queue name for the job submission #SBATCH --partition=gp # partition/queue name for the job submission #SBATCH --ntasks=1 # number of tasks/processes #SBATCH --ntasks=1 # number of tasks/processes #SBATCH --threads-per-core=1 # do not use hyperthreads (i.e. CPUs = physical cores below) #SBATCH --threads-per-core=1 # do not use hyperthreads (i.e. CPUs = physical cores below) Loading Loading @@ -184,7 +184,7 @@ For example: #SBATCH --output=log.%x.job_%j # file name for stdout/stderr (%x will be replaced with the job name, %j with the jobid) #SBATCH --output=log.%x.job_%j # file name for stdout/stderr (%x will be replaced with the job name, %j with the jobid) #SBATCH --time=0:01:00 # maximum wall time allocated for the job (D-H:MM:SS) #SBATCH --time=0:01:00 # maximum wall time allocated for the job (D-H:MM:SS) #SBATCH --partition=gpXY # partition/queue name for the job submission #SBATCH --partition=gp # partition/queue name for the job submission #SBATCH --nodes=2 # number of nodes #SBATCH --nodes=2 # number of nodes #SBATCH --ntasks-per-node=2 # MPI processes per node #SBATCH --ntasks-per-node=2 # MPI processes per node Loading