User Tools

Site Tools


expanse

Here is the instructions to run NAMD on Expanse.

Your work directory is as /expanse/lustre/scratch/$USER/temp_project/

Example files for equilibrium run (first-dyn.csh and first-dyn.scr) and production run (dyn.csh, dyn.scr, dyn-1.inp, and dyn-2.inp) could be found as follows:
For CPU nodes @ /home/mkhsieh/script/namd_run/cpu/
OR download cpu.tgz
Note:
Each standard compute node has ~256 GB of memory and 128 cores
Each standard node core will be allocated 1 GB of memory, users should explicitly include the –mem directive to request additional memory; Max. memory per compute node –mem = 248G

Example of SBATCH lines

#SBATCH --job-name=your_job_name
#SBATCH --account=**ask Dr. Klauda**
#SBATCH --partition=shared
#SBATCH --nodes=1
#SBATCH --ntasks-per-node=32
#SBATCH --cpus-per-task=1
#SBATCH --mem=64G
#SBATCH --time=48:00:00

Example of execution line

namd2 +p32 +setcpuaffinity dyn.inp >& dyn.out

For GPU nodes @ /home/mkhsieh/script/namd_run/gpu
OR download gpu.tgz
Note:
Each GPU node has 4 GPUs, ~384GB of memory and 40 cores
Default resource allocation for 1 GPU = 1 GPU, 1 CPU, and 1G of memory, users will need to explicitly ask for additional resources in their job script.
For max memory on a GPU node, users should request –mem = 374G; A GPU SU is equivalent to 1GPU, <10CPUs, and <96G of memory.

Example of SBATCH lines

#SBATCH --job-name=your_job_name
#SBATCH --account=**ask Dr. Klauda**
#SBATCH --partition=gpu-shared
#SBATCH --nodes=1
#SBATCH --ntasks-per-node=1
#SBATCH --cpus-per-task=10
#SBATCH --mem=95G
#SBATCH --gpus=1
#SBATCH --time=48:00:00

Example of execution line

mpirun --mca btl_openib_allow_ib true -np 1 --map-by ppr:1:node namd2 +ppn10 +setcpuaffinity +devices 0 dyn.inp >& dyn.out
expanse.txt · Last modified: 2021/08/22 18:49 by edit