LAMMPS (Large-scale Atomic/Molecular Massively Parallel Simulator) is a classical molecular dynamics program from Sandia National Laboratories.
It has potentials for soft materials (biomolecules, polymers) and solid-state materials (metals, semiconductors) and coarse-grained or mesoscopic systems. It can be used to model atoms or, more generically, as a parallel particle simulator at the atomic, meso, or continuum scale.
LAMMPS 16 March 2018 version is compiled and available on the HPC systems. Please use the following template script (lammps_run) to submit your LAMMPS jobs to either parallel12 or parallel24 queues:
#!/bin/bash
#PBS –P project_lmp
#PBS -j oe
#PBS -N lmp_job1
###Submit jobs to queue parallel24 with 2*24 threads.
#PBS -q parallel24
#PBS -l select=2:ncpus=24:mpiprocs=24:mem=20GB
###For submitting jobs to parallel12 queue to run on two nodes
###PBS -q parallel12
###PBS -l select=2:ncpus=12:mpiprocs=12:mem=10GB
cd $PBS_O_WORKDIR; ## this line is needed, do not delete.
np=$(cat ${PBS_NODEFILE} | wc -l ); ### get number of CPUs
source /etc/profile.d/rec_modules.sh
module load lammps_gmp_2018;
export LAM=`which lmp_mpi`
mpirun -np $np -f ${PBS_NODEFILE} $LAM < lmp_job1.in > lmp_job1.logThe GPU version of LAMMPS 16 March 2018 is also compiled and available, please use this sample script (lmpgpu_run) to create and submit your LAMMPS GPU jobs. Please take note of the number of mpiprocs specified.
#!/bin/bash
#PBS –P project_lmp
#PBS -j oe
#PBS -N lmp_gpu
###Submit jobs to queue gpu using four GPUs on two nodes
#PBS -q gpu
#PBS -l select=2:ncpus=12:mpiprocs=2:mem=10GB
cd $PBS_O_WORKDIR; ## this line is needed, do not delete.
np=$( cat ${PBS_NODEFILE} | wc -l ); ### get number of CPUs
####---- LAMMPS Job Execution ---
source /etc/profile.d/rec_modules.sh
module load lammps_gpu
module load cuda7.5
mpirun -np $np -f ${PBS_NODEFILE} lmp_mpi -sf gpu -pk gpu 2 < in.gpujob > out.gpujobSubmit your batch job using the following command:
$ qsub lammps_run
All batch jobs should be submit from the high performance workspace /hpctmp or /hpctmp2.
Please contact us at nusit-hpc@nus.edu.sg if you have any queries.