Orca
To run Orca, you should register at https://orcaforum.cec.mpg.de/. For now we are allowed to have Orca installed centrally in the SCC. This may change in the future!
Using ORCA
Load the following modules:
module load openmpi module load orca
For a serial job, create a jobscript like:
#!/bin/bash #SBATCH -p medium #SBATCH -n 1 #SBATCH -t 1-00:00:00 INPUTFILE=test.inp $ORCA_PATH/orca ${INPUTFILE}
This tells the batch system to submit the job to queue medium and require 1 processor for 24 hours.
For parallel jobs, this needs a little trick, since orca can't run on shared filesystems like NFS, CVFS or FHGFS. We need to use /local as a local filesystem for the run:
#!/bin/bash #SBATCH -p medium #SBATCH -J ORCA #SBATCH -n 20 #SBATCH -N 1 #SBATCH -t 1-00:00:00 #SBATCH --ntasks-per-node=20 #SBATCH --signal=B:12@600 #SBATCH -C local INPUTFILE=test.inp work=$PWD trap 'srun -n ${SLURM_JOB_NUM_NODES} --ntasks-per-node=1 cp -af ${TMP_LOCAL}/* ${work}/; exit 12' 12 cp -af ${INPUTFILE} ${work}/*.gbw ${work}/*.pot ${TMP_LOCAL}/ cd $TMP_LOCAL $ORCA_PATH/orca ${INPUTFILE} & wait srun -n ${SLURM_JOB_NUM_NODES} --ntasks-per-node=1 cp -af ${TMP_LOCAL}/* ${work}/ >/dev/null 2>&1
This tells the batch system to submit the job to partition medium and require 20 processors on one node for 24 hours. Please make sure that your input file in this case (-n 20) contains the line '%pal nprocs 20 end' (without quotes)! '%pal nprocs 20 end' must equal the number of processes you reserve with the '-n' option.
Save the script as myjob.job, for example, and submit with
sbatch myjob.job