Orca

To run Orca, you should register at https://orcaforum.cec.mpg.de/. For now we are allowed to have Orca installed centrally in the SCC. This may change in the future!

Using ORCA

Load the following modules:

module load openmpi
module load orca


For a serial job, create a jobscript like:

#!/bin/bash
#SBATCH -p medium
#SBATCH -n 1
#SBATCH -t 1-00:00:00

INPUTFILE=test.inp

$ORCA_PATH/orca ${INPUTFILE}


This tells the batch system to submit the job to queue medium and require 1 processor for 24 hours.

For parallel jobs, this needs a little trick, since orca can't run on shared filesystems like NFS, CVFS or FHGFS. We need to use /local as a local filesystem for the run:

#!/bin/bash
#SBATCH -p medium
#SBATCH -J ORCA
#SBATCH -n 20
#SBATCH -N 1
#SBATCH -t 1-00:00:00
#SBATCH --ntasks-per-node=20
#SBATCH --signal=B:12@600
#SBATCH -C local

INPUTFILE=test.inp

work=$PWD

trap 'srun -n ${SLURM_JOB_NUM_NODES} --ntasks-per-node=1 cp -af ${TMP_LOCAL}/* ${work}/; exit 12' 12

cp -af ${INPUTFILE} ${work}/*.gbw ${work}/*.pot ${TMP_LOCAL}/

cd $TMP_LOCAL

$ORCA_PATH/orca ${INPUTFILE} &
wait

srun -n ${SLURM_JOB_NUM_NODES} --ntasks-per-node=1 cp -af ${TMP_LOCAL}/* ${work}/ >/dev/null 2>&1


This tells the batch system to submit the job to partition medium and require 20 processors on one node for 24 hours. Please make sure that your input file in this case (-n 20) contains the line '%pal nprocs 20 end' (without quotes)! '%pal nprocs 20 end' must equal the number of processes you reserve with the '-n' option.

Save the script as myjob.job, for example, and submit with

sbatch myjob.job