======= Orca =======
To run Orca, you should register at [[https://orcaforum.cec.mpg.de/]]. For now we are allowed to have Orca installed centrally in the SCC. This may change in the future!
===== Using ORCA =====
Load the following modules:
module load openmpi
module load orca
\\
For a serial job, create a jobscript like:
#!/bin/bash
#SBATCH -p medium
#SBATCH -n 1
#SBATCH -t 1-00:00:00
INPUTFILE=test.inp
$ORCA_PATH/orca ${INPUTFILE}
\\
This tells the batch system to submit the job to queue medium and require 1 processor for 24 hours.
For parallel jobs, this needs a little trick, since orca can't run on shared filesystems like NFS, CVFS or FHGFS. We need to use /local as a local filesystem for the run:
#!/bin/bash
#SBATCH -p medium
#SBATCH -J ORCA
#SBATCH -n 20
#SBATCH -N 1
#SBATCH -t 1-00:00:00
#SBATCH --ntasks-per-node=20
#SBATCH --signal=B:12@600
#SBATCH -C local
INPUTFILE=test.inp
work=$PWD
trap 'srun -n ${SLURM_JOB_NUM_NODES} --ntasks-per-node=1 cp -af ${TMP_LOCAL}/* ${work}/; exit 12' 12
cp -af ${INPUTFILE} ${work}/*.gbw ${work}/*.pot ${TMP_LOCAL}/
cd $TMP_LOCAL
$ORCA_PATH/orca ${INPUTFILE} &
wait
srun -n ${SLURM_JOB_NUM_NODES} --ntasks-per-node=1 cp -af ${TMP_LOCAL}/* ${work}/ >/dev/null 2>&1
\\
This tells the batch system to submit the job to partition medium and require 20 processors on one node for 24 hours. **Please make sure that your input file in this case (-n 20) contains the line '%pal nprocs 20 end' (without quotes)!** '%pal nprocs 20 end' must equal the number of processes you reserve with the '-n' option.
Save the script as myjob.job, for example, and submit with
sbatch myjob.job