File Systems

This page is still being worked on!

We have a number of different file systems with different characteristics. This page aims to give an overview of these, their usage and availability.

Overview over File systems

Home

Your home directory is available everywhere, permanent, and comes with backup. Your disk space is limited by your quota, but can be increased. It is comparably slow, however. Please do not use it for I/O intensive workloads. You can find more information about the $HOME file system here.

scratch1

This is the shared scratch space, available on amp, agq and agt nodes and frontend login-mdc.hpc.gwdg.de. You can use -C scratch to make sure to get a node with access to shared /scratch. It is very fast, there is currently no automatic file deletion, but also no backup! We may have to delete files manually when we run out of space. You will receive a warning before this happens.
To copy data there, you can use the machine transfer-mdc.hpc.gwdg.de, but have a look at Transfer Data first.

scratch2

This space is the same as scratch described above except it is ONLY available on the nodes dfa, dge, dmp, dsu and dte and on the frontend login-fas.hpc.gwdg.de. You can use -C scratch2 to make sure to get a node with access to that space.
To copy data there, you can use the machine transfer-fas.hpc.gwdg.de, but have a look at Transfer Data first.

local

This is the local hard disk of the node. It is a fast, SSD based option for storing temporary data. Currently, all our nodes have local SSD storage.
A directory is automatically created for each job at /local/jobs/<jobID> and the path is exported as the environment variable $TMP_LOCAL.
Any date saved on the local disk will be removed at the end of the job.

scratch1 SSD Pool

The scratch1 system has directory backed by a very fast SSD pool.

Temporary directories

There are two different temporary directories created in your jobs for different use cases.

$TMP_LOCAL points to the local SSD. This should be used for temporary data, that does not need to be shared across nodes.

$TMP_SCRATCH points to a temporary directory on scratch (even SSD-backed on scratch1). This should be used for temporary data, that have to be shared between nodes of a job, or which size exceed the size of our local SSDs.

Please keep in mind, that these directories will automatically be deleted at the end of the job. If you want to keep data written to this directory, make sure to copy it at the end of the job. Using job signals can also be helpful to prevent accidental data loss.

Scratch Cleanup

To keep our scratch file systems usable, we will soon start a regular clean up. Our current plan is, to delete top-level directories in your scratch directories, that have not been accessed for 30 days.

You will receive sufficient warning before any files are deleted.

This website uses cookies. By using the website, you agree with storing cookies on your computer. Also you acknowledge that you have read and understand our Privacy Policy. If you do not agree leave the website.More information about cookies