Overview
The DCC has several shared file systems available for all users of the cluster. General partitions are on Isilon, 40Gbps or 10Gbps network attached storage arrays.
Sensitive data is not permitted on cluster storage.
Path | Size | Description | Backups |
---|---|---|---|
/work/<netid> |
650 TB | Unpartitioned, high speed volume, shared across all users. Files older than 60 days are purged automatically. | None |
/cwork/<netid> |
830 TB | high speed volume, availabe to all users. Files older than 90 days are purged automatically. | None |
/hpc/home/<netid> |
25 GB | Use for personal scripts, and other environment setup. When a users Duke ID is deactivated, their access and home directory is automatically removed from the cluster. | None |
/hpc/group/<groupname> |
1 TB, Expandable for a fee in Research Toolkits | Private to each lab group, persistent working space for applications and projects that last longer than 75 days. | 7 day snapshot |
/hpc/dctrl/<netid> |
500 GB | Private to each PhD student, persistent working space for applications and projects. | 7 day snapshot |
/datacommons/<groupname> |
Fee-based | Archival storage that can be mounted to the DCC to ease transfer of data to computational space and results out to long term storage. Because I/O will not be as performant as with cluster storage, job file access should not be configured that will cause excessive read/write to Data Commons storage. | Optional 30 day backup |
OIT storage rates are set annually. Group storage is "Standard" and Data Commons is "Archive".
How should I use DCC storage?
To optimize the performance of the cluster and make your utilization efficient, we recommend the following:
- /home -> personal scripts and configuration files, environment setup information
- /group -> software installations, lab specific scripts, moderately sized data sets or intermediate results that are needed for longer than 75 days.
-
/work -> large data sets under analysis, intermediate results. In the root folder, create your own folder for your use with:
mkdir <netid>
***Remember: Files older than 75 days are automatically purged! -
/cwork/
-> experimental version of /work for large data sets under analysis, intermediate data, and preliminary results. ***Remember: Files older than 90 days are automatically purged!
-
/datacommons -> long term storage for source data and results data
DCC compute nodes also have a /scratch volume that is local to the compute node. This can be used when highly performant storage is needed during a job, but data should be deleted at the completion of the job. Sizes vary, so use of this requires strong working knowledge of the nodes on the DCC.