| 29 | |
| 30 | = Storage on Cypress = |
| 31 | Every Cypress user has two locations in which to store data: A small, high security, low performance, personal home directory and a large, secure, group shared Lustre directory. |
| 32 | |
| 33 | == Storage: home directory == |
| 34 | |
| 35 | Your home directory on Cypress is intended to store customized source code, binaries, scripts, analyzed results, manuscripts, and other small but important files. This directory is limited to 10 GB (10,000 MB), and is backed up. To view your current quota and usage, run the command: |
| 36 | |
| 37 | {{{ |
| 38 | quota -f /home |
| 39 | }}} |
| 40 | |
| 41 | Please do not use your home directory to perform simulations with heavy I/O (file read/write) usage. Instead, please use your group's Lustre project directory. |
| 42 | |
| 43 | == Storage: Lustre group project directory == |
| 44 | |
| 45 | Cypress has a 699 TB Lustre filesystem available for use by active jobs, and for sharing large files for active jobs within a research group. The Lustre filesystem has 2 Object Storage Servers (OSSs) which provide file I/O for 24 Object Storage Targets (OSTs). The Lustre filesystem is available to compute nodes via the 40 Gigabit Ethernet network. The default stripe count is set to 1. |
| 46 | |
| 47 | Allocations on this filesystem are provided per project/research group. Each group is given a space allocation of 1 TB and an inode allocation of 1 million (i.e. up to 1 million files or directories) on the Lustre filesystem. If you need additional disk space to run your computations, your PI may request a quota adjustment. To request a quota adjustment, please provide details and an estimate of the disk space used/required by your computations. Allocations are based on demonstrated need and available resources. |
| 48 | |
| 49 | The Lustre filesystem is not for bulk or archival storage of data. The filesystem is configured for redundancy and hardware fault tolerance, but is not backed up. If you require space for bulk / archival storage, please contact us, and we will take a look at the available options. |
| 50 | |
| 51 | Your group's Lustre project directory will be at: |
| 52 | |
| 53 | {{{ |
| 54 | /lustre/project/<your-group-name> |
| 55 | }}} |
| 56 | |
| 57 | "your-group-name" is your Linux group name, as returned by the command "id -gn". Your group is free to organize your project directory as desired, but it is recommended to create separate subfolders for different sets of data, or for different groups of simulations. |
| 58 | |
| 59 | To view your group's current usage and quota, run the command: |
| 60 | |
| 61 | {{{ |
| 62 | lfs quota -g `id -gn` /lustre |
| 63 | }}} |
| 64 | |
| 65 | To view your own usage, you can run: |
| 66 | |
| 67 | {{{ |
| 68 | lfs quota -u `id -un` /lustre |
| 69 | }}} |