oph:cluster:storage
Differenze
Queste sono le differenze tra la revisione selezionata e la versione attuale della pagina.
Entrambe le parti precedenti la revisioneRevisione precedenteProssima revisione | Revisione precedente | ||
oph:cluster:storage [2024/09/27 08:56] – [/home/temp (deprecated, pending removal)] diego.zuccato@unibo.it | oph:cluster:storage [2025/02/06 16:07] (versione attuale) – [/home/] nuovo sito per /home/web diego.zuccato@unibo.it | ||
---|---|---|---|
Linea 1: | Linea 1: | ||
====== Storage types ====== | ====== Storage types ====== | ||
- | The DIFA-OPH cluster offers different storage areas with very different | + | The DIFA-OPH cluster offers |
- | ==== $HOME ==== | + | ===== /home/ ===== |
- | * This is a storage area available from every node. Should | + | This is a storage area available from every node and this is the space that you access when you connect to the cluster frontend. This is meant to store non-reproducible data (e.g. source codes) and is regularly backed up. It should |
+ | The /home is the area where your home folders are stored, as well as other shared areas such as ''/ | ||
- | * NFS-mount via Ethernet (1Gbps): not very fast but quite responsive. | + | * **/ |
- | * Quota limit: 50GB (soft) / 100GB (hard); check how much you're using with '' | + | * **/ |
+ | * This space is web-accessible at '' | ||
+ | * web access is read-only and it is not possible to create dynamic pages. | ||
+ | * **Per-sector quota** of 1TB (soft) / 2TB (hard) | ||
+ | * Requires either index.html or .htaccess file with '' | ||
- | ==== /home/temp (deprecated, | + | ==== Technical characteristics |
- | <WRAP center round important 80%> | + | * NFS-mount via Ethernet (1Gbps which is not very fast but quite responsive); |
- | Migrate data you want to keep to /scratch: /home/temp <wrap em>will be deleted soon</ | + | * Quota limit to 50 GB (100GB as hard limit); check how much you're using with '' |
- | </ | + | |
- | * **Temporary** replacement for /scratch | + | ===== /scratch/ ==== |
- | * Do not use for jobs' output | + | |
- | ==== /home/web ==== | + | |
- | * This share is **web-accessible (read-only)** at '' | + | This is the fast Input/ |
- | ==== / | + | :!: folders inside sectors areas **must** use the account as name or you won't get important mails => possibe data loss. |
+ | ==== Technical characteristics: | ||
- | * Used as **work area for jobs that do not need very big datasets** or need to have lots of files in a single directory (**not recommended**, might degrade performances for all users!). | + | * Parallel filesystem, |
+ | * No quota, but **files | ||
+ | | ||
- | ==== /archive ==== | ||
- | * Main archive | + | ===== /archive/ ==== |
- | **IMPORTANT: | + | |
- | If you need some big files (say a dataset) and your code do not allow for specifying a different path, just **use symlinks from** '' | + | This is the main archive area to be used for large files, big datasets or archives; it is designed |
- | * Max size for a single file is **8TB**: when archiving big dataset use splitting (preferably chunks should be less than 1TB) | + | |
- | * **Readonly access** from compute nodes, read/write only from frontends | + | |
- | * Quota (both on file size and number of files): **per-sector**, with extra space allocated for specific projects (in / | + | |
- | ==== / | + | :!: folders inside sectors areas **must** use the account as name or you won't get important mails => possibe data loss. |
+ | ==== Technical characteristics: | ||
- | * Parallel filesystem, for **quick** (SSD-backed) access to the data you're working on. | + | * **To not be used to store a large number of small files**, this will compromise |
- | * No quota, but **files older than TBD((Decision to be finalized by the board)) FIXME days** will be automatically deleted without further notice. | + | * Max size for a single file is 8TB. When archiving big datasets please split them into sub-folders, compress and store them separately |
- | * No cross-server redundancy, just local RAID: if (when) a server | + | * Read-only access from compute nodes, read/write only from frontends and filetransfer nodes |
+ | * Quota imposed | ||
+ | | ||
+ | ==== Monitoring system of space usage ==== | ||
- | ==== $TMPDIR ==== | + | To allow users/ |
+ | |||
+ | In particular: | ||
+ | * every sector/ | ||
+ | * every sector/ | ||
+ | * individual users will receive an email only if their sector/ | ||
+ | |||
+ | |||
+ | ====== $TMPDIR | ||
+ | |||
+ | Every node does have some available space on local storage in '' | ||
+ | |||
+ | ==== Technical characteristics: | ||
+ | |||
+ | * local space: not shared between multiple nodes, not even for a single multi-node job | ||
+ | * quite fast | ||
+ | * automatically cleaned when job ends | ||
- | Every node does have some available space on local storage. It's useful to store temporary files that do not need to be shared between nodes. Being local, latency is very low. But local disks aren't big. $TMPDIR can usually store around 200GB of data. It gets **automatically cleaned when the job terminates**: |
oph/cluster/storage.1727427399.txt.gz · Ultima modifica: 2024/09/27 08:56 da diego.zuccato@unibo.it