oph:cluster:storage
Differenze
Queste sono le differenze tra la revisione selezionata e la versione attuale della pagina.
Entrambe le parti precedenti la revisioneRevisione precedenteProssima revisione | Revisione precedente | ||
oph:cluster:storage [2024/08/19 11:16] – [/home/temp] aggiunti avvisi diego.zuccato@unibo.it | oph:cluster:storage [2025/02/06 16:07] (versione attuale) – [/home/] nuovo sito per /home/web diego.zuccato@unibo.it | ||
---|---|---|---|
Linea 1: | Linea 1: | ||
====== Storage types ====== | ====== Storage types ====== | ||
- | The DIFA-OPH cluster offers different storage areas with very different | + | The DIFA-OPH cluster offers |
- | ==== $HOME ==== | + | ===== /home/ ===== |
- | * This is a storage area available from every node. Should | + | This is a storage area available from every node and this is the space that you access when you connect to the cluster frontend. This is meant to store non-reproducible data (e.g. source codes) and is regularly backed up. It should |
+ | The /home is the area where your home folders are stored, as well as other shared areas such as ''/ | ||
- | * NFS-mount via Ethernet (1Gbps): not very fast but quite responsive. | + | * **/ |
- | * Quota limit: 50GB (soft) / 100GB (hard); check how much you're using with '' | + | * **/ |
+ | * This space is web-accessible at '' | ||
+ | * web access is read-only and it is not possible to create dynamic pages. | ||
+ | * **Per-sector quota** of 1TB (soft) / 2TB (hard) | ||
+ | * Requires either index.html or .htaccess file with '' | ||
- | ==== / | + | ==== Technical characteristics |
- | * **Temporary** replacement for /scratch :!: please <wrap hi> | + | * NFS-mount via Ethernet (1Gbps which is not very fast but quite responsive); |
- | * Will be deleted shortly after / | + | * Quota limit to 50 GB (100GB as hard limit); check how much you're using with '' |
- | * <wrap em>do not use for jobs' | + | |
- | ==== /home/web ==== | + | |
- | * This share is **web-accessible (read-only)** at '' | + | ===== /scratch/ ==== |
- | ==== /home/work ==== | + | This is the fast Input/Output area to be used for direct read/write operations from the compute nodes. There is no quota in this area, but an automatic cleaning procedure is enforced on all files older than 40 days to avoid the disk space being exhausted, as this would make running jobs crash when trying to write outputs on disk. Therefore, once your jobs are finished you are recommended to archive the relevant data to ''/ |
- | | + | :!: folders inside sectors areas **must** use the account as name or you won't get important mails => possibe data loss. |
+ | ==== Technical characteristics: | ||
- | ==== / | + | * Parallel filesystem, for quick (SSD-backed) access to the data you are working on; |
+ | * No quota, but **files older than 40 days will be automatically deleted** without further notice; | ||
+ | * No cross-server redundancy, just local RAID. Hence, if (when) a server (or two disks in the same RAID) fails, all data becomes unavailable. | ||
- | * :!: **PLANNED** :!: => not useable (yet) | ||
- | * Main archive area. **To be used for large files, big datasets and archives**. | ||
- | * **Readonly access** from compute nodes, read/write only from frontends and filetransfer nodes | ||
- | * **IMPORTANT: | ||
- | * Quota: **per-sector** (see '' | ||
- | ==== /scratch | + | ===== /archive/ |
- | * Parallel filesystem, for **quick** (SSD-backed) access to the data you're working on. | + | This is the main archive area to be used for large files, big datasets or archives; it is designed to be a distributed storage area for long-term data preservation. Data in this area should |
- | * No quota, but **files older than TBD((Decision to be finalized by the board)) FIXME days** | + | |
- | * No cross-server redundancy, just local RAID: if (when) a server (or two disks in the same RAID) fails, all data becomes unavailable -- always keep a copy of your important data archived elsewhere (maybe | + | |
- | ==== $TMPDIR ==== | + | :!: folders inside sectors areas **must** use the account as name or you won't get important mails => possibe data loss. |
+ | ==== Technical characteristics: | ||
+ | |||
+ | * **To not be used to store a large number of small files**, this will compromise the functionality of the storage space eventually blocking all the reading/ | ||
+ | * Max size for a single file is 8TB. When archiving big datasets please split them into sub-folders, | ||
+ | * Read-only access from compute nodes, read/write only from frontends and filetransfer nodes | ||
+ | * Quota imposed (both on file size and number of files) per sector, with extra space allocated for specific projects (in / | ||
+ | * Currently ACLs (setfacl) are notsupported (cephfs exported via NFS-Ganesha does not allow to set/get ACLs) | ||
+ | ==== Monitoring system of space usage ==== | ||
+ | |||
+ | To allow users/ | ||
+ | |||
+ | In particular: | ||
+ | * every sector/ | ||
+ | * every sector/ | ||
+ | * individual users will receive an email only if their sector/ | ||
+ | |||
+ | |||
+ | ====== $TMPDIR | ||
+ | |||
+ | Every node does have some available space on local storage in '' | ||
+ | |||
+ | ==== Technical characteristics: | ||
+ | |||
+ | * local space: not shared between multiple nodes, not even for a single multi-node job | ||
+ | * quite fast | ||
+ | * automatically cleaned when job ends | ||
- | Every node does have some available space on local storage. It's useful to store temporary files that do not need to be shared between nodes. Being local, latency is very low. But local disks aren't big. $TMPDIR can usually store around 200GB of data. It gets **automatically cleaned when the job terminates**: |
oph/cluster/storage.1724066213.txt.gz · Ultima modifica: 2024/08/19 11:16 da diego.zuccato@unibo.it