oph:cluster:storage
Differenze
Queste sono le differenze tra la revisione selezionata e la versione attuale della pagina.
Entrambe le parti precedenti la revisioneRevisione precedenteProssima revisione | Revisione precedente | ||
oph:cluster:storage [2024/06/03 11:31] – Aggiornato a nuova /scratch e preparato per /archive diego.zuccato@unibo.it | oph:cluster:storage [2025/02/06 16:07] (versione attuale) – [/home/] nuovo sito per /home/web diego.zuccato@unibo.it | ||
---|---|---|---|
Linea 1: | Linea 1: | ||
====== Storage types ====== | ====== Storage types ====== | ||
- | The DIFA-OPH cluster offers different storage areas with very different | + | The DIFA-OPH cluster offers |
- | ==== $HOME ==== | + | ===== /home/ ===== |
- | * This is a storage area available from every node. Should | + | This is a storage area available from every node and this is the space that you access when you connect to the cluster frontend. This is meant to store non-reproducible data (e.g. source codes) and is regularly backed up. It should |
+ | The /home is the area where your home folders are stored, as well as other shared areas such as ''/ | ||
- | * NFS-mount via Ethernet (1Gbps): not very fast but quite responsive. | + | * **/ |
- | * Quota limit: 50GB (soft) / 100GB (hard); check how much you're using with '' | + | * **/ |
+ | * This space is web-accessible at '' | ||
+ | * web access is read-only and it is not possible to create dynamic pages. | ||
+ | * **Per-sector quota** of 1TB (soft) / 2TB (hard) | ||
+ | * Requires either index.html or .htaccess file with '' | ||
- | ==== / | + | ==== Technical characteristics |
- | * **Temporary** replacement for /scratch :!: please migrate | + | * NFS-mount via Ethernet (1Gbps which is not very fast but quite responsive); |
+ | | ||
- | ==== /home/web ==== | + | ===== /scratch/ ==== |
- | * This share is **web-accessible (read-only)** at '' | + | This is the fast Input/ |
- | ==== / | + | :!: folders inside sectors areas **must** use the account as name or you won't get important mails => possibe data loss. |
+ | ==== Technical characteristics: | ||
- | * Used as **work area for jobs that do not need very big datasets** or need to have lots of files in a single directory (**not recommended**, might degrade performances for all users!). | + | * Parallel filesystem, |
+ | * No quota, but **files | ||
+ | | ||
- | ==== /archive (PLANNED) ==== | ||
- | * :!: **PLANNED** :!: => not useable (yet) | + | ===== / |
- | * Main archive area. **To be used for large files, big datasets and archives**. | + | |
- | * **Readonly access** from compute nodes, read/write only from frontends and filetransfer nodes | + | |
- | * **IMPORTANT: | + | |
- | * Quota: **per-sector** (see '' | + | |
- | ==== /scratch ==== | + | This is the main archive area to be used for large files, big datasets or archives; it is designed to be a distributed storage area for long-term data preservation. Data in this area should be stored in the form of compressed folders because the presence of a large number of small files will compromise its functionality. Every sector or project has a dedicated area with an associated quota on '' |
- | | + | :!: folders inside sectors areas **must** use the account as name or you won't get important mails => possibe data loss. |
- | * No quota, but **files older than TBD((Decision to be finalized by the board)) FIXME days** will be automatically deleted without further notice. | + | ==== Technical characteristics: ==== |
- | * No cross-server redundancy, just local RAID: if (when) a server (or two disks in the same RAID) fails, all data becomes unavailable -- always keep a copy of your important data archived elsewhere (maybe /archive, but for very important data **offsite is better**) | + | |
- | ==== $TMPDIR ==== | + | * **To not be used to store a large number of small files**, this will compromise the functionality of the storage space eventually blocking all the reading/ |
+ | * Max size for a single file is 8TB. When archiving big datasets please split them into sub-folders, | ||
+ | * Read-only access from compute nodes, read/write only from frontends and filetransfer nodes | ||
+ | * Quota imposed (both on file size and number of files) per sector, with extra space allocated for specific projects (in / | ||
+ | * Currently ACLs (setfacl) are notsupported (cephfs exported via NFS-Ganesha does not allow to set/get ACLs) | ||
+ | ==== Monitoring system of space usage ==== | ||
+ | |||
+ | To allow users/ | ||
+ | |||
+ | In particular: | ||
+ | * every sector/ | ||
+ | * every sector/ | ||
+ | * individual users will receive an email only if their sector/ | ||
+ | |||
+ | |||
+ | ====== $TMPDIR | ||
+ | |||
+ | Every node does have some available space on local storage in '' | ||
+ | |||
+ | ==== Technical characteristics: | ||
+ | |||
+ | * local space: not shared between multiple nodes, not even for a single multi-node job | ||
+ | * quite fast | ||
+ | * automatically cleaned when job ends | ||
- | Every node does have some available space on local storage. It's useful to store temporary files that do not need to be shared between nodes. Being local, latency is very low. But local disks aren't big. $TMPDIR can usually store around 200GB of data. It gets **automatically cleaned when the job terminates**: |
oph/cluster/storage.1717414272.txt.gz · Ultima modifica: 2024/06/03 11:31 da diego.zuccato@unibo.it