|
# Compute Resources |
|
|
|
## Login Instance |
|
|
|
This is the shell you get into when ssh'ng from outside |
|
|
|
- Networked (except ssh to outside) |
|
- 1 core per user |
|
- 5 GB of RAM per user |
|
- 30 min of CPU time per process |
|
|
|
## Pre/post processing Instance |
|
|
|
Activated with `--partition=prepost` |
|
|
|
- Networked |
|
- only 4 nodes |
|
- 2 to 20 hours |
|
- No limitations of the login shell |
|
- 1x V100-16GB |
|
- The computing hours are not deducted from your allocation |
|
|
|
to request: |
|
``` |
|
srun --pty --partition=prepost --account=six@cpu --nodes=1 --ntasks=1 --cpus-per-task=10 --hint=nomultithread --time=1:00:00 bash --rcfile $six_ALL_CCFRWORK/start-prod |
|
``` |
|
|
|
or to work interactively there, `srun` into the box (though no control which of the 4 you get): |
|
|
|
``` |
|
srun -p prepost -A six@cpu --time=20:00:00 --pty bash |
|
``` |
|
|
|
To choose a specific box (if some are too overload by other users), one could ssh directly to that partition via: |
|
``` |
|
ssh jean-zay-pp # from inside |
|
ssh jean-zay-pp.idris.fr # from outside |
|
``` |
|
There are 4 boxes, so `jean-zay-pp1`, ..., `jean-zay-pp4`. It's possible that larger numbers have less users, but not necessarily. |
|
|
|
In this case there is no need to do SLURM. |
|
|
|
But in this approach only 30min will be given before any running process will be killed. Just like the login shell. I think the only difference is more CPU usage is given here before the process is killed than on the login shell. |
|
|
|
Note: `--partition=compil` too has internet, but can't ssh there. |
|
|
|
In general the `compil` partition is usually less busy than `prepost`. |
|
|
|
|
|
## GPU Instances |
|
|
|
- No network to outside world |
|
- 160 GB of usable memory. The memory allocation is 4 GB per reserved CPU core if hyperthreading is deactivated (`--hint=nomultithread`). So max per node is `--cpus-per-task=40` |
|
|
|
To select this type of partition use `--account=six@gpu`. |
|
|
|
|
|
## CPU Instances |
|
|
|
- All cpus of the same partition are the same |
|
- Different partitions are likely to have different cpus |
|
|
|
For example on `gpu_p1` partitions (4x v100-32gb) |
|
|
|
``` |
|
$ lscpu | grep name |
|
Model name: Intel(R) Xeon(R) Gold 6248 CPU @ 2.50GHz |
|
``` |
|
|
|
To select this type of partition use `--account=six@cpu`. |
|
|
|
|
|
## Quotas |
|
|
|
Group/project (`six`): |
|
|
|
- `$six_ALL_CCFRSCRATCH` - 400TB / ??? inodes fastest (full SSD), → files removed after 30 days without access |
|
- `$six_ALL_CCFRWORK` - 25TB / 500k inodes (slower than SCRATCH) → sources, constantly used input/output files |
|
- `$six_ALL_CCFRSTORE` - 100TB / 100k inodes (slow) → for long term storage in tar files (very few inodes!) |
|
- `/gpfsssd/worksf/projects/rech/six/commun/` - 1TB / 3M inodes → for conda and python git clones that take tens of thousands of inodes |
|
|
|
Personal: |
|
|
|
- `$HOME` - 3GB / 150k inodes (for small files) |
|
- `$SCRATCH` - fastest (full SSD), no quota, files removed after 30 days without access |
|
- `$WORK` - Shared with the `$six_ALL_CCFRWORK` quota, that is `du -sh $six_ALL_CCFRWORK/..` |
|
- `$STORE` - Shared with the `$six_ALL_CCFRSTORE` quota, that is `du -sh $six_ALL_CCFRSTORE/..` |
|
|
|
Note that WORK and STORE group quotas of the project include all project's users' WORK and STORE usage correspondingly. |
|
|
|
[Detailed information](http://www.idris.fr/eng/jean-zay/cpu/jean-zay-cpu-calculateurs-disques-eng.html) |
|
|
|
Checking usage: |
|
``` |
|
idrquota -m # $HOME @ user |
|
idrquota -s -p six # $STORE @ shared (this is updated every 30min) |
|
idrquota -w -p six # $WORK @ shared |
|
``` |
|
|
|
|
|
if you prefer it the easy way here is an alias to add to `~/.bashrc`: |
|
``` |
|
alias dfi=' \ |
|
echo \"*** Total \(six\) ***\"; \ |
|
idrquota -w -p six; \ |
|
idrquota -s -p six; \ |
|
echo SCRATCH: $(du -hs /gpfsscratch/rech/six/ | cut -f1) \(out of 400TB\); \ |
|
echo WORKSF: $(du -hs /gpfsssd/worksf/projects/rech/six | cut -f1) \(out of 2TB\); \ |
|
echo WORKSF: $(du -hs --inodes /gpfsssd/worksf/projects/rech/six | cut -f1) inodes \(out of 3M\); \ |
|
echo; \ |
|
echo \"*** Personal ***\"; \ |
|
idrquota -m; \ |
|
echo WORK: $(du -hs $WORK | cut -f1); \ |
|
echo WORK: $(du -hs --inodes $WORK | cut -f1) inodes; \ |
|
echo STORE: $(du -hs $STORE | cut -f1); \ |
|
echo STORE: $(du -hs --inodes $STORE | cut -f1) inodes; \ |
|
echo SCRATCH: $(du -hs $SCRATCH | cut -f1); \ |
|
echo SCRATCH: $(du -hs --inodes $SCRATCH | cut -f1) inodes; \ |
|
' |
|
``` |
|
This includes the report on usage of personal WORK and SCRATCH partitions. |
|
|
|
|
|
|
|
## Directories |
|
|
|
- `$six_ALL_CCFRSCRATCH` - for checkpoints - make sure to copy important ones to WORK or tarball to STORE |
|
- `$six_ALL_CCFRWORK` - for everything else |
|
- `$six_ALL_CCFRSTORE` - for long term storage in tar files (very few inodes!) |
|
- `/gpfsssd/worksf/projects/rech/six/commun/` - for conda and python git clones that take tens of thousands of inodes - it's a small partition with a huge number of inodes. 1TB and 3M inodes. |
|
XXX: update this and above once env var was created. |
|
|
|
|
|
More specifically: |
|
|
|
- `$six_ALL_CCFRWORK/cache_dir` - `CACHE_DIR` points here |
|
- `$six_ALL_CCFRWORK/checkpoints` - symlink to `$six_ALL_CCFRWORK/checkpoints` - point slurm scripts here |
|
- `$six_ALL_CCFRWORK/code` - clones of repos we use as source (`transformers`, `megatron-lm`, etc.) |
|
- `$six_ALL_CCFRWORK/conda` - our production conda environment |
|
- `$six_ALL_CCFRWORK/datasets` - cached datasets (normally under `~/.cache/huggingface/datasets`) |
|
- `$six_ALL_CCFRWORK/datasets-custom` - Manually created datasets are here (do not delete these - some take many hours to build): |
|
- `$six_ALL_CCFRWORK/downloads` - (normally under `~/.cache/huggingface/downloads`) |
|
- `$six_ALL_CCFRWORK/envs` - custom scripts to create easy to use environments |
|
- `$six_ALL_CCFRWORK/models-custom` - manually created or converted models |
|
- `$six_ALL_CCFRWORK/modules` - (normally under `~/.cache/huggingface/modules`) |
|
|
|
|
|
|
|
## Diagnosing the Lack of Disc Space |
|
|
|
To help diagnose the situations when we are short of disc space here are some tools: |
|
|
|
Useful commands: |
|
|
|
* Get current dir's sub-dir usage breakdown sorted by highest usage first: |
|
``` |
|
du -ahd1 | sort -rh |
|
``` |
|
|
|
* Check that users don't consume too much of their personal `$WORK` space, which goes towards the total WORK space limit. |
|
|
|
``` |
|
du -ahd1 $six_ALL_CCFRWORK/.. | sort -rh |
|
``` |
|
|
|
|
|
## Efficient tar-balling to STORE |
|
|
|
When short on space you don't want to create large tarballs in the WORK dir, instead tar directly to the destination, e.g. |
|
|
|
e.g. w/o gzip since we already have arrow binary files |
|
|
|
``` |
|
mkdir -p $six_ALL_CCFRSTORE/datasets |
|
cd $six_ALL_CCFRWORK/datasets |
|
tar -cvf $six_ALL_CCFRSTORE/datasets/openwebtext.tar openwebtext |
|
``` |
|
|
|
|
|
e.g. w/ gzip for non-binary data |
|
``` |
|
tar -czvf $six_ALL_CCFRSTORE/datasets/openwebtext.tgz openwebtext |
|
``` |
|
|
|
If the file is large and takes some resources to build, `tar` will get killed, in such case you can't do it from the login instance and have to use one of the beefier instances. e.g.: |
|
``` |
|
srun --pty --nodes=1 --ntasks=1 -A six@cpu --cpus-per-task=40 --hint=nomultithread --time=2:00:00 bash --rcfile $six_ALL_CCFRWORK/start-prod |
|
tar ... |
|
``` |
|
and if that's not enough do a slurm job |
|
|