Compute Resources
Login Instance
This is the shell you get into when ssh'ng from outside
- Networked (except ssh to outside)
- 1 core per user
- 5 GB of RAM per user
- 30 min of CPU time per process
Pre/post processing Instance
Activated with --partition=prepost
- Networked
- only 4 nodes
- 2 to 20 hours
- No limitations of the login shell
- 1x V100-16GB
- The computing hours are not deducted from your allocation
to request:
srun --pty --partition=prepost --account=six@cpu --nodes=1 --ntasks=1 --cpus-per-task=10 --hint=nomultithread --time=1:00:00 bash --rcfile $six_ALL_CCFRWORK/start-prod
or to work interactively there, srun
into the box (though no control which of the 4 you get):
srun -p prepost -A six@cpu --time=20:00:00 --pty bash
To choose a specific box (if some are too overload by other users), one could ssh directly to that partition via:
ssh jean-zay-pp # from inside
ssh jean-zay-pp.idris.fr # from outside
There are 4 boxes, so jean-zay-pp1
, ..., jean-zay-pp4
. It's possible that larger numbers have less users, but not necessarily.
In this case there is no need to do SLURM.
But in this approach only 30min will be given before any running process will be killed. Just like the login shell. I think the only difference is more CPU usage is given here before the process is killed than on the login shell.
Note: --partition=compil
too has internet, but can't ssh there.
In general the compil
partition is usually less busy than prepost
.
GPU Instances
- No network to outside world
- 160 GB of usable memory. The memory allocation is 4 GB per reserved CPU core if hyperthreading is deactivated (
--hint=nomultithread
). So max per node is--cpus-per-task=40
To select this type of partition use --account=six@gpu
.
CPU Instances
- All cpus of the same partition are the same
- Different partitions are likely to have different cpus
For example on gpu_p1
partitions (4x v100-32gb)
$ lscpu | grep name
Model name: Intel(R) Xeon(R) Gold 6248 CPU @ 2.50GHz
To select this type of partition use --account=six@cpu
.
Quotas
Group/project (six
):
$six_ALL_CCFRSCRATCH
- 400TB / ??? inodes fastest (full SSD), → files removed after 30 days without access$six_ALL_CCFRWORK
- 25TB / 500k inodes (slower than SCRATCH) → sources, constantly used input/output files$six_ALL_CCFRSTORE
- 100TB / 100k inodes (slow) → for long term storage in tar files (very few inodes!)/gpfsssd/worksf/projects/rech/six/commun/
- 1TB / 3M inodes → for conda and python git clones that take tens of thousands of inodes
Personal:
$HOME
- 3GB / 150k inodes (for small files)$SCRATCH
- fastest (full SSD), no quota, files removed after 30 days without access$WORK
- Shared with the$six_ALL_CCFRWORK
quota, that isdu -sh $six_ALL_CCFRWORK/..
$STORE
- Shared with the$six_ALL_CCFRSTORE
quota, that isdu -sh $six_ALL_CCFRSTORE/..
Note that WORK and STORE group quotas of the project include all project's users' WORK and STORE usage correspondingly.
Checking usage:
idrquota -m # $HOME @ user
idrquota -s -p six # $STORE @ shared (this is updated every 30min)
idrquota -w -p six # $WORK @ shared
if you prefer it the easy way here is an alias to add to ~/.bashrc
:
alias dfi=' \
echo \"*** Total \(six\) ***\"; \
idrquota -w -p six; \
idrquota -s -p six; \
echo SCRATCH: $(du -hs /gpfsscratch/rech/six/ | cut -f1) \(out of 400TB\); \
echo WORKSF: $(du -hs /gpfsssd/worksf/projects/rech/six | cut -f1) \(out of 2TB\); \
echo WORKSF: $(du -hs --inodes /gpfsssd/worksf/projects/rech/six | cut -f1) inodes \(out of 3M\); \
echo; \
echo \"*** Personal ***\"; \
idrquota -m; \
echo WORK: $(du -hs $WORK | cut -f1); \
echo WORK: $(du -hs --inodes $WORK | cut -f1) inodes; \
echo STORE: $(du -hs $STORE | cut -f1); \
echo STORE: $(du -hs --inodes $STORE | cut -f1) inodes; \
echo SCRATCH: $(du -hs $SCRATCH | cut -f1); \
echo SCRATCH: $(du -hs --inodes $SCRATCH | cut -f1) inodes; \
'
This includes the report on usage of personal WORK and SCRATCH partitions.
Directories
$six_ALL_CCFRSCRATCH
- for checkpoints - make sure to copy important ones to WORK or tarball to STORE$six_ALL_CCFRWORK
- for everything else$six_ALL_CCFRSTORE
- for long term storage in tar files (very few inodes!)/gpfsssd/worksf/projects/rech/six/commun/
- for conda and python git clones that take tens of thousands of inodes - it's a small partition with a huge number of inodes. 1TB and 3M inodes. XXX: update this and above once env var was created.
More specifically:
$six_ALL_CCFRWORK/cache_dir
-CACHE_DIR
points here$six_ALL_CCFRWORK/checkpoints
- symlink to$six_ALL_CCFRWORK/checkpoints
- point slurm scripts here$six_ALL_CCFRWORK/code
- clones of repos we use as source (transformers
,megatron-lm
, etc.)$six_ALL_CCFRWORK/conda
- our production conda environment$six_ALL_CCFRWORK/datasets
- cached datasets (normally under~/.cache/huggingface/datasets
)$six_ALL_CCFRWORK/datasets-custom
- Manually created datasets are here (do not delete these - some take many hours to build):$six_ALL_CCFRWORK/downloads
- (normally under~/.cache/huggingface/downloads
)$six_ALL_CCFRWORK/envs
- custom scripts to create easy to use environments$six_ALL_CCFRWORK/models-custom
- manually created or converted models$six_ALL_CCFRWORK/modules
- (normally under~/.cache/huggingface/modules
)
Diagnosing the Lack of Disc Space
To help diagnose the situations when we are short of disc space here are some tools:
Useful commands:
- Get current dir's sub-dir usage breakdown sorted by highest usage first:
du -ahd1 | sort -rh
- Check that users don't consume too much of their personal
$WORK
space, which goes towards the total WORK space limit.
du -ahd1 $six_ALL_CCFRWORK/.. | sort -rh
Efficient tar-balling to STORE
When short on space you don't want to create large tarballs in the WORK dir, instead tar directly to the destination, e.g.
e.g. w/o gzip since we already have arrow binary files
mkdir -p $six_ALL_CCFRSTORE/datasets
cd $six_ALL_CCFRWORK/datasets
tar -cvf $six_ALL_CCFRSTORE/datasets/openwebtext.tar openwebtext
e.g. w/ gzip for non-binary data
tar -czvf $six_ALL_CCFRSTORE/datasets/openwebtext.tgz openwebtext
If the file is large and takes some resources to build, tar
will get killed, in such case you can't do it from the login instance and have to use one of the beefier instances. e.g.:
srun --pty --nodes=1 --ntasks=1 -A six@cpu --cpus-per-task=40 --hint=nomultithread --time=2:00:00 bash --rcfile $six_ALL_CCFRWORK/start-prod
tar ...
and if that's not enough do a slurm job