applied-ai-018's picture
Add files using upload-large-folder tool
f9bc5ed verified

Slurm scripts

Mainly here as indicative. Adapt to current traning.

  • cpu.slurm -> for data preprocessing
  • gpu.slurm -> arguments are adapted to maximize the gpu mem of the 8 32GB GPU requested

We are using common disk spaces for datasets, caches and experiment dumps:

  • Experiment dumps -> $six_ALL_CCFRWORK/experiments

SCRATCH disk spaces are wiped regularly (wiping every file that was not accessed in the past 30 days) so we have S3 buckets (https://console.cloud.google.com/storage/browser/bigscience-experiments and https://console.cloud.google.com/storage/browser/bigscience-datasets) as shared storage that is accessible from JZ but from others instances too.