File size: 3,124 Bytes
2024260
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
# tr10 13B ML


## setup/tune up


To interactively tune up the setup:

```
salloc --constraint=v100-32g --account=six@v100 --nodes=4 --ntasks=4 --cpus-per-task=40 --gres=gpu:4 --hint=nomultithread --time=120 bash --rcfile $six_ALL_CCFRWORK/code/tr10-13B/bigscience/train/tr10-13B-ml/start-tr10-13B
```


Conda setup:

```
export CONDA_ENVS_PATH=$six_ALL_CCFRWORK/conda

conda create -y -n tr10-13B python=3.8
conda activate tr10-13B

pip3 install torch==1.10.0+cu113 torchvision==0.11.1+cu113 torchaudio==0.10.0+cu113 -f https://download.pytorch.org/whl/cu113/torch_stable.html

mkdir $six_ALL_CCFRWORK/code/tr10-13B
cd $six_ALL_CCFRWORK/code/tr10-13B

cd $six_ALL_CCFRWORK/code/tr10-13B/apex
./build.sh

pip install deepspeed

cd $six_ALL_CCFRWORK/code/tr10-13B/DeepSpeed
./build.sh

pip install transformers

cd $six_ALL_CCFRWORK/code/tr10-13B/transformers
pip install -e .

cd $six_ALL_CCFRWORK/code/tr10-13B/megatron-lm
pip install -r requirements.txt
```

Env setup script to be `source start-tr10-13B` [start-tr10-13B](./start-tr10-13B)



configs:

works:
```
NNODES=4
TP_SIZE=4
PP_SIZE=4
```


tokenizer

It's at https://huggingface.co/teven/test_150k_vocab_tokenizer/tree/main !

So instead of running with :
```
--vocab-file $VOCAB_FILE \
--merge-file $MERGE_FILE \
```

You should run with:
```
--tokenizer-type PretrainedFromHF \
--tokenizer-name-or-path teven/test_150k_vocab_tokenizer \
```

Preprocessed a c4 10k samples, you can use it with:
```
DATA_PATH=$six_ALL_CCFRSCRATCH/datasets-custom/150k_vocab_size_test/c4_10k_samples_150k_vocab_size
```

## Config


Julien Launay:

(1) the main difference will be multilinguality, and the larger vocabulary.
(2) For PrefixLM, we are not sure yet, as for now prefix is underperforming the vanilla model + it has some quirks. Thomas is working on a potential fix. We will keep you updated, but I think you can start working without prefix.
(3) Embeddings. ALiBi is still underperforming all others. Maybe we could consider going with rotary? @Iz Beltagy what's your opinion on this? Rotary probably won't change significantly your benchmark, but will degrade performance by a few percents across the board.
we don’t have a conclusive answer yet but both shouldn’t affect model size. If any, they will make the model a tiny bit smaller
(4) Activation. We need to evaluate the GeGLU run. GeGLU would bring a significant change to the size of the MLPs, which would be significant for your benchmark.
it shouldn’t change the overall model size but will change the size of some of the FF layers so might change how TP works

### `--init-method-std`

`--init-method-std 0.00884`

We derived this from: `NHIDDEN=5120`

`0.00884 = sqrt(2/(5120*5))` (from the ScaleNorm paper https://arxiv.org/abs/1910.05895)

### `NHEADS`

NHEADS=40, why...

### `--embed-layernorm`

We want this because it solved the problems with the 104B training instabilities.

If we choose to train with it, we will need to add the additional code to for the HF GPT2 converted model.

### `--partition-activations`

can be used to shard activations across gpus to save more gpu memory