|
# Experiments and Results |
|
|
|
|
|
- [GPT2 Experiments](./gpt2.md) - main working doc with all the experiment results |
|
|
|
|
|
|
|
|
|
The following is mostly outdated and should be recycled/updated: |
|
|
|
|
|
Model |
|
|
|
* Basically GPT, but with slightly different attention/loss mask (see agenda above) |
|
* Targeting 175-200B parameters, seq length 512 or 1024, adam optimizer |
|
* Note: our nodes have 128GB total gpu memory each -> need sharding (model weights alone are 350GB in fp16!) |
|
* Note 2: may need to adjust expectations based on estimated training time (pending) |
|
* Subsequent runs: consider encoder-decoder arch? longer sequences? adafactor? |
|
|
|
Making it train: |
|
|
|
* Candidates for run1: deepspeed-gpt and megatron-gpt. Save fairscale/t5/custom for later. Use reference training code instead of integrating. |
|
* Only consider training nodes with 4x V100-32GB because there are too few 8x nodes in JZ |
|
* Current strategy: fully sharded data-parallel between nodes, arbitrary black magic within one node. |
|
|
|
For each framework, we need to measure (hackathon participants + Sidd + Laurel) |
|
* maximum number of parameters that we can train on 1 node |
|
* minimum number of nodes to fit 175B-parameter model (num_layers/dmodel/dhead: from gpt3 table 2.1) |
|
* training throughput (i.e. training tokens per second in both configurations) |
|
* Once we get through the initial debugging, ask microsoft/nvidia teams to sanity-check our configs for DS/Megatron respectively |
|
|
|
|
|
Investigate: |
|
* @Stas measured allreduce bandwidth between two nodes as ~135Gbps, now he's is investigating if we can get closer to the theoretical max = 400Gbps |
|
* Estimated training time for 175B parameters - based on training throughput in megatron/deepspeed tech reports; estimated time to "reasonable performance". (TODO @Yozh and @Max Ryabinin). Sanity-check whether the 8-node servers are indeed too few to train in reasonable time. |
|
* Figure out the how to introduce semi-autoregressive attention mask (and loss mask) to the code of deepspeed and megatron-LM training configs (looking for volunteers!) |
|
|
|
Next steps: |
|
* After Stas figures out the bandwidth, we need to measure how performance scales with #nodes using data-parallel training. |
|
* Ask/measure how many server nodes can we reliably provision with JZ. Find the optimal trade-off between #nodes vs training time. |
|
* a quick preliminary run with the same configuration, but fewer parameters ... and we can unleash the beast! |
|
|