File size: 918 Bytes
370453e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 |
# Finetuning 176B
Finetuning 176B at the end of the training might be necessary to ensure exact logits match between Megatron-DS trained model and HF model.
For now, there are 2 main bottlenecks that are responsible of not giving 100% logits match between HF model and Megatron model
## Diverging bottlenecks
### TP merging strategy
See [this issue](https://github.com/pytorch/pytorch/issues/76232). When merging TP ranks the logits exactness is lost. The idea would be to finetune the 176B model with TP=1
### Use `torch_softmax` instead of `fused_softmax`
`fused_softmax` and `torch_softmax` does not give the same results (ie, `torch.testing.assert_allclose(atol=0.0, rtol=0.0)` does not pass). The main model could be finetuned with `torch_softmax`.
See [this line](https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/cb48bd2c8bc182fb9872f127ef7c2267fbf9cc2e/megatron/model/fused_softmax.py#L204)
|