Update README.md
Browse files
README.md
CHANGED
|
@@ -1,8 +1,4 @@
|
|
| 1 |
-
This is my reproduction of the Microsoft team's work, WarriorCoder: Learning from Expert Battles to Augment Code Large Language Models. It is fully based on open-source models to construct training data and adopt supervised fine-tuning (SFT) to train the model.
|
| 2 |
-
Original paper link: https://arxiv.org/pdf/2412.17395
|
| 3 |
-
I have also published the training data constructed during my reproduction of the paper in another repository: https://huggingface.co/datasets/HuggingMicah/warrior_reproduce .
|
| 4 |
-
|
| 5 |
-
Also, I reproduced the experimental results in the paper. There are some differences from the original results, and I have marked the distinctions with underlines.
|
| 6 |
| Models | Matplotlib (155) | NumPy (220) | Pandas (291) | PyTorch (68) | SciPy (106) | Sklearn (115) | TensorFlow (45) | Overall (1000) |
|
| 7 |
| --- | --- | --- | --- | --- | --- | --- | --- | --- |
|
| 8 |
| INCODER (6.7B) | 28.3 | 4.4 | 3.1 | 4.4 | 2.8 | 2.8 | 3.8 | 7.4 |
|
|
@@ -23,4 +19,4 @@ Also, I reproduced the experimental results in the paper. There are some differe
|
|
| 23 |
| WizardCoder-SC (15B) | 51.4 | 45.3 | 61.6 | 50.7 |
|
| 24 |
| Magicoder-CL (6.7B) | 60.4 | 55.7 | 64.2 | 52.5 |
|
| 25 |
| MagicoderS-CL (6.7B) | 70.7 | 66.4 | 68.3 | 56.4 |
|
| 26 |
-
| WarriorCoder (6.7B) |
|
|
|
|
| 1 |
+
This is my reproduction of the Microsoft team's work, WarriorCoder: Learning from Expert Battles to Augment Code Large Language Models. It is fully based on open-source models to construct training data and adopt supervised fine-tuning (SFT) to train the model. Also, I reproduced the experimental results in the paper. These results are excellent, confirming that the idea of 'learning from expert battles' proposed in the paper has great potential. I have also published the training data constructed during my reproduction of the paper in another repository, and everyone is welcome to use it. Original paper link: https://arxiv.org/pdf/2412.17395 I have also published the training data constructed during my reproduction of the paper in another repository: https://huggingface.co/datasets/HuggingMicah/warrior_reproduce .
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
| Models | Matplotlib (155) | NumPy (220) | Pandas (291) | PyTorch (68) | SciPy (106) | Sklearn (115) | TensorFlow (45) | Overall (1000) |
|
| 3 |
| --- | --- | --- | --- | --- | --- | --- | --- | --- |
|
| 4 |
| INCODER (6.7B) | 28.3 | 4.4 | 3.1 | 4.4 | 2.8 | 2.8 | 3.8 | 7.4 |
|
|
|
|
| 19 |
| WizardCoder-SC (15B) | 51.4 | 45.3 | 61.6 | 50.7 |
|
| 20 |
| Magicoder-CL (6.7B) | 60.4 | 55.7 | 64.2 | 52.5 |
|
| 21 |
| MagicoderS-CL (6.7B) | 70.7 | 66.4 | 68.3 | 56.4 |
|
| 22 |
+
| WarriorCoder (6.7B) | 79.9 | 75.4 | 75.8 | 64.5 |
|