Update README.md (#11)
Browse files- Update README.md (e5103930e082d6cb05c400eeaff765a84e2eddd8)
Co-authored-by: haipengluo <[email protected]>
README.md
CHANGED
|
@@ -6,9 +6,8 @@ license: llama2
|
|
| 6 |
## WizardMath: Empowering Mathematical Reasoning for Large Language Models via Reinforced Evol-Instruct (RLEIF)
|
| 7 |
|
| 8 |
|
| 9 |
-
|
| 10 |
<p align="center">
|
| 11 |
-
π€ <a href="https://huggingface.co/WizardLM" target="_blank">HF Repo</a>
|
| 12 |
</p>
|
| 13 |
<p align="center">
|
| 14 |
π Join our <a href="https://discord.gg/VZjjHtWrKs" target="_blank">Discord</a>
|
|
@@ -24,13 +23,13 @@ license: llama2
|
|
| 24 |
| WizardCoder-1B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardCoder-1B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2306.08568" target="_blank">[WizardCoder]</a> | 23.8 |28.6 | -- | <a href="https://huggingface.co/spaces/bigcode/bigcode-model-license-agreement" target="_blank">OpenRAIL-M</a> |
|
| 25 |
|
| 26 |
|
| 27 |
-
|
| 28 |
| Model | Checkpoint | Paper | GSM8k | MATH |Online Demo| License|
|
| 29 |
| ----- |------| ---- |------|-------| ----- | ----- |
|
| 30 |
| WizardMath-70B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardMath-70B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2308.09583" target="_blank">[WizardMath]</a>| **81.6** | **22.7** |[Demo](http://47.103.63.15:50083/)| <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 </a> |
|
| 31 |
| WizardMath-13B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardMath-13B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2308.09583" target="_blank">[WizardMath]</a>| **63.9** | **14.0** |[Demo](http://47.103.63.15:50082/)| <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 </a> |
|
| 32 |
| WizardMath-7B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardMath-7B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2308.09583" target="_blank">[WizardMath]</a>| **54.9** | **10.7** | [Demo](http://47.103.63.15:50080/)| <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 </a>|
|
| 33 |
|
|
|
|
| 34 |
|
| 35 |
<font size=4>
|
| 36 |
|
|
@@ -50,7 +49,13 @@ license: llama2
|
|
| 50 |
|
| 51 |
**Discord**: https://discord.gg/VZjjHtWrKs
|
| 52 |
|
|
|
|
|
|
|
|
|
|
| 53 |
|
|
|
|
|
|
|
|
|
|
| 54 |
|
| 55 |
β<b>Note for model system prompts usage:</b>
|
| 56 |
|
|
@@ -77,3 +82,21 @@ Recently, there have been clear changes in the open-source policy and regulation
|
|
| 77 |
Despite this, we have still worked hard to obtain opening the weights of the model first, but the data involves stricter auditing and is in review with our legal team .
|
| 78 |
Our researchers have no authority to publicly release them without authorization.
|
| 79 |
Thank you for your understanding.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
## WizardMath: Empowering Mathematical Reasoning for Large Language Models via Reinforced Evol-Instruct (RLEIF)
|
| 7 |
|
| 8 |
|
|
|
|
| 9 |
<p align="center">
|
| 10 |
+
π€ <a href="https://huggingface.co/WizardLM" target="_blank">HF Repo</a> β’π± <a href="https://github.com/nlpxucan/WizardLM" target="_blank">Github Repo</a> β’ π¦ <a href="https://twitter.com/WizardLM_AI" target="_blank">Twitter</a> β’ π <a href="https://arxiv.org/abs/2304.12244" target="_blank">[WizardLM]</a> β’ π <a href="https://arxiv.org/abs/2306.08568" target="_blank">[WizardCoder]</a> β’ π <a href="https://arxiv.org/abs/2308.09583" target="_blank">[WizardMath]</a> <br>
|
| 11 |
</p>
|
| 12 |
<p align="center">
|
| 13 |
π Join our <a href="https://discord.gg/VZjjHtWrKs" target="_blank">Discord</a>
|
|
|
|
| 23 |
| WizardCoder-1B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardCoder-1B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2306.08568" target="_blank">[WizardCoder]</a> | 23.8 |28.6 | -- | <a href="https://huggingface.co/spaces/bigcode/bigcode-model-license-agreement" target="_blank">OpenRAIL-M</a> |
|
| 24 |
|
| 25 |
|
|
|
|
| 26 |
| Model | Checkpoint | Paper | GSM8k | MATH |Online Demo| License|
|
| 27 |
| ----- |------| ---- |------|-------| ----- | ----- |
|
| 28 |
| WizardMath-70B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardMath-70B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2308.09583" target="_blank">[WizardMath]</a>| **81.6** | **22.7** |[Demo](http://47.103.63.15:50083/)| <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 </a> |
|
| 29 |
| WizardMath-13B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardMath-13B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2308.09583" target="_blank">[WizardMath]</a>| **63.9** | **14.0** |[Demo](http://47.103.63.15:50082/)| <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 </a> |
|
| 30 |
| WizardMath-7B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardMath-7B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2308.09583" target="_blank">[WizardMath]</a>| **54.9** | **10.7** | [Demo](http://47.103.63.15:50080/)| <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 </a>|
|
| 31 |
|
| 32 |
+
|
| 33 |
|
| 34 |
<font size=4>
|
| 35 |
|
|
|
|
| 49 |
|
| 50 |
**Discord**: https://discord.gg/VZjjHtWrKs
|
| 51 |
|
| 52 |
+
## Comparing WizardMath-V1.0 with Other LLMs.
|
| 53 |
+
|
| 54 |
+
π₯ The following figure shows that our **WizardMath-70B-V1.0 attains the fifth position in this benchmark**, surpassing ChatGPT (81.6 vs. 80.8) , Claude Instant (81.6 vs. 80.9), PaLM 2 540B (81.6 vs. 80.7).
|
| 55 |
|
| 56 |
+
<p align="center" width="100%">
|
| 57 |
+
<a ><img src="https://raw.githubusercontent.com/nlpxucan/WizardLM/main/WizardMath/images/wizardmath_gsm8k.png" alt="WizardMath" style="width: 96%; min-width: 300px; display: block; margin: auto;"></a>
|
| 58 |
+
</p>
|
| 59 |
|
| 60 |
β<b>Note for model system prompts usage:</b>
|
| 61 |
|
|
|
|
| 82 |
Despite this, we have still worked hard to obtain opening the weights of the model first, but the data involves stricter auditing and is in review with our legal team .
|
| 83 |
Our researchers have no authority to publicly release them without authorization.
|
| 84 |
Thank you for your understanding.
|
| 85 |
+
|
| 86 |
+
|
| 87 |
+
## Inference Demo Script
|
| 88 |
+
|
| 89 |
+
We provide the inference demo code [here](https://github.com/nlpxucan/WizardLM/tree/main/demo).
|
| 90 |
+
|
| 91 |
+
## Citation
|
| 92 |
+
|
| 93 |
+
Please cite the repo if you use the data, method or code in this repo.
|
| 94 |
+
|
| 95 |
+
```
|
| 96 |
+
@article{luo2023wizardmath,
|
| 97 |
+
title={WizardMath: Empowering Mathematical Reasoning for Large Language Models via Reinforced Evol-Instruct},
|
| 98 |
+
author={Luo, Haipeng and Sun, Qingfeng and Xu, Can and Zhao, Pu and Lou, Jianguang and Tao, Chongyang and Geng, Xiubo and Lin, Qingwei and Chen, Shifeng and Zhang, Dongmei},
|
| 99 |
+
journal={arXiv preprint arXiv:2308.09583},
|
| 100 |
+
year={2023}
|
| 101 |
+
}
|
| 102 |
+
```
|