haipeng1 commited on
Commit
93cdade
1 Parent(s): c24fd5c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -0
README.md CHANGED
@@ -49,7 +49,13 @@ license: llama2
49
 
50
  **Discord**: https://discord.gg/VZjjHtWrKs
51
 
 
52
 
 
 
 
 
 
53
 
54
  ❗<b>Note for model system prompts usage:</b>
55
 
@@ -76,3 +82,21 @@ Recently, there have been clear changes in the open-source policy and regulation
76
  Despite this, we have still worked hard to obtain opening the weights of the model first, but the data involves stricter auditing and is in review with our legal team .
77
  Our researchers have no authority to publicly release them without authorization.
78
  Thank you for your understanding.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
49
 
50
  **Discord**: https://discord.gg/VZjjHtWrKs
51
 
52
+ ## Comparing WizardMath-V1.0 with Other LLMs.
53
 
54
+ 🔥 The following figure shows that our **WizardMath-70B-V1.0 attains the fifth position in this benchmark**, surpassing ChatGPT (81.6 vs. 80.8) , Claude Instant (81.6 vs. 80.9), PaLM 2 540B (81.6 vs. 80.7).
55
+
56
+ <p align="center" width="100%">
57
+ <a ><img src="https://raw.githubusercontent.com/nlpxucan/WizardLM/main/WizardMath/images/wizardmath_gsm8k.png" alt="WizardMath" style="width: 96%; min-width: 300px; display: block; margin: auto;"></a>
58
+ </p>
59
 
60
  ❗<b>Note for model system prompts usage:</b>
61
 
 
82
  Despite this, we have still worked hard to obtain opening the weights of the model first, but the data involves stricter auditing and is in review with our legal team .
83
  Our researchers have no authority to publicly release them without authorization.
84
  Thank you for your understanding.
85
+
86
+
87
+ ## Inference Demo Script
88
+
89
+ We provide the inference demo code [here](https://github.com/nlpxucan/WizardLM/tree/main/demo).
90
+
91
+ ## Citation
92
+
93
+ Please cite the repo if you use the data, method or code in this repo.
94
+
95
+ ```
96
+ @article{luo2023wizardmath,
97
+ title={WizardMath: Empowering Mathematical Reasoning for Large Language Models via Reinforced Evol-Instruct},
98
+ author={Luo, Haipeng and Sun, Qingfeng and Xu, Can and Zhao, Pu and Lou, Jianguang and Tao, Chongyang and Geng, Xiubo and Lin, Qingwei and Chen, Shifeng and Zhang, Dongmei},
99
+ journal={arXiv preprint arXiv:2308.09583},
100
+ year={2023}
101
+ }
102
+ ```