Update README.md
Browse files
README.md
CHANGED
@@ -28,6 +28,9 @@ license: llama2
|
|
28 |
|Xwin-Math-70B-V1.0| 87.0 | 31.8 | π€ <a href="https://huggingface.co/Xwin-LM/Xwin-Math-70B-V1.0" target="_blank">HF Link</a> | <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 License|
|
29 |
|Xwin-Math-70B-V1.1| 90.6 | 51.9 | π€ <a href="https://huggingface.co/Xwin-LM/Xwin-Math-70B-V1.1" target="_blank">HF Link</a> | <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 License|
|
30 |
|
|
|
|
|
|
|
31 |
## π Benchmarks
|
32 |
### Xwin-Math performance on [MATH](https://github.com/hendrycks/math) and [GSM8K](https://github.com/openai/grade-school-math).
|
33 |
|
@@ -124,7 +127,7 @@ To generate the model's responses, you can use the `generate.py` script. Please
|
|
124 |
|
125 |
For the generation process, we use the Vicuna-v1.1 system prompt with chain-of-thought and format instruction. We also employ a greedy decoding strategy and set the maximum sequence length to 2048.
|
126 |
```
|
127 |
-
"A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: {instruction} Give your solution in detail. In the end, write your final answer in the format of 'The answer is: <ANSWER>.'. ASSISTANT:
|
128 |
```
|
129 |
|
130 |
Here is an simple example to generate using [vLLM](https://docs.vllm.ai/en/latest/).
|
|
|
28 |
|Xwin-Math-70B-V1.0| 87.0 | 31.8 | π€ <a href="https://huggingface.co/Xwin-LM/Xwin-Math-70B-V1.0" target="_blank">HF Link</a> | <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 License|
|
29 |
|Xwin-Math-70B-V1.1| 90.6 | 51.9 | π€ <a href="https://huggingface.co/Xwin-LM/Xwin-Math-70B-V1.1" target="_blank">HF Link</a> | <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 License|
|
30 |
|
31 |
+
* Xwin-Math-7B-V1.1 uses 1.92M GSM8K and 960K MATH synthetic data.
|
32 |
+
* Xwin-Math-70B-V1.1 uses 960K GSM8K and 480K MATH synthetic data.
|
33 |
+
|
34 |
## π Benchmarks
|
35 |
### Xwin-Math performance on [MATH](https://github.com/hendrycks/math) and [GSM8K](https://github.com/openai/grade-school-math).
|
36 |
|
|
|
127 |
|
128 |
For the generation process, we use the Vicuna-v1.1 system prompt with chain-of-thought and format instruction. We also employ a greedy decoding strategy and set the maximum sequence length to 2048.
|
129 |
```
|
130 |
+
"A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: {instruction} Give your solution in detail. In the end, write your final answer in the format of 'The answer is: <ANSWER>.'. ASSISTANT:"
|
131 |
```
|
132 |
|
133 |
Here is an simple example to generate using [vLLM](https://docs.vllm.ai/en/latest/).
|