Update README.md
Browse files
README.md
CHANGED
@@ -31,6 +31,20 @@ You can also visit the [homepage](https://code-reward-model.github.io/) and the
|
|
31 |
|
32 |
The model is trained based on [Llama3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct).
|
33 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
34 |
# Performance
|
35 |
|
36 |
## Best-of-N
|
@@ -86,20 +100,6 @@ utilizing Llama3.1-8B as the policy model. The top two performances are marked i
|
|
86 |
| Llama3.1-70B | <u>78.30</u> | <u>78.76</u> | <u>17.19</u> | <u>25.97</u> |
|
87 |
| *CodeRM-8B (Ours)* | **80.46** | **81.27** | **16.48** | **22.71** |
|
88 |
|
89 |
-
# Prompt Format
|
90 |
-
|
91 |
-
```
|
92 |
-
Below is a question and it's corresponding code answer. Please write test cases to check the correctness of the code answer. You need to use the unittest library in Python and create a test class for testing.
|
93 |
-
|
94 |
-
### question
|
95 |
-
{question}
|
96 |
-
|
97 |
-
### code solution
|
98 |
-
{code in function format}
|
99 |
-
|
100 |
-
Please add detailed comments to the test cases you write. You do not need to test the function's ability to throw exceptions.
|
101 |
-
```
|
102 |
-
|
103 |
# Citation
|
104 |
|
105 |
If you find our model helpful, please cite the original paper:
|
|
|
31 |
|
32 |
The model is trained based on [Llama3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct).
|
33 |
|
34 |
+
# Prompt Format
|
35 |
+
|
36 |
+
```
|
37 |
+
Below is a question and it's corresponding code answer. Please write test cases to check the correctness of the code answer. You need to use the unittest library in Python and create a test class for testing.
|
38 |
+
|
39 |
+
### question
|
40 |
+
{question}
|
41 |
+
|
42 |
+
### code solution
|
43 |
+
{code in function format}
|
44 |
+
|
45 |
+
Please add detailed comments to the test cases you write. You do not need to test the function's ability to throw exceptions.
|
46 |
+
```
|
47 |
+
|
48 |
# Performance
|
49 |
|
50 |
## Best-of-N
|
|
|
100 |
| Llama3.1-70B | <u>78.30</u> | <u>78.76</u> | <u>17.19</u> | <u>25.97</u> |
|
101 |
| *CodeRM-8B (Ours)* | **80.46** | **81.27** | **16.48** | **22.71** |
|
102 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
103 |
# Citation
|
104 |
|
105 |
If you find our model helpful, please cite the original paper:
|