KAKA22 commited on
Commit
dd9b7c5
·
verified ·
1 Parent(s): 9dc8592

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -14
README.md CHANGED
@@ -31,6 +31,20 @@ You can also visit the [homepage](https://code-reward-model.github.io/) and the
31
 
32
  The model is trained based on [Llama3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct).
33
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
  # Performance
35
 
36
  ## Best-of-N
@@ -86,20 +100,6 @@ utilizing Llama3.1-8B as the policy model. The top two performances are marked i
86
  | Llama3.1-70B | <u>78.30</u> | <u>78.76</u> | <u>17.19</u> | <u>25.97</u> |
87
  | *CodeRM-8B (Ours)* | **80.46** | **81.27** | **16.48** | **22.71** |
88
 
89
- # Prompt Format
90
-
91
- ```
92
- Below is a question and it's corresponding code answer. Please write test cases to check the correctness of the code answer. You need to use the unittest library in Python and create a test class for testing.
93
-
94
- ### question
95
- {question}
96
-
97
- ### code solution
98
- {code in function format}
99
-
100
- Please add detailed comments to the test cases you write. You do not need to test the function's ability to throw exceptions.
101
- ```
102
-
103
  # Citation
104
 
105
  If you find our model helpful, please cite the original paper:
 
31
 
32
  The model is trained based on [Llama3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct).
33
 
34
+ # Prompt Format
35
+
36
+ ```
37
+ Below is a question and it's corresponding code answer. Please write test cases to check the correctness of the code answer. You need to use the unittest library in Python and create a test class for testing.
38
+
39
+ ### question
40
+ {question}
41
+
42
+ ### code solution
43
+ {code in function format}
44
+
45
+ Please add detailed comments to the test cases you write. You do not need to test the function's ability to throw exceptions.
46
+ ```
47
+
48
  # Performance
49
 
50
  ## Best-of-N
 
100
  | Llama3.1-70B | <u>78.30</u> | <u>78.76</u> | <u>17.19</u> | <u>25.97</u> |
101
  | *CodeRM-8B (Ours)* | **80.46** | **81.27** | **16.48** | **22.71** |
102
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
103
  # Citation
104
 
105
  If you find our model helpful, please cite the original paper: