kaikaidai commited on
Commit
c514098
·
verified ·
1 Parent(s): ebdee62

Updated language

Browse files
Files changed (1) hide show
  1. common.py +3 -3
common.py CHANGED
@@ -12,7 +12,7 @@ BATTLE_RULES = """
12
  - Both AIs stay anonymous - if either reveals its identity, the duel is void
13
  - Choose the LLM judge that most aligns with your judgement
14
  - If both score the same - choose the critique that you prefer more!
15
- <br><br>
16
  """
17
 
18
  # CSS Styles
@@ -39,12 +39,12 @@ CSS_STYLES = """
39
  # Default Eval Prompt
40
  EVAL_DESCRIPTION = """
41
  ## 📝 Instructions
42
- **Precise evaluation criteria leads to more consistent and reliable judgments.** A good evaluation prompt should include the following elements:
43
  - Evaluation criteria
44
  - Scoring rubric
45
  - (Optional) Examples\n
46
 
47
- **Any variables you define in your prompt using {{double curly braces}} will automatically map to the corresponding input fields under "Sample to evaluate" section on the right.**
48
 
49
  <br><br>
50
  """
 
12
  - Both AIs stay anonymous - if either reveals its identity, the duel is void
13
  - Choose the LLM judge that most aligns with your judgement
14
  - If both score the same - choose the critique that you prefer more!
15
+ <br>
16
  """
17
 
18
  # CSS Styles
 
39
  # Default Eval Prompt
40
  EVAL_DESCRIPTION = """
41
  ## 📝 Instructions
42
+ **Precise evaluation criteria lead to more consistent and reliable judgments.** A good Evaluator Prompt should include the following elements:
43
  - Evaluation criteria
44
  - Scoring rubric
45
  - (Optional) Examples\n
46
 
47
+ **Any variables you define in your prompt using {{double curly braces}} will automatically map to the corresponding input fields under the "Sample to evaluate" section on the right.**
48
 
49
  <br><br>
50
  """