BubbleQ commited on
Commit
c12901a
·
verified ·
1 Parent(s): 3e4e7aa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -0
README.md CHANGED
@@ -127,6 +127,7 @@ Note:
127
  | | MT-Eval | 8.13 | 7.36 | 6.7 | 8.18 | 8.45 | 8.12 | - |
128
  | | AlignBench v1.1 | 7 | 6.13 | 5.99 | 6.95 | 6.3 | 6.33 | 7.06 |
129
  | | Average | 53.74 | - | 46.54 | 52.61 | 50.54 | 48.95 | - |
 
130
  Note:
131
  1. For InternLM3-8B-Instruct, the results marked with `*` are sourced from their official website, other evaluations are conducted based on internal evaluation frameworks.
132
  2. For Multi-IF, we report the overall average computed across all three rounds, pooling the Chinese and English metrics.
 
127
  | | MT-Eval | 8.13 | 7.36 | 6.7 | 8.18 | 8.45 | 8.12 | - |
128
  | | AlignBench v1.1 | 7 | 6.13 | 5.99 | 6.95 | 6.3 | 6.33 | 7.06 |
129
  | | Average | 53.74 | - | 46.54 | 52.61 | 50.54 | 48.95 | - |
130
+
131
  Note:
132
  1. For InternLM3-8B-Instruct, the results marked with `*` are sourced from their official website, other evaluations are conducted based on internal evaluation frameworks.
133
  2. For Multi-IF, we report the overall average computed across all three rounds, pooling the Chinese and English metrics.