Spaces:
Running
Running
Commit ·
8eecd10
0
Parent(s):
Deploy Chinese Writing Bench leaderboard with dual judge support
Browse files- README.md +8 -0
- data/all-scores.json +236 -0
- data/complicated-writing-scores.json +236 -0
- data/gpt5.4-judge-all-scores.json +236 -0
- data/gpt5.4-judge-complicated-writing-scores.json +236 -0
- index.html +101 -0
- script.js +206 -0
- styles.css +458 -0
README.md
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
title: Chinese Writing Bench
|
| 3 |
+
emoji: 🔬
|
| 4 |
+
colorFrom: blue
|
| 5 |
+
colorTo: purple
|
| 6 |
+
sdk: static
|
| 7 |
+
pinned: false
|
| 8 |
+
---
|
data/all-scores.json
ADDED
|
@@ -0,0 +1,236 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"Mistral-3.1-24B-2503": {
|
| 3 |
+
"comprehension_score": 4.707142857142857,
|
| 4 |
+
"structure_score": 4.796428571428572,
|
| 5 |
+
"prose_style_score": 4.2178571428571425,
|
| 6 |
+
"creativity_score": 3.914285714285714,
|
| 7 |
+
"depth_score": 3.9035714285714285,
|
| 8 |
+
"helpfulness_score": 4.242857142857143,
|
| 9 |
+
"overall_score": 4.203571428571428
|
| 10 |
+
},
|
| 11 |
+
"Phi-4-14B": {
|
| 12 |
+
"comprehension_score": 4.735714285714286,
|
| 13 |
+
"structure_score": 4.889285714285714,
|
| 14 |
+
"prose_style_score": 4.460714285714285,
|
| 15 |
+
"creativity_score": 4.242857142857143,
|
| 16 |
+
"depth_score": 4.178571428571429,
|
| 17 |
+
"helpfulness_score": 4.317857142857143,
|
| 18 |
+
"overall_score": 4.296428571428572
|
| 19 |
+
},
|
| 20 |
+
"gpt-4o-mini-2024-07-18": {
|
| 21 |
+
"comprehension_score": 4.760714285714286,
|
| 22 |
+
"structure_score": 4.9,
|
| 23 |
+
"prose_style_score": 4.542857142857143,
|
| 24 |
+
"creativity_score": 4.289285714285715,
|
| 25 |
+
"depth_score": 4.082142857142857,
|
| 26 |
+
"helpfulness_score": 4.35,
|
| 27 |
+
"overall_score": 4.321428571428571
|
| 28 |
+
},
|
| 29 |
+
"gpt-4.1-mini-2025-04-14": {
|
| 30 |
+
"comprehension_score": 5.042857142857143,
|
| 31 |
+
"structure_score": 4.935714285714286,
|
| 32 |
+
"prose_style_score": 4.675,
|
| 33 |
+
"creativity_score": 4.335714285714285,
|
| 34 |
+
"depth_score": 4.2,
|
| 35 |
+
"helpfulness_score": 4.564285714285714,
|
| 36 |
+
"overall_score": 4.435714285714286
|
| 37 |
+
},
|
| 38 |
+
"Qwen3-30B-A3B": {
|
| 39 |
+
"comprehension_score": 5.146428571428571,
|
| 40 |
+
"structure_score": 5.289285714285715,
|
| 41 |
+
"prose_style_score": 4.853571428571429,
|
| 42 |
+
"creativity_score": 4.714285714285714,
|
| 43 |
+
"depth_score": 4.5321428571428575,
|
| 44 |
+
"helpfulness_score": 4.925,
|
| 45 |
+
"overall_score": 4.814285714285714
|
| 46 |
+
},
|
| 47 |
+
"Qwen3-8B": {
|
| 48 |
+
"comprehension_score": 5.146428571428571,
|
| 49 |
+
"structure_score": 5.260714285714286,
|
| 50 |
+
"prose_style_score": 4.910714285714286,
|
| 51 |
+
"creativity_score": 4.689285714285714,
|
| 52 |
+
"depth_score": 4.496428571428571,
|
| 53 |
+
"helpfulness_score": 4.8464285714285715,
|
| 54 |
+
"overall_score": 4.817857142857143
|
| 55 |
+
},
|
| 56 |
+
"gpt-4o-2024-11-20": {
|
| 57 |
+
"comprehension_score": 5.2785714285714285,
|
| 58 |
+
"structure_score": 5.3464285714285715,
|
| 59 |
+
"prose_style_score": 5.207142857142857,
|
| 60 |
+
"creativity_score": 5.082142857142857,
|
| 61 |
+
"depth_score": 4.746428571428571,
|
| 62 |
+
"helpfulness_score": 5.010714285714286,
|
| 63 |
+
"overall_score": 4.985714285714286
|
| 64 |
+
},
|
| 65 |
+
"gemma3-27b": {
|
| 66 |
+
"comprehension_score": 5.428571428571429,
|
| 67 |
+
"structure_score": 5.535714285714286,
|
| 68 |
+
"prose_style_score": 5.0285714285714285,
|
| 69 |
+
"creativity_score": 5.121428571428571,
|
| 70 |
+
"depth_score": 4.8464285714285715,
|
| 71 |
+
"helpfulness_score": 5.2178571428571425,
|
| 72 |
+
"overall_score": 5.082142857142857
|
| 73 |
+
},
|
| 74 |
+
"Qwen3-32B": {
|
| 75 |
+
"comprehension_score": 5.457142857142857,
|
| 76 |
+
"structure_score": 5.628571428571429,
|
| 77 |
+
"prose_style_score": 5.285714285714286,
|
| 78 |
+
"creativity_score": 5.078571428571428,
|
| 79 |
+
"depth_score": 4.882142857142857,
|
| 80 |
+
"helpfulness_score": 5.314285714285714,
|
| 81 |
+
"overall_score": 5.2214285714285715
|
| 82 |
+
},
|
| 83 |
+
"Mistral-3.2-24B-2506": {
|
| 84 |
+
"comprehension_score": 5.5321428571428575,
|
| 85 |
+
"structure_score": 5.817857142857143,
|
| 86 |
+
"prose_style_score": 5.5321428571428575,
|
| 87 |
+
"creativity_score": 5.607142857142857,
|
| 88 |
+
"depth_score": 5.010714285714286,
|
| 89 |
+
"helpfulness_score": 5.3464285714285715,
|
| 90 |
+
"overall_score": 5.3464285714285715
|
| 91 |
+
},
|
| 92 |
+
"o4-mini-2025-04-16": {
|
| 93 |
+
"comprehension_score": 5.803571428571429,
|
| 94 |
+
"structure_score": 5.696428571428571,
|
| 95 |
+
"prose_style_score": 5.589285714285714,
|
| 96 |
+
"creativity_score": 5.442857142857143,
|
| 97 |
+
"depth_score": 5.1,
|
| 98 |
+
"helpfulness_score": 5.582142857142857,
|
| 99 |
+
"overall_score": 5.542857142857143
|
| 100 |
+
},
|
| 101 |
+
"Qwen3-235B": {
|
| 102 |
+
"comprehension_score": 5.735714285714286,
|
| 103 |
+
"structure_score": 5.875,
|
| 104 |
+
"prose_style_score": 5.628571428571429,
|
| 105 |
+
"creativity_score": 5.514285714285714,
|
| 106 |
+
"depth_score": 5.214285714285714,
|
| 107 |
+
"helpfulness_score": 5.55,
|
| 108 |
+
"overall_score": 5.5928571428571425
|
| 109 |
+
},
|
| 110 |
+
"Gemini-2.5-Flash": {
|
| 111 |
+
"comprehension_score": 6.196428571428571,
|
| 112 |
+
"structure_score": 6.2214285714285715,
|
| 113 |
+
"prose_style_score": 6.0,
|
| 114 |
+
"creativity_score": 5.828571428571428,
|
| 115 |
+
"depth_score": 5.75,
|
| 116 |
+
"helpfulness_score": 6.0321428571428575,
|
| 117 |
+
"overall_score": 6.007142857142857
|
| 118 |
+
},
|
| 119 |
+
"Qwen3-30B-A3B-Thinking": {
|
| 120 |
+
"comprehension_score": 5.914285714285715,
|
| 121 |
+
"structure_score": 6.053571428571429,
|
| 122 |
+
"prose_style_score": 6.5321428571428575,
|
| 123 |
+
"creativity_score": 6.5285714285714285,
|
| 124 |
+
"depth_score": 5.65,
|
| 125 |
+
"helpfulness_score": 5.817857142857143,
|
| 126 |
+
"overall_score": 6.064285714285714
|
| 127 |
+
},
|
| 128 |
+
"Deepseek-V3-0324": {
|
| 129 |
+
"comprehension_score": 6.082142857142857,
|
| 130 |
+
"structure_score": 6.242857142857143,
|
| 131 |
+
"prose_style_score": 6.560714285714286,
|
| 132 |
+
"creativity_score": 6.610714285714286,
|
| 133 |
+
"depth_score": 5.796428571428572,
|
| 134 |
+
"helpfulness_score": 5.957142857142857,
|
| 135 |
+
"overall_score": 6.203571428571428
|
| 136 |
+
},
|
| 137 |
+
"Qwen3-32B-Thinking": {
|
| 138 |
+
"comprehension_score": 6.339285714285714,
|
| 139 |
+
"structure_score": 6.425,
|
| 140 |
+
"prose_style_score": 7.089285714285714,
|
| 141 |
+
"creativity_score": 7.185714285714286,
|
| 142 |
+
"depth_score": 6.192857142857143,
|
| 143 |
+
"helpfulness_score": 6.182142857142857,
|
| 144 |
+
"overall_score": 6.578571428571428
|
| 145 |
+
},
|
| 146 |
+
"Qwen3-235B-Thinking": {
|
| 147 |
+
"comprehension_score": 6.385714285714286,
|
| 148 |
+
"structure_score": 6.45,
|
| 149 |
+
"prose_style_score": 7.296428571428572,
|
| 150 |
+
"creativity_score": 7.435714285714286,
|
| 151 |
+
"depth_score": 6.321428571428571,
|
| 152 |
+
"helpfulness_score": 6.285714285714286,
|
| 153 |
+
"overall_score": 6.660714285714286
|
| 154 |
+
},
|
| 155 |
+
"Deepseek-R1-0528": {
|
| 156 |
+
"comprehension_score": 6.7178571428571425,
|
| 157 |
+
"structure_score": 6.703571428571428,
|
| 158 |
+
"prose_style_score": 7.0964285714285715,
|
| 159 |
+
"creativity_score": 7.021428571428571,
|
| 160 |
+
"depth_score": 6.5964285714285715,
|
| 161 |
+
"helpfulness_score": 6.575,
|
| 162 |
+
"overall_score": 6.771428571428571
|
| 163 |
+
},
|
| 164 |
+
"o3-2025-04-16": {
|
| 165 |
+
"comprehension_score": 6.914285714285715,
|
| 166 |
+
"structure_score": 6.746428571428571,
|
| 167 |
+
"prose_style_score": 7.05,
|
| 168 |
+
"creativity_score": 6.896428571428571,
|
| 169 |
+
"depth_score": 6.439285714285714,
|
| 170 |
+
"helpfulness_score": 6.742857142857143,
|
| 171 |
+
"overall_score": 6.946428571428571
|
| 172 |
+
},
|
| 173 |
+
"gpt-5.4": {
|
| 174 |
+
"comprehension_score": 7.035714285714286,
|
| 175 |
+
"structure_score": 6.975,
|
| 176 |
+
"prose_style_score": 7.128571428571429,
|
| 177 |
+
"creativity_score": 6.9035714285714285,
|
| 178 |
+
"depth_score": 6.760714285714286,
|
| 179 |
+
"helpfulness_score": 7.025,
|
| 180 |
+
"overall_score": 7.060714285714286
|
| 181 |
+
},
|
| 182 |
+
"Gemini-3.1-Flash": {
|
| 183 |
+
"comprehension_score": 6.414285714285715,
|
| 184 |
+
"structure_score": 6.625,
|
| 185 |
+
"prose_style_score": 6.75,
|
| 186 |
+
"creativity_score": 6.660714285714286,
|
| 187 |
+
"depth_score": 6.057142857142857,
|
| 188 |
+
"helpfulness_score": 6.414285714285715,
|
| 189 |
+
"overall_score": 6.5285714285714285
|
| 190 |
+
},
|
| 191 |
+
"Gemini-3.1-Pro": {
|
| 192 |
+
"comprehension_score": 6.767857142857143,
|
| 193 |
+
"structure_score": 6.867857142857143,
|
| 194 |
+
"prose_style_score": 7.114285714285714,
|
| 195 |
+
"creativity_score": 6.875,
|
| 196 |
+
"depth_score": 6.325,
|
| 197 |
+
"helpfulness_score": 6.757142857142857,
|
| 198 |
+
"overall_score": 6.864285714285714
|
| 199 |
+
},
|
| 200 |
+
"MiniMax-M2.5": {
|
| 201 |
+
"comprehension_score": 6.364285714285714,
|
| 202 |
+
"structure_score": 6.428571428571429,
|
| 203 |
+
"prose_style_score": 6.021428571428571,
|
| 204 |
+
"creativity_score": 6.075,
|
| 205 |
+
"depth_score": 5.853571428571429,
|
| 206 |
+
"helpfulness_score": 6.275,
|
| 207 |
+
"overall_score": 6.260714285714286
|
| 208 |
+
},
|
| 209 |
+
"Qwen3.5-Plus": {
|
| 210 |
+
"comprehension_score": 6.7821428571428575,
|
| 211 |
+
"structure_score": 6.75,
|
| 212 |
+
"prose_style_score": 6.925,
|
| 213 |
+
"creativity_score": 6.735714285714286,
|
| 214 |
+
"depth_score": 6.296428571428572,
|
| 215 |
+
"helpfulness_score": 6.714285714285714,
|
| 216 |
+
"overall_score": 6.796428571428572
|
| 217 |
+
},
|
| 218 |
+
"GLM-5": {
|
| 219 |
+
"comprehension_score": 6.410714285714286,
|
| 220 |
+
"structure_score": 6.5285714285714285,
|
| 221 |
+
"prose_style_score": 6.689285714285714,
|
| 222 |
+
"creativity_score": 6.507142857142857,
|
| 223 |
+
"depth_score": 5.996428571428571,
|
| 224 |
+
"helpfulness_score": 6.360714285714286,
|
| 225 |
+
"overall_score": 6.492857142857143
|
| 226 |
+
},
|
| 227 |
+
"step-3.5-flash": {
|
| 228 |
+
"comprehension_score": 6.696428571428571,
|
| 229 |
+
"structure_score": 6.814285714285714,
|
| 230 |
+
"prose_style_score": 7.175,
|
| 231 |
+
"creativity_score": 7.260714285714286,
|
| 232 |
+
"depth_score": 6.714285714285714,
|
| 233 |
+
"helpfulness_score": 6.628571428571429,
|
| 234 |
+
"overall_score": 6.828571428571428
|
| 235 |
+
}
|
| 236 |
+
}
|
data/complicated-writing-scores.json
ADDED
|
@@ -0,0 +1,236 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"Mistral-3.1-24B-2503": {
|
| 3 |
+
"comprehension_score": 4.518867924528302,
|
| 4 |
+
"structure_score": 4.518867924528302,
|
| 5 |
+
"prose_style_score": 3.7452830188679247,
|
| 6 |
+
"creativity_score": 3.5849056603773586,
|
| 7 |
+
"depth_score": 3.556603773584906,
|
| 8 |
+
"helpfulness_score": 4.0,
|
| 9 |
+
"overall_score": 3.9339622641509435
|
| 10 |
+
},
|
| 11 |
+
"Phi-4-14B": {
|
| 12 |
+
"comprehension_score": 4.622641509433962,
|
| 13 |
+
"structure_score": 4.688679245283019,
|
| 14 |
+
"prose_style_score": 4.09433962264151,
|
| 15 |
+
"creativity_score": 3.990566037735849,
|
| 16 |
+
"depth_score": 3.858490566037736,
|
| 17 |
+
"helpfulness_score": 4.19811320754717,
|
| 18 |
+
"overall_score": 4.10377358490566
|
| 19 |
+
},
|
| 20 |
+
"gpt-4o-mini-2024-07-18": {
|
| 21 |
+
"comprehension_score": 4.688679245283019,
|
| 22 |
+
"structure_score": 4.783018867924528,
|
| 23 |
+
"prose_style_score": 4.216981132075472,
|
| 24 |
+
"creativity_score": 4.066037735849057,
|
| 25 |
+
"depth_score": 3.792452830188679,
|
| 26 |
+
"helpfulness_score": 4.245283018867925,
|
| 27 |
+
"overall_score": 4.2075471698113205
|
| 28 |
+
},
|
| 29 |
+
"gpt-4.1-mini-2025-04-14": {
|
| 30 |
+
"comprehension_score": 4.981132075471698,
|
| 31 |
+
"structure_score": 4.877358490566038,
|
| 32 |
+
"prose_style_score": 4.509433962264151,
|
| 33 |
+
"creativity_score": 4.283018867924528,
|
| 34 |
+
"depth_score": 4.018867924528302,
|
| 35 |
+
"helpfulness_score": 4.509433962264151,
|
| 36 |
+
"overall_score": 4.330188679245283
|
| 37 |
+
},
|
| 38 |
+
"Qwen3-8B": {
|
| 39 |
+
"comprehension_score": 4.8584905660377355,
|
| 40 |
+
"structure_score": 5.018867924528302,
|
| 41 |
+
"prose_style_score": 4.556603773584905,
|
| 42 |
+
"creativity_score": 4.433962264150943,
|
| 43 |
+
"depth_score": 4.179245283018868,
|
| 44 |
+
"helpfulness_score": 4.566037735849057,
|
| 45 |
+
"overall_score": 4.462264150943396
|
| 46 |
+
},
|
| 47 |
+
"Qwen3-30B-A3B": {
|
| 48 |
+
"comprehension_score": 4.952830188679245,
|
| 49 |
+
"structure_score": 5.122641509433962,
|
| 50 |
+
"prose_style_score": 4.556603773584905,
|
| 51 |
+
"creativity_score": 4.547169811320755,
|
| 52 |
+
"depth_score": 4.283018867924528,
|
| 53 |
+
"helpfulness_score": 4.745283018867925,
|
| 54 |
+
"overall_score": 4.528301886792453
|
| 55 |
+
},
|
| 56 |
+
"gemma3-27b": {
|
| 57 |
+
"comprehension_score": 5.264150943396227,
|
| 58 |
+
"structure_score": 5.330188679245283,
|
| 59 |
+
"prose_style_score": 4.59433962264151,
|
| 60 |
+
"creativity_score": 4.820754716981132,
|
| 61 |
+
"depth_score": 4.584905660377358,
|
| 62 |
+
"helpfulness_score": 5.066037735849057,
|
| 63 |
+
"overall_score": 4.783018867924528
|
| 64 |
+
},
|
| 65 |
+
"Qwen3-32B": {
|
| 66 |
+
"comprehension_score": 5.226415094339623,
|
| 67 |
+
"structure_score": 5.386792452830188,
|
| 68 |
+
"prose_style_score": 4.915094339622642,
|
| 69 |
+
"creativity_score": 4.90566037735849,
|
| 70 |
+
"depth_score": 4.613207547169812,
|
| 71 |
+
"helpfulness_score": 5.037735849056604,
|
| 72 |
+
"overall_score": 4.80188679245283
|
| 73 |
+
},
|
| 74 |
+
"gpt-4o-2024-11-20": {
|
| 75 |
+
"comprehension_score": 5.264150943396227,
|
| 76 |
+
"structure_score": 5.320754716981132,
|
| 77 |
+
"prose_style_score": 5.066037735849057,
|
| 78 |
+
"creativity_score": 5.122641509433962,
|
| 79 |
+
"depth_score": 4.6415094339622645,
|
| 80 |
+
"helpfulness_score": 4.990566037735849,
|
| 81 |
+
"overall_score": 4.943396226415095
|
| 82 |
+
},
|
| 83 |
+
"Mistral-3.2-24B-2506": {
|
| 84 |
+
"comprehension_score": 5.4245283018867925,
|
| 85 |
+
"structure_score": 5.622641509433962,
|
| 86 |
+
"prose_style_score": 5.1415094339622645,
|
| 87 |
+
"creativity_score": 5.481132075471698,
|
| 88 |
+
"depth_score": 4.688679245283019,
|
| 89 |
+
"helpfulness_score": 5.113207547169812,
|
| 90 |
+
"overall_score": 5.056603773584905
|
| 91 |
+
},
|
| 92 |
+
"o4-mini-2025-04-16": {
|
| 93 |
+
"comprehension_score": 5.509433962264151,
|
| 94 |
+
"structure_score": 5.481132075471698,
|
| 95 |
+
"prose_style_score": 5.330188679245283,
|
| 96 |
+
"creativity_score": 5.264150943396227,
|
| 97 |
+
"depth_score": 4.7924528301886795,
|
| 98 |
+
"helpfulness_score": 5.235849056603773,
|
| 99 |
+
"overall_score": 5.132075471698113
|
| 100 |
+
},
|
| 101 |
+
"Qwen3-235B": {
|
| 102 |
+
"comprehension_score": 5.7075471698113205,
|
| 103 |
+
"structure_score": 5.811320754716981,
|
| 104 |
+
"prose_style_score": 5.367924528301887,
|
| 105 |
+
"creativity_score": 5.5,
|
| 106 |
+
"depth_score": 5.0754716981132075,
|
| 107 |
+
"helpfulness_score": 5.452830188679245,
|
| 108 |
+
"overall_score": 5.415094339622642
|
| 109 |
+
},
|
| 110 |
+
"Gemini-2.5-Flash": {
|
| 111 |
+
"comprehension_score": 6.19811320754717,
|
| 112 |
+
"structure_score": 6.1415094339622645,
|
| 113 |
+
"prose_style_score": 5.9245283018867925,
|
| 114 |
+
"creativity_score": 5.80188679245283,
|
| 115 |
+
"depth_score": 5.7075471698113205,
|
| 116 |
+
"helpfulness_score": 6.0,
|
| 117 |
+
"overall_score": 5.943396226415095
|
| 118 |
+
},
|
| 119 |
+
"Qwen3-30B-A3B-Thinking": {
|
| 120 |
+
"comprehension_score": 5.773584905660377,
|
| 121 |
+
"structure_score": 6.047169811320755,
|
| 122 |
+
"prose_style_score": 6.547169811320755,
|
| 123 |
+
"creativity_score": 6.735849056603773,
|
| 124 |
+
"depth_score": 5.669811320754717,
|
| 125 |
+
"helpfulness_score": 5.783018867924528,
|
| 126 |
+
"overall_score": 6.028301886792453
|
| 127 |
+
},
|
| 128 |
+
"Deepseek-V3-0324": {
|
| 129 |
+
"comprehension_score": 6.113207547169812,
|
| 130 |
+
"structure_score": 6.122641509433962,
|
| 131 |
+
"prose_style_score": 6.547169811320755,
|
| 132 |
+
"creativity_score": 6.820754716981132,
|
| 133 |
+
"depth_score": 5.820754716981132,
|
| 134 |
+
"helpfulness_score": 5.8584905660377355,
|
| 135 |
+
"overall_score": 6.150943396226415
|
| 136 |
+
},
|
| 137 |
+
"Qwen3-32B-Thinking": {
|
| 138 |
+
"comprehension_score": 6.2075471698113205,
|
| 139 |
+
"structure_score": 6.216981132075472,
|
| 140 |
+
"prose_style_score": 7.150943396226415,
|
| 141 |
+
"creativity_score": 7.443396226415095,
|
| 142 |
+
"depth_score": 6.179245283018868,
|
| 143 |
+
"helpfulness_score": 6.09433962264151,
|
| 144 |
+
"overall_score": 6.547169811320755
|
| 145 |
+
},
|
| 146 |
+
"Qwen3-235B-Thinking": {
|
| 147 |
+
"comprehension_score": 6.216981132075472,
|
| 148 |
+
"structure_score": 6.254716981132075,
|
| 149 |
+
"prose_style_score": 7.4245283018867925,
|
| 150 |
+
"creativity_score": 7.773584905660377,
|
| 151 |
+
"depth_score": 6.320754716981132,
|
| 152 |
+
"helpfulness_score": 6.169811320754717,
|
| 153 |
+
"overall_score": 6.584905660377358
|
| 154 |
+
},
|
| 155 |
+
"Deepseek-R1-0528": {
|
| 156 |
+
"comprehension_score": 6.773584905660377,
|
| 157 |
+
"structure_score": 6.735849056603773,
|
| 158 |
+
"prose_style_score": 7.320754716981132,
|
| 159 |
+
"creativity_score": 7.367924528301887,
|
| 160 |
+
"depth_score": 6.849056603773585,
|
| 161 |
+
"helpfulness_score": 6.726415094339623,
|
| 162 |
+
"overall_score": 6.9245283018867925
|
| 163 |
+
},
|
| 164 |
+
"o3-2025-04-16": {
|
| 165 |
+
"comprehension_score": 7.226415094339623,
|
| 166 |
+
"structure_score": 7.018867924528302,
|
| 167 |
+
"prose_style_score": 7.452830188679245,
|
| 168 |
+
"creativity_score": 7.2924528301886795,
|
| 169 |
+
"depth_score": 6.688679245283019,
|
| 170 |
+
"helpfulness_score": 6.915094339622642,
|
| 171 |
+
"overall_score": 7.179245283018868
|
| 172 |
+
},
|
| 173 |
+
"gpt-5.4": {
|
| 174 |
+
"comprehension_score": 7.023529411764706,
|
| 175 |
+
"structure_score": 6.870588235294117,
|
| 176 |
+
"prose_style_score": 7.311764705882353,
|
| 177 |
+
"creativity_score": 6.947058823529412,
|
| 178 |
+
"depth_score": 6.8882352941176475,
|
| 179 |
+
"helpfulness_score": 6.9,
|
| 180 |
+
"overall_score": 7.052941176470588
|
| 181 |
+
},
|
| 182 |
+
"Gemini-3.1-Flash": {
|
| 183 |
+
"comprehension_score": 6.211764705882353,
|
| 184 |
+
"structure_score": 6.411764705882353,
|
| 185 |
+
"prose_style_score": 6.711764705882353,
|
| 186 |
+
"creativity_score": 6.511764705882353,
|
| 187 |
+
"depth_score": 5.8882352941176475,
|
| 188 |
+
"helpfulness_score": 6.123529411764705,
|
| 189 |
+
"overall_score": 6.341176470588235
|
| 190 |
+
},
|
| 191 |
+
"Gemini-3.1-Pro": {
|
| 192 |
+
"comprehension_score": 6.58235294117647,
|
| 193 |
+
"structure_score": 6.635294117647059,
|
| 194 |
+
"prose_style_score": 7.1117647058823525,
|
| 195 |
+
"creativity_score": 6.735294117647059,
|
| 196 |
+
"depth_score": 6.2,
|
| 197 |
+
"helpfulness_score": 6.488235294117647,
|
| 198 |
+
"overall_score": 6.7176470588235295
|
| 199 |
+
},
|
| 200 |
+
"MiniMax-M2.5": {
|
| 201 |
+
"comprehension_score": 6.205882352941177,
|
| 202 |
+
"structure_score": 6.2,
|
| 203 |
+
"prose_style_score": 5.9,
|
| 204 |
+
"creativity_score": 5.9941176470588236,
|
| 205 |
+
"depth_score": 5.694117647058824,
|
| 206 |
+
"helpfulness_score": 6.011764705882353,
|
| 207 |
+
"overall_score": 6.035294117647059
|
| 208 |
+
},
|
| 209 |
+
"Qwen3.5-Plus": {
|
| 210 |
+
"comprehension_score": 6.594117647058823,
|
| 211 |
+
"structure_score": 6.488235294117647,
|
| 212 |
+
"prose_style_score": 6.982352941176471,
|
| 213 |
+
"creativity_score": 6.629411764705883,
|
| 214 |
+
"depth_score": 6.135294117647059,
|
| 215 |
+
"helpfulness_score": 6.405882352941177,
|
| 216 |
+
"overall_score": 6.605882352941176
|
| 217 |
+
},
|
| 218 |
+
"GLM-5": {
|
| 219 |
+
"comprehension_score": 6.2176470588235295,
|
| 220 |
+
"structure_score": 6.3352941176470585,
|
| 221 |
+
"prose_style_score": 6.658823529411765,
|
| 222 |
+
"creativity_score": 6.411764705882353,
|
| 223 |
+
"depth_score": 5.882352941176471,
|
| 224 |
+
"helpfulness_score": 6.1117647058823525,
|
| 225 |
+
"overall_score": 6.323529411764706
|
| 226 |
+
},
|
| 227 |
+
"step-3.5-flash": {
|
| 228 |
+
"comprehension_score": 6.6,
|
| 229 |
+
"structure_score": 6.682352941176471,
|
| 230 |
+
"prose_style_score": 7.2,
|
| 231 |
+
"creativity_score": 7.288235294117647,
|
| 232 |
+
"depth_score": 6.670588235294118,
|
| 233 |
+
"helpfulness_score": 6.458823529411765,
|
| 234 |
+
"overall_score": 6.723529411764706
|
| 235 |
+
}
|
| 236 |
+
}
|
data/gpt5.4-judge-all-scores.json
ADDED
|
@@ -0,0 +1,236 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"o3-2025-04-16": {
|
| 3 |
+
"comprehension_score": 7.389285714285714,
|
| 4 |
+
"structure_score": 7.417857142857143,
|
| 5 |
+
"prose_style_score": 7.575,
|
| 6 |
+
"creativity_score": 7.564285714285714,
|
| 7 |
+
"depth_score": 7.196428571428571,
|
| 8 |
+
"helpfulness_score": 7.357142857142857,
|
| 9 |
+
"overall_score": 7.307142857142857
|
| 10 |
+
},
|
| 11 |
+
"Deepseek-R1-0528": {
|
| 12 |
+
"comprehension_score": 7.146428571428571,
|
| 13 |
+
"structure_score": 7.0964285714285715,
|
| 14 |
+
"prose_style_score": 7.085714285714285,
|
| 15 |
+
"creativity_score": 7.5964285714285715,
|
| 16 |
+
"depth_score": 7.4,
|
| 17 |
+
"helpfulness_score": 7.053571428571429,
|
| 18 |
+
"overall_score": 7.139285714285714
|
| 19 |
+
},
|
| 20 |
+
"Deepseek-V3-0324": {
|
| 21 |
+
"comprehension_score": 6.453571428571428,
|
| 22 |
+
"structure_score": 6.578571428571428,
|
| 23 |
+
"prose_style_score": 6.832142857142857,
|
| 24 |
+
"creativity_score": 7.142857142857143,
|
| 25 |
+
"depth_score": 6.353571428571429,
|
| 26 |
+
"helpfulness_score": 6.364285714285714,
|
| 27 |
+
"overall_score": 6.457142857142857
|
| 28 |
+
},
|
| 29 |
+
"Qwen3-235B-Thinking": {
|
| 30 |
+
"comprehension_score": 6.4071428571428575,
|
| 31 |
+
"structure_score": 6.446428571428571,
|
| 32 |
+
"prose_style_score": 7.2214285714285715,
|
| 33 |
+
"creativity_score": 7.75,
|
| 34 |
+
"depth_score": 6.739285714285714,
|
| 35 |
+
"helpfulness_score": 6.239285714285714,
|
| 36 |
+
"overall_score": 6.517857142857143
|
| 37 |
+
},
|
| 38 |
+
"Qwen3-32B-Thinking": {
|
| 39 |
+
"comprehension_score": 6.457142857142857,
|
| 40 |
+
"structure_score": 6.4714285714285715,
|
| 41 |
+
"prose_style_score": 7.128571428571429,
|
| 42 |
+
"creativity_score": 7.525,
|
| 43 |
+
"depth_score": 6.614285714285714,
|
| 44 |
+
"helpfulness_score": 6.317857142857143,
|
| 45 |
+
"overall_score": 6.546428571428572
|
| 46 |
+
},
|
| 47 |
+
"Qwen3-30B-A3B-Thinking": {
|
| 48 |
+
"comprehension_score": 5.964285714285714,
|
| 49 |
+
"structure_score": 6.110714285714286,
|
| 50 |
+
"prose_style_score": 6.521428571428571,
|
| 51 |
+
"creativity_score": 6.8428571428571425,
|
| 52 |
+
"depth_score": 5.985714285714286,
|
| 53 |
+
"helpfulness_score": 5.882142857142857,
|
| 54 |
+
"overall_score": 6.010714285714286
|
| 55 |
+
},
|
| 56 |
+
"Gemini-2.5-Flash": {
|
| 57 |
+
"comprehension_score": 6.503571428571429,
|
| 58 |
+
"structure_score": 6.571428571428571,
|
| 59 |
+
"prose_style_score": 5.942857142857143,
|
| 60 |
+
"creativity_score": 6.228571428571429,
|
| 61 |
+
"depth_score": 6.257142857142857,
|
| 62 |
+
"helpfulness_score": 6.428571428571429,
|
| 63 |
+
"overall_score": 6.2785714285714285
|
| 64 |
+
},
|
| 65 |
+
"Qwen3-235B": {
|
| 66 |
+
"comprehension_score": 5.896428571428571,
|
| 67 |
+
"structure_score": 6.2,
|
| 68 |
+
"prose_style_score": 5.667857142857143,
|
| 69 |
+
"creativity_score": 5.785714285714286,
|
| 70 |
+
"depth_score": 5.503571428571429,
|
| 71 |
+
"helpfulness_score": 5.8464285714285715,
|
| 72 |
+
"overall_score": 5.767857142857143
|
| 73 |
+
},
|
| 74 |
+
"Qwen3-32B": {
|
| 75 |
+
"comprehension_score": 5.303571428571429,
|
| 76 |
+
"structure_score": 5.582142857142857,
|
| 77 |
+
"prose_style_score": 4.960714285714285,
|
| 78 |
+
"creativity_score": 5.189285714285714,
|
| 79 |
+
"depth_score": 4.896428571428571,
|
| 80 |
+
"helpfulness_score": 5.117857142857143,
|
| 81 |
+
"overall_score": 5.042857142857143
|
| 82 |
+
},
|
| 83 |
+
"Qwen3-8B": {
|
| 84 |
+
"comprehension_score": 4.860714285714286,
|
| 85 |
+
"structure_score": 5.0928571428571425,
|
| 86 |
+
"prose_style_score": 4.55,
|
| 87 |
+
"creativity_score": 4.710714285714285,
|
| 88 |
+
"depth_score": 4.457142857142857,
|
| 89 |
+
"helpfulness_score": 4.639285714285714,
|
| 90 |
+
"overall_score": 4.560714285714286
|
| 91 |
+
},
|
| 92 |
+
"Qwen3-30B-A3B": {
|
| 93 |
+
"comprehension_score": 4.953571428571428,
|
| 94 |
+
"structure_score": 5.2214285714285715,
|
| 95 |
+
"prose_style_score": 4.6,
|
| 96 |
+
"creativity_score": 4.7785714285714285,
|
| 97 |
+
"depth_score": 4.521428571428571,
|
| 98 |
+
"helpfulness_score": 4.760714285714286,
|
| 99 |
+
"overall_score": 4.65
|
| 100 |
+
},
|
| 101 |
+
"gemma3-27b": {
|
| 102 |
+
"comprehension_score": 5.164285714285715,
|
| 103 |
+
"structure_score": 5.364285714285714,
|
| 104 |
+
"prose_style_score": 4.553571428571429,
|
| 105 |
+
"creativity_score": 5.064285714285714,
|
| 106 |
+
"depth_score": 4.710714285714285,
|
| 107 |
+
"helpfulness_score": 4.992857142857143,
|
| 108 |
+
"overall_score": 4.814285714285714
|
| 109 |
+
},
|
| 110 |
+
"Phi-4-14B": {
|
| 111 |
+
"comprehension_score": 4.2785714285714285,
|
| 112 |
+
"structure_score": 4.514285714285714,
|
| 113 |
+
"prose_style_score": 3.7714285714285714,
|
| 114 |
+
"creativity_score": 4.310714285714286,
|
| 115 |
+
"depth_score": 4.0285714285714285,
|
| 116 |
+
"helpfulness_score": 4.017857142857143,
|
| 117 |
+
"overall_score": 3.9857142857142858
|
| 118 |
+
},
|
| 119 |
+
"Mistral-3.2-24B-2506": {
|
| 120 |
+
"comprehension_score": 5.475,
|
| 121 |
+
"structure_score": 5.746428571428571,
|
| 122 |
+
"prose_style_score": 5.35,
|
| 123 |
+
"creativity_score": 5.885714285714286,
|
| 124 |
+
"depth_score": 5.214285714285714,
|
| 125 |
+
"helpfulness_score": 5.3464285714285715,
|
| 126 |
+
"overall_score": 5.335714285714285
|
| 127 |
+
},
|
| 128 |
+
"Mistral-3.1-24B-2503": {
|
| 129 |
+
"comprehension_score": 4.303571428571429,
|
| 130 |
+
"structure_score": 4.521428571428571,
|
| 131 |
+
"prose_style_score": 3.6857142857142855,
|
| 132 |
+
"creativity_score": 3.9642857142857144,
|
| 133 |
+
"depth_score": 3.8285714285714287,
|
| 134 |
+
"helpfulness_score": 4.021428571428571,
|
| 135 |
+
"overall_score": 3.9214285714285713
|
| 136 |
+
},
|
| 137 |
+
"gpt-4o-2024-11-20": {
|
| 138 |
+
"comprehension_score": 5.275,
|
| 139 |
+
"structure_score": 5.239285714285714,
|
| 140 |
+
"prose_style_score": 4.996428571428571,
|
| 141 |
+
"creativity_score": 5.321428571428571,
|
| 142 |
+
"depth_score": 4.892857142857143,
|
| 143 |
+
"helpfulness_score": 4.957142857142857,
|
| 144 |
+
"overall_score": 4.985714285714286
|
| 145 |
+
},
|
| 146 |
+
"gpt-4.1-mini-2025-04-14": {
|
| 147 |
+
"comprehension_score": 4.8,
|
| 148 |
+
"structure_score": 4.746428571428571,
|
| 149 |
+
"prose_style_score": 4.339285714285714,
|
| 150 |
+
"creativity_score": 4.489285714285714,
|
| 151 |
+
"depth_score": 4.275,
|
| 152 |
+
"helpfulness_score": 4.389285714285714,
|
| 153 |
+
"overall_score": 4.35
|
| 154 |
+
},
|
| 155 |
+
"gpt-4o-mini-2024-07-18": {
|
| 156 |
+
"comprehension_score": 4.417857142857143,
|
| 157 |
+
"structure_score": 4.575,
|
| 158 |
+
"prose_style_score": 4.064285714285714,
|
| 159 |
+
"creativity_score": 4.375,
|
| 160 |
+
"depth_score": 4.085714285714285,
|
| 161 |
+
"helpfulness_score": 4.139285714285714,
|
| 162 |
+
"overall_score": 4.15
|
| 163 |
+
},
|
| 164 |
+
"o4-mini-2025-04-16": {
|
| 165 |
+
"comprehension_score": 5.803571428571429,
|
| 166 |
+
"structure_score": 5.864285714285714,
|
| 167 |
+
"prose_style_score": 5.428571428571429,
|
| 168 |
+
"creativity_score": 5.760714285714286,
|
| 169 |
+
"depth_score": 5.310714285714286,
|
| 170 |
+
"helpfulness_score": 5.617857142857143,
|
| 171 |
+
"overall_score": 5.557142857142857
|
| 172 |
+
},
|
| 173 |
+
"gpt-5.4": {
|
| 174 |
+
"comprehension_score": 7.4678571428571425,
|
| 175 |
+
"structure_score": 7.517857142857143,
|
| 176 |
+
"prose_style_score": 7.585714285714285,
|
| 177 |
+
"creativity_score": 7.182142857142857,
|
| 178 |
+
"depth_score": 7.5,
|
| 179 |
+
"helpfulness_score": 7.664285714285715,
|
| 180 |
+
"overall_score": 7.421428571428572
|
| 181 |
+
},
|
| 182 |
+
"Gemini-3.1-Flash": {
|
| 183 |
+
"comprehension_score": 7.0964285714285715,
|
| 184 |
+
"structure_score": 7.232142857142857,
|
| 185 |
+
"prose_style_score": 7.317857142857143,
|
| 186 |
+
"creativity_score": 7.417857142857143,
|
| 187 |
+
"depth_score": 7.053571428571429,
|
| 188 |
+
"helpfulness_score": 7.185714285714286,
|
| 189 |
+
"overall_score": 7.117857142857143
|
| 190 |
+
},
|
| 191 |
+
"Gemini-3.1-Pro": {
|
| 192 |
+
"comprehension_score": 7.510714285714286,
|
| 193 |
+
"structure_score": 7.5321428571428575,
|
| 194 |
+
"prose_style_score": 7.5285714285714285,
|
| 195 |
+
"creativity_score": 7.614285714285714,
|
| 196 |
+
"depth_score": 7.321428571428571,
|
| 197 |
+
"helpfulness_score": 7.65,
|
| 198 |
+
"overall_score": 7.460714285714285
|
| 199 |
+
},
|
| 200 |
+
"MiniMax-M2.5": {
|
| 201 |
+
"comprehension_score": 6.292857142857143,
|
| 202 |
+
"structure_score": 6.442857142857143,
|
| 203 |
+
"prose_style_score": 5.614285714285714,
|
| 204 |
+
"creativity_score": 6.171428571428572,
|
| 205 |
+
"depth_score": 5.975,
|
| 206 |
+
"helpfulness_score": 6.242857142857143,
|
| 207 |
+
"overall_score": 6.071428571428571
|
| 208 |
+
},
|
| 209 |
+
"Qwen3.5-Plus": {
|
| 210 |
+
"comprehension_score": 7.228571428571429,
|
| 211 |
+
"structure_score": 7.3,
|
| 212 |
+
"prose_style_score": 7.232142857142857,
|
| 213 |
+
"creativity_score": 7.353571428571429,
|
| 214 |
+
"depth_score": 7.196428571428571,
|
| 215 |
+
"helpfulness_score": 7.289285714285715,
|
| 216 |
+
"overall_score": 7.189285714285714
|
| 217 |
+
},
|
| 218 |
+
"step-3.5-flash": {
|
| 219 |
+
"comprehension_score": 7.285714285714286,
|
| 220 |
+
"structure_score": 7.335714285714285,
|
| 221 |
+
"prose_style_score": 7.053571428571429,
|
| 222 |
+
"creativity_score": 7.789285714285715,
|
| 223 |
+
"depth_score": 7.525,
|
| 224 |
+
"helpfulness_score": 7.2178571428571425,
|
| 225 |
+
"overall_score": 7.253571428571429
|
| 226 |
+
},
|
| 227 |
+
"GLM-5": {
|
| 228 |
+
"comprehension_score": 7.103571428571429,
|
| 229 |
+
"structure_score": 7.182142857142857,
|
| 230 |
+
"prose_style_score": 7.114285714285714,
|
| 231 |
+
"creativity_score": 7.185714285714286,
|
| 232 |
+
"depth_score": 6.964285714285714,
|
| 233 |
+
"helpfulness_score": 7.146428571428571,
|
| 234 |
+
"overall_score": 7.042857142857143
|
| 235 |
+
}
|
| 236 |
+
}
|
data/gpt5.4-judge-complicated-writing-scores.json
ADDED
|
@@ -0,0 +1,236 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"o3-2025-04-16": {
|
| 3 |
+
"comprehension_score": 7.5,
|
| 4 |
+
"structure_score": 7.541176470588235,
|
| 5 |
+
"prose_style_score": 7.788235294117647,
|
| 6 |
+
"creativity_score": 7.688235294117647,
|
| 7 |
+
"depth_score": 7.358823529411764,
|
| 8 |
+
"helpfulness_score": 7.476470588235294,
|
| 9 |
+
"overall_score": 7.4411764705882355
|
| 10 |
+
},
|
| 11 |
+
"Deepseek-R1-0528": {
|
| 12 |
+
"comprehension_score": 7.070588235294117,
|
| 13 |
+
"structure_score": 7.0,
|
| 14 |
+
"prose_style_score": 7.352941176470588,
|
| 15 |
+
"creativity_score": 7.647058823529412,
|
| 16 |
+
"depth_score": 7.341176470588235,
|
| 17 |
+
"helpfulness_score": 6.923529411764706,
|
| 18 |
+
"overall_score": 7.105882352941176
|
| 19 |
+
},
|
| 20 |
+
"Deepseek-V3-0324": {
|
| 21 |
+
"comprehension_score": 6.188235294117647,
|
| 22 |
+
"structure_score": 6.3,
|
| 23 |
+
"prose_style_score": 6.758823529411765,
|
| 24 |
+
"creativity_score": 7.0588235294117645,
|
| 25 |
+
"depth_score": 6.223529411764706,
|
| 26 |
+
"helpfulness_score": 6.035294117647059,
|
| 27 |
+
"overall_score": 6.223529411764706
|
| 28 |
+
},
|
| 29 |
+
"Qwen3-235B-Thinking": {
|
| 30 |
+
"comprehension_score": 6.376470588235295,
|
| 31 |
+
"structure_score": 6.252941176470588,
|
| 32 |
+
"prose_style_score": 7.435294117647059,
|
| 33 |
+
"creativity_score": 7.829411764705882,
|
| 34 |
+
"depth_score": 6.7176470588235295,
|
| 35 |
+
"helpfulness_score": 6.176470588235294,
|
| 36 |
+
"overall_score": 6.517647058823529
|
| 37 |
+
},
|
| 38 |
+
"Qwen3-32B-Thinking": {
|
| 39 |
+
"comprehension_score": 6.3882352941176475,
|
| 40 |
+
"structure_score": 6.323529411764706,
|
| 41 |
+
"prose_style_score": 7.323529411764706,
|
| 42 |
+
"creativity_score": 7.570588235294117,
|
| 43 |
+
"depth_score": 6.523529411764706,
|
| 44 |
+
"helpfulness_score": 6.258823529411765,
|
| 45 |
+
"overall_score": 6.523529411764706
|
| 46 |
+
},
|
| 47 |
+
"Qwen3-30B-A3B-Thinking": {
|
| 48 |
+
"comprehension_score": 5.823529411764706,
|
| 49 |
+
"structure_score": 5.952941176470588,
|
| 50 |
+
"prose_style_score": 6.570588235294117,
|
| 51 |
+
"creativity_score": 6.829411764705882,
|
| 52 |
+
"depth_score": 5.7823529411764705,
|
| 53 |
+
"helpfulness_score": 5.735294117647059,
|
| 54 |
+
"overall_score": 5.876470588235295
|
| 55 |
+
},
|
| 56 |
+
"Gemini-2.5-Flash": {
|
| 57 |
+
"comprehension_score": 6.429411764705883,
|
| 58 |
+
"structure_score": 6.482352941176471,
|
| 59 |
+
"prose_style_score": 5.847058823529411,
|
| 60 |
+
"creativity_score": 6.147058823529412,
|
| 61 |
+
"depth_score": 6.094117647058823,
|
| 62 |
+
"helpfulness_score": 6.2823529411764705,
|
| 63 |
+
"overall_score": 6.147058823529412
|
| 64 |
+
},
|
| 65 |
+
"Qwen3-235B": {
|
| 66 |
+
"comprehension_score": 5.788235294117647,
|
| 67 |
+
"structure_score": 6.070588235294117,
|
| 68 |
+
"prose_style_score": 5.511764705882353,
|
| 69 |
+
"creativity_score": 5.647058823529412,
|
| 70 |
+
"depth_score": 5.4,
|
| 71 |
+
"helpfulness_score": 5.623529411764705,
|
| 72 |
+
"overall_score": 5.58235294117647
|
| 73 |
+
},
|
| 74 |
+
"Qwen3-32B": {
|
| 75 |
+
"comprehension_score": 5.117647058823529,
|
| 76 |
+
"structure_score": 5.364705882352941,
|
| 77 |
+
"prose_style_score": 4.694117647058824,
|
| 78 |
+
"creativity_score": 4.982352941176471,
|
| 79 |
+
"depth_score": 4.647058823529412,
|
| 80 |
+
"helpfulness_score": 4.794117647058823,
|
| 81 |
+
"overall_score": 4.729411764705882
|
| 82 |
+
},
|
| 83 |
+
"Qwen3-8B": {
|
| 84 |
+
"comprehension_score": 4.6647058823529415,
|
| 85 |
+
"structure_score": 4.870588235294117,
|
| 86 |
+
"prose_style_score": 4.2823529411764705,
|
| 87 |
+
"creativity_score": 4.5,
|
| 88 |
+
"depth_score": 4.241176470588235,
|
| 89 |
+
"helpfulness_score": 4.347058823529411,
|
| 90 |
+
"overall_score": 4.3
|
| 91 |
+
},
|
| 92 |
+
"Qwen3-30B-A3B": {
|
| 93 |
+
"comprehension_score": 4.805882352941176,
|
| 94 |
+
"structure_score": 4.988235294117647,
|
| 95 |
+
"prose_style_score": 4.405882352941177,
|
| 96 |
+
"creativity_score": 4.617647058823529,
|
| 97 |
+
"depth_score": 4.364705882352941,
|
| 98 |
+
"helpfulness_score": 4.523529411764706,
|
| 99 |
+
"overall_score": 4.429411764705883
|
| 100 |
+
},
|
| 101 |
+
"gemma3-27b": {
|
| 102 |
+
"comprehension_score": 5.070588235294117,
|
| 103 |
+
"structure_score": 5.241176470588235,
|
| 104 |
+
"prose_style_score": 4.470588235294118,
|
| 105 |
+
"creativity_score": 5.011764705882353,
|
| 106 |
+
"depth_score": 4.58235294117647,
|
| 107 |
+
"helpfulness_score": 4.817647058823529,
|
| 108 |
+
"overall_score": 4.6647058823529415
|
| 109 |
+
},
|
| 110 |
+
"Phi-4-14B": {
|
| 111 |
+
"comprehension_score": 4.235294117647059,
|
| 112 |
+
"structure_score": 4.529411764705882,
|
| 113 |
+
"prose_style_score": 3.652941176470588,
|
| 114 |
+
"creativity_score": 4.258823529411765,
|
| 115 |
+
"depth_score": 4.023529411764706,
|
| 116 |
+
"helpfulness_score": 4.017647058823529,
|
| 117 |
+
"overall_score": 3.9647058823529413
|
| 118 |
+
},
|
| 119 |
+
"Mistral-3.2-24B-2506": {
|
| 120 |
+
"comprehension_score": 5.3352941176470585,
|
| 121 |
+
"structure_score": 5.570588235294117,
|
| 122 |
+
"prose_style_score": 5.188235294117647,
|
| 123 |
+
"creativity_score": 5.7176470588235295,
|
| 124 |
+
"depth_score": 5.047058823529412,
|
| 125 |
+
"helpfulness_score": 5.147058823529412,
|
| 126 |
+
"overall_score": 5.147058823529412
|
| 127 |
+
},
|
| 128 |
+
"Mistral-3.1-24B-2503": {
|
| 129 |
+
"comprehension_score": 4.276470588235294,
|
| 130 |
+
"structure_score": 4.523529411764706,
|
| 131 |
+
"prose_style_score": 3.523529411764706,
|
| 132 |
+
"creativity_score": 3.9411764705882355,
|
| 133 |
+
"depth_score": 3.8117647058823527,
|
| 134 |
+
"helpfulness_score": 4.029411764705882,
|
| 135 |
+
"overall_score": 3.9058823529411764
|
| 136 |
+
},
|
| 137 |
+
"gpt-4o-2024-11-20": {
|
| 138 |
+
"comprehension_score": 5.223529411764706,
|
| 139 |
+
"structure_score": 5.176470588235294,
|
| 140 |
+
"prose_style_score": 4.882352941176471,
|
| 141 |
+
"creativity_score": 5.211764705882353,
|
| 142 |
+
"depth_score": 4.788235294117647,
|
| 143 |
+
"helpfulness_score": 4.852941176470588,
|
| 144 |
+
"overall_score": 4.858823529411764
|
| 145 |
+
},
|
| 146 |
+
"gpt-4.1-mini-2025-04-14": {
|
| 147 |
+
"comprehension_score": 4.7,
|
| 148 |
+
"structure_score": 4.694117647058824,
|
| 149 |
+
"prose_style_score": 4.247058823529412,
|
| 150 |
+
"creativity_score": 4.405882352941177,
|
| 151 |
+
"depth_score": 4.176470588235294,
|
| 152 |
+
"helpfulness_score": 4.264705882352941,
|
| 153 |
+
"overall_score": 4.241176470588235
|
| 154 |
+
},
|
| 155 |
+
"gpt-4o-mini-2024-07-18": {
|
| 156 |
+
"comprehension_score": 4.41764705882353,
|
| 157 |
+
"structure_score": 4.588235294117647,
|
| 158 |
+
"prose_style_score": 4.0,
|
| 159 |
+
"creativity_score": 4.329411764705882,
|
| 160 |
+
"depth_score": 4.076470588235294,
|
| 161 |
+
"helpfulness_score": 4.117647058823529,
|
| 162 |
+
"overall_score": 4.147058823529412
|
| 163 |
+
},
|
| 164 |
+
"o4-mini-2025-04-16": {
|
| 165 |
+
"comprehension_score": 5.694117647058824,
|
| 166 |
+
"structure_score": 5.764705882352941,
|
| 167 |
+
"prose_style_score": 5.341176470588235,
|
| 168 |
+
"creativity_score": 5.623529411764705,
|
| 169 |
+
"depth_score": 5.152941176470589,
|
| 170 |
+
"helpfulness_score": 5.488235294117647,
|
| 171 |
+
"overall_score": 5.41764705882353
|
| 172 |
+
},
|
| 173 |
+
"gpt-5.4": {
|
| 174 |
+
"comprehension_score": 7.529411764705882,
|
| 175 |
+
"structure_score": 7.535294117647059,
|
| 176 |
+
"prose_style_score": 7.858823529411764,
|
| 177 |
+
"creativity_score": 7.3,
|
| 178 |
+
"depth_score": 7.7176470588235295,
|
| 179 |
+
"helpfulness_score": 7.5588235294117645,
|
| 180 |
+
"overall_score": 7.482352941176471
|
| 181 |
+
},
|
| 182 |
+
"Gemini-3.1-Flash": {
|
| 183 |
+
"comprehension_score": 7.041176470588235,
|
| 184 |
+
"structure_score": 7.1647058823529415,
|
| 185 |
+
"prose_style_score": 7.411764705882353,
|
| 186 |
+
"creativity_score": 7.294117647058823,
|
| 187 |
+
"depth_score": 7.0,
|
| 188 |
+
"helpfulness_score": 7.064705882352941,
|
| 189 |
+
"overall_score": 7.064705882352941
|
| 190 |
+
},
|
| 191 |
+
"Gemini-3.1-Pro": {
|
| 192 |
+
"comprehension_score": 7.511764705882353,
|
| 193 |
+
"structure_score": 7.5,
|
| 194 |
+
"prose_style_score": 7.682352941176471,
|
| 195 |
+
"creativity_score": 7.5,
|
| 196 |
+
"depth_score": 7.305882352941176,
|
| 197 |
+
"helpfulness_score": 7.552941176470588,
|
| 198 |
+
"overall_score": 7.452941176470588
|
| 199 |
+
},
|
| 200 |
+
"MiniMax-M2.5": {
|
| 201 |
+
"comprehension_score": 6.176470588235294,
|
| 202 |
+
"structure_score": 6.323529411764706,
|
| 203 |
+
"prose_style_score": 5.511764705882353,
|
| 204 |
+
"creativity_score": 6.011764705882353,
|
| 205 |
+
"depth_score": 5.788235294117647,
|
| 206 |
+
"helpfulness_score": 6.035294117647059,
|
| 207 |
+
"overall_score": 5.905882352941177
|
| 208 |
+
},
|
| 209 |
+
"Qwen3.5-Plus": {
|
| 210 |
+
"comprehension_score": 7.205882352941177,
|
| 211 |
+
"structure_score": 7.2823529411764705,
|
| 212 |
+
"prose_style_score": 7.41764705882353,
|
| 213 |
+
"creativity_score": 7.394117647058824,
|
| 214 |
+
"depth_score": 7.205882352941177,
|
| 215 |
+
"helpfulness_score": 7.229411764705882,
|
| 216 |
+
"overall_score": 7.2
|
| 217 |
+
},
|
| 218 |
+
"step-3.5-flash": {
|
| 219 |
+
"comprehension_score": 7.258823529411765,
|
| 220 |
+
"structure_score": 7.252941176470588,
|
| 221 |
+
"prose_style_score": 7.105882352941176,
|
| 222 |
+
"creativity_score": 7.8352941176470585,
|
| 223 |
+
"depth_score": 7.552941176470588,
|
| 224 |
+
"helpfulness_score": 7.1647058823529415,
|
| 225 |
+
"overall_score": 7.229411764705882
|
| 226 |
+
},
|
| 227 |
+
"GLM-5": {
|
| 228 |
+
"comprehension_score": 7.064705882352941,
|
| 229 |
+
"structure_score": 7.147058823529412,
|
| 230 |
+
"prose_style_score": 7.176470588235294,
|
| 231 |
+
"creativity_score": 7.1,
|
| 232 |
+
"depth_score": 6.8882352941176475,
|
| 233 |
+
"helpfulness_score": 7.064705882352941,
|
| 234 |
+
"overall_score": 7.011764705882353
|
| 235 |
+
}
|
| 236 |
+
}
|
index.html
ADDED
|
@@ -0,0 +1,101 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
<!DOCTYPE html>
|
| 2 |
+
<html lang="en">
|
| 3 |
+
<head>
|
| 4 |
+
<meta charset="UTF-8">
|
| 5 |
+
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
| 6 |
+
<title>Zhiyin</title>
|
| 7 |
+
<link rel="stylesheet" href="styles.css">
|
| 8 |
+
<link href="https://fonts.googleapis.com/css2?family=Inter:wght@300;400;500;600;700&display=swap" rel="stylesheet">
|
| 9 |
+
<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/font-awesome/6.0.0/css/all.min.css">
|
| 10 |
+
</head>
|
| 11 |
+
<body>
|
| 12 |
+
<div class="container">
|
| 13 |
+
<header class="header">
|
| 14 |
+
<h1><i class="fas fa-feather"></i> Zhiyin</h1>
|
| 15 |
+
<p>Exploring the Frontier of Chinese LLM Writing</p>
|
| 16 |
+
</header>
|
| 17 |
+
|
| 18 |
+
<div class="dashboard card">
|
| 19 |
+
<section class="overview-section">
|
| 20 |
+
<h2 class="section-title">
|
| 21 |
+
<span class="accent-bar"></span>
|
| 22 |
+
Benchmark Overview
|
| 23 |
+
<span class="section-title-spacer"></span>
|
| 24 |
+
<span class="external-links-text">
|
| 25 |
+
<a href="https://github.com/zake7749/Chinese-Writing-Bench" target="_blank" rel="noopener" class="external-link-text">GitHub</a>
|
| 26 |
+
<span class="divider">•</span>
|
| 27 |
+
<a href="https://huggingface.co/datasets/zake7749/chinese-writing-benchmark" target="_blank" rel="noopener" class="external-link-text">Hugging Face</a>
|
| 28 |
+
</span>
|
| 29 |
+
</h2>
|
| 30 |
+
<div class="overview-content">
|
| 31 |
+
<div class="overview-text">
|
| 32 |
+
<p>
|
| 33 |
+
<strong>Zhiyin</strong> is an LLM-as-a-judge benchmark for Chinese writing evaluation, featuring 280 test cases across 18 diverse writing tasks in this V1 release.
|
| 34 |
+
</p>
|
| 35 |
+
<p>
|
| 36 |
+
Our method relies on <strong>pairwise comparison</strong>. A powerful language model acts as the judge, scoring a model's response relative to a fixed baseline (GPT-4.1), which is anchored at a score of 5.
|
| 37 |
+
</p>
|
| 38 |
+
|
| 39 |
+
<h4>Scoring System</h4>
|
| 40 |
+
<p>The judge assigns the model's response an integer score from 0 to 10, where:</p>
|
| 41 |
+
<ul>
|
| 42 |
+
<li>A score > 5 indicates the response is <strong>superior</strong> to the baseline.</li>
|
| 43 |
+
<li>A score = 5 indicates the response is <strong>on par</strong> with the baseline.</li>
|
| 44 |
+
<li>A score < 5 indicates the response is <strong>inferior</strong> to the baseline.</li>
|
| 45 |
+
</ul>
|
| 46 |
+
|
| 47 |
+
<h4>Evaluation Dimensions</h4>
|
| 48 |
+
<p>To ensure a comprehensive analysis, the final score is informed by a multi-dimensional assessment. The judge evaluates the response across six key criteria:</p>
|
| 49 |
+
<ol>
|
| 50 |
+
<li><strong>Comprehension & Relevance:</strong> How well the response understands the prompt's intent and stays on topic.</li>
|
| 51 |
+
<li><strong>Structure & Coherence:</strong> How clear, logical, and well-organized the writing is.</li>
|
| 52 |
+
<li><strong>Prose & Style:</strong> The quality of the language, grammar, and adherence to the requested tone.</li>
|
| 53 |
+
<li><strong>Creativity & Originality:</strong> The novelty of the ideas and the uniqueness of the perspective.</li>
|
| 54 |
+
<li><strong>Depth & Insight:</strong> The level of detail, analysis, and substance provided.</li>
|
| 55 |
+
<li><strong>Helpfulness:</strong> How effectively the response fulfills the user's overall goal.</li>
|
| 56 |
+
</ol>
|
| 57 |
+
</div>
|
| 58 |
+
</div>
|
| 59 |
+
</section>
|
| 60 |
+
|
| 61 |
+
<section class="judge-section">
|
| 62 |
+
<h2 class="section-title"><span class="accent-bar"></span>Judge Model</h2>
|
| 63 |
+
<div class="judge-toggle">
|
| 64 |
+
<button class="judge-btn active" data-judge="o3">O3</button>
|
| 65 |
+
<button class="judge-btn" data-judge="gpt5.4">GPT-5.4</button>
|
| 66 |
+
</div>
|
| 67 |
+
</section>
|
| 68 |
+
|
| 69 |
+
<section class="table-section">
|
| 70 |
+
<h2 class="section-title"><span class="accent-bar"></span>All Writing Tasks</h2>
|
| 71 |
+
<div id="generalTable" class="table-container">
|
| 72 |
+
<!-- General Writing table will be populated here -->
|
| 73 |
+
</div>
|
| 74 |
+
</section>
|
| 75 |
+
|
| 76 |
+
<section class="table-section">
|
| 77 |
+
<h2 class="section-title"><span class="accent-bar"></span>Complicated Writing Tasks</h2>
|
| 78 |
+
<div id="complicatedTable" class="table-container">
|
| 79 |
+
<!-- Complicated Writing table will be populated here -->
|
| 80 |
+
</div>
|
| 81 |
+
</section>
|
| 82 |
+
|
| 83 |
+
<section class="citation-section">
|
| 84 |
+
<h2 class="section-title"><span class="accent-bar"></span>Citation</h2>
|
| 85 |
+
<div class="citation-content">
|
| 86 |
+
<p>
|
| 87 |
+
If you use these results, please cite our paper:<br>
|
| 88 |
+
<em>"Zhiyin: Exploring the Frontier of Chinese LLM Writing, 2025. https://github.com/zake7749/Chinese-Writing-Bench"</em>
|
| 89 |
+
</p>
|
| 90 |
+
</div>
|
| 91 |
+
</section>
|
| 92 |
+
</div>
|
| 93 |
+
|
| 94 |
+
<div id="loading" class="loading">
|
| 95 |
+
<div class="spinner"></div>
|
| 96 |
+
<p>Loading benchmark data...</p>
|
| 97 |
+
</div>
|
| 98 |
+
</div>
|
| 99 |
+
<script src="script.js"></script>
|
| 100 |
+
</body>
|
| 101 |
+
</html>
|
script.js
ADDED
|
@@ -0,0 +1,206 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
class LLMBenchmarkDashboard {
|
| 2 |
+
constructor() {
|
| 3 |
+
this.currentJudge = 'o3';
|
| 4 |
+
this.judgeData = {
|
| 5 |
+
o3: { general: null, complicated: null },
|
| 6 |
+
'gpt5.4': { general: null, complicated: null }
|
| 7 |
+
};
|
| 8 |
+
this.generalSort = { column: 'overall_score', direction: 'desc' };
|
| 9 |
+
this.complicatedSort = { column: 'overall_score', direction: 'desc' };
|
| 10 |
+
this.metricDisplayNames = {
|
| 11 |
+
comprehension_score: 'Comprehension',
|
| 12 |
+
structure_score: 'Coherence',
|
| 13 |
+
prose_style_score: 'Style',
|
| 14 |
+
creativity_score: 'Creativity',
|
| 15 |
+
depth_score: 'Depth',
|
| 16 |
+
helpfulness_score: 'Helpfulness',
|
| 17 |
+
overall_score: 'Overall'
|
| 18 |
+
};
|
| 19 |
+
this.judgePaths = {
|
| 20 |
+
o3: {
|
| 21 |
+
general: 'data/all-scores.json',
|
| 22 |
+
complicated: 'data/complicated-writing-scores.json'
|
| 23 |
+
},
|
| 24 |
+
'gpt5.4': {
|
| 25 |
+
general: 'data/gpt5.4-judge-all-scores.json',
|
| 26 |
+
complicated: 'data/gpt5.4-judge-complicated-writing-scores.json'
|
| 27 |
+
}
|
| 28 |
+
};
|
| 29 |
+
this.modelLinks = {
|
| 30 |
+
'Monomer-24B-Writer': 'https://huggingface.co/zake7749/Monomer-24B-Writer-Preview',
|
| 31 |
+
'Monomer-8B-Writer': 'https://huggingface.co/zake7749/Monomer-8B-Writer-Preview'
|
| 32 |
+
};
|
| 33 |
+
this.init();
|
| 34 |
+
}
|
| 35 |
+
|
| 36 |
+
async init() {
|
| 37 |
+
this.showLoading(true);
|
| 38 |
+
const promises = [];
|
| 39 |
+
for (const judge of ['o3', 'gpt5.4']) {
|
| 40 |
+
for (const type of ['general', 'complicated']) {
|
| 41 |
+
promises.push(this.loadData(this.judgePaths[judge][type], judge, type));
|
| 42 |
+
}
|
| 43 |
+
}
|
| 44 |
+
await Promise.all(promises);
|
| 45 |
+
this.renderTable('general');
|
| 46 |
+
this.renderTable('complicated');
|
| 47 |
+
this.setupJudgeToggle();
|
| 48 |
+
this.showLoading(false);
|
| 49 |
+
}
|
| 50 |
+
|
| 51 |
+
async loadData(path, judge, type) {
|
| 52 |
+
try {
|
| 53 |
+
const response = await fetch(path);
|
| 54 |
+
if (!response.ok) {
|
| 55 |
+
throw new Error(`HTTP error! status: ${response.status}`);
|
| 56 |
+
}
|
| 57 |
+
this.judgeData[judge][type] = await response.json();
|
| 58 |
+
} catch (error) {
|
| 59 |
+
console.error(`Error loading ${judge}/${type} data:`, error);
|
| 60 |
+
}
|
| 61 |
+
}
|
| 62 |
+
|
| 63 |
+
get generalData() {
|
| 64 |
+
return this.judgeData[this.currentJudge].general;
|
| 65 |
+
}
|
| 66 |
+
|
| 67 |
+
get complicatedData() {
|
| 68 |
+
return this.judgeData[this.currentJudge].complicated;
|
| 69 |
+
}
|
| 70 |
+
|
| 71 |
+
setupJudgeToggle() {
|
| 72 |
+
const buttons = document.querySelectorAll('.judge-btn');
|
| 73 |
+
buttons.forEach(btn => {
|
| 74 |
+
btn.addEventListener('click', () => {
|
| 75 |
+
const judge = btn.dataset.judge;
|
| 76 |
+
if (judge === this.currentJudge) return;
|
| 77 |
+
|
| 78 |
+
this.currentJudge = judge;
|
| 79 |
+
buttons.forEach(b => b.classList.remove('active'));
|
| 80 |
+
btn.classList.add('active');
|
| 81 |
+
|
| 82 |
+
this.renderTable('general');
|
| 83 |
+
this.renderTable('complicated');
|
| 84 |
+
});
|
| 85 |
+
});
|
| 86 |
+
}
|
| 87 |
+
|
| 88 |
+
renderTable(type) {
|
| 89 |
+
const data = type === 'general' ? this.generalData : this.complicatedData;
|
| 90 |
+
const sortState = type === 'general' ? this.generalSort : this.complicatedSort;
|
| 91 |
+
const tableContainer = document.getElementById(type === 'general' ? 'generalTable' : 'complicatedTable');
|
| 92 |
+
if (!data) return;
|
| 93 |
+
|
| 94 |
+
const models = Object.keys(data);
|
| 95 |
+
const metrics = Object.keys(data[models[0]]);
|
| 96 |
+
|
| 97 |
+
const tableHTML = `
|
| 98 |
+
<table>
|
| 99 |
+
<thead>
|
| 100 |
+
<tr>
|
| 101 |
+
<th class="sortable${sortState.column === 'model' ? ' sort-' + sortState.direction : ''}" data-type="${type}" data-column="model">Model</th>
|
| 102 |
+
${metrics.map(metric => `
|
| 103 |
+
<th class="sortable${sortState.column === metric ? ' sort-' + sortState.direction : ''}" data-type="${type}" data-column="${metric}">${this.metricDisplayNames[metric] || metric}</th>
|
| 104 |
+
`).join('')}
|
| 105 |
+
</tr>
|
| 106 |
+
</thead>
|
| 107 |
+
<tbody>
|
| 108 |
+
${this.getSortedTableData(data, sortState, metrics).map(row => {
|
| 109 |
+
const isMonomer = this.modelLinks[row.model];
|
| 110 |
+
return `
|
| 111 |
+
<tr${isMonomer ? ' class="highlight-row"' : ''}>
|
| 112 |
+
<td class="model-cell">${isMonomer ? `<a href="${this.modelLinks[row.model]}" target="_blank" rel="noopener" class="model-link">${row.model}</a>` : row.model}</td>
|
| 113 |
+
${metrics.map(metric => `
|
| 114 |
+
<td class="score-cell">${this.formatScore(row[metric])}</td>
|
| 115 |
+
`).join('')}
|
| 116 |
+
</tr>
|
| 117 |
+
`;
|
| 118 |
+
}).join('')}
|
| 119 |
+
</tbody>
|
| 120 |
+
</table>
|
| 121 |
+
`;
|
| 122 |
+
|
| 123 |
+
tableContainer.innerHTML = tableHTML;
|
| 124 |
+
this.setupTableSorting(type);
|
| 125 |
+
}
|
| 126 |
+
|
| 127 |
+
getSortedTableData(data, sortState, metrics) {
|
| 128 |
+
const models = Object.keys(data);
|
| 129 |
+
let tableData = models.map(model => ({
|
| 130 |
+
model,
|
| 131 |
+
...data[model]
|
| 132 |
+
}));
|
| 133 |
+
|
| 134 |
+
if (sortState.column) {
|
| 135 |
+
tableData.sort((a, b) => {
|
| 136 |
+
let aVal = a[sortState.column];
|
| 137 |
+
let bVal = b[sortState.column];
|
| 138 |
+
if (sortState.column === 'model') {
|
| 139 |
+
aVal = aVal.toLowerCase();
|
| 140 |
+
bVal = bVal.toLowerCase();
|
| 141 |
+
}
|
| 142 |
+
if (aVal < bVal) return sortState.direction === 'asc' ? -1 : 1;
|
| 143 |
+
if (aVal > bVal) return sortState.direction === 'asc' ? 1 : -1;
|
| 144 |
+
return 0;
|
| 145 |
+
});
|
| 146 |
+
}
|
| 147 |
+
return tableData;
|
| 148 |
+
}
|
| 149 |
+
|
| 150 |
+
formatScore(value) {
|
| 151 |
+
if (typeof value === 'number') {
|
| 152 |
+
return value.toFixed(2);
|
| 153 |
+
}
|
| 154 |
+
return value;
|
| 155 |
+
}
|
| 156 |
+
|
| 157 |
+
setupTableSorting(type) {
|
| 158 |
+
const tableContainer = document.getElementById(type === 'general' ? 'generalTable' : 'complicatedTable');
|
| 159 |
+
const headers = tableContainer.querySelectorAll('th.sortable');
|
| 160 |
+
headers.forEach(header => {
|
| 161 |
+
header.addEventListener('click', () => {
|
| 162 |
+
const column = header.dataset.column;
|
| 163 |
+
this.handleSort(type, column, header);
|
| 164 |
+
});
|
| 165 |
+
});
|
| 166 |
+
}
|
| 167 |
+
|
| 168 |
+
handleSort(type, column, headerElement) {
|
| 169 |
+
const sortState = type === 'general' ? this.generalSort : this.complicatedSort;
|
| 170 |
+
if (sortState.column === column) {
|
| 171 |
+
sortState.direction = sortState.direction === 'asc' ? 'desc' : 'asc';
|
| 172 |
+
} else {
|
| 173 |
+
sortState.column = column;
|
| 174 |
+
sortState.direction = 'asc';
|
| 175 |
+
}
|
| 176 |
+
const tableContainer = document.getElementById(type === 'general' ? 'generalTable' : 'complicatedTable');
|
| 177 |
+
tableContainer.querySelectorAll('th.sortable').forEach(th => {
|
| 178 |
+
th.classList.remove('sort-asc', 'sort-desc');
|
| 179 |
+
});
|
| 180 |
+
headerElement.classList.add(`sort-${sortState.direction}`);
|
| 181 |
+
this.renderTable(type);
|
| 182 |
+
}
|
| 183 |
+
|
| 184 |
+
showLoading(show) {
|
| 185 |
+
const loading = document.getElementById('loading');
|
| 186 |
+
if (show) {
|
| 187 |
+
loading.classList.remove('hidden');
|
| 188 |
+
} else {
|
| 189 |
+
loading.classList.add('hidden');
|
| 190 |
+
}
|
| 191 |
+
}
|
| 192 |
+
|
| 193 |
+
showError(message) {
|
| 194 |
+
const loading = document.getElementById('loading');
|
| 195 |
+
loading.innerHTML = `
|
| 196 |
+
<div class="no-data">
|
| 197 |
+
<i class="fas fa-exclamation-triangle"></i>
|
| 198 |
+
<p>${message}</p>
|
| 199 |
+
</div>
|
| 200 |
+
`;
|
| 201 |
+
}
|
| 202 |
+
}
|
| 203 |
+
|
| 204 |
+
document.addEventListener('DOMContentLoaded', () => {
|
| 205 |
+
new LLMBenchmarkDashboard();
|
| 206 |
+
});
|
styles.css
ADDED
|
@@ -0,0 +1,458 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
* {
|
| 2 |
+
margin: 0;
|
| 3 |
+
padding: 0;
|
| 4 |
+
box-sizing: border-box;
|
| 5 |
+
}
|
| 6 |
+
|
| 7 |
+
body {
|
| 8 |
+
font-family: 'Inter', sans-serif;
|
| 9 |
+
background: linear-gradient(135deg, #667eea 0%, #764ba2 100%);
|
| 10 |
+
min-height: 100vh;
|
| 11 |
+
color: #333;
|
| 12 |
+
}
|
| 13 |
+
|
| 14 |
+
.container {
|
| 15 |
+
max-width: 1400px;
|
| 16 |
+
margin: 0 auto;
|
| 17 |
+
padding: 20px;
|
| 18 |
+
}
|
| 19 |
+
|
| 20 |
+
.header {
|
| 21 |
+
text-align: center;
|
| 22 |
+
margin-bottom: 24px;
|
| 23 |
+
color: white;
|
| 24 |
+
padding: 48px 0 36px 0;
|
| 25 |
+
position: relative;
|
| 26 |
+
background: none;
|
| 27 |
+
box-shadow: 0 2px 12px 0 rgba(76, 75, 162, 0.06);
|
| 28 |
+
}
|
| 29 |
+
|
| 30 |
+
.header h1 {
|
| 31 |
+
font-size: 3.2rem;
|
| 32 |
+
font-weight: 800;
|
| 33 |
+
margin-bottom: 18px;
|
| 34 |
+
letter-spacing: 0.12em;
|
| 35 |
+
display: flex;
|
| 36 |
+
align-items: center;
|
| 37 |
+
justify-content: center;
|
| 38 |
+
gap: 18px;
|
| 39 |
+
}
|
| 40 |
+
|
| 41 |
+
.header h1 .fas {
|
| 42 |
+
font-size: 2.2rem;
|
| 43 |
+
background: linear-gradient(135deg, #764ba2 60%, #667eea 100%);
|
| 44 |
+
color: white;
|
| 45 |
+
border-radius: 50%;
|
| 46 |
+
padding: 16px 18px;
|
| 47 |
+
box-shadow: 0 2px 12px #764ba244;
|
| 48 |
+
vertical-align: middle;
|
| 49 |
+
}
|
| 50 |
+
|
| 51 |
+
.header p {
|
| 52 |
+
font-size: 1.25rem;
|
| 53 |
+
font-weight: 400;
|
| 54 |
+
color: #ece6fa;
|
| 55 |
+
opacity: 0.92;
|
| 56 |
+
margin-top: 0;
|
| 57 |
+
margin-bottom: 0;
|
| 58 |
+
letter-spacing: 0.04em;
|
| 59 |
+
}
|
| 60 |
+
|
| 61 |
+
.card {
|
| 62 |
+
background: white;
|
| 63 |
+
border-radius: 16px;
|
| 64 |
+
box-shadow: 0 6px 32px 0 rgba(76, 75, 162, 0.08);
|
| 65 |
+
border: 1.5px solid #ece6fa;
|
| 66 |
+
padding: 32px 32px 24px 32px;
|
| 67 |
+
}
|
| 68 |
+
|
| 69 |
+
.section-title {
|
| 70 |
+
display: flex;
|
| 71 |
+
align-items: center;
|
| 72 |
+
font-size: 1.35rem;
|
| 73 |
+
font-weight: 700;
|
| 74 |
+
color: #4b2996;
|
| 75 |
+
margin-bottom: 28px;
|
| 76 |
+
letter-spacing: 0.5px;
|
| 77 |
+
text-transform: none;
|
| 78 |
+
background: none;
|
| 79 |
+
border: none;
|
| 80 |
+
padding-left: 0;
|
| 81 |
+
position: relative;
|
| 82 |
+
padding-bottom: 8px;
|
| 83 |
+
border-bottom: 2px solid #ecf0f1;
|
| 84 |
+
}
|
| 85 |
+
|
| 86 |
+
.section-title-spacer {
|
| 87 |
+
flex: 1 1 auto;
|
| 88 |
+
}
|
| 89 |
+
|
| 90 |
+
.accent-bar {
|
| 91 |
+
display: inline-block;
|
| 92 |
+
width: 7px;
|
| 93 |
+
height: 28px;
|
| 94 |
+
border-radius: 4px;
|
| 95 |
+
background: linear-gradient(180deg, #764ba2 0%, #667eea 100%);
|
| 96 |
+
margin-right: 16px;
|
| 97 |
+
}
|
| 98 |
+
|
| 99 |
+
.overview-section, .table-section, .citation-section {
|
| 100 |
+
margin-bottom: 56px;
|
| 101 |
+
}
|
| 102 |
+
|
| 103 |
+
.overview-content {
|
| 104 |
+
font-size: 1.08rem;
|
| 105 |
+
color: #3a3550;
|
| 106 |
+
line-height: 1.7;
|
| 107 |
+
padding-left: 2px;
|
| 108 |
+
border-left: 4px solid #ece6fa;
|
| 109 |
+
padding-left: 24px;
|
| 110 |
+
background: none;
|
| 111 |
+
}
|
| 112 |
+
|
| 113 |
+
.overview-content p {
|
| 114 |
+
margin-bottom: 18px;
|
| 115 |
+
margin-top: 0;
|
| 116 |
+
}
|
| 117 |
+
|
| 118 |
+
.overview-content h4 {
|
| 119 |
+
font-size: 1.13rem;
|
| 120 |
+
font-weight: 700;
|
| 121 |
+
color: #764ba2;
|
| 122 |
+
margin-top: 28px;
|
| 123 |
+
margin-bottom: 10px;
|
| 124 |
+
letter-spacing: 0.02em;
|
| 125 |
+
}
|
| 126 |
+
|
| 127 |
+
.overview-content ul,
|
| 128 |
+
.overview-content ol {
|
| 129 |
+
margin: 0 0 18px 24px;
|
| 130 |
+
padding-left: 18px;
|
| 131 |
+
}
|
| 132 |
+
|
| 133 |
+
.overview-content ul li,
|
| 134 |
+
.overview-content ol li {
|
| 135 |
+
margin-bottom: 8px;
|
| 136 |
+
line-height: 1.6;
|
| 137 |
+
font-size: 1.05rem;
|
| 138 |
+
}
|
| 139 |
+
|
| 140 |
+
.overview-content ul li:last-child,
|
| 141 |
+
.overview-content ol li:last-child {
|
| 142 |
+
margin-bottom: 0;
|
| 143 |
+
}
|
| 144 |
+
|
| 145 |
+
.citation-content {
|
| 146 |
+
font-size: 1.08rem;
|
| 147 |
+
color: #3a3550;
|
| 148 |
+
line-height: 1.7;
|
| 149 |
+
padding-left: 2px;
|
| 150 |
+
}
|
| 151 |
+
|
| 152 |
+
.citation-content em {
|
| 153 |
+
color: #764ba2;
|
| 154 |
+
font-size: 1rem;
|
| 155 |
+
font-style: italic;
|
| 156 |
+
background: #f3f0fa;
|
| 157 |
+
padding: 2px 6px;
|
| 158 |
+
border-radius: 4px;
|
| 159 |
+
}
|
| 160 |
+
|
| 161 |
+
.table-container {
|
| 162 |
+
overflow-x: auto;
|
| 163 |
+
background: #f8f9fa;
|
| 164 |
+
border-radius: 12px;
|
| 165 |
+
padding: 20px;
|
| 166 |
+
max-height: 480px;
|
| 167 |
+
overflow-y: auto;
|
| 168 |
+
}
|
| 169 |
+
|
| 170 |
+
table {
|
| 171 |
+
width: 100%;
|
| 172 |
+
border-collapse: collapse;
|
| 173 |
+
background: white;
|
| 174 |
+
border-radius: 8px;
|
| 175 |
+
overflow: hidden;
|
| 176 |
+
box-shadow: 0 2px 8px rgba(0,0,0,0.1);
|
| 177 |
+
}
|
| 178 |
+
|
| 179 |
+
th, td {
|
| 180 |
+
padding: 14px 18px;
|
| 181 |
+
border-bottom: 1px solid #e9ecef;
|
| 182 |
+
}
|
| 183 |
+
|
| 184 |
+
th {
|
| 185 |
+
background: #764ba2;
|
| 186 |
+
color: white;
|
| 187 |
+
font-weight: 700;
|
| 188 |
+
font-size: 1rem;
|
| 189 |
+
cursor: pointer;
|
| 190 |
+
transition: background-color 0.3s ease;
|
| 191 |
+
position: sticky;
|
| 192 |
+
top: 0;
|
| 193 |
+
z-index: 2;
|
| 194 |
+
text-transform: none;
|
| 195 |
+
letter-spacing: 0.1em;
|
| 196 |
+
padding-right: 32px;
|
| 197 |
+
}
|
| 198 |
+
|
| 199 |
+
th:hover {
|
| 200 |
+
background: #5e388e;
|
| 201 |
+
}
|
| 202 |
+
|
| 203 |
+
th.sortable::after {
|
| 204 |
+
content: '';
|
| 205 |
+
display: inline-block;
|
| 206 |
+
position: absolute;
|
| 207 |
+
right: 12px;
|
| 208 |
+
top: 50%;
|
| 209 |
+
transform: translateY(-50%);
|
| 210 |
+
width: 16px;
|
| 211 |
+
height: 16px;
|
| 212 |
+
background-repeat: no-repeat;
|
| 213 |
+
background-position: center;
|
| 214 |
+
background-size: 16px 16px;
|
| 215 |
+
opacity: 0.7;
|
| 216 |
+
/* Default: double chevron (unsorted) */
|
| 217 |
+
background-image: url('data:image/svg+xml;utf8,<svg width="16" height="16" fill="none" xmlns="http://www.w3.org/2000/svg"><path d="M4 6l4-4 4 4" stroke="white" stroke-width="2" stroke-linecap="round" stroke-linejoin="round"/><path d="M4 10l4 4 4-4" stroke="white" stroke-width="2" stroke-linecap="round" stroke-linejoin="round"/></svg>');
|
| 218 |
+
}
|
| 219 |
+
|
| 220 |
+
th.sort-asc::after {
|
| 221 |
+
/* Up chevron */
|
| 222 |
+
background-image: url('data:image/svg+xml;utf8,<svg width="16" height="16" fill="none" xmlns="http://www.w3.org/2000/svg"><path d="M4 10l4-4 4 4" stroke="white" stroke-width="2" stroke-linecap="round" stroke-linejoin="round"/></svg>');
|
| 223 |
+
opacity: 1;
|
| 224 |
+
}
|
| 225 |
+
|
| 226 |
+
th.sort-desc::after {
|
| 227 |
+
/* Down chevron */
|
| 228 |
+
background-image: url('data:image/svg+xml;utf8,<svg width="16" height="16" fill="none" xmlns="http://www.w3.org/2000/svg"><path d="M4 6l4 4 4-4" stroke="white" stroke-width="2" stroke-linecap="round" stroke-linejoin="round"/></svg>');
|
| 229 |
+
opacity: 1;
|
| 230 |
+
}
|
| 231 |
+
|
| 232 |
+
.model-cell, th:first-child {
|
| 233 |
+
text-align: left;
|
| 234 |
+
}
|
| 235 |
+
|
| 236 |
+
td:not(.model-cell), th:not(:first-child) {
|
| 237 |
+
text-align: right;
|
| 238 |
+
}
|
| 239 |
+
|
| 240 |
+
td {
|
| 241 |
+
font-size: 1rem;
|
| 242 |
+
color: #4b2996;
|
| 243 |
+
font-weight: 500;
|
| 244 |
+
}
|
| 245 |
+
|
| 246 |
+
tr:hover {
|
| 247 |
+
background: #f3f0fa;
|
| 248 |
+
}
|
| 249 |
+
|
| 250 |
+
.score-cell {
|
| 251 |
+
font-weight: 700;
|
| 252 |
+
color: #764ba2;
|
| 253 |
+
}
|
| 254 |
+
|
| 255 |
+
.model-cell {
|
| 256 |
+
font-weight: 700;
|
| 257 |
+
color: #495057;
|
| 258 |
+
}
|
| 259 |
+
|
| 260 |
+
.loading {
|
| 261 |
+
position: fixed;
|
| 262 |
+
top: 0;
|
| 263 |
+
left: 0;
|
| 264 |
+
width: 100%;
|
| 265 |
+
height: 100%;
|
| 266 |
+
background: rgba(0,0,0,0.8);
|
| 267 |
+
display: flex;
|
| 268 |
+
flex-direction: column;
|
| 269 |
+
justify-content: center;
|
| 270 |
+
align-items: center;
|
| 271 |
+
z-index: 1000;
|
| 272 |
+
color: white;
|
| 273 |
+
}
|
| 274 |
+
|
| 275 |
+
.loading.hidden {
|
| 276 |
+
display: none;
|
| 277 |
+
}
|
| 278 |
+
|
| 279 |
+
.spinner {
|
| 280 |
+
width: 50px;
|
| 281 |
+
height: 50px;
|
| 282 |
+
border: 4px solid #f3f3f3;
|
| 283 |
+
border-top: 4px solid #764ba2;
|
| 284 |
+
border-radius: 50%;
|
| 285 |
+
animation: spin 1s linear infinite;
|
| 286 |
+
margin-bottom: 20px;
|
| 287 |
+
}
|
| 288 |
+
|
| 289 |
+
@keyframes spin {
|
| 290 |
+
0% { transform: rotate(0deg); }
|
| 291 |
+
100% { transform: rotate(360deg); }
|
| 292 |
+
}
|
| 293 |
+
|
| 294 |
+
.no-data {
|
| 295 |
+
text-align: center;
|
| 296 |
+
padding: 40px;
|
| 297 |
+
color: #6c757d;
|
| 298 |
+
font-size: 1.1rem;
|
| 299 |
+
}
|
| 300 |
+
|
| 301 |
+
.no-data i {
|
| 302 |
+
font-size: 3rem;
|
| 303 |
+
margin-bottom: 20px;
|
| 304 |
+
opacity: 0.5;
|
| 305 |
+
}
|
| 306 |
+
|
| 307 |
+
@media (max-width: 768px) {
|
| 308 |
+
.container {
|
| 309 |
+
padding: 10px;
|
| 310 |
+
}
|
| 311 |
+
.header h1 {
|
| 312 |
+
font-size: 2rem;
|
| 313 |
+
}
|
| 314 |
+
.dashboard.card {
|
| 315 |
+
padding: 12px 4px 12px 4px;
|
| 316 |
+
}
|
| 317 |
+
.card {
|
| 318 |
+
padding: 12px 4px 12px 4px;
|
| 319 |
+
}
|
| 320 |
+
.table-container {
|
| 321 |
+
padding: 10px;
|
| 322 |
+
}
|
| 323 |
+
th, td {
|
| 324 |
+
padding: 8px 8px;
|
| 325 |
+
font-size: 0.92rem;
|
| 326 |
+
}
|
| 327 |
+
.overview-section, .table-section, .citation-section {
|
| 328 |
+
margin-bottom: 36px;
|
| 329 |
+
}
|
| 330 |
+
}
|
| 331 |
+
|
| 332 |
+
.judge-section {
|
| 333 |
+
margin-bottom: 36px;
|
| 334 |
+
}
|
| 335 |
+
|
| 336 |
+
.judge-toggle {
|
| 337 |
+
display: flex;
|
| 338 |
+
gap: 0;
|
| 339 |
+
border-radius: 10px;
|
| 340 |
+
overflow: hidden;
|
| 341 |
+
border: 2px solid #764ba2;
|
| 342 |
+
width: fit-content;
|
| 343 |
+
}
|
| 344 |
+
|
| 345 |
+
.judge-btn {
|
| 346 |
+
padding: 10px 28px;
|
| 347 |
+
font-size: 1rem;
|
| 348 |
+
font-weight: 700;
|
| 349 |
+
font-family: 'Inter', sans-serif;
|
| 350 |
+
cursor: pointer;
|
| 351 |
+
border: none;
|
| 352 |
+
background: white;
|
| 353 |
+
color: #764ba2;
|
| 354 |
+
transition: background 0.2s, color 0.2s;
|
| 355 |
+
letter-spacing: 0.03em;
|
| 356 |
+
}
|
| 357 |
+
|
| 358 |
+
.judge-btn:not(:last-child) {
|
| 359 |
+
border-right: 2px solid #764ba2;
|
| 360 |
+
}
|
| 361 |
+
|
| 362 |
+
.judge-btn:hover {
|
| 363 |
+
background: #f3eaff;
|
| 364 |
+
}
|
| 365 |
+
|
| 366 |
+
.judge-btn.active {
|
| 367 |
+
background: linear-gradient(135deg, #764ba2 0%, #667eea 100%);
|
| 368 |
+
color: white;
|
| 369 |
+
}
|
| 370 |
+
|
| 371 |
+
.highlight-row {
|
| 372 |
+
background: #f3eaff !important;
|
| 373 |
+
box-shadow: 0 2px 8px 0 rgba(118, 75, 162, 0.08);
|
| 374 |
+
}
|
| 375 |
+
|
| 376 |
+
.model-link {
|
| 377 |
+
color: #764ba2;
|
| 378 |
+
font-weight: 700;
|
| 379 |
+
text-decoration: none;
|
| 380 |
+
transition: color 0.2s;
|
| 381 |
+
}
|
| 382 |
+
|
| 383 |
+
.model-link:hover {
|
| 384 |
+
color: #4b2996;
|
| 385 |
+
text-decoration: none;
|
| 386 |
+
}
|
| 387 |
+
|
| 388 |
+
.external-links {
|
| 389 |
+
display: flex;
|
| 390 |
+
justify-content: center;
|
| 391 |
+
align-items: center;
|
| 392 |
+
margin-top: 1.2rem;
|
| 393 |
+
margin-bottom: 0.5rem;
|
| 394 |
+
gap: 1.5rem;
|
| 395 |
+
}
|
| 396 |
+
|
| 397 |
+
.external-link {
|
| 398 |
+
display: flex;
|
| 399 |
+
align-items: center;
|
| 400 |
+
color: #764ba2;
|
| 401 |
+
text-decoration: none;
|
| 402 |
+
font-weight: 700;
|
| 403 |
+
transition: color 0.2s, transform 0.2s;
|
| 404 |
+
}
|
| 405 |
+
|
| 406 |
+
.external-link svg {
|
| 407 |
+
width: 2.1rem;
|
| 408 |
+
height: 2.1rem;
|
| 409 |
+
display: block;
|
| 410 |
+
}
|
| 411 |
+
|
| 412 |
+
.external-link:hover {
|
| 413 |
+
color: #4b2996;
|
| 414 |
+
transform: scale(1.08);
|
| 415 |
+
}
|
| 416 |
+
|
| 417 |
+
.external-links-text {
|
| 418 |
+
display: flex;
|
| 419 |
+
justify-content: flex-end;
|
| 420 |
+
align-items: center;
|
| 421 |
+
margin: 0;
|
| 422 |
+
font-size: 1.05rem;
|
| 423 |
+
font-weight: 700;
|
| 424 |
+
gap: 0.5rem;
|
| 425 |
+
width: auto;
|
| 426 |
+
}
|
| 427 |
+
|
| 428 |
+
.external-link-text {
|
| 429 |
+
color: #4b2996;
|
| 430 |
+
background: none;
|
| 431 |
+
text-decoration: none;
|
| 432 |
+
font-weight: 700;
|
| 433 |
+
letter-spacing: 0.01em;
|
| 434 |
+
padding: 0.18em 0.7em;
|
| 435 |
+
border-radius: 18px;
|
| 436 |
+
transition: color 0.2s, background 0.2s, box-shadow 0.2s;
|
| 437 |
+
box-shadow: none;
|
| 438 |
+
display: inline-block;
|
| 439 |
+
}
|
| 440 |
+
|
| 441 |
+
.external-link-text:hover {
|
| 442 |
+
color: #fff;
|
| 443 |
+
background: linear-gradient(90deg, #764ba2 60%, #667eea 100%);
|
| 444 |
+
text-decoration: none;
|
| 445 |
+
box-shadow: 0 2px 8px 0 rgba(118, 75, 162, 0.13);
|
| 446 |
+
}
|
| 447 |
+
|
| 448 |
+
.divider {
|
| 449 |
+
display: inline-block;
|
| 450 |
+
color: #bba6d6;
|
| 451 |
+
font-size: 1.1em;
|
| 452 |
+
font-weight: 400;
|
| 453 |
+
margin: 0 0.0em;
|
| 454 |
+
user-select: none;
|
| 455 |
+
vertical-align: middle;
|
| 456 |
+
line-height: 0.5;
|
| 457 |
+
letter-spacing: 0;
|
| 458 |
+
}
|