Update README.md
Browse files
README.md
CHANGED
|
@@ -4,7 +4,6 @@ license: apache-2.0
|
|
| 4 |
|
| 5 |
|
| 6 |
### Accuracy Comparison
|
| 7 |
-
|
| 8 |
<table>
|
| 9 |
<thead>
|
| 10 |
<tr>
|
|
@@ -122,7 +121,7 @@ license: apache-2.0
|
|
| 122 |
</tbody>
|
| 123 |
</table>
|
| 124 |
|
| 125 |
-
|
| 126 |
### Accuracy
|
| 127 |
<table>
|
| 128 |
<thead>
|
|
@@ -135,7 +134,6 @@ license: apache-2.0
|
|
| 135 |
</tr>
|
| 136 |
</thead>
|
| 137 |
<tbody>
|
| 138 |
-
<!-- OpenLLM Leaderboard V1 -->
|
| 139 |
<tr>
|
| 140 |
<td rowspan="7"><b>OpenLLM V1</b></td>
|
| 141 |
<td>ARC-Challenge (Acc-Norm, 25-shot)</td>
|
|
@@ -179,7 +177,6 @@ license: apache-2.0
|
|
| 179 |
<td><b>76.59</b></td>
|
| 180 |
<td><b>99.96</b></td>
|
| 181 |
</tr>
|
| 182 |
-
<!-- OpenLLM Leaderboard V2 -->
|
| 183 |
<tr>
|
| 184 |
<td rowspan="7"><b>OpenLLM V2</b></td>
|
| 185 |
<td>IFEval (Inst Level Strict Acc, 0-shot)</td>
|
|
@@ -241,7 +238,6 @@ license: apache-2.0
|
|
| 241 |
</tr>
|
| 242 |
</thead>
|
| 243 |
<tbody>
|
| 244 |
-
<!-- OpenLLM Leaderboard V1 -->
|
| 245 |
<tr>
|
| 246 |
<td rowspan="7"><b>OpenLLM V1</b></td>
|
| 247 |
<td>ARC-Challenge (Acc-Norm, 25-shot)</td>
|
|
@@ -285,7 +281,6 @@ license: apache-2.0
|
|
| 285 |
<td><b>76.62</b></td>
|
| 286 |
<td><b>100.00</b></td>
|
| 287 |
</tr>
|
| 288 |
-
<!-- OpenLLM Leaderboard V2 -->
|
| 289 |
<tr>
|
| 290 |
<td rowspan="7"><b>OpenLLM V2</b></td>
|
| 291 |
<td>IFEval (Inst Level Strict Acc, 0-shot)</td>
|
|
@@ -345,7 +340,6 @@ license: apache-2.0
|
|
| 345 |
</tr>
|
| 346 |
</thead>
|
| 347 |
<tbody>
|
| 348 |
-
<!-- OpenLLM Leaderboard V1 -->
|
| 349 |
<tr>
|
| 350 |
<td rowspan="7"><b>OpenLLM V1</b></td>
|
| 351 |
<td>ARC-Challenge (Acc-Norm, 25-shot)</td>
|
|
@@ -389,7 +383,6 @@ license: apache-2.0
|
|
| 389 |
<td><b>76.51</b></td>
|
| 390 |
<td><b>99.86</b></td>
|
| 391 |
</tr>
|
| 392 |
-
<!-- OpenLLM Leaderboard V2 -->
|
| 393 |
<tr>
|
| 394 |
<td rowspan="7"><b>OpenLLM V2</b></td>
|
| 395 |
<td>IFEval (Inst Level Strict Acc, 0-shot)</td>
|
|
@@ -434,5 +427,5 @@ license: apache-2.0
|
|
| 434 |
<td><b>97.77</b></td>
|
| 435 |
</tr>
|
| 436 |
</tbody>
|
| 437 |
-
</table>
|
| 438 |
|
|
|
|
| 4 |
|
| 5 |
|
| 6 |
### Accuracy Comparison
|
|
|
|
| 7 |
<table>
|
| 8 |
<thead>
|
| 9 |
<tr>
|
|
|
|
| 121 |
</tbody>
|
| 122 |
</table>
|
| 123 |
|
| 124 |
+
<!--
|
| 125 |
### Accuracy
|
| 126 |
<table>
|
| 127 |
<thead>
|
|
|
|
| 134 |
</tr>
|
| 135 |
</thead>
|
| 136 |
<tbody>
|
|
|
|
| 137 |
<tr>
|
| 138 |
<td rowspan="7"><b>OpenLLM V1</b></td>
|
| 139 |
<td>ARC-Challenge (Acc-Norm, 25-shot)</td>
|
|
|
|
| 177 |
<td><b>76.59</b></td>
|
| 178 |
<td><b>99.96</b></td>
|
| 179 |
</tr>
|
|
|
|
| 180 |
<tr>
|
| 181 |
<td rowspan="7"><b>OpenLLM V2</b></td>
|
| 182 |
<td>IFEval (Inst Level Strict Acc, 0-shot)</td>
|
|
|
|
| 238 |
</tr>
|
| 239 |
</thead>
|
| 240 |
<tbody>
|
|
|
|
| 241 |
<tr>
|
| 242 |
<td rowspan="7"><b>OpenLLM V1</b></td>
|
| 243 |
<td>ARC-Challenge (Acc-Norm, 25-shot)</td>
|
|
|
|
| 281 |
<td><b>76.62</b></td>
|
| 282 |
<td><b>100.00</b></td>
|
| 283 |
</tr>
|
|
|
|
| 284 |
<tr>
|
| 285 |
<td rowspan="7"><b>OpenLLM V2</b></td>
|
| 286 |
<td>IFEval (Inst Level Strict Acc, 0-shot)</td>
|
|
|
|
| 340 |
</tr>
|
| 341 |
</thead>
|
| 342 |
<tbody>
|
|
|
|
| 343 |
<tr>
|
| 344 |
<td rowspan="7"><b>OpenLLM V1</b></td>
|
| 345 |
<td>ARC-Challenge (Acc-Norm, 25-shot)</td>
|
|
|
|
| 383 |
<td><b>76.51</b></td>
|
| 384 |
<td><b>99.86</b></td>
|
| 385 |
</tr>
|
|
|
|
| 386 |
<tr>
|
| 387 |
<td rowspan="7"><b>OpenLLM V2</b></td>
|
| 388 |
<td>IFEval (Inst Level Strict Acc, 0-shot)</td>
|
|
|
|
| 427 |
<td><b>97.77</b></td>
|
| 428 |
</tr>
|
| 429 |
</tbody>
|
| 430 |
+
</table> -->
|
| 431 |
|