Update README.md
Browse files
README.md
CHANGED
|
@@ -103,7 +103,7 @@ The training set is constructed via a multi-model annotation framework:
|
|
| 103 |
| 4 | **Eva-4B (Ours)** | **81.3%** | **0.807** |
|
| 104 |
| 5 | GPT-5.2 | 80.5% | 0.805 |
|
| 105 |
|
| 106 |
-
Note:
|
| 107 |
|
| 108 |
### Per-class F1 (Eva-4B)
|
| 109 |
|
|
@@ -151,7 +151,7 @@ Response format:
|
|
| 151 |
{"reason": "brief explanation under 100 characters", "label": "direct|intermediate|fully_evasive"}
|
| 152 |
```
|
| 153 |
|
| 154 |
-
Answer in
|
| 155 |
|
| 156 |
question = "What are your revenue expectations for next quarter?"
|
| 157 |
answer = "We remain optimistic about our business trajectory and will continue to focus on executing our strategic priorities."
|
|
@@ -216,4 +216,4 @@ If you use this model, please cite the accompanying paper:
|
|
| 216 |
|
| 217 |
---
|
| 218 |
|
| 219 |
-
Last updated: 2026-01-12
|
|
|
|
| 103 |
| 4 | **Eva-4B (Ours)** | **81.3%** | **0.807** |
|
| 104 |
| 5 | GPT-5.2 | 80.5% | 0.805 |
|
| 105 |
|
| 106 |
+
Note: based on the accuracy values, Eva-4B is **2nd among open-source models**, after GLM-4.7 (82.6%).
|
| 107 |
|
| 108 |
### Per-class F1 (Eva-4B)
|
| 109 |
|
|
|
|
| 151 |
{"reason": "brief explanation under 100 characters", "label": "direct|intermediate|fully_evasive"}
|
| 152 |
```
|
| 153 |
|
| 154 |
+
Answer in json block content, no other text"""
|
| 155 |
|
| 156 |
question = "What are your revenue expectations for next quarter?"
|
| 157 |
answer = "We remain optimistic about our business trajectory and will continue to focus on executing our strategic priorities."
|
|
|
|
| 216 |
|
| 217 |
---
|
| 218 |
|
| 219 |
+
Last updated: 2026-01-12
|