FutureMa commited on
Commit
21b2c0c
·
verified ·
1 Parent(s): 0eb3939

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -103,7 +103,7 @@ The training set is constructed via a multi-model annotation framework:
103
  | 4 | **Eva-4B (Ours)** | **81.3%** | **0.807** |
104
  | 5 | GPT-5.2 | 80.5% | 0.805 |
105
 
106
- Note: by the accuracy values in the paper’s table, Eva-4B is above GPT-5.2. The paper also states Eva-4B **“ranks 5th overall and 2nd among open-source models (after GLM-4.7)”**, which appears inconsistent with the raw ordering implied by the accuracies.
107
 
108
  ### Per-class F1 (Eva-4B)
109
 
@@ -151,7 +151,7 @@ Response format:
151
  {"reason": "brief explanation under 100 characters", "label": "direct|intermediate|fully_evasive"}
152
  ```
153
 
154
- Answer in ```json content, no other text"""
155
 
156
  question = "What are your revenue expectations for next quarter?"
157
  answer = "We remain optimistic about our business trajectory and will continue to focus on executing our strategic priorities."
@@ -216,4 +216,4 @@ If you use this model, please cite the accompanying paper:
216
 
217
  ---
218
 
219
- Last updated: 2026-01-12
 
103
  | 4 | **Eva-4B (Ours)** | **81.3%** | **0.807** |
104
  | 5 | GPT-5.2 | 80.5% | 0.805 |
105
 
106
+ Note: based on the accuracy values, Eva-4B is **2nd among open-source models**, after GLM-4.7 (82.6%).
107
 
108
  ### Per-class F1 (Eva-4B)
109
 
 
151
  {"reason": "brief explanation under 100 characters", "label": "direct|intermediate|fully_evasive"}
152
  ```
153
 
154
+ Answer in json block content, no other text"""
155
 
156
  question = "What are your revenue expectations for next quarter?"
157
  answer = "We remain optimistic about our business trajectory and will continue to focus on executing our strategic priorities."
 
216
 
217
  ---
218
 
219
+ Last updated: 2026-01-12