Malikeh1375 commited on
Commit
92cd6c9
·
verified ·
1 Parent(s): ca4f565

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -1
README.md CHANGED
@@ -122,7 +122,39 @@ TokSuite–GPT-2 is evaluated on the **TokSuite robustness benchmark**, which me
122
  - OCR and spacing artifacts,
123
  - LaTeX and STEM-style formatting.
124
 
125
- Performance is reported as **relative accuracy drop** with respect to canonical inputs.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
126
 
127
  ---
128
 
 
122
  - OCR and spacing artifacts,
123
  - LaTeX and STEM-style formatting.
124
 
125
+
126
+ **Tokenization robustness under multilingual text perturbations**
127
+ Values represent **relative performance drop**, computed as `(Acc_clean − Acc_perturbed) / Acc_clean`, where **lower values indicate greater robustness**.
128
+
129
+ Perturbation types include:
130
+ - **Input:** non-native keyboard input and romanization
131
+ - **Diacr.:** optional diacritics
132
+ - **Orth.:** orthographic errors
133
+ - **Morph:** morphological variations including derivations, inflections, and contractions
134
+ - **Noise:** homoglyph substitutions, OCR artifacts, typos, and spacing errors
135
+ - **LaTeX:** LaTeX-style mathematical formatting
136
+ - **STEM:** scientific diagrams and notational conventions
137
+ - **Unic.:** Unicode styling characters
138
+
139
+ **NEN** denotes non-English inputs and **EN** denotes English inputs. The **Avg** column reports the average relative performance drop across all perturbation categories.
140
+
141
+
142
+ | Model | Input | Diacr. | Orth. | Morph | Noise | LaTeX | STEM | Unic. | Avg ↓ |
143
+ |---------------|-------|--------|-------|-------|-------|-------|------|-------|-------|
144
+ | TokenMonster | **0.23** | **0.33** | 0.08 | **0.01** | **-0.07** | **0.10** | 0.18 | 0.21 | **0.17** |
145
+ | XGLM | 0.34 | 0.49 | 0.10 | 0.11 | 0.07 | 0.12 | 0.22 | 0.29 | 0.22 |
146
+ | BLOOM | 0.30 | 0.34 | 0.13 | 0.07 | 0.11 | 0.18 | 0.18 | 0.24 | 0.22 |
147
+ | ByT5 | 0.30 | 0.44 | **0.04** | 0.06 | 0.04 | 0.14 | **0.18** | 0.17 | 0.22 |
148
+ | Comma | 0.28 | 0.43 | 0.05 | 0.07 | **0.00** | 0.11 | 0.20 | 0.23 | 0.22 |
149
+ | mBERT | 0.33 | 0.44 | 0.11 | 0.11 | 0.06 | 0.18 | 0.22 | **0.14** | 0.24 |
150
+ | GPT-4o | 0.30 | 0.51 | 0.08 | 0.05 | 0.05 | 0.16 | 0.19 | 0.24 | 0.24 |
151
+ | GPT-2 | 0.34 | 0.46 | 0.07 | 0.10 | 0.06 | 0.14 | 0.21 | 0.24 | 0.25 |
152
+ | Phi-3 | 0.33 | 0.46 | 0.16 | 0.09 | 0.08 | 0.17 | 0.21 | 0.24 | 0.25 |
153
+ | Gemma-2 | 0.32 | 0.42 | 0.14 | **0.15** | 0.03 | 0.16 | 0.25 | 0.22 | 0.26 |
154
+ | Qwen-3 | **0.36** | 0.42 | 0.14 | 0.11 | 0.06 | 0.16 | 0.23 | 0.26 | 0.26 |
155
+ | Llama-3.2 | 0.33 | **0.55** | 0.11 | 0.10 | 0.08 | 0.15 | 0.24 | 0.17 | 0.26 |
156
+ | Aya | 0.31 | 0.46 | 0.14 | 0.10 | 0.03 | **0.19** | **0.25** | 0.21 | 0.26 |
157
+ | Tekken | 0.33 | 0.47 | **0.18** | 0.03 | **0.31** | 0.10 | 0.21 | **0.27** | **0.27** |
158
 
159
  ---
160