rausch commited on
Commit
2fddeac
·
verified ·
1 Parent(s): 73fdbf8

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +49 -0
README.md ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - de
4
+ license: mit
5
+ library_name: transformers
6
+ pipeline_tag: text-generation
7
+ tags:
8
+ - t5
9
+ - german
10
+ - scientific
11
+ - wechsel
12
+ datasets:
13
+ - unpaywall-scientific
14
+ ---
15
+
16
+ # DE-T5-Sci-Transfer-15k
17
+
18
+ Final German scientific model: WECHSEL-initialized EN-T5-Sci → continued for 15 000 steps on German scientific data (same regimen as DE-T5-Base-15k). Checkpoint: `cross_lingual_transfer/logs/train/.../step-step=015000.ckpt`.
19
+
20
+ ## Model Details
21
+ - Base: EN-T5-Sci weights, German tokenizer
22
+ - Optimizer: Adafactor, lr=1e-3, inverse sqrt schedule, warmup 1.5k, grad clip 1.0
23
+ - Effective batch: 48 (per-GPU 48, grad accumulation 1)
24
+ - Objective: Span corruption (15 % noise, mean span 3)
25
+
26
+ ## Training Data
27
+ German subset of the Unpaywall-derived scientific corpus (sliding windows 512 tokens, 50 % overlap). Same cleaning pipeline as the English run.
28
+
29
+ ## Evaluation (Global-MMLU, zero-shot)
30
+ | Metric | EN | DE |
31
+ | --- | --- | --- |
32
+ | Overall accuracy | **0.2738** | **0.2700** |
33
+ | Humanities | 0.2559 | 0.2536 |
34
+ | STEM | 0.2867 | 0.2851 |
35
+ | Social Sciences | 0.3058 | 0.3055 |
36
+ | Other | 0.2562 | 0.2443 |
37
+
38
+ This is the best-performing German checkpoint across both languages in the final evaluation (`evaluation_results/scientific_crosslingual_transfer_eval_full_15k`).
39
+
40
+ ## Intended Use
41
+ Zero-shot QA in German/English scientific domains, or as a strong starting point for German task-specific fine-tuning (NER, relation extraction, etc.).
42
+
43
+ ## Limitations
44
+ - Still inherits T5-base context length/parameter budget.
45
+ - Evaluated only on Global-MMLU; downstream fine-tuning recommended for specialized tasks.
46
+ - Training corpus is domain-specific (scientific); may underperform on casual text.
47
+
48
+ ## Citation
49
+ Please cite the thesis (Nikolas Rauscher, 2025) and the WECHSEL paper (Minixhofer et al. 2022).