al1808th commited on
Commit
cb20085
·
1 Parent(s): 8117ca9

refactor: only models here, training pipeline archive on gh

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. README.md +231 -17
  2. parsed/old-conll17-1/svediakorp-letter141673-Stalhammar.conllu → conll17_baseline_sv_only.pt +2 -2
  3. parsed/old-conll17-1/svediakorp-letter141673-Stalhammar_preprocessed.conllu → conll17_bm.pt +2 -2
  4. eval/gold_as_val.conllu → conll17_bm_dk.pt +2 -2
  5. parsed/old-conll17-1/svediakorp-sec1018-spf175.conllu → conll17_de_lit.pt +2 -2
  6. conll17_icepahc.pt +3 -0
  7. conll17_is-modern.pt +3 -0
  8. conll17_isPUD-pahc-gc.pt +3 -0
  9. conll17_isPUD.pt +3 -0
  10. conll17_nn.pt +3 -0
  11. conll17_sv_diachron.pt +3 -0
  12. digphil_inference_pipeline.py +0 -114
  13. eval/execute_eval.py +0 -273
  14. eval/mod_conll18_ud_eval.py +0 -672
  15. eval/scores/__init__.py +0 -0
  16. eval/scores/scores_talbanken.txt +0 -162
  17. eval/scores/scores_transformer_no_silver.txt +0 -162
  18. eval/scores/scores_transformer_silver.txt +0 -162
  19. logs/latest.txt +0 -1
  20. logs/log_bert-base-swedish-cased_sv_20260219_135555.txt +0 -114
  21. logs/log_bert-base-swedish-cased_sv_20260219_135610.txt +0 -138
  22. logs/log_bert-base-swedish-cased_sv_20260219_135635.txt +0 -0
  23. logs/log_bert-base-swedish-cased_sv_20260318_131623.txt +0 -271
  24. logs/log_bert-base-swedish-cased_sv_20260318_132248.txt +0 -109
  25. logs/log_bert-base-swedish-cased_sv_20260318_132404.txt +0 -109
  26. logs/log_bert-base-swedish-cased_sv_20260318_132733.txt +0 -325
  27. logs/log_bert-base-swedish-cased_sv_20260318_133119.txt +0 -0
  28. logs/log_bert-base-swedish-cased_sv_20260318_141654.txt +0 -0
  29. logs/log_bert-base-swedish-cased_sv_20260318_215701.txt +0 -0
  30. logs/log_bert-base-swedish-cased_sv_20260319_200948.txt +0 -0
  31. logs/log_bert-base-swedish-cased_sv_20260319_201349.txt +0 -0
  32. logs/log_bert-base-swedish-cased_sv_20260319_214621.txt +0 -0
  33. logs/log_conll17.pt_sv_20251203_232257.txt +0 -719
  34. logs/log_conll17.pt_sv_20260218_190034.txt +0 -120
  35. logs/log_conll17.pt_sv_20260218_190242.txt +0 -143
  36. logs/log_conll17.pt_sv_20260218_190400.txt +0 -38
  37. logs/log_conll17.pt_sv_20260218_190808.txt +0 -38
  38. logs/log_conll17.pt_sv_20260218_190956.txt +0 -38
  39. logs/log_conll17.pt_sv_20260218_191305.txt +0 -38
  40. logs/log_conll17.pt_sv_20260218_191719.txt +0 -38
  41. logs/log_conll17.pt_sv_20260219_104841.txt +0 -44
  42. logs/log_conll17.pt_sv_20260219_105228.txt +0 -315
  43. logs/log_conll17.pt_sv_20260219_105756.txt +0 -0
  44. logs/log_conll17.pt_sv_20260219_114115.txt +0 -0
  45. logs/log_conll17.pt_sv_diachron_20251203_223822.txt +0 -731
  46. logs/log_conll17.pt_sv_diachron_20251212_145854.txt +0 -161
  47. logs/log_conll17.pt_sv_diachron_20251212_150001.txt +0 -0
  48. logs/log_conll17.pt_sv_diachron_20260116_065352.txt +0 -830
  49. logs/log_conll17.pt_sv_diachron_20260116_073035.txt +0 -868
  50. logs/log_conll17.pt_sv_diachron_de_lit_20251203_154509.txt +0 -740
README.md CHANGED
@@ -15,35 +15,249 @@ library_name: stanza
15
 
16
  # Retraining Stanza to optimize dependency parsing on a diachronic Swedish corpus
17
 
18
- This repository contains forked code from the official Stanza github, with scripts that help prepare for and train models on different combinations of treebanks relevant to historical Swedish.
19
 
20
- ## Guide
21
 
22
- Dev/test for all models is 10/90 of our human validated gold sentences (https://github.com/alanev52/Diachronic_Treebanks_DigPhil/tree/main/parsed_data/validated)
 
23
 
24
- Example work flow, training a model with pretrained vectors from kubhist2 1880 incremental and training data from Swedish and Bokmål treebanks, as well as our own machine-parsed diachronic corpus:
25
 
26
- ```
27
- python prepare-train-val-test.py sv diachron bm
 
 
 
 
 
 
 
 
 
 
 
 
28
 
29
- source scripts/config_alvis.sh
30
 
31
- python -m stanza.utils.datasets.prepare_depparse_treebank UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/diachronic.pt
32
 
33
- python -m stanza.utils.training.run_depparse UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/diachronic.pt --batch_size 32 --dropout 0.33
34
  ```
 
 
 
 
 
 
 
 
35
 
36
- All of the above can be done with a single command by using:
 
 
 
 
 
37
 
38
- ```
39
- ./make_new_model.sh {vectors} {language codes}
40
- ```
41
 
42
- which for the example becomes:
43
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
44
  ```
45
- ./make_new_model.sh diachronic.pt sv diachron bm
46
- ```
47
 
48
- ## Results
 
 
49
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
 
16
  # Retraining Stanza to optimize dependency parsing on a diachronic Swedish corpus
17
 
18
+ This repository contains Stanza BiLSTM models retrained on different combinations of UD treebanks relevant to historical Swedish. The models prefixed `conll17_` are trained with static embeddings, and the models prefixed `transformer_` are trained with dynamic embeddings from the transformer encoder "KBLab/bert-base-swedish-cased".
19
 
20
+ ## LAS Scores
21
 
22
+ LAS scores for the models are computed against a gold set of 109 manually annoted sentences divided into five different periods.
23
+ For the models trained on static vectors, only the overall test scores is given:
24
 
25
+ ### Models with static vector embeddings (`conll17.pt`)
26
 
27
+ | Languages | LAS |
28
+ |--------------------------------|-------|
29
+ | **Swedish (with diachronic)** | **61.95** |
30
+ | Icelandic (PUD) | 61.49 |
31
+ | German (LIT) | 61.43 |
32
+ | Icelandic (GC) | 61.43 |
33
+ | Bokmaal, Danish | 60.13 |
34
+ | Nynorsk | 50.46 |
35
+ | *Swedish (without diachronic)* | *50.34* |
36
+ | Icelandic (Modern) | 46.47 |
37
+ | Bokmaal | 45.96 |
38
+ | Icelandic (IcePaHC) | 44.60 |
39
+
40
+ For the transformer-fed models, more fine-grained scores on each period are given as a histogram. The model `transformer_seen_gold_no_silver.pt` was given the gold set during training and hence has no score, but is intuitively the best model. As a benchmark, an "out-of-the-box" Stanza trained only on Talbanken is given.
41
 
42
+ <img src="plot/las_f1_scores_comparison.png" width="40%">
43
 
44
+ ## Inference
45
 
46
+ Example for how the models can be run:
47
  ```
48
+ import os
49
+ from pathlib import Path
50
+ import stanza
51
+ from stanza.utils.conll import CoNLL
52
+ import time
53
+ import gc
54
+ import torch
55
+ from tqdm import tqdm
56
 
57
+ if torch.xpu.is_available():
58
+ device = torch.device("xpu")
59
+ elif torch.cuda.is_available():
60
+ device = torch.device("cuda")
61
+ else:
62
+ device = torch.device("cpu")
63
 
64
+ ###################################################################################################
65
+ ##### SETTINGS ####################################################################################
66
+ ###################################################################################################
67
 
68
+ vanilla = False
69
 
70
+ conllu_in_dir = Path("YOUR/CONLLUS/")
71
+ conllu_out_dir = Path("OUT/DIR")
72
+
73
+ depparse_model_path = "transformer_seen_gold_no_silver.pt"
74
+
75
+ os.makedirs(conllu_out_dir, exist_ok=True)
76
+
77
+ ##### PREPARE #####################################################################################
78
+
79
+ if vanilla:
80
+ print("Using vanilla Swedish Stanza pipeline with default models.")
81
+ nlp = stanza.Pipeline(
82
+ "sv",
83
+ processors="tokenize,pos,lemma,depparse",
84
+ tokenize_pretokenized=True, # to keep original tokens
85
+ use_gpu=True,
86
+ pos_batch_size=3000,
87
+ package=None,
88
+ device=device,
89
+ )
90
+
91
+ else:
92
+ nlp = stanza.Pipeline(
93
+ "sv",
94
+ processors="tokenize,pos,lemma,depparse",
95
+ tokenize_pretokenized=True, # to keep original tokens
96
+ use_gpu=True,
97
+ pos_batch_size=3000,
98
+ package=None,
99
+ download_method=None,
100
+ depparse_model_path=depparse_model_path, # depparse_model_path becomes model_path inside the depparse processor config. explicit model_path is to control where normal Stanza models are stored
101
+ device=device,
102
+ )
103
+
104
+ ##### INFERENCE ###################################################################################
105
+
106
+ t0 = time.time()
107
+ total_sentences = 0
108
+
109
+ for fname in tqdm(sorted(os.listdir(conllu_in_dir)), desc="Files"):
110
+ if not fname.endswith(".conllu"):
111
+ continue
112
+
113
+ in_path = conllu_in_dir / fname
114
+ out_path = conllu_out_dir / fname
115
+
116
+ if out_path.exists():
117
+ print(f"File already parsed, skipping: {out_path}")
118
+ continue
119
+
120
+ doc = CoNLL.conll2doc(in_path)
121
+
122
+ total_sentences += len(doc.sentences)
123
+
124
+ with torch.inference_mode():
125
+ parsed_doc = nlp(doc)
126
+
127
+ with open(out_path, "w", encoding="utf-8") as f:
128
+ CoNLL.write_doc2conll(parsed_doc, f)
129
+
130
+ # Release per-file objects and ask both Python and the torch allocator to
131
+ # reclaim memory that is no longer needed.
132
+ del doc
133
+ del parsed_doc
134
+ gc.collect()
135
+ if device.type == "cuda":
136
+ torch.cuda.empty_cache()
137
+ memory_max = torch.cuda.max_memory_allocated()
138
+ print(f"Max memory allocated so far: {memory_max / (1024 ** 3):.2f} GB")
139
+ elif device.type == "xpu":
140
+ torch.xpu.empty_cache()
141
+ memory_max = torch.xpu.max_memory_allocated()
142
+ print(f"Max memory allocated so far: {memory_max / (1024 ** 3):.2f} GB")
143
+
144
+ t = time.time() - t0
145
+
146
+ ###################################################################################################
147
+
148
+ if device.type == "cuda":
149
+ memory_max = torch.cuda.max_memory_allocated()
150
+ torch.cuda.reset_peak_memory_stats()
151
+ elif device.type == "xpu":
152
+ memory_max = torch.xpu.max_memory_allocated()
153
+ torch.xpu.reset_peak_memory_stats()
154
+ else:
155
+ memory_max = 0
156
+
157
+ print(
158
+ "\nFinished parsing.\n"
159
+ f"Total sentences parsed: {total_sentences}\n"
160
+ f"Total time: {t / 60:.2f} minutes\n"
161
+ f"Max memory allocated: {memory_max / (1024 ** 3):.2f} GB\n"
162
+ )
163
  ```
 
 
164
 
165
+ ## Training args
166
+
167
+ Full list of training args:
168
 
169
+ ```
170
+ batch_size: 32
171
+ bert_finetune: False
172
+ bert_finetune_layers: None
173
+ bert_hidden_layers: 4
174
+ bert_learning_rate: 1.0
175
+ bert_model: KBLab/bert-base-swedish-cased
176
+ bert_start_finetuning: 200
177
+ bert_warmup_steps: 200
178
+ bert_weight_decay: 0.0
179
+ beta2: 0.999
180
+ char: True
181
+ char_emb_dim: 100
182
+ char_hidden_dim: 400
183
+ char_num_layers: 1
184
+ char_rec_dropout: 0
185
+ charlm: True
186
+ charlm_backward_file: /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
187
+ charlm_forward_file: /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt
188
+ charlm_save_dir: saved_models/charlm
189
+ charlm_shorthand: sv_conll17
190
+ checkpoint: True
191
+ checkpoint_interval: 500
192
+ checkpoint_save_name: None
193
+ continue_from: None
194
+ data_dir: data/depparse
195
+ deep_biaff_hidden_dim: 400
196
+ deep_biaff_output_dim: 160
197
+ device: xpu:0
198
+ distance: True
199
+ dropout: 0.33
200
+ eval_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
201
+ eval_interval: 100
202
+ gold_labels: True
203
+ hidden_dim: 400
204
+ lang: sv
205
+ linearization: True
206
+ log_norms: False
207
+ log_step: 20
208
+ lora_alpha: 128
209
+ lora_dropout: 0.1
210
+ lora_modules_to_save: []
211
+ lora_rank: 64
212
+ lora_target_modules: ['query', 'value', 'output.dense', 'intermediate.dense']
213
+ lr: 2.0
214
+ max_grad_norm: 1.0
215
+ max_steps: 50000
216
+ max_steps_before_stop: 2000
217
+ mode: train
218
+ model_type: graph
219
+ num_layers: 3
220
+ optim: adadelta
221
+ output_file: None
222
+ output_latex: False
223
+ pretrain: True
224
+ pretrain_max_vocab: 250000
225
+ rec_dropout: 0
226
+ reversed: False
227
+ sample_train: 1.0
228
+ save_dir: saved_models/depparse
229
+ save_name: {shorthand}_{embedding}_parser.pt
230
+ second_batch_size: None
231
+ second_bert_learning_rate: 0.001
232
+ second_lr: 0.0002
233
+ second_optim: adam
234
+ second_optim_start_step: 10000
235
+ second_warmup_steps: 200
236
+ seed: 1234
237
+ shorthand: sv_diachronic
238
+ silver_file: None
239
+ silver_weight: 0.5
240
+ tag_emb_dim: 50
241
+ train_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
242
+ train_size: None
243
+ transformed_dim: 125
244
+ transition_embedding_dim: 20
245
+ transition_hidden_dim: 20
246
+ transition_merge_hidden_dim: 200
247
+ transition_subtree_combination: SubtreeCombination.NONE
248
+ transition_subtree_nonlinearity: none
249
+ use_arc_embedding: False
250
+ use_peft: False
251
+ use_ufeats: True
252
+ use_upos: True
253
+ use_xpos: True
254
+ wandb: False
255
+ wandb_name: None
256
+ weight_decay: 1e-05
257
+ word_cutoff: 7
258
+ word_dropout: 0.33
259
+ word_emb_dim: 75
260
+ wordvec_dir: /home/urdatorn/stanza_resources/sv/pretrain
261
+ wordvec_file: None
262
+ wordvec_pretrain_file: /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
263
+ ```
parsed/old-conll17-1/svediakorp-letter141673-Stalhammar.conllu → conll17_baseline_sv_only.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e4779224fe6e1d1580faa201af29e0500110596035652fc788e4a5b0131dd9b
3
- size 4846318
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50aee8f0eab8b4b0e28958c0e428f1aad29cf118e4320332862b8701fc6dd6e3
3
+ size 141291582
parsed/old-conll17-1/svediakorp-letter141673-Stalhammar_preprocessed.conllu → conll17_bm.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5dbe992db0d973c529d09bfd34bf1c647db3c17aed5e51039e044e859323b472
3
- size 16025
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b0848c6832ca155fcfbf78040530aeb27598562cda430821afddbead91ea0b1
3
+ size 148183635
eval/gold_as_val.conllu → conll17_bm_dk.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f4e442db9582dfc10263eb2cf6cbed4719194ea0f6a1ab0ff82fa652b7a414c7
3
- size 111668
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab11092dd5795e51add850b4aa325a474b72f374b93df53a78fd247ad456497c
3
+ size 150762047
parsed/old-conll17-1/svediakorp-sec1018-spf175.conllu → conll17_de_lit.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d50cf732cd496a2e203b2fd7e799147d2b6e6919711b84d0d4b702523833dc0
3
- size 753506
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b84862d0f728ba53e2076208a94d5e90bd1dc7e0f2884a12db3f21ffcaf073dc
3
+ size 146969172
conll17_icepahc.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f59e82a5e94514bdf086d19931a26f5cd6274c6ad29c667f3366e93be0bec585
3
+ size 151645170
conll17_is-modern.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e02dffc9ae0d419373d1ba597f4d7e0459d70d6909e18211d40c33472e5a6d46
3
+ size 148251198
conll17_isPUD-pahc-gc.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fe5f0e97a6a4e257b124f8b229c59cc61d58124512421d56795ac6ffb67359e
3
+ size 146969174
conll17_isPUD.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcdc23474d269b8f4dba7d246f184b42a446cae7294f1134121cf5da8511efa6
3
+ size 146117067
conll17_nn.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d74ed335831d36ee6d4569d1496a62860779c378ed391a693a8aa09159f388e
3
+ size 147986243
conll17_sv_diachron.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a67914c27dcc9cdc33034cdbebed9eec4c55b10ab2aeffc5f5d6053a96c5107a
3
+ size 145283635
digphil_inference_pipeline.py DELETED
@@ -1,114 +0,0 @@
1
- import os
2
- from pathlib import Path
3
- import stanza
4
- from stanza.utils.conll import CoNLL
5
- import time
6
- import gc
7
- import torch
8
- from tqdm import tqdm
9
-
10
- if torch.xpu.is_available():
11
- device = torch.device("xpu")
12
- elif torch.cuda.is_available():
13
- device = torch.device("cuda")
14
-
15
- ###################################################################################################
16
- ##### SETTINGS ####################################################################################
17
- ###################################################################################################
18
-
19
- vanilla = False
20
-
21
- conllu_in_dir = Path("/home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated")
22
- #conllu_out_dir = Path("parsed/testset_talbanken_not_seen_gold")
23
- conllu_out_dir = Path("parsed/testset_transformer_not_seen_gold_no_silver")
24
-
25
- depparse_model_path = "saved_models/depparse/transformer_not_seen_gold_no_silver.pt"
26
-
27
- os.makedirs(conllu_out_dir, exist_ok=True)
28
-
29
- ##### PREPARE #####################################################################################
30
-
31
- if vanilla:
32
- print("Using vanilla Swedish Stanza pipeline with default models.")
33
- nlp = stanza.Pipeline(
34
- "sv",
35
- processors="tokenize,pos,lemma,depparse",
36
- tokenize_pretokenized=True, # to keep original tokens
37
- use_gpu=True,
38
- pos_batch_size=3000,
39
- package=None,
40
- device=device,
41
- )
42
-
43
- else:
44
- nlp = stanza.Pipeline(
45
- "sv",
46
- processors="tokenize,pos,lemma,depparse",
47
- tokenize_pretokenized=True, # to keep original tokens
48
- use_gpu=True,
49
- pos_batch_size=3000,
50
- package=None,
51
- download_method=None,
52
- depparse_model_path=depparse_model_path, # depparse_model_path becomes model_path inside the depparse processor config. explicit model_path is to control where normal Stanza models are stored
53
- device=device,
54
- )
55
-
56
- ##### INFERENCE ###################################################################################
57
-
58
- t0 = time.time()
59
- total_sentences = 0
60
-
61
- for fname in tqdm(sorted(os.listdir(conllu_in_dir)), desc="Files"):
62
- if not fname.endswith(".conllu"):
63
- continue
64
-
65
- in_path = conllu_in_dir / fname
66
- out_path = conllu_out_dir / fname
67
-
68
- if out_path.exists():
69
- print(f"File already parsed, skipping: {out_path}")
70
- continue
71
-
72
- doc = CoNLL.conll2doc(in_path)
73
-
74
- total_sentences += len(doc.sentences)
75
-
76
- with torch.inference_mode():
77
- parsed_doc = nlp(doc)
78
-
79
- with open(out_path, "w", encoding="utf-8") as f:
80
- CoNLL.write_doc2conll(parsed_doc, f)
81
-
82
- # Release per-file objects and ask both Python and the torch allocator to
83
- # reclaim memory that is no longer needed.
84
- del doc
85
- del parsed_doc
86
- gc.collect()
87
- if device.type == "cuda":
88
- torch.cuda.empty_cache()
89
- memory_max = torch.cuda.max_memory_allocated()
90
- print(f"Max memory allocated so far: {memory_max / (1024 ** 3):.2f} GB")
91
- elif device.type == "xpu":
92
- torch.xpu.empty_cache()
93
- memory_max = torch.xpu.max_memory_allocated()
94
- print(f"Max memory allocated so far: {memory_max / (1024 ** 3):.2f} GB")
95
-
96
- t = time.time() - t0
97
-
98
- ###################################################################################################
99
-
100
- if device.type == "cuda":
101
- memory_max = torch.cuda.max_memory_allocated()
102
- torch.cuda.reset_peak_memory_stats()
103
- elif device.type == "xpu":
104
- memory_max = torch.xpu.max_memory_allocated()
105
- torch.xpu.reset_peak_memory_stats()
106
- else:
107
- memory_max = 0
108
-
109
- print(
110
- "\nFinished parsing.\n"
111
- f"Total sentences parsed: {total_sentences}\n"
112
- f"Total time: {t / 60:.2f} minutes\n"
113
- f"Max memory allocated: {memory_max / (1024 ** 3):.2f} GB\n"
114
- )
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
eval/execute_eval.py DELETED
@@ -1,273 +0,0 @@
1
- from mod_conll18_ud_eval import load_conllu, evaluate
2
- from collections import defaultdict
3
- from pathlib import Path
4
- import pyconll
5
- import sys
6
- import os
7
- import numpy as np
8
-
9
- ROOT = Path.cwd()
10
- print(ROOT)
11
- gold_dir = Path("/home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated") # mapp med vårt manuellt annoterade guldstandardset
12
- val_file = ROOT / "eval/gold_as_val.conllu" # sentences used during validation - must be excluded
13
-
14
- def load_val_sentence_ids(val_file_path):
15
- """Load sentence IDs from the validation file to exclude from evaluation."""
16
- val_ids = set()
17
- if not os.path.exists(val_file_path):
18
- print(f"Warning: validation file {val_file_path} not found. No sentences will be excluded.")
19
- return val_ids
20
- val_data = pyconll.load_from_file(str(val_file_path))
21
- for sentence in val_data:
22
- val_ids.add(sentence.id)
23
- print(f"Loaded {len(val_ids)} sentence IDs to exclude from evaluation (used in validation).")
24
- return val_ids
25
-
26
- def filter_gold_file(gold_file_path, excluded_ids):
27
- """Create a filtered gold file excluding sentences used during validation.
28
- Returns path to filtered file and the number of excluded sentences."""
29
- gold_data = pyconll.load_from_file(gold_file_path)
30
- filtered_sentences = [s for s in gold_data if s.id not in excluded_ids]
31
- n_excluded = len(gold_data) - len(filtered_sentences)
32
-
33
- if n_excluded > 0:
34
- print(f" Excluded {n_excluded} validation sentences from {os.path.basename(gold_file_path)}")
35
-
36
- # Write filtered data to a temp file
37
- filtered_path = gold_file_path.replace(".conllu", "_eval_filtered.conllu")
38
- with open(filtered_path, "w") as f:
39
- for sentence in filtered_sentences:
40
- f.write(sentence.conll())
41
- f.write("\n\n")
42
- return filtered_path, len(filtered_sentences)
43
-
44
- # Load validation sentence IDs at startup
45
- val_sentence_ids = load_val_sentence_ids(val_file)
46
-
47
- # Model configurations: (predicted_dir, output_filename)
48
- MODEL_CONFIGS = [
49
- (ROOT / "parsed/testset_talbanken_not_seen_gold", "eval/scores/scores_talbanken.txt"),
50
- (ROOT / "parsed/testset_transformer_silver_not_seen_gold", "eval/scores/scores_transformer_silver.txt"),
51
- (ROOT / "parsed/testset_transformer_not_seen_gold_no_silver", "eval/scores/scores_transformer_no_silver.txt"),
52
- ]
53
-
54
- def check_valid_gold(gold_file_path):
55
- """ Check that the gold trees are valid """
56
- gold_data = pyconll.load_from_file(gold_file_path)
57
- errors = []
58
- for sentence in gold_data:
59
- for token in sentence:
60
- if token.head is None:
61
- e = ValueError(f"Invalid gold tree in file {gold_file_path}, sentence ID {sentence.id}: token {token.id} '{token.form}' has no head.")
62
- errors.append(e)
63
- elif token.head == token.id:
64
- e = ValueError(f"Invalid gold tree in file {gold_file_path}, sentence ID {sentence.id}: token {token.id} '{token.form}' has itself as head.")
65
- errors.append(e)
66
- roots = [token for token in sentence if token.head == '0']
67
- if len(roots) != 1:
68
- e = ValueError(f"Invalid gold tree in file {gold_file_path}, sentence ID {sentence.id}: expected 1 root, found {len(roots)}.")
69
- errors.append(e)
70
- return errors if len(errors) > 0 else None
71
-
72
- def preprocess_system_file(system_file_path, gold_file_path):
73
- """ Make a new base file with only the trees present in the gold file """
74
- gold_conllu = pyconll.load_from_file(gold_file_path)
75
- system_conllu = pyconll.load_from_file(system_file_path)
76
- gold_ids = {sentence.id for sentence in gold_conllu}
77
- filtered_sentences = [sentence for sentence in system_conllu if sentence.id in gold_ids]
78
- preprocessed_file_path = system_file_path.replace(".conllu", "_preprocessed.conllu")
79
- with open(preprocessed_file_path, "w") as outfile:
80
- for sentence in filtered_sentences:
81
- outfile.write(sentence.conll())
82
- outfile.write("\n\n")
83
- return preprocessed_file_path
84
-
85
- def execute_evaluation(gold_file_path, predicted_file_path):
86
- # Read goldfile
87
- try:
88
- gold_data = load_conllu(open(gold_file_path, "r", encoding="utf-8"))
89
- except Exception as e:
90
- print(f"Skipping text because of error reading gold file {gold_file_path}: {e}")
91
- #raise e
92
- return None
93
- # Read predicted file and ignore some format errors
94
- try:
95
- predicted_data = load_conllu(open(predicted_file_path, "r", encoding="utf-8"), ignore_invalid_format=True)
96
- except Exception as e:
97
- print(f"Skipping text because of error reading predicted file {predicted_file_path}: {e}")
98
- #raise e
99
- return None
100
- # Evaluate the predictions against the gold standard
101
- try:
102
- results = evaluate(gold_data, predicted_data)
103
- except Exception as e:
104
- print(f"Error during evaluation between {gold_file_path} and {predicted_file_path}: {e}")
105
- raise e
106
- return results
107
-
108
- time_periods = ["1700-1750", "1750-1800", "1800-1850", "1850-1900", "1900-1950"]
109
- metrics = ["UPOS", "UAS", "LAS"] # the relevant metrics
110
-
111
- # To look up time period index for a sample
112
- sample_period = {"svediakorp-letter141673-Stalhammar": 0, "svediakorp-sec25-Runius": 0, "svediakorp-sec330-GyllenborgC_SwenskaSpratthoken": 0,
113
- "svediakorp-sec277-EnbomPU_MedborgeligtSkalde": 1, "svediakorp-sec268-DulciU_VitterhetsNojen3": 1,
114
- "svediakorp-sec991-spf148": 2, "svediakorp-sec252-BremerF_Teckningar1": 2, "svediakorp-sec324-GranbergPA_Enslighetsalskaren": 2,
115
- "svediakorp-sec254-CederborghF_BerattelseOmJohnHall": 2, "svediakorp-sec987-spf144": 2, "svediakorp-sec988-spf145": 2,
116
- "svediakorp-sec452-NyblomH_FantasierFyra": 3, "svediakorp-sec486-SchwartzMS_BellmansSkor": 3, "svediakorp-sec1102-spf259": 3,
117
- "svediakorp-sec208-Anonym_DetGrasligaMordet": 3, "svediakorp-sec1063-spf220": 3,
118
- "svediakorp-sec631-HasselskogN_HallaHallaGronkoping": 4, "svediakorp-sec1033-spf190": 4, "svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol": 4,
119
- "svediakorp-sec613-EngstromA_StrindbergOchJag": 4, "svediakorp-sec639-HeidenstamV_Proletarfilosofiens": 4}
120
-
121
-
122
- def run_evaluation(predicted_dir, output_file):
123
- """Run evaluation for a single model configuration."""
124
- print(f"\n{'='*70}")
125
- print(f"Evaluating: {predicted_dir}")
126
- print(f"Output: {output_file}")
127
- print(f"{'='*70}\n")
128
-
129
- # Check if predicted directory exists
130
- if not os.path.exists(predicted_dir):
131
- print(f"ERROR: Predicted directory does not exist: {predicted_dir}")
132
- return
133
-
134
- # Reset per-model state
135
- score_arrays = [np.zeros((4, len(metrics))), np.zeros((4, len(metrics))), np.zeros((4, len(metrics))), np.zeros((4, len(metrics))), np.zeros((4, len(metrics)))]
136
- skipped_samples = []
137
- sample_nsents = defaultdict()
138
- processed_files = [] # Track (gold_path, pred_path) pairs for overall evaluation
139
-
140
- with open(output_file, "w") as outfile:
141
- # Loop through each of the validated samples
142
- for sample in os.listdir(gold_dir):
143
- sample_name = sample.split("/")[-1].replace(".conllu", "")
144
-
145
- print("DEBUG sample_name:", sample_name)
146
-
147
- period = sample_period.get(sample_name) # int 0-4
148
- print("DEBUG period value:", period)
149
-
150
- # Filter gold file to exclude sentences used during validation
151
- gold_file_path_orig = os.path.join(gold_dir, sample)
152
- gold_file_path, n_filtered_sents = filter_gold_file(gold_file_path_orig, val_sentence_ids)
153
-
154
- # Skip if all sentences were excluded
155
- if n_filtered_sents == 0:
156
- print(f"Skipping {sample_name}: all sentences were used in validation")
157
- skipped_samples.append(sample_name)
158
- continue
159
-
160
- # Check that gold trees are valid (using filtered file)
161
- e = check_valid_gold(gold_file_path)
162
- if e: # skip invalid gold files
163
- for error in e:
164
- print(error)
165
- skipped_samples.append(sample_name)
166
- continue
167
- # Always regenerate preprocessed predicted file to match filtered gold
168
- unprocessed_predicted_file_path = os.path.join(predicted_dir, sample)
169
- if os.path.exists(unprocessed_predicted_file_path):
170
- # Use filtered gold file to preprocess predictions
171
- predicted_file_path = preprocess_system_file(unprocessed_predicted_file_path, gold_file_path)
172
- else:
173
- print(f"Skipping {sample_name}: missing predicted file {unprocessed_predicted_file_path}")
174
- skipped_samples.append(sample_name)
175
- continue
176
- # Score base tree (parser output) against validated tree (gold standard)
177
- results = execute_evaluation(gold_file_path, predicted_file_path)
178
-
179
- if not results: # skip invalid files
180
- skipped_samples.append(sample_name)
181
- # Write results to scores file and save to time period arrays
182
- else:
183
- # Save number of sentences in lookup dict (using filtered count)
184
- sample_nsents[sample_name] = n_filtered_sents
185
- # Track file paths for overall evaluation
186
- processed_files.append((gold_file_path, predicted_file_path))
187
- print("DEBUG period value:", period)
188
- print("DEBUG time_periods:", time_periods)
189
-
190
- outfile.write(f"Sample: {sample_name}, time period: {time_periods[period]}\n")
191
- outfile.write("Metric\tPrecision\tRecall\tF1\tAligned Accuracy\n")
192
- metric_n = 0 # counter for loop below
193
- # Loop over each metric: UPOS, UAS, LAS etc.
194
- for metric, score in results.items(): # get the metric and its 4 scores
195
- if metric in metrics: # only save the relevant metrics
196
- # Save scores weighted by the number of sentences to arrays
197
- score_arrays[period][0][metric_n] += (score.precision * sample_nsents[sample_name])
198
- score_arrays[period][1][metric_n] += (score.recall * sample_nsents[sample_name])
199
- score_arrays[period][2][metric_n] += (score.f1 * sample_nsents[sample_name])
200
- # Write the metric and its file-average scores
201
- outfile.write(f"{metric}\t{score.precision:.4f}\t{score.recall:.4f}\t{score.f1:.4f}\t")
202
- if score.aligned_accuracy is None:
203
- outfile.write(f"{score.aligned_accuracy}\n")
204
- else:
205
- outfile.write(f"{score.aligned_accuracy:.4f}\n")
206
- # Save aligned accuracy to score arrays
207
- score_arrays[period][3][metric_n] += (score.aligned_accuracy * sample_nsents[sample_name])
208
- metric_n += 1 # move to next metric in array
209
- outfile.write("\n")
210
- outfile.write(f"{'-'*70}\n")
211
-
212
- # After all sample scores are written, compute average scores for each time period
213
- for period, period_scores in enumerate(score_arrays):
214
- # Get number of samples and sentences for the period
215
- samples_per_period = [sample_name for sample_name in sample_period.keys() if sample_period[sample_name] == period and sample_name not in skipped_samples]
216
- n_sents_per_period = sum([sample_nsents[sample_name] for sample_name in samples_per_period])
217
- # Write
218
- outfile.write(f"Average scores for time period {time_periods[period]} ({len(samples_per_period)} samples with a total of {n_sents_per_period} sentences):\n")
219
- outfile.write("Metric\tPrecision\tRecall\tF1\tAligned Accuracy\n")
220
- for metric_n, metric in enumerate(metrics):
221
- # Normalize weighted averages by the total number of sentences for the time period
222
- precision_avg = period_scores[0][metric_n] / n_sents_per_period
223
- recall_avg = period_scores[1][metric_n] / n_sents_per_period
224
- f1_avg = period_scores[2][metric_n] / n_sents_per_period
225
- aligned_acc_avg = period_scores[3][metric_n] / n_sents_per_period
226
- outfile.write(f"{metric}\t{precision_avg:.4f}\t{recall_avg:.4f}\t{f1_avg:.4f}\t{aligned_acc_avg:.4f}\n")
227
- outfile.write("\n")
228
-
229
- # Compute overall scores by concatenating all files and running single evaluation
230
- valid_samples = [sample_name for sample_name in sample_period.keys() if sample_name in sample_nsents and sample_name not in skipped_samples]
231
- total_sents = sum(sample_nsents[sample_name] for sample_name in valid_samples)
232
-
233
- if len(processed_files) == 0:
234
- outfile.write(f"Overall scores across all time periods ({len(valid_samples)} samples with a total of {total_sents} sentences):\n")
235
- outfile.write("Metric\tPrecision\tRecall\tF1\tAligned Accuracy\n")
236
- outfile.write("No valid samples available for overall evaluation.\n")
237
- else:
238
- # Concatenate all gold and predicted files
239
- all_gold_path = os.path.join(predicted_dir, "_all_gold_combined.conllu")
240
- all_pred_path = os.path.join(predicted_dir, "_all_pred_combined.conllu")
241
-
242
- with open(all_gold_path, "w") as gold_out, open(all_pred_path, "w") as pred_out:
243
- for gold_path, pred_path in processed_files:
244
- with open(gold_path) as f:
245
- gold_out.write(f.read())
246
- with open(pred_path) as f:
247
- pred_out.write(f.read())
248
-
249
- # Run single evaluation on combined files
250
- overall_results = execute_evaluation(all_gold_path, all_pred_path)
251
-
252
- outfile.write(f"Overall scores across all time periods ({len(valid_samples)} samples with a total of {total_sents} sentences):\n")
253
- outfile.write("Metric\tPrecision\tRecall\tF1\tAligned Accuracy\n")
254
-
255
- if overall_results:
256
- for metric, score in overall_results.items():
257
- if metric in metrics:
258
- outfile.write(f"{metric}\t{score.precision:.4f}\t{score.recall:.4f}\t{score.f1:.4f}\t")
259
- if score.aligned_accuracy is None:
260
- outfile.write(f"{score.aligned_accuracy}\n")
261
- else:
262
- outfile.write(f"{score.aligned_accuracy:.4f}\n")
263
- else:
264
- outfile.write("Error computing overall scores.\n")
265
-
266
- print(f"Results saved to {output_file}")
267
-
268
-
269
- # Run evaluation for all model configurations
270
- if __name__ == "__main__":
271
- for predicted_dir, output_file in MODEL_CONFIGS:
272
- run_evaluation(predicted_dir, output_file)
273
- print("\nAll evaluations complete!")
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
eval/mod_conll18_ud_eval.py DELETED
@@ -1,672 +0,0 @@
1
- # This script is from the CoNLL 2018 Shared Task.
2
- # It was downloaded from here: https://universaldependencies.org/conll18/evaluation.html
3
- # in October 2025, and modified by Astrid B. Ingelstam for the group project about diachronic treebanks.
4
-
5
- #!/usr/bin/env python3
6
-
7
- # Compatible with Python 2.7 and 3.2+, can be used either as a module
8
- # or a standalone executable.
9
- #
10
- # Copyright 2017, 2018 Institute of Formal and Applied Linguistics (UFAL),
11
- # Faculty of Mathematics and Physics, Charles University, Czech Republic.
12
- #
13
- # This Source Code Form is subject to the terms of the Mozilla Public
14
- # License, v. 2.0. If a copy of the MPL was not distributed with this
15
- # file, You can obtain one at http://mozilla.org/MPL/2.0/.
16
- #
17
- # Authors: Milan Straka, Martin Popel <surname@ufal.mff.cuni.cz>
18
- #
19
- # Changelog:
20
- # - [12 Apr 2018] Version 0.9: Initial release.
21
- # - [19 Apr 2018] Version 1.0: Fix bug in MLAS (duplicate entries in functional_children).
22
- # Add --counts option.
23
- # - [02 May 2018] Version 1.1: When removing spaces to match gold and system characters,
24
- # consider all Unicode characters of category Zs instead of
25
- # just ASCII space.
26
- # - [25 Jun 2018] Version 1.2: Use python3 in the she-bang (instead of python).
27
- # In Python2, make the whole computation use `unicode` strings.
28
-
29
- # Command line usage
30
- # ------------------
31
- # conll18_ud_eval.py [-v] gold_conllu_file system_conllu_file
32
- #
33
- # - if no -v is given, only the official CoNLL18 UD Shared Task evaluation metrics
34
- # are printed
35
- # - if -v is given, more metrics are printed (as precision, recall, F1 score,
36
- # and in case the metric is computed on aligned words also accuracy on these):
37
- # - Tokens: how well do the gold tokens match system tokens
38
- # - Sentences: how well do the gold sentences match system sentences
39
- # - Words: how well can the gold words be aligned to system words
40
- # - UPOS: using aligned words, how well does UPOS match
41
- # - XPOS: using aligned words, how well does XPOS match
42
- # - UFeats: using aligned words, how well does universal FEATS match
43
- # - AllTags: using aligned words, how well does UPOS+XPOS+FEATS match
44
- # - Lemmas: using aligned words, how well does LEMMA match
45
- # - UAS: using aligned words, how well does HEAD match
46
- # - LAS: using aligned words, how well does HEAD+DEPREL(ignoring subtypes) match
47
- # - CLAS: using aligned words with content DEPREL, how well does
48
- # HEAD+DEPREL(ignoring subtypes) match
49
- # - MLAS: using aligned words with content DEPREL, how well does
50
- # HEAD+DEPREL(ignoring subtypes)+UPOS+UFEATS+FunctionalChildren(DEPREL+UPOS+UFEATS) match
51
- # - BLEX: using aligned words with content DEPREL, how well does
52
- # HEAD+DEPREL(ignoring subtypes)+LEMMAS match
53
- # - if -c is given, raw counts of correct/gold_total/system_total/aligned words are printed
54
- # instead of precision/recall/F1/AlignedAccuracy for all metrics.
55
-
56
- # API usage
57
- # ---------
58
- # - load_conllu(file)
59
- # - loads CoNLL-U file from given file object to an internal representation
60
- # - the file object should return str in both Python 2 and Python 3
61
- # - raises UDError exception if the given file cannot be loaded
62
- # - evaluate(gold_ud, system_ud)
63
- # - evaluate the given gold and system CoNLL-U files (loaded with load_conllu)
64
- # - raises UDError if the concatenated tokens of gold and system file do not match
65
- # - returns a dictionary with the metrics described above, each metric having
66
- # three fields: precision, recall and f1
67
-
68
- # Description of token matching
69
- # -----------------------------
70
- # In order to match tokens of gold file and system file, we consider the text
71
- # resulting from concatenation of gold tokens and text resulting from
72
- # concatenation of system tokens. These texts should match -- if they do not,
73
- # the evaluation fails.
74
- #
75
- # If the texts do match, every token is represented as a range in this original
76
- # text, and tokens are equal only if their range is the same.
77
-
78
- # Description of word matching
79
- # ----------------------------
80
- # When matching words of gold file and system file, we first match the tokens.
81
- # The words which are also tokens are matched as tokens, but words in multi-word
82
- # tokens have to be handled differently.
83
- #
84
- # To handle multi-word tokens, we start by finding "multi-word spans".
85
- # Multi-word span is a span in the original text such that
86
- # - it contains at least one multi-word token
87
- # - all multi-word tokens in the span (considering both gold and system ones)
88
- # are completely inside the span (i.e., they do not "stick out")
89
- # - the multi-word span is as small as possible
90
- #
91
- # For every multi-word span, we align the gold and system words completely
92
- # inside this span using LCS on their FORMs. The words not intersecting
93
- # (even partially) any multi-word span are then aligned as tokens.
94
-
95
-
96
- from __future__ import division
97
- from __future__ import print_function
98
-
99
- import argparse
100
- import io
101
- import sys
102
- import unicodedata
103
- import unittest
104
- from collections import defaultdict
105
-
106
- CONFUSIONS = defaultdict(lambda: defaultdict(int)) #ADDED
107
-
108
- # CoNLL-U column names
109
- ID, FORM, LEMMA, UPOS, XPOS, FEATS, HEAD, DEPREL, DEPS, MISC = range(10)
110
-
111
- # Content and functional relations
112
- CONTENT_DEPRELS = {
113
- "acl", "advcl", "advmod", "amod", "appos", "aux", "case", "cc", "ccomp",
114
- "clf", "compound", "conj", "cop", "csubj", "dep", "det", "discourse",
115
- "dislocated", "expl", "fixed", "flat", "goeswith", "iobj", "list", "mark",
116
- "nmod", "nsubj", "nummod", "obj", "obl", "orphan", "parataxis", "punct",
117
- "reparandum", "root", "vocative", "xcomp"
118
- }
119
- # Add for detailed scores
120
- UPOS_TAGS = {
121
- "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM",
122
- "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X"
123
- }
124
- ##
125
- FUNCTIONAL_DEPRELS = {
126
- "aux", "cop", "mark", "det", "clf", "case", "cc"
127
- }
128
-
129
- UNIVERSAL_FEATURES = {
130
- "PronType", "NumType", "Poss", "Reflex", "Foreign", "Abbr", "Gender",
131
- "Animacy", "Number", "Case", "Definite", "Degree", "VerbForm", "Mood",
132
- "Tense", "Aspect", "Voice", "Evident", "Polarity", "Person", "Polite"
133
- }
134
-
135
- # UD Error is used when raising exceptions in this module
136
- class UDError(Exception):
137
- pass
138
-
139
- # Conversion methods handling `str` <-> `unicode` conversions in Python2
140
- def _decode(text):
141
- return text if sys.version_info[0] >= 3 or not isinstance(text, str) else text.decode("utf-8")
142
-
143
- def _encode(text):
144
- return text if sys.version_info[0] >= 3 or not isinstance(text, unicode) else text.encode("utf-8")
145
-
146
- # Load given CoNLL-U file into internal representation
147
- def load_conllu(file, ignore_invalid_format=False):
148
- # Internal representation classes
149
- class UDRepresentation:
150
- def __init__(self):
151
- # Characters of all the tokens in the whole file.
152
- # Whitespace between tokens is not included.
153
- self.characters = []
154
- # List of UDSpan instances with start&end indices into `characters`.
155
- self.tokens = []
156
- # List of UDWord instances.
157
- self.words = []
158
- # List of UDSpan instances with start&end indices into `characters`.
159
- self.sentences = []
160
- class UDSpan:
161
- def __init__(self, start, end):
162
- self.start = start
163
- # Note that self.end marks the first position **after the end** of span,
164
- # so we can use characters[start:end] or range(start, end).
165
- self.end = end
166
- class UDWord:
167
- def __init__(self, span, columns, is_multiword):
168
- # Span of this word (or MWT, see below) within ud_representation.characters.
169
- self.span = span
170
- # 10 columns of the CoNLL-U file: ID, FORM, LEMMA,...
171
- self.columns = columns
172
- # is_multiword==True means that this word is part of a multi-word token.
173
- # In that case, self.span marks the span of the whole multi-word token.
174
- self.is_multiword = is_multiword
175
- # Reference to the UDWord instance representing the HEAD (or None if root).
176
- self.parent = None
177
- # List of references to UDWord instances representing functional-deprel children.
178
- self.functional_children = []
179
- # Only consider universal FEATS.
180
- self.columns[FEATS] = "|".join(sorted(feat for feat in columns[FEATS].split("|")
181
- if feat.split("=", 1)[0] in UNIVERSAL_FEATURES))
182
- # Let's ignore language-specific deprel subtypes.
183
- self.columns[DEPREL] = columns[DEPREL].split(":")[0]
184
- # Precompute which deprels are CONTENT_DEPRELS and which FUNCTIONAL_DEPRELS
185
- self.is_content_deprel = self.columns[DEPREL] in CONTENT_DEPRELS
186
- self.is_functional_deprel = self.columns[DEPREL] in FUNCTIONAL_DEPRELS
187
-
188
- ud = UDRepresentation()
189
-
190
- # Load the CoNLL-U file
191
- index, sentence_start = 0, None
192
- while True:
193
- line = file.readline()
194
- if not line:
195
- break
196
- line = _decode(line.rstrip("\r\n"))
197
-
198
- # Handle sentence start boundaries
199
- if sentence_start is None:
200
- # Skip comments
201
- if line.startswith("#"):
202
- continue
203
- # Start a new sentence
204
- ud.sentences.append(UDSpan(index, 0))
205
- sentence_start = len(ud.words)
206
- if not line:
207
- # Add parent and children UDWord links and check there are no cycles
208
- def process_word(word):
209
- if word.parent == "remapping":
210
- if not ignore_invalid_format:
211
- raise UDError("There is a cycle in a sentence")
212
- if word.parent is None:
213
- try:
214
- head = int(word.columns[HEAD])
215
- except ValueError:
216
- head = 0 # Astrid: handle headless nodes as roots
217
- if head < 0 or head > len(ud.words) - sentence_start:
218
- if not ignore_invalid_format:
219
- raise UDError("HEAD '{}' points outside of the sentence".format(_encode(word.columns[HEAD])))
220
- if head:
221
- parent = ud.words[sentence_start + head - 1]
222
- word.parent = "remapping"
223
- process_word(parent)
224
- word.parent = parent
225
-
226
- for word in ud.words[sentence_start:]:
227
- process_word(word)
228
- # func_children cannot be assigned within process_word
229
- # because it is called recursively and may result in adding one child twice.
230
- for word in ud.words[sentence_start:]:
231
- if word.parent and word.is_functional_deprel:
232
- word.parent.functional_children.append(word)
233
-
234
- # Check there is a single root node
235
- if len([word for word in ud.words[sentence_start:] if word.parent is None]) != 1:
236
- #print(f"Warning: There are {len([word for word in ud.words[sentence_start:] if word.parent is None])} roots in a sentence.")
237
- if not ignore_invalid_format:
238
- raise UDError("There are multiple roots in a sentence")
239
-
240
- # End the sentence
241
- ud.sentences[-1].end = index
242
- sentence_start = None
243
- continue
244
-
245
- # Read next token/word
246
- columns = line.split("\t")
247
- if len(columns) != 10:
248
- raise UDError("The CoNLL-U line does not contain 10 tab-separated columns: '{}'".format(_encode(line)))
249
-
250
- # Skip empty nodes
251
- if "." in columns[ID]:
252
- continue
253
-
254
- # Delete spaces from FORM, so gold.characters == system.characters
255
- # even if one of them tokenizes the space. Use any Unicode character
256
- # with category Zs.
257
- columns[FORM] = "".join(filter(lambda c: unicodedata.category(c) != "Zs", columns[FORM]))
258
- if not columns[FORM]:
259
- raise UDError("There is an empty FORM in the CoNLL-U file")
260
-
261
- # Save token
262
- ud.characters.extend(columns[FORM])
263
- ud.tokens.append(UDSpan(index, index + len(columns[FORM])))
264
- index += len(columns[FORM])
265
-
266
- # Handle multi-word tokens to save word(s)
267
- if "-" in columns[ID]:
268
- try:
269
- start, end = map(int, columns[ID].split("-"))
270
- except:
271
- raise UDError("Cannot parse multi-word token ID '{}'".format(_encode(columns[ID])))
272
-
273
- for _ in range(start, end + 1):
274
- word_line = _decode(file.readline().rstrip("\r\n"))
275
- word_columns = word_line.split("\t")
276
- if len(word_columns) != 10:
277
- raise UDError("The CoNLL-U line does not contain 10 tab-separated columns: '{}'".format(_encode(word_line)))
278
- ud.words.append(UDWord(ud.tokens[-1], word_columns, is_multiword=True))
279
- # Basic tokens/words
280
- else:
281
- try:
282
- word_id = int(columns[ID])
283
- except:
284
- raise UDError("Cannot parse word ID '{}'".format(_encode(columns[ID])))
285
- if word_id != len(ud.words) - sentence_start + 1:
286
- raise UDError("Incorrect word ID '{}' for word '{}', expected '{}'".format(
287
- _encode(columns[ID]), _encode(columns[FORM]), len(ud.words) - sentence_start + 1))
288
-
289
- try:
290
- head_id = int(columns[HEAD])
291
- except:
292
- if ignore_invalid_format:
293
- head_id = 0 # Astrid: handle headless nodes as roots
294
- else:
295
- raise UDError("Cannot parse HEAD '{}'".format(_encode(columns[HEAD])))
296
- if head_id < 0:
297
- raise UDError("HEAD cannot be negative")
298
-
299
- ud.words.append(UDWord(ud.tokens[-1], columns, is_multiword=False))
300
-
301
- if sentence_start is not None:
302
- raise UDError("The CoNLL-U file does not end with empty line")
303
-
304
- return ud
305
-
306
- # Evaluate the gold and system treebanks (loaded using load_conllu).
307
- def evaluate(gold_ud, system_ud, detailed=False, cm=False): #ADDED cm=False
308
- class Score:
309
- def __init__(self, gold_total, system_total, correct, aligned_total=None):
310
- self.correct = correct
311
- self.gold_total = gold_total
312
- self.system_total = system_total
313
- self.aligned_total = aligned_total
314
- self.precision = correct / system_total if system_total else 0.0
315
- self.recall = correct / gold_total if gold_total else 0.0
316
- self.f1 = 2 * correct / (system_total + gold_total) if system_total + gold_total else 0.0
317
- self.aligned_accuracy = correct / aligned_total if aligned_total else aligned_total
318
- class AlignmentWord:
319
- def __init__(self, gold_word, system_word):
320
- self.gold_word = gold_word
321
- self.system_word = system_word
322
- class Alignment:
323
- def __init__(self, gold_words, system_words):
324
- self.gold_words = gold_words
325
- self.system_words = system_words
326
- self.matched_words = []
327
- self.matched_words_map = {}
328
- def append_aligned_words(self, gold_word, system_word):
329
- self.matched_words.append(AlignmentWord(gold_word, system_word))
330
- self.matched_words_map[system_word] = gold_word
331
-
332
- def spans_score(gold_spans, system_spans):
333
- correct, gi, si = 0, 0, 0
334
- while gi < len(gold_spans) and si < len(system_spans):
335
- if system_spans[si].start < gold_spans[gi].start:
336
- si += 1
337
- elif gold_spans[gi].start < system_spans[si].start:
338
- gi += 1
339
- else:
340
- correct += gold_spans[gi].end == system_spans[si].end
341
- si += 1
342
- gi += 1
343
-
344
- return Score(len(gold_spans), len(system_spans), correct)
345
-
346
- def alignment_score(alignment, key_fn=None, filter_fn=None, name=None): # added "name" flag for confusion matrix
347
- if filter_fn is not None:
348
- gold = sum(1 for gold in alignment.gold_words if filter_fn(gold))
349
- system = sum(1 for system in alignment.system_words if filter_fn(system))
350
- aligned = sum(
351
- 1 for pair in alignment.matched_words
352
- if filter_fn(pair.gold_word) and filter_fn(pair.system_word))
353
- else:
354
- gold = len(alignment.gold_words)
355
- system = len(alignment.system_words)
356
- aligned = len(alignment.matched_words)
357
-
358
- if key_fn is None:
359
- # Return score for whole aligned words
360
- return Score(gold, system, aligned)
361
-
362
- def gold_aligned_gold(word):
363
- return word
364
- def gold_aligned_system(word):
365
- return alignment.matched_words_map.get(word, "NotAligned") if word is not None else None
366
- correct = 0
367
- for words in alignment.matched_words:
368
- if filter_fn is None or filter_fn(words.gold_word):
369
- gold_label = key_fn(words.gold_word, gold_aligned_gold) # ADDED
370
- system_label = key_fn(words.system_word, gold_aligned_system) # ADDED
371
- if key_fn(words.gold_word, gold_aligned_gold) == key_fn(words.system_word, gold_aligned_system):
372
- correct += 1
373
-
374
- if name is not None:
375
- CONFUSIONS[name][(gold_label, system_label)] += 1
376
-
377
- return Score(gold, system, correct, aligned)
378
-
379
- def beyond_end(words, i, multiword_span_end):
380
- if i >= len(words):
381
- return True
382
- if words[i].is_multiword:
383
- return words[i].span.start >= multiword_span_end
384
- return words[i].span.end > multiword_span_end
385
-
386
- def extend_end(word, multiword_span_end):
387
- if word.is_multiword and word.span.end > multiword_span_end:
388
- return word.span.end
389
- return multiword_span_end
390
-
391
- def find_multiword_span(gold_words, system_words, gi, si):
392
- # We know gold_words[gi].is_multiword or system_words[si].is_multiword.
393
- # Find the start of the multiword span (gs, ss), so the multiword span is minimal.
394
- # Initialize multiword_span_end characters index.
395
- if gold_words[gi].is_multiword:
396
- multiword_span_end = gold_words[gi].span.end
397
- if not system_words[si].is_multiword and system_words[si].span.start < gold_words[gi].span.start:
398
- si += 1
399
- else: # if system_words[si].is_multiword
400
- multiword_span_end = system_words[si].span.end
401
- if not gold_words[gi].is_multiword and gold_words[gi].span.start < system_words[si].span.start:
402
- gi += 1
403
- gs, ss = gi, si
404
-
405
- # Find the end of the multiword span
406
- # (so both gi and si are pointing to the word following the multiword span end).
407
- while not beyond_end(gold_words, gi, multiword_span_end) or \
408
- not beyond_end(system_words, si, multiword_span_end):
409
- if gi < len(gold_words) and (si >= len(system_words) or
410
- gold_words[gi].span.start <= system_words[si].span.start):
411
- multiword_span_end = extend_end(gold_words[gi], multiword_span_end)
412
- gi += 1
413
- else:
414
- multiword_span_end = extend_end(system_words[si], multiword_span_end)
415
- si += 1
416
- return gs, ss, gi, si
417
-
418
- def compute_lcs(gold_words, system_words, gi, si, gs, ss):
419
- lcs = [[0] * (si - ss) for i in range(gi - gs)]
420
- for g in reversed(range(gi - gs)):
421
- for s in reversed(range(si - ss)):
422
- if gold_words[gs + g].columns[FORM].lower() == system_words[ss + s].columns[FORM].lower():
423
- lcs[g][s] = 1 + (lcs[g+1][s+1] if g+1 < gi-gs and s+1 < si-ss else 0)
424
- lcs[g][s] = max(lcs[g][s], lcs[g+1][s] if g+1 < gi-gs else 0)
425
- lcs[g][s] = max(lcs[g][s], lcs[g][s+1] if s+1 < si-ss else 0)
426
- return lcs
427
-
428
- def align_words(gold_words, system_words):
429
- alignment = Alignment(gold_words, system_words)
430
-
431
- gi, si = 0, 0
432
- while gi < len(gold_words) and si < len(system_words):
433
- if gold_words[gi].is_multiword or system_words[si].is_multiword:
434
- # A: Multi-word tokens => align via LCS within the whole "multiword span".
435
- gs, ss, gi, si = find_multiword_span(gold_words, system_words, gi, si)
436
-
437
- if si > ss and gi > gs:
438
- lcs = compute_lcs(gold_words, system_words, gi, si, gs, ss)
439
-
440
- # Store aligned words
441
- s, g = 0, 0
442
- while g < gi - gs and s < si - ss:
443
- if gold_words[gs + g].columns[FORM].lower() == system_words[ss + s].columns[FORM].lower():
444
- alignment.append_aligned_words(gold_words[gs+g], system_words[ss+s])
445
- g += 1
446
- s += 1
447
- elif lcs[g][s] == (lcs[g+1][s] if g+1 < gi-gs else 0):
448
- g += 1
449
- else:
450
- s += 1
451
- else:
452
- # B: No multi-word token => align according to spans.
453
- if (gold_words[gi].span.start, gold_words[gi].span.end) == (system_words[si].span.start, system_words[si].span.end):
454
- alignment.append_aligned_words(gold_words[gi], system_words[si])
455
- gi += 1
456
- si += 1
457
- elif gold_words[gi].span.start <= system_words[si].span.start:
458
- gi += 1
459
- else:
460
- si += 1
461
-
462
- return alignment
463
-
464
- # Check that the underlying character sequences do match.
465
- if gold_ud.characters != system_ud.characters:
466
- """index = 0
467
- while index < len(gold_ud.characters) and index < len(system_ud.characters) and \
468
- gold_ud.characters[index] == system_ud.characters[index]:
469
- index += 1
470
-
471
- raise UDError(
472
- "The concatenation of tokens in gold file and in system file differ!\n" +
473
- "First 20 differing characters in gold file: '{}' and system file: '{}'".format(
474
- "".join(map(_encode, gold_ud.characters[index:index + 20])),
475
- "".join(map(_encode, system_ud.characters[index:index + 20]))
476
- )
477
- )"""
478
- gold_ud.characters = system_ud.characters # Astrid: token form is altered - go with base tree form
479
-
480
-
481
- # Added to compute detailed UPOS scores:
482
- def detailed_score(alignment, labels, column_ind):
483
- r = {}
484
- total_gold, total_system, total_correct = 0, 0, 0
485
- for label in labels:
486
- score = alignment_score(
487
- alignment,
488
- filter_fn=lambda w, label=label: w.columns[column_ind] == label
489
- )
490
-
491
- r[label] = {
492
- "precision": score.precision,
493
- "recall": score.recall,
494
- "f1": score.f1
495
- }
496
- total_gold += score.gold_total
497
- total_system += score.system_total
498
- total_correct += score.correct
499
-
500
- r[f'TOTAL'] = {
501
- "precision": total_correct / total_system if total_system else 0.0,
502
- "recall": total_correct / total_gold if total_gold else 0.0,
503
- "f1": 2 * total_correct / (total_system + total_gold) if (total_system + total_gold) else 0.0
504
- }
505
- return r
506
-
507
- # Added to compute detailed UPOS scores:
508
-
509
- ## ADDED
510
- def dump_confusion(confusion, filename):
511
- labels = sorted(set(g for g, s in confusion) | set(s for g, s in confusion))
512
- with open(filename, "w", encoding="utf-8") as f:
513
- f.write("gold\\system\t" + "\t".join(map(str, labels)) + "\n")
514
- for g in labels:
515
- row = [str(confusion.get((g, s), 0)) for s in labels]
516
- f.write(str(g) + "\t" + "\t".join(row) + "\n")
517
-
518
- #for name, matrix in CONFUSIONS.items():
519
- # dump_confusion(matrix, f"{name}_confusion.tsv")
520
- ## ADDED
521
-
522
-
523
- # Align words
524
- alignment = align_words(gold_ud.words, system_ud.words)
525
-
526
- # Compute the F1-scores
527
- if detailed: #The initial F1-scores
528
- upos_scores = detailed_score(alignment, UPOS_TAGS, UPOS)
529
- deprel_scores = detailed_score(alignment, CONTENT_DEPRELS, DEPREL)
530
- return upos_scores, deprel_scores
531
-
532
- # For Confusion matrix
533
- if cm:
534
- sc = {
535
- "UPOS": alignment_score(alignment,lambda w, _: w.columns[UPOS], name = "UPOS"),
536
- "DEPREL": alignment_score(alignment, lambda w, _: w.columns[DEPREL], name ="DEPREL"),
537
- }
538
- return CONFUSIONS
539
-
540
-
541
- return {
542
- "Tokens": spans_score(gold_ud.tokens, system_ud.tokens),
543
- "Sentences": spans_score(gold_ud.sentences, system_ud.sentences),
544
- "Words": alignment_score(alignment),
545
- "UPOS": alignment_score(alignment, lambda w, _: w.columns[UPOS]),
546
- "XPOS": alignment_score(alignment, lambda w, _: w.columns[XPOS]),
547
- "UFeats": alignment_score(alignment, lambda w, _: w.columns[FEATS]),
548
- "AllTags": alignment_score(alignment, lambda w, _: (w.columns[UPOS], w.columns[XPOS], w.columns[FEATS])),
549
- "Lemmas": alignment_score(alignment, lambda w, ga: w.columns[LEMMA] if ga(w).columns[LEMMA] != "_" else "_"),
550
- "UAS": alignment_score(alignment, lambda w, ga: ga(w.parent)),
551
- "LAS": alignment_score(alignment, lambda w, ga: (ga(w.parent), w.columns[DEPREL])),
552
- "CLAS": alignment_score(alignment, lambda w, ga: (ga(w.parent), w.columns[DEPREL]),
553
- filter_fn=lambda w: w.is_content_deprel),
554
- "MLAS": alignment_score(alignment, lambda w, ga: (ga(w.parent), w.columns[DEPREL], w.columns[UPOS], w.columns[FEATS],
555
- [(ga(c), c.columns[DEPREL], c.columns[UPOS], c.columns[FEATS])
556
- for c in w.functional_children]),
557
- filter_fn=lambda w: w.is_content_deprel),
558
- "BLEX": alignment_score(alignment, lambda w, ga: (ga(w.parent), w.columns[DEPREL],
559
- w.columns[LEMMA] if ga(w).columns[LEMMA] != "_" else "_"),
560
- filter_fn=lambda w: w.is_content_deprel),
561
- }
562
-
563
-
564
- def load_conllu_file(path, ignore_invalid_format=False):
565
- _file = open(path, mode="r", **({"encoding": "utf-8"} if sys.version_info >= (3, 0) else {}))
566
- return load_conllu(_file, ignore_invalid_format=ignore_invalid_format)
567
-
568
- def evaluate_wrapper(args):
569
- # Load CoNLL-U files
570
- gold_ud = load_conllu_file(args.gold_file)
571
- system_ud = load_conllu_file(args.system_file, ignore_invalid_format=args.ignore_invalid_format)
572
- return evaluate(gold_ud, system_ud)
573
-
574
- def main():
575
- # Parse arguments
576
- parser = argparse.ArgumentParser()
577
- parser.add_argument("gold_file", type=str,
578
- help="Name of the CoNLL-U file with the gold data.")
579
- parser.add_argument("system_file", type=str,
580
- help="Name of the CoNLL-U file with the predicted data.")
581
- parser.add_argument("--verbose", "-v", default=False, action="store_true",
582
- help="Print all metrics.")
583
- parser.add_argument("--counts", "-c", default=False, action="store_true",
584
- help="Print raw counts of correct/gold/system/aligned words instead of prec/rec/F1 for all metrics.")
585
- parser.add_argument("--ignore-invalid-format", "-i", default=False, action="store_true",
586
- help="Ignore invalid format errors, such as multiple roots, when loading CoNLL-U files.")
587
- args = parser.parse_args()
588
-
589
- # Evaluate
590
- evaluation = evaluate_wrapper(args)
591
-
592
- # Print the evaluation
593
- if not args.verbose and not args.counts:
594
- print("LAS F1 Score: {:.2f}".format(100 * evaluation["LAS"].f1))
595
- print("MLAS Score: {:.2f}".format(100 * evaluation["MLAS"].f1))
596
- print("BLEX Score: {:.2f}".format(100 * evaluation["BLEX"].f1))
597
- else:
598
- if args.counts:
599
- print("Metric | Correct | Gold | Predicted | Aligned")
600
- else:
601
- print("Metric | Precision | Recall | F1 Score | AligndAcc")
602
- print("-----------+-----------+-----------+-----------+-----------")
603
- for metric in["Tokens", "Sentences", "Words", "UPOS", "XPOS", "UFeats", "AllTags", "Lemmas", "UAS", "LAS", "CLAS", "MLAS", "BLEX"]:
604
- if args.counts:
605
- print("{:11}|{:10} |{:10} |{:10} |{:10}".format(
606
- metric,
607
- evaluation[metric].correct,
608
- evaluation[metric].gold_total,
609
- evaluation[metric].system_total,
610
- evaluation[metric].aligned_total or (evaluation[metric].correct if metric == "Words" else "")
611
- ))
612
- else:
613
- print("{:11}|{:10.2f} |{:10.2f} |{:10.2f} |{}".format(
614
- metric,
615
- 100 * evaluation[metric].precision,
616
- 100 * evaluation[metric].recall,
617
- 100 * evaluation[metric].f1,
618
- "{:10.2f}".format(100 * evaluation[metric].aligned_accuracy) if evaluation[metric].aligned_accuracy is not None else ""
619
- ))
620
-
621
- if __name__ == "__main__":
622
- main()
623
-
624
- # Tests, which can be executed with `python -m unittest conll18_ud_eval`.
625
- class TestAlignment(unittest.TestCase):
626
- @staticmethod
627
- def _load_words(words):
628
- """Prepare fake CoNLL-U files with fake HEAD to prevent multiple roots errors."""
629
- lines, num_words = [], 0
630
- for w in words:
631
- parts = w.split(" ")
632
- if len(parts) == 1:
633
- num_words += 1
634
- lines.append("{}\t{}\t_\t_\t_\t_\t{}\t_\t_\t_".format(num_words, parts[0], int(num_words>1)))
635
- else:
636
- lines.append("{}-{}\t{}\t_\t_\t_\t_\t_\t_\t_\t_".format(num_words + 1, num_words + len(parts) - 1, parts[0]))
637
- for part in parts[1:]:
638
- num_words += 1
639
- lines.append("{}\t{}\t_\t_\t_\t_\t{}\t_\t_\t_".format(num_words, part, int(num_words>1)))
640
- return load_conllu((io.StringIO if sys.version_info >= (3, 0) else io.BytesIO)("\n".join(lines+["\n"])))
641
-
642
- def _test_exception(self, gold, system):
643
- self.assertRaises(UDError, evaluate, self._load_words(gold), self._load_words(system))
644
-
645
- def _test_ok(self, gold, system, correct):
646
- metrics = evaluate(self._load_words(gold), self._load_words(system))
647
- gold_words = sum((max(1, len(word.split(" ")) - 1) for word in gold))
648
- system_words = sum((max(1, len(word.split(" ")) - 1) for word in system))
649
- self.assertEqual((metrics["Words"].precision, metrics["Words"].recall, metrics["Words"].f1),
650
- (correct / system_words, correct / gold_words, 2 * correct / (gold_words + system_words)))
651
-
652
- def test_exception(self):
653
- self._test_exception(["a"], ["b"])
654
-
655
- def test_equal(self):
656
- self._test_ok(["a"], ["a"], 1)
657
- self._test_ok(["a", "b", "c"], ["a", "b", "c"], 3)
658
-
659
- def test_equal_with_multiword(self):
660
- self._test_ok(["abc a b c"], ["a", "b", "c"], 3)
661
- self._test_ok(["a", "bc b c", "d"], ["a", "b", "c", "d"], 4)
662
- self._test_ok(["abcd a b c d"], ["ab a b", "cd c d"], 4)
663
- self._test_ok(["abc a b c", "de d e"], ["a", "bcd b c d", "e"], 5)
664
-
665
- def test_alignment(self):
666
- self._test_ok(["abcd"], ["a", "b", "c", "d"], 0)
667
- self._test_ok(["abc", "d"], ["a", "b", "c", "d"], 1)
668
- self._test_ok(["a", "bc", "d"], ["a", "b", "c", "d"], 2)
669
- self._test_ok(["a", "bc b c", "d"], ["a", "b", "cd"], 2)
670
- self._test_ok(["abc a BX c", "def d EX f"], ["ab a b", "cd c d", "ef e f"], 4)
671
- self._test_ok(["ab a b", "cd bc d"], ["a", "bc", "d"], 2)
672
- self._test_ok(["a", "bc b c", "d"], ["ab AX BX", "cd CX a"], 1)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
eval/scores/__init__.py DELETED
File without changes
eval/scores/scores_talbanken.txt DELETED
@@ -1,162 +0,0 @@
1
- Sample: svediakorp-sec987-spf144, time period: 1800-1850
2
- Metric Precision Recall F1 Aligned Accuracy
3
- UPOS 1.0000 1.0000 1.0000 1.0000
4
- UAS 1.0000 1.0000 1.0000 1.0000
5
- LAS 0.9615 0.9615 0.9615 0.9615
6
-
7
- Sample: svediakorp-sec613-EngstromA_StrindbergOchJag, time period: 1900-1950
8
- Metric Precision Recall F1 Aligned Accuracy
9
- UPOS 1.0000 1.0000 1.0000 1.0000
10
- UAS 0.9146 0.9146 0.9146 0.9146
11
- LAS 0.8659 0.8659 0.8659 0.8659
12
-
13
- Sample: svediakorp-sec268-DulciU_VitterhetsNojen3, time period: 1750-1800
14
- Metric Precision Recall F1 Aligned Accuracy
15
- UPOS 0.9663 0.9663 0.9663 0.9663
16
- UAS 0.8798 0.8798 0.8798 0.8798
17
- LAS 0.8365 0.8365 0.8365 0.8365
18
-
19
- Sample: svediakorp-sec25-Runius, time period: 1700-1750
20
- Metric Precision Recall F1 Aligned Accuracy
21
- UPOS 0.8750 0.8750 0.8750 0.8750
22
- UAS 0.7188 0.7188 0.7188 0.7188
23
- LAS 0.6562 0.6562 0.6562 0.6562
24
-
25
- Sample: svediakorp-sec208-Anonym_DetGrasligaMordet, time period: 1850-1900
26
- Metric Precision Recall F1 Aligned Accuracy
27
- UPOS 0.9865 0.9865 0.9865 0.9865
28
- UAS 0.8784 0.8784 0.8784 0.8784
29
- LAS 0.8243 0.8243 0.8243 0.8243
30
-
31
- Sample: svediakorp-sec324-GranbergPA_Enslighetsalskaren, time period: 1800-1850
32
- Metric Precision Recall F1 Aligned Accuracy
33
- UPOS 0.9400 0.9400 0.9400 0.9400
34
- UAS 0.8400 0.8400 0.8400 0.8400
35
- LAS 0.7400 0.7400 0.7400 0.7400
36
-
37
- Sample: svediakorp-sec254-CederborghF_BerattelseOmJohnHall, time period: 1800-1850
38
- Metric Precision Recall F1 Aligned Accuracy
39
- UPOS 0.9706 0.9706 0.9706 0.9706
40
- UAS 0.5412 0.5412 0.5412 0.5412
41
- LAS 0.5000 0.5000 0.5000 0.5000
42
-
43
- Sample: svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol, time period: 1900-1950
44
- Metric Precision Recall F1 Aligned Accuracy
45
- UPOS 0.9800 0.9800 0.9800 0.9800
46
- UAS 0.7200 0.7200 0.7200 0.7200
47
- LAS 0.7200 0.7200 0.7200 0.7200
48
-
49
- Sample: svediakorp-sec452-NyblomH_FantasierFyra, time period: 1850-1900
50
- Metric Precision Recall F1 Aligned Accuracy
51
- UPOS 0.9464 0.9464 0.9464 0.9464
52
- UAS 0.7500 0.7500 0.7500 0.7500
53
- LAS 0.6786 0.6786 0.6786 0.6786
54
-
55
- Sample: svediakorp-sec252-BremerF_Teckningar1, time period: 1800-1850
56
- Metric Precision Recall F1 Aligned Accuracy
57
- UPOS 1.0000 1.0000 1.0000 1.0000
58
- UAS 0.8750 0.8750 0.8750 0.8750
59
- LAS 0.8750 0.8750 0.8750 0.8750
60
-
61
- Sample: svediakorp-sec486-SchwartzMS_BellmansSkor, time period: 1850-1900
62
- Metric Precision Recall F1 Aligned Accuracy
63
- UPOS 0.9873 0.9873 0.9873 0.9873
64
- UAS 0.7722 0.7722 0.7722 0.7722
65
- LAS 0.6962 0.6962 0.6962 0.6962
66
-
67
- Sample: svediakorp-sec330-GyllenborgC_SwenskaSpratthoken, time period: 1700-1750
68
- Metric Precision Recall F1 Aligned Accuracy
69
- UPOS 0.8650 0.8650 0.8650 0.8650
70
- UAS 0.6687 0.6687 0.6687 0.6687
71
- LAS 0.5828 0.5828 0.5828 0.5828
72
-
73
- Sample: svediakorp-sec631-HasselskogN_HallaHallaGronkoping, time period: 1900-1950
74
- Metric Precision Recall F1 Aligned Accuracy
75
- UPOS 0.9605 0.9605 0.9605 0.9605
76
- UAS 0.7368 0.7368 0.7368 0.7368
77
- LAS 0.6974 0.6974 0.6974 0.6974
78
-
79
- Sample: svediakorp-sec277-EnbomPU_MedborgeligtSkalde, time period: 1750-1800
80
- Metric Precision Recall F1 Aligned Accuracy
81
- UPOS 0.9464 0.9464 0.9464 0.9464
82
- UAS 0.7500 0.7500 0.7500 0.7500
83
- LAS 0.6786 0.6786 0.6786 0.6786
84
-
85
- Sample: svediakorp-sec1102-spf259, time period: 1850-1900
86
- Metric Precision Recall F1 Aligned Accuracy
87
- UPOS 0.9444 0.9444 0.9444 0.9444
88
- UAS 0.8889 0.8889 0.8889 0.8889
89
- LAS 0.8889 0.8889 0.8889 0.8889
90
-
91
- Sample: svediakorp-sec991-spf148, time period: 1800-1850
92
- Metric Precision Recall F1 Aligned Accuracy
93
- UPOS 0.9561 0.9561 0.9561 0.9561
94
- UAS 0.7982 0.7982 0.7982 0.7982
95
- LAS 0.7368 0.7368 0.7368 0.7368
96
-
97
- Sample: svediakorp-sec639-HeidenstamV_Proletarfilosofiens, time period: 1900-1950
98
- Metric Precision Recall F1 Aligned Accuracy
99
- UPOS 0.9861 0.9861 0.9861 0.9861
100
- UAS 1.0000 1.0000 1.0000 1.0000
101
- LAS 1.0000 1.0000 1.0000 1.0000
102
-
103
- Sample: svediakorp-sec988-spf145, time period: 1800-1850
104
- Metric Precision Recall F1 Aligned Accuracy
105
- UPOS 0.9147 0.9147 0.9147 0.9147
106
- UAS 0.8450 0.8450 0.8450 0.8450
107
- LAS 0.7829 0.7829 0.7829 0.7829
108
-
109
- Sample: svediakorp-sec1063-spf220, time period: 1850-1900
110
- Metric Precision Recall F1 Aligned Accuracy
111
- UPOS 0.9660 0.9660 0.9660 0.9660
112
- UAS 0.9184 0.9184 0.9184 0.9184
113
- LAS 0.8707 0.8707 0.8707 0.8707
114
-
115
- Sample: svediakorp-sec1033-spf190, time period: 1900-1950
116
- Metric Precision Recall F1 Aligned Accuracy
117
- UPOS 0.9692 0.9692 0.9692 0.9692
118
- UAS 0.8923 0.8923 0.8923 0.8923
119
- LAS 0.8462 0.8462 0.8462 0.8462
120
-
121
- Sample: svediakorp-letter141673-Stalhammar, time period: 1700-1750
122
- Metric Precision Recall F1 Aligned Accuracy
123
- UPOS 0.7059 0.7059 0.7059 0.7059
124
- UAS 0.5294 0.5294 0.5294 0.5294
125
- LAS 0.4118 0.4118 0.4118 0.4118
126
-
127
- ----------------------------------------------------------------------
128
- Average scores for time period 1700-1750 (3 samples with a total of 8 sentences):
129
- Metric Precision Recall F1 Aligned Accuracy
130
- UPOS 0.8302 0.8302 0.8302 0.8302
131
- UAS 0.6589 0.6589 0.6589 0.6589
132
- LAS 0.5768 0.5768 0.5768 0.5768
133
-
134
- Average scores for time period 1750-1800 (2 samples with a total of 5 sentences):
135
- Metric Precision Recall F1 Aligned Accuracy
136
- UPOS 0.9584 0.9584 0.9584 0.9584
137
- UAS 0.8279 0.8279 0.8279 0.8279
138
- LAS 0.7734 0.7734 0.7734 0.7734
139
-
140
- Average scores for time period 1800-1850 (6 samples with a total of 19 sentences):
141
- Metric Precision Recall F1 Aligned Accuracy
142
- UPOS 0.9517 0.9517 0.9517 0.9517
143
- UAS 0.7774 0.7774 0.7774 0.7774
144
- LAS 0.7164 0.7164 0.7164 0.7164
145
-
146
- Average scores for time period 1850-1900 (5 samples with a total of 14 sentences):
147
- Metric Precision Recall F1 Aligned Accuracy
148
- UPOS 0.9692 0.9692 0.9692 0.9692
149
- UAS 0.8552 0.8552 0.8552 0.8552
150
- LAS 0.8006 0.8006 0.8006 0.8006
151
-
152
- Average scores for time period 1900-1950 (5 samples with a total of 13 sentences):
153
- Metric Precision Recall F1 Aligned Accuracy
154
- UPOS 0.9805 0.9805 0.9805 0.9805
155
- UAS 0.8719 0.8719 0.8719 0.8719
156
- LAS 0.8439 0.8439 0.8439 0.8439
157
-
158
- Overall scores across all time periods (21 samples with a total of 59 sentences):
159
- Metric Precision Recall F1 Aligned Accuracy
160
- UPOS 0.9511 0.9511 0.9511 0.9511
161
- UAS 0.7992 0.7992 0.7992 0.7992
162
- LAS 0.7452 0.7452 0.7452 0.7452
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
eval/scores/scores_transformer_no_silver.txt DELETED
@@ -1,162 +0,0 @@
1
- Sample: svediakorp-sec987-spf144, time period: 1800-1850
2
- Metric Precision Recall F1 Aligned Accuracy
3
- UPOS 1.0000 1.0000 1.0000 1.0000
4
- UAS 0.9615 0.9615 0.9615 0.9615
5
- LAS 0.9231 0.9231 0.9231 0.9231
6
-
7
- Sample: svediakorp-sec613-EngstromA_StrindbergOchJag, time period: 1900-1950
8
- Metric Precision Recall F1 Aligned Accuracy
9
- UPOS 1.0000 1.0000 1.0000 1.0000
10
- UAS 0.9024 0.9024 0.9024 0.9024
11
- LAS 0.8537 0.8537 0.8537 0.8537
12
-
13
- Sample: svediakorp-sec268-DulciU_VitterhetsNojen3, time period: 1750-1800
14
- Metric Precision Recall F1 Aligned Accuracy
15
- UPOS 0.9663 0.9663 0.9663 0.9663
16
- UAS 0.8317 0.8317 0.8317 0.8317
17
- LAS 0.7837 0.7837 0.7837 0.7837
18
-
19
- Sample: svediakorp-sec25-Runius, time period: 1700-1750
20
- Metric Precision Recall F1 Aligned Accuracy
21
- UPOS 0.8750 0.8750 0.8750 0.8750
22
- UAS 0.6875 0.6875 0.6875 0.6875
23
- LAS 0.5625 0.5625 0.5625 0.5625
24
-
25
- Sample: svediakorp-sec208-Anonym_DetGrasligaMordet, time period: 1850-1900
26
- Metric Precision Recall F1 Aligned Accuracy
27
- UPOS 0.9865 0.9865 0.9865 0.9865
28
- UAS 0.7973 0.7973 0.7973 0.7973
29
- LAS 0.7432 0.7432 0.7432 0.7432
30
-
31
- Sample: svediakorp-sec324-GranbergPA_Enslighetsalskaren, time period: 1800-1850
32
- Metric Precision Recall F1 Aligned Accuracy
33
- UPOS 0.9400 0.9400 0.9400 0.9400
34
- UAS 0.8800 0.8800 0.8800 0.8800
35
- LAS 0.8100 0.8100 0.8100 0.8100
36
-
37
- Sample: svediakorp-sec254-CederborghF_BerattelseOmJohnHall, time period: 1800-1850
38
- Metric Precision Recall F1 Aligned Accuracy
39
- UPOS 0.9706 0.9706 0.9706 0.9706
40
- UAS 0.5882 0.5882 0.5882 0.5882
41
- LAS 0.5118 0.5118 0.5118 0.5118
42
-
43
- Sample: svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol, time period: 1900-1950
44
- Metric Precision Recall F1 Aligned Accuracy
45
- UPOS 0.9800 0.9800 0.9800 0.9800
46
- UAS 0.7800 0.7800 0.7800 0.7800
47
- LAS 0.7800 0.7800 0.7800 0.7800
48
-
49
- Sample: svediakorp-sec452-NyblomH_FantasierFyra, time period: 1850-1900
50
- Metric Precision Recall F1 Aligned Accuracy
51
- UPOS 0.9464 0.9464 0.9464 0.9464
52
- UAS 0.8393 0.8393 0.8393 0.8393
53
- LAS 0.8036 0.8036 0.8036 0.8036
54
-
55
- Sample: svediakorp-sec252-BremerF_Teckningar1, time period: 1800-1850
56
- Metric Precision Recall F1 Aligned Accuracy
57
- UPOS 1.0000 1.0000 1.0000 1.0000
58
- UAS 0.8750 0.8750 0.8750 0.8750
59
- LAS 0.8750 0.8750 0.8750 0.8750
60
-
61
- Sample: svediakorp-sec486-SchwartzMS_BellmansSkor, time period: 1850-1900
62
- Metric Precision Recall F1 Aligned Accuracy
63
- UPOS 0.9873 0.9873 0.9873 0.9873
64
- UAS 0.7848 0.7848 0.7848 0.7848
65
- LAS 0.6835 0.6835 0.6835 0.6835
66
-
67
- Sample: svediakorp-sec330-GyllenborgC_SwenskaSpratthoken, time period: 1700-1750
68
- Metric Precision Recall F1 Aligned Accuracy
69
- UPOS 0.8650 0.8650 0.8650 0.8650
70
- UAS 0.7301 0.7301 0.7301 0.7301
71
- LAS 0.6135 0.6135 0.6135 0.6135
72
-
73
- Sample: svediakorp-sec631-HasselskogN_HallaHallaGronkoping, time period: 1900-1950
74
- Metric Precision Recall F1 Aligned Accuracy
75
- UPOS 0.9605 0.9605 0.9605 0.9605
76
- UAS 0.8684 0.8684 0.8684 0.8684
77
- LAS 0.7632 0.7632 0.7632 0.7632
78
-
79
- Sample: svediakorp-sec277-EnbomPU_MedborgeligtSkalde, time period: 1750-1800
80
- Metric Precision Recall F1 Aligned Accuracy
81
- UPOS 0.9464 0.9464 0.9464 0.9464
82
- UAS 0.7857 0.7857 0.7857 0.7857
83
- LAS 0.7500 0.7500 0.7500 0.7500
84
-
85
- Sample: svediakorp-sec1102-spf259, time period: 1850-1900
86
- Metric Precision Recall F1 Aligned Accuracy
87
- UPOS 0.9444 0.9444 0.9444 0.9444
88
- UAS 0.8889 0.8889 0.8889 0.8889
89
- LAS 0.8889 0.8889 0.8889 0.8889
90
-
91
- Sample: svediakorp-sec991-spf148, time period: 1800-1850
92
- Metric Precision Recall F1 Aligned Accuracy
93
- UPOS 0.9561 0.9561 0.9561 0.9561
94
- UAS 0.7982 0.7982 0.7982 0.7982
95
- LAS 0.7807 0.7807 0.7807 0.7807
96
-
97
- Sample: svediakorp-sec639-HeidenstamV_Proletarfilosofiens, time period: 1900-1950
98
- Metric Precision Recall F1 Aligned Accuracy
99
- UPOS 0.9861 0.9861 0.9861 0.9861
100
- UAS 0.9583 0.9583 0.9583 0.9583
101
- LAS 0.9444 0.9444 0.9444 0.9444
102
-
103
- Sample: svediakorp-sec988-spf145, time period: 1800-1850
104
- Metric Precision Recall F1 Aligned Accuracy
105
- UPOS 0.9147 0.9147 0.9147 0.9147
106
- UAS 0.8682 0.8682 0.8682 0.8682
107
- LAS 0.7984 0.7984 0.7984 0.7984
108
-
109
- Sample: svediakorp-sec1063-spf220, time period: 1850-1900
110
- Metric Precision Recall F1 Aligned Accuracy
111
- UPOS 0.9660 0.9660 0.9660 0.9660
112
- UAS 0.9116 0.9116 0.9116 0.9116
113
- LAS 0.8503 0.8503 0.8503 0.8503
114
-
115
- Sample: svediakorp-sec1033-spf190, time period: 1900-1950
116
- Metric Precision Recall F1 Aligned Accuracy
117
- UPOS 0.9692 0.9692 0.9692 0.9692
118
- UAS 0.9077 0.9077 0.9077 0.9077
119
- LAS 0.8154 0.8154 0.8154 0.8154
120
-
121
- Sample: svediakorp-letter141673-Stalhammar, time period: 1700-1750
122
- Metric Precision Recall F1 Aligned Accuracy
123
- UPOS 0.7059 0.7059 0.7059 0.7059
124
- UAS 0.6471 0.6471 0.6471 0.6471
125
- LAS 0.4706 0.4706 0.4706 0.4706
126
-
127
- ----------------------------------------------------------------------
128
- Average scores for time period 1700-1750 (3 samples with a total of 8 sentences):
129
- Metric Precision Recall F1 Aligned Accuracy
130
- UPOS 0.8302 0.8302 0.8302 0.8302
131
- UAS 0.6880 0.6880 0.6880 0.6880
132
- LAS 0.5523 0.5523 0.5523 0.5523
133
-
134
- Average scores for time period 1750-1800 (2 samples with a total of 5 sentences):
135
- Metric Precision Recall F1 Aligned Accuracy
136
- UPOS 0.9584 0.9584 0.9584 0.9584
137
- UAS 0.8133 0.8133 0.8133 0.8133
138
- LAS 0.7702 0.7702 0.7702 0.7702
139
-
140
- Average scores for time period 1800-1850 (6 samples with a total of 19 sentences):
141
- Metric Precision Recall F1 Aligned Accuracy
142
- UPOS 0.9517 0.9517 0.9517 0.9517
143
- UAS 0.7986 0.7986 0.7986 0.7986
144
- LAS 0.7464 0.7464 0.7464 0.7464
145
-
146
- Average scores for time period 1850-1900 (5 samples with a total of 14 sentences):
147
- Metric Precision Recall F1 Aligned Accuracy
148
- UPOS 0.9692 0.9692 0.9692 0.9692
149
- UAS 0.8561 0.8561 0.8561 0.8561
150
- LAS 0.7954 0.7954 0.7954 0.7954
151
-
152
- Average scores for time period 1900-1950 (5 samples with a total of 13 sentences):
153
- Metric Precision Recall F1 Aligned Accuracy
154
- UPOS 0.9805 0.9805 0.9805 0.9805
155
- UAS 0.8925 0.8925 0.8925 0.8925
156
- LAS 0.8405 0.8405 0.8405 0.8405
157
-
158
- Overall scores across all time periods (21 samples with a total of 59 sentences):
159
- Metric Precision Recall F1 Aligned Accuracy
160
- UPOS 0.9511 0.9511 0.9511 0.9511
161
- UAS 0.8140 0.8140 0.8140 0.8140
162
- LAS 0.7503 0.7503 0.7503 0.7503
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
eval/scores/scores_transformer_silver.txt DELETED
@@ -1,162 +0,0 @@
1
- Sample: svediakorp-sec987-spf144, time period: 1800-1850
2
- Metric Precision Recall F1 Aligned Accuracy
3
- UPOS 1.0000 1.0000 1.0000 1.0000
4
- UAS 0.9615 0.9615 0.9615 0.9615
5
- LAS 0.8846 0.8846 0.8846 0.8846
6
-
7
- Sample: svediakorp-sec613-EngstromA_StrindbergOchJag, time period: 1900-1950
8
- Metric Precision Recall F1 Aligned Accuracy
9
- UPOS 1.0000 1.0000 1.0000 1.0000
10
- UAS 0.8780 0.8780 0.8780 0.8780
11
- LAS 0.8171 0.8171 0.8171 0.8171
12
-
13
- Sample: svediakorp-sec268-DulciU_VitterhetsNojen3, time period: 1750-1800
14
- Metric Precision Recall F1 Aligned Accuracy
15
- UPOS 0.9663 0.9663 0.9663 0.9663
16
- UAS 0.8221 0.8221 0.8221 0.8221
17
- LAS 0.7692 0.7692 0.7692 0.7692
18
-
19
- Sample: svediakorp-sec25-Runius, time period: 1700-1750
20
- Metric Precision Recall F1 Aligned Accuracy
21
- UPOS 0.8750 0.8750 0.8750 0.8750
22
- UAS 0.6562 0.6562 0.6562 0.6562
23
- LAS 0.5625 0.5625 0.5625 0.5625
24
-
25
- Sample: svediakorp-sec208-Anonym_DetGrasligaMordet, time period: 1850-1900
26
- Metric Precision Recall F1 Aligned Accuracy
27
- UPOS 0.9865 0.9865 0.9865 0.9865
28
- UAS 0.7568 0.7568 0.7568 0.7568
29
- LAS 0.7027 0.7027 0.7027 0.7027
30
-
31
- Sample: svediakorp-sec324-GranbergPA_Enslighetsalskaren, time period: 1800-1850
32
- Metric Precision Recall F1 Aligned Accuracy
33
- UPOS 0.9400 0.9400 0.9400 0.9400
34
- UAS 0.8500 0.8500 0.8500 0.8500
35
- LAS 0.7500 0.7500 0.7500 0.7500
36
-
37
- Sample: svediakorp-sec254-CederborghF_BerattelseOmJohnHall, time period: 1800-1850
38
- Metric Precision Recall F1 Aligned Accuracy
39
- UPOS 0.9706 0.9706 0.9706 0.9706
40
- UAS 0.5765 0.5765 0.5765 0.5765
41
- LAS 0.5000 0.5000 0.5000 0.5000
42
-
43
- Sample: svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol, time period: 1900-1950
44
- Metric Precision Recall F1 Aligned Accuracy
45
- UPOS 0.9800 0.9800 0.9800 0.9800
46
- UAS 0.8000 0.8000 0.8000 0.8000
47
- LAS 0.7600 0.7600 0.7600 0.7600
48
-
49
- Sample: svediakorp-sec452-NyblomH_FantasierFyra, time period: 1850-1900
50
- Metric Precision Recall F1 Aligned Accuracy
51
- UPOS 0.9464 0.9464 0.9464 0.9464
52
- UAS 0.8571 0.8571 0.8571 0.8571
53
- LAS 0.7679 0.7679 0.7679 0.7679
54
-
55
- Sample: svediakorp-sec252-BremerF_Teckningar1, time period: 1800-1850
56
- Metric Precision Recall F1 Aligned Accuracy
57
- UPOS 1.0000 1.0000 1.0000 1.0000
58
- UAS 0.8333 0.8333 0.8333 0.8333
59
- LAS 0.7917 0.7917 0.7917 0.7917
60
-
61
- Sample: svediakorp-sec486-SchwartzMS_BellmansSkor, time period: 1850-1900
62
- Metric Precision Recall F1 Aligned Accuracy
63
- UPOS 0.9873 0.9873 0.9873 0.9873
64
- UAS 0.7722 0.7722 0.7722 0.7722
65
- LAS 0.6582 0.6582 0.6582 0.6582
66
-
67
- Sample: svediakorp-sec330-GyllenborgC_SwenskaSpratthoken, time period: 1700-1750
68
- Metric Precision Recall F1 Aligned Accuracy
69
- UPOS 0.8650 0.8650 0.8650 0.8650
70
- UAS 0.7178 0.7178 0.7178 0.7178
71
- LAS 0.6074 0.6074 0.6074 0.6074
72
-
73
- Sample: svediakorp-sec631-HasselskogN_HallaHallaGronkoping, time period: 1900-1950
74
- Metric Precision Recall F1 Aligned Accuracy
75
- UPOS 0.9605 0.9605 0.9605 0.9605
76
- UAS 0.8026 0.8026 0.8026 0.8026
77
- LAS 0.7105 0.7105 0.7105 0.7105
78
-
79
- Sample: svediakorp-sec277-EnbomPU_MedborgeligtSkalde, time period: 1750-1800
80
- Metric Precision Recall F1 Aligned Accuracy
81
- UPOS 0.9464 0.9464 0.9464 0.9464
82
- UAS 0.8214 0.8214 0.8214 0.8214
83
- LAS 0.7500 0.7500 0.7500 0.7500
84
-
85
- Sample: svediakorp-sec1102-spf259, time period: 1850-1900
86
- Metric Precision Recall F1 Aligned Accuracy
87
- UPOS 0.9444 0.9444 0.9444 0.9444
88
- UAS 0.8889 0.8889 0.8889 0.8889
89
- LAS 0.8889 0.8889 0.8889 0.8889
90
-
91
- Sample: svediakorp-sec991-spf148, time period: 1800-1850
92
- Metric Precision Recall F1 Aligned Accuracy
93
- UPOS 0.9561 0.9561 0.9561 0.9561
94
- UAS 0.7982 0.7982 0.7982 0.7982
95
- LAS 0.7544 0.7544 0.7544 0.7544
96
-
97
- Sample: svediakorp-sec639-HeidenstamV_Proletarfilosofiens, time period: 1900-1950
98
- Metric Precision Recall F1 Aligned Accuracy
99
- UPOS 0.9861 0.9861 0.9861 0.9861
100
- UAS 0.9306 0.9306 0.9306 0.9306
101
- LAS 0.8889 0.8889 0.8889 0.8889
102
-
103
- Sample: svediakorp-sec988-spf145, time period: 1800-1850
104
- Metric Precision Recall F1 Aligned Accuracy
105
- UPOS 0.9147 0.9147 0.9147 0.9147
106
- UAS 0.7442 0.7442 0.7442 0.7442
107
- LAS 0.6822 0.6822 0.6822 0.6822
108
-
109
- Sample: svediakorp-sec1063-spf220, time period: 1850-1900
110
- Metric Precision Recall F1 Aligned Accuracy
111
- UPOS 0.9660 0.9660 0.9660 0.9660
112
- UAS 0.8435 0.8435 0.8435 0.8435
113
- LAS 0.7551 0.7551 0.7551 0.7551
114
-
115
- Sample: svediakorp-sec1033-spf190, time period: 1900-1950
116
- Metric Precision Recall F1 Aligned Accuracy
117
- UPOS 0.9692 0.9692 0.9692 0.9692
118
- UAS 0.9077 0.9077 0.9077 0.9077
119
- LAS 0.8000 0.8000 0.8000 0.8000
120
-
121
- Sample: svediakorp-letter141673-Stalhammar, time period: 1700-1750
122
- Metric Precision Recall F1 Aligned Accuracy
123
- UPOS 0.7059 0.7059 0.7059 0.7059
124
- UAS 0.6471 0.6471 0.6471 0.6471
125
- LAS 0.4706 0.4706 0.4706 0.4706
126
-
127
- ----------------------------------------------------------------------
128
- Average scores for time period 1700-1750 (3 samples with a total of 8 sentences):
129
- Metric Precision Recall F1 Aligned Accuracy
130
- UPOS 0.8302 0.8302 0.8302 0.8302
131
- UAS 0.6693 0.6693 0.6693 0.6693
132
- LAS 0.5507 0.5507 0.5507 0.5507
133
-
134
- Average scores for time period 1750-1800 (2 samples with a total of 5 sentences):
135
- Metric Precision Recall F1 Aligned Accuracy
136
- UPOS 0.9584 0.9584 0.9584 0.9584
137
- UAS 0.8218 0.8218 0.8218 0.8218
138
- LAS 0.7615 0.7615 0.7615 0.7615
139
-
140
- Average scores for time period 1800-1850 (6 samples with a total of 19 sentences):
141
- Metric Precision Recall F1 Aligned Accuracy
142
- UPOS 0.9517 0.9517 0.9517 0.9517
143
- UAS 0.7615 0.7615 0.7615 0.7615
144
- LAS 0.6935 0.6935 0.6935 0.6935
145
-
146
- Average scores for time period 1850-1900 (5 samples with a total of 14 sentences):
147
- Metric Precision Recall F1 Aligned Accuracy
148
- UPOS 0.9692 0.9692 0.9692 0.9692
149
- UAS 0.8210 0.8210 0.8210 0.8210
150
- LAS 0.7382 0.7382 0.7382 0.7382
151
-
152
- Average scores for time period 1900-1950 (5 samples with a total of 13 sentences):
153
- Metric Precision Recall F1 Aligned Accuracy
154
- UPOS 0.9805 0.9805 0.9805 0.9805
155
- UAS 0.8734 0.8734 0.8734 0.8734
156
- LAS 0.8045 0.8045 0.8045 0.8045
157
-
158
- Overall scores across all time periods (21 samples with a total of 59 sentences):
159
- Metric Precision Recall F1 Aligned Accuracy
160
- UPOS 0.9511 0.9511 0.9511 0.9511
161
- UAS 0.7878 0.7878 0.7878 0.7878
162
- LAS 0.7122 0.7122 0.7122 0.7122
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/latest.txt DELETED
@@ -1 +0,0 @@
1
- log_bert-base-swedish-cased_sv_20260319_214621.txt
 
 
logs/log_bert-base-swedish-cased_sv_20260219_135555.txt DELETED
@@ -1,114 +0,0 @@
1
- === LOGFILE: logs/log_bert-base-swedish-cased_sv_20260219_135555.txt ===
2
- Language codes: sv
3
- Using pretrained model: KBLab/bert-base-swedish-cased
4
-
5
- Running: python prepare-train-val-SILVER.py sv
6
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
7
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
8
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
9
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
10
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
11
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
12
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
13
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
14
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
15
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
16
- Loading DigPhil MACHINE for SILVER (minus gold overlap)…
17
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
18
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
19
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
20
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
21
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
22
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
23
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
24
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
25
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
26
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
27
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
28
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
29
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
30
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
31
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
32
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
33
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
34
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
35
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
36
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
37
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
38
- Adding 79 remaining gold sentences to TRAIN
39
- Cleaning TRAIN...
40
- Cleaning SILVER...
41
- [REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
42
- [REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
43
- [REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
44
- [REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
45
- [REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
46
- [REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
47
- [REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
48
- [REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
49
- [REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
50
- [REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
51
- [REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
52
- [REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
53
- [REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
54
- [REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
55
- [REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
56
- [REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
57
- [REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
58
- [REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
59
- [REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
60
- [REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
61
- [REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
62
- [REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
63
- [REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
64
- [REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
65
- [REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
66
- [REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
67
- [REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
68
- [REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
69
- [REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
70
- [REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
71
- [REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
72
- [REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
73
- [REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
74
- [REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
75
- [REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
76
- [REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
77
- [REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
78
- [REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
79
- [REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
80
- [REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
81
- [REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
82
- [REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
83
- [REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
84
- [REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
85
- [REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
86
- [REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
87
- [REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
88
- [REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
89
- [REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
90
- [REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
91
- [REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
92
- [REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
93
- [REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
94
- [REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
95
- [REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
96
- [REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
97
- [REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
98
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
99
- [REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
100
- [REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
101
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
102
- [REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
103
- [REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
104
- Cleaning DEV...
105
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
106
- Cleaning TEST...
107
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
108
- Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13873 valid sentences)
109
- Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (46432 valid sentences)
110
- Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (29 valid sentences)
111
- Writing TEST → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (all DIGPHIL_GOLD, 108 valid sentences)
112
- Done.
113
- Sourcing scripts/config.sh
114
- Running stanza dataset preparation…
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_bert-base-swedish-cased_sv_20260219_135610.txt DELETED
@@ -1,138 +0,0 @@
1
- === LOGFILE: logs/log_bert-base-swedish-cased_sv_20260219_135610.txt ===
2
- Language codes: sv
3
- Using pretrained model: KBLab/bert-base-swedish-cased
4
-
5
- Running: python prepare-train-val-SILVER.py sv
6
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
7
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
8
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
9
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
10
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
11
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
12
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
13
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
14
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
15
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
16
- Loading DigPhil MACHINE for SILVER (minus gold overlap)…
17
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
18
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
19
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
20
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
21
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
22
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
23
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
24
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
25
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
26
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
27
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
28
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
29
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
30
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
31
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
32
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
33
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
34
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
35
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
36
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
37
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
38
- Adding 75 remaining gold sentences to TRAIN
39
- Cleaning TRAIN...
40
- Cleaning SILVER...
41
- [REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
42
- [REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
43
- [REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
44
- [REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
45
- [REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
46
- [REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
47
- [REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
48
- [REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
49
- [REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
50
- [REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
51
- [REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
52
- [REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
53
- [REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
54
- [REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
55
- [REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
56
- [REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
57
- [REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
58
- [REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
59
- [REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
60
- [REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
61
- [REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
62
- [REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
63
- [REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
64
- [REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
65
- [REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
66
- [REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
67
- [REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
68
- [REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
69
- [REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
70
- [REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
71
- [REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
72
- [REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
73
- [REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
74
- [REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
75
- [REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
76
- [REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
77
- [REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
78
- [REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
79
- [REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
80
- [REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
81
- [REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
82
- [REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
83
- [REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
84
- [REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
85
- [REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
86
- [REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
87
- [REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
88
- [REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
89
- [REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
90
- [REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
91
- [REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
92
- [REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
93
- [REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
94
- [REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
95
- [REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
96
- [REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
97
- [REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
98
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
99
- [REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
100
- [REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
101
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
102
- [REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
103
- [REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
104
- Cleaning DEV...
105
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
106
- Cleaning TEST...
107
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
108
- Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13869 valid sentences)
109
- Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (46432 valid sentences)
110
- Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (33 valid sentences)
111
- Writing TEST → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (all DIGPHIL_GOLD, 108 valid sentences)
112
- Done.
113
- Sourcing scripts/config.sh
114
- Running stanza dataset preparation…
115
- 2026-02-19 13:56:13 INFO: Datasets program called with:
116
- /home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --gold
117
- Augmented 60 quotes: Counter({'《》': 9, '»«': 8, '″″': 8, '““': 7, '„”': 6, '""': 6, '„“': 6, '«»': 4, '””': 4, '「」': 2})
118
- Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
119
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /tmp/tmpzh8h6jmq/sv_diachronic.train.gold.conllu
120
- Swapped 'w1, w2' for 'w1 ,w2' 94 times
121
- Added 85 new sentences with asdf, zzzz -> asdf,zzzz
122
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /tmp/tmpzh8h6jmq/sv_diachronic.dev.gold.conllu
123
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /tmp/tmpzh8h6jmq/sv_diachronic.test.gold.conllu
124
- Copying from /tmp/tmpzh8h6jmq/sv_diachronic.train.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
125
- Copying from /tmp/tmpzh8h6jmq/sv_diachronic.dev.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
126
- Copying from /tmp/tmpzh8h6jmq/sv_diachronic.test.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
127
- Running stanza dependency parser training…
128
- 2026-02-19 13:56:16 INFO: Training program called with:
129
- /home/urdatorn/git/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --batch_size 32 --dropout 0.33 --use_bert --bert_model KBLab/bert-base-swedish-cased --silver_file ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu
130
- 2026-02-19 13:56:16 DEBUG: UD_Swedish-diachronic: sv_diachronic
131
- 2026-02-19 13:56:16 INFO: Using model /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
132
- 2026-02-19 13:56:16 INFO: Using model /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
133
- 2026-02-19 13:56:16 INFO: Using default pretrain for language sv, found in /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt To use a different pretrain, specify --wordvec_pretrain_file
134
- 2026-02-19 13:56:16 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_transformer_parser.pt exists, skipping!
135
- DONE.
136
- Full log saved to: logs/log_bert-base-swedish-cased_sv_20260219_135610.txt
137
- Symlink updated: logs/latest.txt → log_bert-base-swedish-cased_sv_20260219_135610.txt
138
- Latest log: log_bert-base-swedish-cased_sv_20260219_135610.txt
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_bert-base-swedish-cased_sv_20260219_135635.txt DELETED
The diff for this file is too large to render. See raw diff
 
logs/log_bert-base-swedish-cased_sv_20260318_131623.txt DELETED
@@ -1,271 +0,0 @@
1
- === LOGFILE: logs/log_bert-base-swedish-cased_sv_20260318_131623.txt ===
2
- Language codes: sv
3
- Using pretrained model: KBLab/bert-base-swedish-cased
4
-
5
- Running: python prepare-train-val-SILVER.py sv
6
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
7
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
8
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
9
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
10
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
11
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
12
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
13
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
14
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
15
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
16
- Loading DigPhil MACHINE for SILVER (minus gold overlap)…
17
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec987-spf144.conllu
18
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
19
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
20
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec25-Runius.conllu
21
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
22
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
23
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
24
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
25
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
26
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec252-BremerF_Teckningar1.conllu
27
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
28
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
29
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
30
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
31
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1102-spf259.conllu
32
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec991-spf148.conllu
33
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
34
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec988-spf145.conllu
35
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1063-spf220.conllu
36
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1033-spf190.conllu
37
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-letter141673-Stalhammar.conllu
38
- Adding 76 remaining gold sentences to TRAIN
39
- Cleaning TRAIN...
40
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
41
- Cleaning SILVER...
42
- [REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
43
- [REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
44
- [REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
45
- [REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
46
- [REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
47
- [REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
48
- [REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
49
- [REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
50
- [REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
51
- [REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
52
- [REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
53
- [REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
54
- [REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
55
- [REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
56
- [REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
57
- [REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
58
- [REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
59
- [REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
60
- [REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
61
- [REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
62
- [REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
63
- [REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
64
- [REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
65
- [REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
66
- [REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
67
- [REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
68
- [REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
69
- [REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
70
- [REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
71
- [REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
72
- [REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
73
- [REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
74
- [REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
75
- [REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
76
- [REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
77
- [REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
78
- [REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
79
- [REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
80
- [REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
81
- [REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
82
- [REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
83
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
84
- [REMOVED] sent_id=1266 ERRORS=['Token 17: Missing form']
85
- [REMOVED] sent_id=4897 ERRORS=['Token 46: Missing form', 'Token 47: Missing form']
86
- Cleaning DEV...
87
- Cleaning TEST...
88
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
89
- Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13869 valid sentences)
90
- Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (51431 valid sentences)
91
- Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (33 valid sentences)
92
- Writing TEST → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (all DIGPHIL_GOLD, 108 valid sentences)
93
- Done.
94
- Sourcing scripts/config.sh
95
- Running stanza dataset preparation…
96
- 2026-03-18 13:16:26 INFO: Datasets program called with:
97
- /home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --gold
98
- Augmented 60 quotes: Counter({'《》': 9, '»«': 8, '″″': 8, '““': 7, '„”': 6, '""': 6, '„“': 6, '«»': 4, '””': 4, '「」': 2})
99
- Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
100
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /tmp/tmpz3gm6g6g/sv_diachronic.train.gold.conllu
101
- Swapped 'w1, w2' for 'w1 ,w2' 94 times
102
- Added 85 new sentences with asdf, zzzz -> asdf,zzzz
103
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /tmp/tmpz3gm6g6g/sv_diachronic.dev.gold.conllu
104
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /tmp/tmpz3gm6g6g/sv_diachronic.test.gold.conllu
105
- Copying from /tmp/tmpz3gm6g6g/sv_diachronic.train.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
106
- Copying from /tmp/tmpz3gm6g6g/sv_diachronic.dev.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
107
- Copying from /tmp/tmpz3gm6g6g/sv_diachronic.test.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
108
- Running stanza dependency parser training…
109
- 2026-03-18 13:16:28 INFO: Training program called with:
110
- /home/urdatorn/git/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --batch_size 32 --dropout 0.33 --use_bert --bert_model KBLab/bert-base-swedish-cased --silver_file ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu
111
- 2026-03-18 13:16:28 DEBUG: UD_Swedish-diachronic: sv_diachronic
112
- 2026-03-18 13:16:28 INFO: Using model /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
113
- 2026-03-18 13:16:28 INFO: Using model /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
114
- 2026-03-18 13:16:28 INFO: Using default pretrain for language sv, found in /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt To use a different pretrain, specify --wordvec_pretrain_file
115
- 2026-03-18 13:16:28 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_transformer_parser.pt does not exist, training new model
116
- 2026-03-18 13:16:28 INFO: Using model /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
117
- 2026-03-18 13:16:28 INFO: Using model /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
118
- 2026-03-18 13:16:28 INFO: Using default pretrain for language sv, found in /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt To use a different pretrain, specify --wordvec_pretrain_file
119
- 2026-03-18 13:16:28 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/home/urdatorn/stanza_resources/sv/pretrain', '--train_file', '/home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--wordvec_pretrain_file', '/home/urdatorn/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt', '--batch_size', '32', '--dropout', '0.33', '--bert_model', 'KBLab/bert-base-swedish-cased', '--silver_file', 'ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu']
120
- 2026-03-18 13:16:28 INFO: Running parser in train mode
121
- 2026-03-18 13:16:29 DEBUG: Loaded pretrain from /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
122
- 2026-03-18 13:16:29 INFO: Using 7 as the word cutoff based on the size of the pretrain (250000)
123
- 2026-03-18 13:16:29 INFO: Using pretrained contextualized char embedding
124
- 2026-03-18 13:16:29 INFO: ARGS USED AT TRAINING TIME:
125
- augment_nopunct: None
126
- batch_size: 32
127
- bert_finetune: False
128
- bert_finetune_layers: None
129
- bert_hidden_layers: 4
130
- bert_learning_rate: 1.0
131
- bert_model: KBLab/bert-base-swedish-cased
132
- bert_start_finetuning: 200
133
- bert_warmup_steps: 200
134
- bert_weight_decay: 0.0
135
- beta2: 0.999
136
- char: True
137
- char_emb_dim: 100
138
- char_hidden_dim: 400
139
- char_num_layers: 1
140
- char_rec_dropout: 0
141
- charlm: True
142
- charlm_backward_file: /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
143
- charlm_forward_file: /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt
144
- charlm_save_dir: saved_models/charlm
145
- charlm_shorthand: sv_conll17
146
- checkpoint: True
147
- checkpoint_interval: 500
148
- checkpoint_save_name: None
149
- continue_from: None
150
- data_dir: data/depparse
151
- deep_biaff_hidden_dim: 400
152
- deep_biaff_output_dim: 160
153
- device: cpu
154
- distance: True
155
- dropout: 0.33
156
- eval_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
157
- eval_interval: 100
158
- gold_labels: True
159
- hidden_dim: 400
160
- lang: sv
161
- linearization: True
162
- log_norms: False
163
- log_step: 20
164
- lora_alpha: 128
165
- lora_dropout: 0.1
166
- lora_modules_to_save: []
167
- lora_rank: 64
168
- lora_target_modules: ['query', 'value', 'output.dense', 'intermediate.dense']
169
- lr: 2.0
170
- max_grad_norm: 1.0
171
- max_steps: 50000
172
- max_steps_before_stop: 2000
173
- mode: train
174
- model_type: graph
175
- num_layers: 3
176
- optim: adadelta
177
- output_file: None
178
- output_latex: False
179
- pretrain: True
180
- pretrain_max_vocab: 250000
181
- rec_dropout: 0
182
- reversed: False
183
- sample_train: 1.0
184
- save_dir: saved_models/depparse
185
- save_name: {shorthand}_{embedding}_parser.pt
186
- second_batch_size: None
187
- second_bert_learning_rate: 0.001
188
- second_lr: 0.0002
189
- second_optim: adam
190
- second_optim_start_step: 10000
191
- second_warmup_steps: 200
192
- seed: 1234
193
- shorthand: sv_diachronic
194
- silver_file: ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu
195
- silver_weight: 0.5
196
- tag_emb_dim: 50
197
- train_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
198
- train_size: None
199
- transformed_dim: 125
200
- transition_embedding_dim: 20
201
- transition_hidden_dim: 20
202
- transition_merge_hidden_dim: 200
203
- transition_subtree_combination: SubtreeCombination.NONE
204
- transition_subtree_nonlinearity: none
205
- use_arc_embedding: False
206
- use_peft: False
207
- use_ufeats: True
208
- use_upos: True
209
- use_xpos: True
210
- wandb: False
211
- wandb_name: None
212
- weight_decay: 1e-05
213
- word_cutoff: 7
214
- word_dropout: 0.33
215
- word_emb_dim: 75
216
- wordvec_dir: /home/urdatorn/stanza_resources/sv/pretrain
217
- wordvec_file: None
218
- wordvec_pretrain_file: /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
219
-
220
- 2026-03-18 13:16:29 INFO: Loading data with batch size 32...
221
- 2026-03-18 13:16:30 INFO: Train File /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 13954
222
- 2026-03-18 13:16:30 INFO: Original data size: 13954
223
- 2026-03-18 13:16:34 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
224
- 2026-03-18 13:16:34 INFO: Original length = 13954
225
- 2026-03-18 13:16:34 INFO: Filtered length = 13954
226
- 2026-03-18 13:16:37 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='-') for the xpos factory for sv_diachronic
227
- 2026-03-18 13:16:41 DEBUG: 9513 batches created.
228
- 2026-03-18 13:16:41 INFO: Augmented data size: 14712
229
- 2026-03-18 13:16:47 DEBUG: 10020 batches created.
230
- 2026-03-18 13:16:48 DEBUG: 32 batches created.
231
- 2026-03-18 13:17:10 DEBUG: 28588 batches created.
232
- 2026-03-18 13:17:10 INFO: Training parser...
233
- /home/urdatorn/.pyenv/versions/xpu/lib/python3.13/site-packages/huggingface_hub/file_download.py:949: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
234
- warnings.warn(
235
- 2026-03-18 13:17:12 DEBUG: Depparse model loading charmodels: /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt and /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
236
- 2026-03-18 13:17:12 DEBUG: Loading charlm from /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt
237
- 2026-03-18 13:17:12 DEBUG: Loading charlm from /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
238
- 2026-03-18 13:17:13 DEBUG: Not using arc embedding enhancement
239
- 2026-03-18 13:17:13 DEBUG: Building Adadelta with lr=2.000000, weight_decay=1e-05
240
- 2026-03-18 13:17:29 INFO: Finished STEP 20/50000, loss = 2.292368 (0.203 sec/batch), lr: 2.000000
241
- 2026-03-18 13:17:43 INFO: Finished STEP 40/50000, loss = 7.419842 (1.468 sec/batch), lr: 2.000000
242
- 2026-03-18 13:17:51 INFO: Finished STEP 60/50000, loss = 7.826311 (1.111 sec/batch), lr: 2.000000
243
- 2026-03-18 13:18:01 INFO: Finished STEP 80/50000, loss = 4.187514 (0.203 sec/batch), lr: 2.000000
244
- 2026-03-18 13:18:09 INFO: Finished STEP 100/50000, loss = 3.687605 (0.196 sec/batch), lr: 2.000000
245
- 2026-03-18 13:18:09 INFO: Evaluating on dev set...
246
- 2026-03-18 13:18:16 INFO: LAS MLAS BLEX
247
- 2026-03-18 13:18:16 INFO: 14.77 8.39 14.11
248
- 2026-03-18 13:18:16 INFO: step 100: train_loss = 5.880744, dev_score = 0.1477
249
- 2026-03-18 13:18:16 DEBUG: training batch usage: Counter({0: 68, 1: 32})
250
- 2026-03-18 13:18:16 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser.pt
251
- 2026-03-18 13:18:16 INFO: new best model saved.
252
- 2026-03-18 13:18:16 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
253
- 2026-03-18 13:18:16 INFO: new model checkpoint saved.
254
- 2026-03-18 13:18:27 INFO: Finished STEP 120/50000, loss = 5.895749 (0.951 sec/batch), lr: 2.000000
255
- 2026-03-18 13:18:38 INFO: Finished STEP 140/50000, loss = 1.609854 (0.200 sec/batch), lr: 2.000000
256
- 2026-03-18 13:18:45 INFO: Finished STEP 160/50000, loss = 1.833139 (0.220 sec/batch), lr: 2.000000
257
- 2026-03-18 13:18:56 INFO: Finished STEP 180/50000, loss = 1.341930 (0.204 sec/batch), lr: 2.000000
258
- 2026-03-18 13:19:04 INFO: Finished STEP 200/50000, loss = 2.078672 (0.224 sec/batch), lr: 2.000000
259
- 2026-03-18 13:19:04 INFO: Evaluating on dev set...
260
- 2026-03-18 13:19:11 INFO: LAS MLAS BLEX
261
- 2026-03-18 13:19:11 INFO: 42.27 32.23 37.34
262
- 2026-03-18 13:19:11 INFO: step 200: train_loss = 3.482911, dev_score = 0.4227
263
- 2026-03-18 13:19:11 DEBUG: training batch usage: Counter({0: 132, 1: 68})
264
- 2026-03-18 13:19:11 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser.pt
265
- 2026-03-18 13:19:11 INFO: new best model saved.
266
- 2026-03-18 13:19:11 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
267
- 2026-03-18 13:19:11 INFO: new model checkpoint saved.
268
- 2026-03-18 13:19:19 INFO: Finished STEP 220/50000, loss = 1.613766 (0.210 sec/batch), lr: 2.000000
269
- 2026-03-18 13:19:27 INFO: Finished STEP 240/50000, loss = 4.277971 (0.822 sec/batch), lr: 2.000000
270
- 2026-03-18 13:19:39 INFO: Finished STEP 260/50000, loss = 4.801499 (0.823 sec/batch), lr: 2.000000
271
- 2026-03-18 13:19:48 INFO: Finished STEP 280/50000, loss = 1.559014 (0.204 sec/batch), lr: 2.000000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_bert-base-swedish-cased_sv_20260318_132248.txt DELETED
@@ -1,109 +0,0 @@
1
- === LOGFILE: logs/log_bert-base-swedish-cased_sv_20260318_132248.txt ===
2
- Language codes: sv
3
- Using pretrained model: KBLab/bert-base-swedish-cased
4
-
5
- Running: python prepare-train-val-SILVER.py sv
6
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
7
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
8
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
9
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
10
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
11
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
12
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
13
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
14
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
15
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
16
- Loading DigPhil MACHINE for SILVER (minus gold overlap)…
17
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec987-spf144.conllu
18
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
19
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
20
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec25-Runius.conllu
21
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
22
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
23
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
24
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
25
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
26
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec252-BremerF_Teckningar1.conllu
27
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
28
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
29
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
30
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
31
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1102-spf259.conllu
32
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec991-spf148.conllu
33
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
34
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec988-spf145.conllu
35
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1063-spf220.conllu
36
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1033-spf190.conllu
37
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-letter141673-Stalhammar.conllu
38
- Adding 77 remaining gold sentences to TRAIN
39
- Cleaning TRAIN...
40
- Cleaning SILVER...
41
- [REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
42
- [REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
43
- [REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
44
- [REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
45
- [REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
46
- [REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
47
- [REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
48
- [REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
49
- [REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
50
- [REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
51
- [REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
52
- [REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
53
- [REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
54
- [REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
55
- [REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
56
- [REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
57
- [REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
58
- [REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
59
- [REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
60
- [REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
61
- [REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
62
- [REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
63
- [REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
64
- [REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
65
- [REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
66
- [REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
67
- [REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
68
- [REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
69
- [REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
70
- [REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
71
- [REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
72
- [REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
73
- [REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
74
- [REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
75
- [REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
76
- [REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
77
- [REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
78
- [REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
79
- [REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
80
- [REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
81
- [REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
82
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
83
- [REMOVED] sent_id=1266 ERRORS=['Token 17: Missing form']
84
- [REMOVED] sent_id=4897 ERRORS=['Token 46: Missing form', 'Token 47: Missing form']
85
- Cleaning DEV...
86
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
87
- Cleaning TEST...
88
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
89
- Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13871 valid sentences)
90
- Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (51431 valid sentences)
91
- Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (31 valid sentences)
92
- Writing TEST → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (all DIGPHIL_GOLD, 108 valid sentences)
93
- Done.
94
- Sourcing scripts/config.sh
95
- Running stanza dataset preparation…
96
- 2026-03-18 13:22:51 INFO: Datasets program called with:
97
- /home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --gold
98
- Augmented 60 quotes: Counter({'《》': 9, '»«': 8, '″″': 8, '““': 7, '„”': 6, '""': 6, '„“': 6, '«»': 4, '””': 4, '「」': 2})
99
- Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
100
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /tmp/tmpw00sc1dk/sv_diachronic.train.gold.conllu
101
- Swapped 'w1, w2' for 'w1 ,w2' 94 times
102
- Added 86 new sentences with asdf, zzzz -> asdf,zzzz
103
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /tmp/tmpw00sc1dk/sv_diachronic.dev.gold.conllu
104
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /tmp/tmpw00sc1dk/sv_diachronic.test.gold.conllu
105
- Copying from /tmp/tmpw00sc1dk/sv_diachronic.train.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
106
- Copying from /tmp/tmpw00sc1dk/sv_diachronic.dev.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
107
- Copying from /tmp/tmpw00sc1dk/sv_diachronic.test.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
108
- Running stanza dependency parser training…
109
- ./make_new_model_BERT_SILVER.sh: line 72: unexpected EOF while looking for matching `"'
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_bert-base-swedish-cased_sv_20260318_132404.txt DELETED
@@ -1,109 +0,0 @@
1
- === LOGFILE: logs/log_bert-base-swedish-cased_sv_20260318_132404.txt ===
2
- Language codes: sv
3
- Using pretrained model: KBLab/bert-base-swedish-cased
4
-
5
- Running: python prepare-train-val-SILVER.py sv
6
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
7
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
8
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
9
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
10
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
11
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
12
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
13
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
14
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
15
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
16
- Loading DigPhil MACHINE for SILVER (minus gold overlap)…
17
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec987-spf144.conllu
18
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
19
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
20
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec25-Runius.conllu
21
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
22
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
23
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
24
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
25
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
26
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec252-BremerF_Teckningar1.conllu
27
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
28
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
29
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
30
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
31
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1102-spf259.conllu
32
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec991-spf148.conllu
33
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
34
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec988-spf145.conllu
35
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1063-spf220.conllu
36
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1033-spf190.conllu
37
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-letter141673-Stalhammar.conllu
38
- Adding 74 remaining gold sentences to TRAIN
39
- Cleaning TRAIN...
40
- Cleaning SILVER...
41
- [REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
42
- [REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
43
- [REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
44
- [REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
45
- [REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
46
- [REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
47
- [REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
48
- [REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
49
- [REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
50
- [REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
51
- [REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
52
- [REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
53
- [REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
54
- [REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
55
- [REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
56
- [REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
57
- [REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
58
- [REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
59
- [REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
60
- [REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
61
- [REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
62
- [REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
63
- [REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
64
- [REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
65
- [REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
66
- [REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
67
- [REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
68
- [REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
69
- [REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
70
- [REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
71
- [REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
72
- [REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
73
- [REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
74
- [REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
75
- [REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
76
- [REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
77
- [REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
78
- [REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
79
- [REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
80
- [REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
81
- [REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
82
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
83
- [REMOVED] sent_id=1266 ERRORS=['Token 17: Missing form']
84
- [REMOVED] sent_id=4897 ERRORS=['Token 46: Missing form', 'Token 47: Missing form']
85
- Cleaning DEV...
86
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
87
- Cleaning TEST...
88
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
89
- Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13868 valid sentences)
90
- Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (51431 valid sentences)
91
- Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (34 valid sentences)
92
- Writing TEST → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (all DIGPHIL_GOLD, 108 valid sentences)
93
- Done.
94
- Sourcing scripts/config.sh
95
- Running stanza dataset preparation…
96
- 2026-03-18 13:24:07 INFO: Datasets program called with:
97
- /home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --gold
98
- Augmented 60 quotes: Counter({'《》': 9, '»«': 8, '″″': 8, '““': 7, '„”': 6, '""': 6, '„“': 6, '«»': 4, '””': 4, '「」': 2})
99
- Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
100
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /tmp/tmp0l99psu3/sv_diachronic.train.gold.conllu
101
- Swapped 'w1, w2' for 'w1 ,w2' 93 times
102
- Added 85 new sentences with asdf, zzzz -> asdf,zzzz
103
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /tmp/tmp0l99psu3/sv_diachronic.dev.gold.conllu
104
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /tmp/tmp0l99psu3/sv_diachronic.test.gold.conllu
105
- Copying from /tmp/tmp0l99psu3/sv_diachronic.train.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
106
- Copying from /tmp/tmp0l99psu3/sv_diachronic.dev.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
107
- Copying from /tmp/tmp0l99psu3/sv_diachronic.test.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
108
- Running stanza dependency parser training…
109
- ./make_new_model_BERT_SILVER.sh: line 72: unexpected EOF while looking for matching `"'
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_bert-base-swedish-cased_sv_20260318_132733.txt DELETED
@@ -1,325 +0,0 @@
1
- === LOGFILE: logs/log_bert-base-swedish-cased_sv_20260318_132733.txt ===
2
- Language codes: sv
3
- Using pretrained model: KBLab/bert-base-swedish-cased
4
-
5
- Running: python prepare-train-val-SILVER.py sv
6
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
7
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
8
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
9
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
10
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
11
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
12
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
13
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
14
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
15
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
16
- Loading DigPhil MACHINE for SILVER (minus gold overlap)…
17
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec987-spf144.conllu
18
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
19
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
20
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec25-Runius.conllu
21
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
22
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
23
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
24
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
25
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
26
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec252-BremerF_Teckningar1.conllu
27
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
28
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
29
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
30
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
31
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1102-spf259.conllu
32
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec991-spf148.conllu
33
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
34
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec988-spf145.conllu
35
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1063-spf220.conllu
36
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1033-spf190.conllu
37
- Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-letter141673-Stalhammar.conllu
38
- Adding 73 remaining gold sentences to TRAIN
39
- Cleaning TRAIN...
40
- Cleaning SILVER...
41
- [REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
42
- [REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
43
- [REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
44
- [REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
45
- [REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
46
- [REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
47
- [REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
48
- [REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
49
- [REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
50
- [REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
51
- [REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
52
- [REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
53
- [REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
54
- [REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
55
- [REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
56
- [REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
57
- [REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
58
- [REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
59
- [REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
60
- [REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
61
- [REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
62
- [REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
63
- [REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
64
- [REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
65
- [REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
66
- [REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
67
- [REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
68
- [REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
69
- [REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
70
- [REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
71
- [REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
72
- [REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
73
- [REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
74
- [REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
75
- [REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
76
- [REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
77
- [REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
78
- [REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
79
- [REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
80
- [REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
81
- [REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
82
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
83
- [REMOVED] sent_id=1266 ERRORS=['Token 17: Missing form']
84
- [REMOVED] sent_id=4897 ERRORS=['Token 46: Missing form', 'Token 47: Missing form']
85
- Cleaning DEV...
86
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
87
- Cleaning TEST...
88
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
89
- Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13867 valid sentences)
90
- Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (51431 valid sentences)
91
- Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (35 valid sentences)
92
- Writing TEST → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (all DIGPHIL_GOLD, 108 valid sentences)
93
- Done.
94
- Sourcing scripts/config.sh
95
- Running stanza dataset preparation…
96
- 2026-03-18 13:27:36 INFO: Datasets program called with:
97
- /home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --gold
98
- Augmented 60 quotes: Counter({'《》': 9, '»«': 8, '″″': 8, '““': 7, '„”': 6, '""': 6, '„“': 6, '«»': 4, '””': 4, '「」': 2})
99
- Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
100
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /tmp/tmpl2s5qxw1/sv_diachronic.train.gold.conllu
101
- Swapped 'w1, w2' for 'w1 ,w2' 98 times
102
- Added 85 new sentences with asdf, zzzz -> asdf,zzzz
103
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /tmp/tmpl2s5qxw1/sv_diachronic.dev.gold.conllu
104
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /tmp/tmpl2s5qxw1/sv_diachronic.test.gold.conllu
105
- Copying from /tmp/tmpl2s5qxw1/sv_diachronic.train.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
106
- Copying from /tmp/tmpl2s5qxw1/sv_diachronic.dev.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
107
- Copying from /tmp/tmpl2s5qxw1/sv_diachronic.test.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
108
- Running stanza dependency parser training…
109
- 2026-03-18 13:27:38 INFO: Training program called with:
110
- /home/urdatorn/git/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --batch_size 32 --dropout 0.33 --use_bert --bert_model KBLab/bert-base-swedish-cased --silver_file ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu --device xpu:0
111
- 2026-03-18 13:27:38 DEBUG: UD_Swedish-diachronic: sv_diachronic
112
- 2026-03-18 13:27:38 INFO: Using model /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
113
- 2026-03-18 13:27:38 INFO: Using model /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
114
- 2026-03-18 13:27:38 INFO: Using default pretrain for language sv, found in /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt To use a different pretrain, specify --wordvec_pretrain_file
115
- 2026-03-18 13:27:38 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_transformer_parser.pt does not exist, training new model
116
- 2026-03-18 13:27:38 INFO: Using model /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
117
- 2026-03-18 13:27:38 INFO: Using model /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
118
- 2026-03-18 13:27:38 INFO: Using default pretrain for language sv, found in /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt To use a different pretrain, specify --wordvec_pretrain_file
119
- 2026-03-18 13:27:38 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/home/urdatorn/stanza_resources/sv/pretrain', '--train_file', '/home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--wordvec_pretrain_file', '/home/urdatorn/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt', '--batch_size', '32', '--dropout', '0.33', '--bert_model', 'KBLab/bert-base-swedish-cased', '--silver_file', 'ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu', '--device', 'xpu:0']
120
- 2026-03-18 13:27:38 INFO: Running parser in train mode
121
- 2026-03-18 13:27:39 DEBUG: Loaded pretrain from /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
122
- 2026-03-18 13:27:39 INFO: Using 7 as the word cutoff based on the size of the pretrain (250000)
123
- 2026-03-18 13:27:39 INFO: Using pretrained contextualized char embedding
124
- 2026-03-18 13:27:39 INFO: ARGS USED AT TRAINING TIME:
125
- augment_nopunct: None
126
- batch_size: 32
127
- bert_finetune: False
128
- bert_finetune_layers: None
129
- bert_hidden_layers: 4
130
- bert_learning_rate: 1.0
131
- bert_model: KBLab/bert-base-swedish-cased
132
- bert_start_finetuning: 200
133
- bert_warmup_steps: 200
134
- bert_weight_decay: 0.0
135
- beta2: 0.999
136
- char: True
137
- char_emb_dim: 100
138
- char_hidden_dim: 400
139
- char_num_layers: 1
140
- char_rec_dropout: 0
141
- charlm: True
142
- charlm_backward_file: /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
143
- charlm_forward_file: /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt
144
- charlm_save_dir: saved_models/charlm
145
- charlm_shorthand: sv_conll17
146
- checkpoint: True
147
- checkpoint_interval: 500
148
- checkpoint_save_name: None
149
- continue_from: None
150
- data_dir: data/depparse
151
- deep_biaff_hidden_dim: 400
152
- deep_biaff_output_dim: 160
153
- device: xpu:0
154
- distance: True
155
- dropout: 0.33
156
- eval_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
157
- eval_interval: 100
158
- gold_labels: True
159
- hidden_dim: 400
160
- lang: sv
161
- linearization: True
162
- log_norms: False
163
- log_step: 20
164
- lora_alpha: 128
165
- lora_dropout: 0.1
166
- lora_modules_to_save: []
167
- lora_rank: 64
168
- lora_target_modules: ['query', 'value', 'output.dense', 'intermediate.dense']
169
- lr: 2.0
170
- max_grad_norm: 1.0
171
- max_steps: 50000
172
- max_steps_before_stop: 2000
173
- mode: train
174
- model_type: graph
175
- num_layers: 3
176
- optim: adadelta
177
- output_file: None
178
- output_latex: False
179
- pretrain: True
180
- pretrain_max_vocab: 250000
181
- rec_dropout: 0
182
- reversed: False
183
- sample_train: 1.0
184
- save_dir: saved_models/depparse
185
- save_name: {shorthand}_{embedding}_parser.pt
186
- second_batch_size: None
187
- second_bert_learning_rate: 0.001
188
- second_lr: 0.0002
189
- second_optim: adam
190
- second_optim_start_step: 10000
191
- second_warmup_steps: 200
192
- seed: 1234
193
- shorthand: sv_diachronic
194
- silver_file: ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu
195
- silver_weight: 0.5
196
- tag_emb_dim: 50
197
- train_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
198
- train_size: None
199
- transformed_dim: 125
200
- transition_embedding_dim: 20
201
- transition_hidden_dim: 20
202
- transition_merge_hidden_dim: 200
203
- transition_subtree_combination: SubtreeCombination.NONE
204
- transition_subtree_nonlinearity: none
205
- use_arc_embedding: False
206
- use_peft: False
207
- use_ufeats: True
208
- use_upos: True
209
- use_xpos: True
210
- wandb: False
211
- wandb_name: None
212
- weight_decay: 1e-05
213
- word_cutoff: 7
214
- word_dropout: 0.33
215
- word_emb_dim: 75
216
- wordvec_dir: /home/urdatorn/stanza_resources/sv/pretrain
217
- wordvec_file: None
218
- wordvec_pretrain_file: /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
219
-
220
- 2026-03-18 13:27:39 INFO: Loading data with batch size 32...
221
- 2026-03-18 13:27:40 INFO: Train File /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 13952
222
- 2026-03-18 13:27:40 INFO: Original data size: 13952
223
- 2026-03-18 13:27:43 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
224
- 2026-03-18 13:27:43 INFO: Original length = 13952
225
- 2026-03-18 13:27:43 INFO: Filtered length = 13952
226
- 2026-03-18 13:27:47 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='-') for the xpos factory for sv_diachronic
227
- 2026-03-18 13:27:50 DEBUG: 9512 batches created.
228
- 2026-03-18 13:27:50 INFO: Augmented data size: 14707
229
- 2026-03-18 13:27:57 DEBUG: 10016 batches created.
230
- 2026-03-18 13:27:57 DEBUG: 34 batches created.
231
- 2026-03-18 13:28:19 DEBUG: 28588 batches created.
232
- 2026-03-18 13:28:19 INFO: Training parser...
233
- /home/urdatorn/.pyenv/versions/xpu/lib/python3.13/site-packages/huggingface_hub/file_download.py:949: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
234
- warnings.warn(
235
- 2026-03-18 13:28:20 DEBUG: Depparse model loading charmodels: /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt and /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
236
- 2026-03-18 13:28:20 DEBUG: Loading charlm from /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt
237
- 2026-03-18 13:28:20 DEBUG: Loading charlm from /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
238
- 2026-03-18 13:28:20 DEBUG: Not using arc embedding enhancement
239
- 2026-03-18 13:28:21 DEBUG: Building Adadelta with lr=2.000000, weight_decay=1e-05
240
- 2026-03-18 13:28:31 INFO: Finished STEP 20/50000, loss = 5.042704 (0.060 sec/batch), lr: 2.000000
241
- 2026-03-18 13:28:38 INFO: Finished STEP 40/50000, loss = 7.420871 (0.629 sec/batch), lr: 2.000000
242
- 2026-03-18 13:28:42 INFO: Finished STEP 60/50000, loss = 7.130751 (0.585 sec/batch), lr: 2.000000
243
- 2026-03-18 13:28:46 INFO: Finished STEP 80/50000, loss = 2.640507 (0.061 sec/batch), lr: 2.000000
244
- 2026-03-18 13:28:49 INFO: Finished STEP 100/50000, loss = 2.198331 (0.062 sec/batch), lr: 2.000000
245
- 2026-03-18 13:28:49 INFO: Evaluating on dev set...
246
- 2026-03-18 13:28:52 INFO: LAS MLAS BLEX
247
- 2026-03-18 13:28:52 INFO: 14.92 7.39 9.17
248
- 2026-03-18 13:28:52 INFO: step 100: train_loss = 6.003559, dev_score = 0.1492
249
- 2026-03-18 13:28:52 DEBUG: training batch usage: Counter({0: 68, 1: 32})
250
- 2026-03-18 13:28:52 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser.pt
251
- 2026-03-18 13:28:52 INFO: new best model saved.
252
- 2026-03-18 13:28:52 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
253
- 2026-03-18 13:28:52 INFO: new model checkpoint saved.
254
- 2026-03-18 13:28:58 INFO: Finished STEP 120/50000, loss = 5.791277 (0.479 sec/batch), lr: 2.000000
255
- 2026-03-18 13:29:02 INFO: Finished STEP 140/50000, loss = 1.561704 (0.065 sec/batch), lr: 2.000000
256
- 2026-03-18 13:29:05 INFO: Finished STEP 160/50000, loss = 2.749855 (0.064 sec/batch), lr: 2.000000
257
- 2026-03-18 13:29:10 INFO: Finished STEP 180/50000, loss = 2.644507 (0.062 sec/batch), lr: 2.000000
258
- 2026-03-18 13:29:14 INFO: Finished STEP 200/50000, loss = 2.221299 (0.062 sec/batch), lr: 2.000000
259
- 2026-03-18 13:29:14 INFO: Evaluating on dev set...
260
- 2026-03-18 13:29:17 INFO: LAS MLAS BLEX
261
- 2026-03-18 13:29:17 INFO: 36.67 23.23 32.82
262
- 2026-03-18 13:29:17 INFO: step 200: train_loss = 3.408124, dev_score = 0.3667
263
- 2026-03-18 13:29:17 DEBUG: training batch usage: Counter({0: 132, 1: 68})
264
- 2026-03-18 13:29:17 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser.pt
265
- 2026-03-18 13:29:17 INFO: new best model saved.
266
- 2026-03-18 13:29:17 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
267
- 2026-03-18 13:29:17 INFO: new model checkpoint saved.
268
- 2026-03-18 13:29:20 INFO: Finished STEP 220/50000, loss = 2.403125 (0.064 sec/batch), lr: 2.000000
269
- 2026-03-18 13:29:24 INFO: Finished STEP 240/50000, loss = 3.965761 (0.412 sec/batch), lr: 2.000000
270
- 2026-03-18 13:29:30 INFO: Finished STEP 260/50000, loss = 4.813926 (0.403 sec/batch), lr: 2.000000
271
- 2026-03-18 13:29:33 INFO: Finished STEP 280/50000, loss = 2.433779 (0.067 sec/batch), lr: 2.000000
272
- 2026-03-18 13:29:37 INFO: Finished STEP 300/50000, loss = 5.255473 (0.374 sec/batch), lr: 2.000000
273
- 2026-03-18 13:29:37 INFO: Evaluating on dev set...
274
- 2026-03-18 13:29:40 INFO: LAS MLAS BLEX
275
- 2026-03-18 13:29:40 INFO: 51.20 39.90 43.72
276
- 2026-03-18 13:29:40 INFO: step 300: train_loss = 3.172424, dev_score = 0.5120
277
- 2026-03-18 13:29:40 DEBUG: training batch usage: Counter({0: 191, 1: 109})
278
- 2026-03-18 13:29:40 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser.pt
279
- 2026-03-18 13:29:40 INFO: new best model saved.
280
- 2026-03-18 13:29:41 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
281
- 2026-03-18 13:29:41 INFO: new model checkpoint saved.
282
- 2026-03-18 13:29:42 INFO: Finished STEP 320/50000, loss = 4.718216 (0.382 sec/batch), lr: 2.000000
283
- 2026-03-18 13:29:46 INFO: Finished STEP 340/50000, loss = 2.242086 (0.065 sec/batch), lr: 2.000000
284
- 2026-03-18 13:29:48 INFO: Finished STEP 360/50000, loss = 1.605185 (0.065 sec/batch), lr: 2.000000
285
- 2026-03-18 13:29:50 INFO: Finished STEP 380/50000, loss = 1.821524 (0.063 sec/batch), lr: 2.000000
286
- 2026-03-18 13:29:54 INFO: Finished STEP 400/50000, loss = 1.597555 (0.065 sec/batch), lr: 2.000000
287
- 2026-03-18 13:29:54 INFO: Evaluating on dev set...
288
- 2026-03-18 13:29:57 INFO: LAS MLAS BLEX
289
- 2026-03-18 13:29:57 INFO: 50.46 38.59 43.93
290
- 2026-03-18 13:29:57 INFO: step 400: train_loss = 2.549639, dev_score = 0.5046
291
- 2026-03-18 13:29:57 DEBUG: training batch usage: Counter({0: 269, 1: 131})
292
- 2026-03-18 13:29:57 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
293
- 2026-03-18 13:29:57 INFO: new model checkpoint saved.
294
- 2026-03-18 13:30:00 INFO: Finished STEP 420/50000, loss = 2.220649 (0.065 sec/batch), lr: 2.000000
295
- 2026-03-18 13:30:04 INFO: Finished STEP 440/50000, loss = 1.550643 (0.064 sec/batch), lr: 2.000000
296
- 2026-03-18 13:30:08 INFO: Finished STEP 460/50000, loss = 3.319951 (0.344 sec/batch), lr: 2.000000
297
- 2026-03-18 13:30:12 INFO: Finished STEP 480/50000, loss = 2.615420 (0.068 sec/batch), lr: 2.000000
298
- 2026-03-18 13:30:14 INFO: Finished STEP 500/50000, loss = 1.976453 (0.068 sec/batch), lr: 2.000000
299
- 2026-03-18 13:30:14 INFO: Evaluating on dev set...
300
- 2026-03-18 13:30:17 INFO: LAS MLAS BLEX
301
- 2026-03-18 13:30:17 INFO: 52.11 42.78 46.03
302
- 2026-03-18 13:30:17 INFO: step 500: train_loss = 2.721934, dev_score = 0.5211
303
- 2026-03-18 13:30:17 DEBUG: training batch usage: Counter({0: 332, 1: 168})
304
- 2026-03-18 13:30:17 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser.pt
305
- 2026-03-18 13:30:17 INFO: new best model saved.
306
- 2026-03-18 13:30:18 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
307
- 2026-03-18 13:30:18 INFO: new model checkpoint saved.
308
- 2026-03-18 13:30:21 INFO: Finished STEP 520/50000, loss = 4.217991 (0.338 sec/batch), lr: 2.000000
309
- 2026-03-18 13:30:24 INFO: Finished STEP 540/50000, loss = 3.852498 (0.337 sec/batch), lr: 2.000000
310
- 2026-03-18 13:30:28 INFO: Finished STEP 560/50000, loss = 1.689470 (0.069 sec/batch), lr: 2.000000
311
- 2026-03-18 13:30:32 INFO: Finished STEP 580/50000, loss = 3.136248 (0.339 sec/batch), lr: 2.000000
312
- 2026-03-18 13:30:35 INFO: Finished STEP 600/50000, loss = 3.013068 (0.068 sec/batch), lr: 2.000000
313
- 2026-03-18 13:30:35 INFO: Evaluating on dev set...
314
- 2026-03-18 13:30:38 INFO: LAS MLAS BLEX
315
- 2026-03-18 13:30:38 INFO: 56.89 46.97 50.02
316
- 2026-03-18 13:30:38 INFO: step 600: train_loss = 2.802572, dev_score = 0.5689
317
- 2026-03-18 13:30:38 DEBUG: training batch usage: Counter({0: 392, 1: 208})
318
- 2026-03-18 13:30:38 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser.pt
319
- 2026-03-18 13:30:38 INFO: new best model saved.
320
- 2026-03-18 13:30:38 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
321
- 2026-03-18 13:30:38 INFO: new model checkpoint saved.
322
- 2026-03-18 13:30:41 INFO: Finished STEP 620/50000, loss = 3.173692 (0.327 sec/batch), lr: 2.000000
323
- 2026-03-18 13:30:44 INFO: Finished STEP 640/50000, loss = 1.843263 (0.070 sec/batch), lr: 2.000000
324
- 2026-03-18 13:30:47 INFO: Finished STEP 660/50000, loss = 1.417565 (0.069 sec/batch), lr: 2.000000
325
- 2026-03-18 13:30:51 INFO: Finished STEP 680/50000, loss = 4.149919 (0.325 sec/batch), lr: 2.000000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_bert-base-swedish-cased_sv_20260318_133119.txt DELETED
The diff for this file is too large to render. See raw diff
 
logs/log_bert-base-swedish-cased_sv_20260318_141654.txt DELETED
The diff for this file is too large to render. See raw diff
 
logs/log_bert-base-swedish-cased_sv_20260318_215701.txt DELETED
The diff for this file is too large to render. See raw diff
 
logs/log_bert-base-swedish-cased_sv_20260319_200948.txt DELETED
The diff for this file is too large to render. See raw diff
 
logs/log_bert-base-swedish-cased_sv_20260319_201349.txt DELETED
The diff for this file is too large to render. See raw diff
 
logs/log_bert-base-swedish-cased_sv_20260319_214621.txt DELETED
The diff for this file is too large to render. See raw diff
 
logs/log_conll17.pt_sv_20251203_232257.txt DELETED
@@ -1,719 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_20251203_232257.txt ===
2
- Language codes: sv
3
- Using pretrained model: conll17.pt
4
-
5
- Running: python prepare-train-val-test.py sv
6
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
7
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
8
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
9
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
10
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
11
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
12
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/ucxn_ud_swedish-talbanken.conllu
13
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
14
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
15
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
16
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
17
- Skipping DigPhil MACHINE (diachron not requested).
18
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
19
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
20
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
21
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
22
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
23
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
24
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
25
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
26
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
27
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
28
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
29
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
30
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
31
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
32
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
33
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
34
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
35
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
36
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
37
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
38
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
39
- Cleaning TRAIN...
40
- Cleaning DEV...
41
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
42
- Cleaning TEST...
43
- Writing TRAIN → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (19820 valid sentences)
44
- Writing DEV → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (9 valid sentences)
45
- Writing TEST → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (99 valid sentences)
46
- Done.
47
- Sourcing scripts/config_alvis.sh
48
- Running stanza dataset preparation…
49
- 2025-12-03 23:23:04 INFO: Datasets program called with:
50
- /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
51
- 2025-12-03 23:23:04 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
52
-
53
- 2025-12-03 23:23:04 INFO: Downloaded file to /cephyr/users/cleland/Alvis/stanza_resources/resources.json
54
- 2025-12-03 23:23:04 DEBUG: Processing parameter "processors"...
55
- 2025-12-03 23:23:04 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
56
- 2025-12-03 23:23:04 INFO: Downloading these customized packages for language: sv (Swedish)...
57
- =======================
58
- | Processor | Package |
59
- -----------------------
60
- =======================
61
-
62
- 2025-12-03 23:23:04 INFO: Finished downloading models and saved to /cephyr/users/cleland/Alvis/stanza_resources
63
- 2025-12-03 23:23:04 INFO: Using tagger model in /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt for sv_diachronic
64
- 2025-12-03 23:23:04 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
65
- 2025-12-03 23:23:04 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
66
- Augmented 56 quotes: Counter({'„”': 9, '″″': 9, '""': 8, '「」': 8, '””': 5, '““': 4, '《》': 4, '»«': 3, '„“': 3, '«»': 3})
67
- 2025-12-03 23:23:05 INFO: Running tagger to retag /local/tmp.5441282/tmp4sg9id1k/sv_diachronic.train.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
68
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5441282/tmp4sg9id1k/sv_diachronic.train.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu']
69
- 2025-12-03 23:23:05 INFO: Running tagger in predict mode
70
- 2025-12-03 23:23:05 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
71
- 2025-12-03 23:23:07 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
72
- 2025-12-03 23:23:07 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
73
- 2025-12-03 23:23:07 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
74
- 2025-12-03 23:23:07 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
75
- 2025-12-03 23:23:07 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
76
- 2025-12-03 23:23:10 INFO: Loading data with batch size 250...
77
- 2025-12-03 23:23:24 INFO: Start evaluation...
78
- 2025-12-03 23:24:31 INFO: UPOS XPOS UFeats AllTags
79
- 2025-12-03 23:24:31 INFO: 98.01 62.93 94.04 60.49
80
- 2025-12-03 23:24:31 INFO: POS Tagger score: sv_diachronic 60.49
81
- 2025-12-03 23:24:32 INFO: Running tagger to retag /local/tmp.5441282/tmp4sg9id1k/sv_diachronic.dev.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
82
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5441282/tmp4sg9id1k/sv_diachronic.dev.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu']
83
- 2025-12-03 23:24:32 INFO: Running tagger in predict mode
84
- 2025-12-03 23:24:32 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
85
- 2025-12-03 23:24:33 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
86
- 2025-12-03 23:24:33 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
87
- 2025-12-03 23:24:33 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
88
- 2025-12-03 23:24:33 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
89
- 2025-12-03 23:24:34 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
90
- 2025-12-03 23:24:34 INFO: Loading data with batch size 250...
91
- 2025-12-03 23:24:34 INFO: Start evaluation...
92
- 2025-12-03 23:24:34 INFO: UPOS XPOS UFeats AllTags
93
- 2025-12-03 23:24:34 INFO: 93.32 90.84 93.32 85.64
94
- 2025-12-03 23:24:34 INFO: POS Tagger score: sv_diachronic 85.64
95
- 2025-12-03 23:24:34 INFO: Running tagger to retag /local/tmp.5441282/tmp4sg9id1k/sv_diachronic.test.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
96
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5441282/tmp4sg9id1k/sv_diachronic.test.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu']
97
- 2025-12-03 23:24:34 INFO: Running tagger in predict mode
98
- 2025-12-03 23:24:34 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
99
- 2025-12-03 23:24:36 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
100
- 2025-12-03 23:24:36 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
101
- 2025-12-03 23:24:36 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
102
- 2025-12-03 23:24:36 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
103
- 2025-12-03 23:24:36 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
104
- 2025-12-03 23:24:36 INFO: Loading data with batch size 250...
105
- 2025-12-03 23:24:36 INFO: Start evaluation...
106
- 2025-12-03 23:24:37 INFO: UPOS XPOS UFeats AllTags
107
- 2025-12-03 23:24:37 INFO: 93.14 96.78 95.32 90.28
108
- 2025-12-03 23:24:37 INFO: POS Tagger score: sv_diachronic 90.28
109
- Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
110
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /local/tmp.5441282/tmp4sg9id1k/sv_diachronic.train.gold.conllu
111
- Swapped 'w1, w2' for 'w1 ,w2' 122 times
112
- Added 100 new sentences with asdf, zzzz -> asdf,zzzz
113
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /local/tmp.5441282/tmp4sg9id1k/sv_diachronic.dev.gold.conllu
114
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /local/tmp.5441282/tmp4sg9id1k/sv_diachronic.test.gold.conllu
115
- Running stanza dependency parser training…
116
- 2025-12-03 23:24:46 INFO: Training program called with:
117
- /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt --batch_size 32 --dropout 0.33
118
- 2025-12-03 23:24:46 DEBUG: UD_Swedish-diachronic: sv_diachronic
119
- 2025-12-03 23:24:46 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
120
- 2025-12-03 23:24:46 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
121
- 2025-12-03 23:24:46 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_charlm_parser.pt does not exist, training new model
122
- 2025-12-03 23:24:46 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
123
- 2025-12-03 23:24:46 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
124
- 2025-12-03 23:24:46 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--train_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
125
- 2025-12-03 23:24:46 INFO: Running parser in train mode
126
- 2025-12-03 23:24:46 INFO: Using pretrained contextualized char embedding
127
- 2025-12-03 23:24:46 INFO: Loading data with batch size 32...
128
- 2025-12-03 23:24:49 INFO: Train File /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 19920
129
- 2025-12-03 23:24:49 INFO: Original data size: 19920
130
- 2025-12-03 23:24:49 INFO: Augmented data size: 20844
131
- 2025-12-03 23:24:57 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
132
- 2025-12-03 23:24:57 INFO: Original length = 20844
133
- 2025-12-03 23:24:57 INFO: Filtered length = 20844
134
- 2025-12-03 23:25:02 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='|') for the xpos factory for sv_diachronic
135
- 2025-12-03 23:25:06 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
136
- 2025-12-03 23:25:11 DEBUG: 13986 batches created.
137
- 2025-12-03 23:25:11 DEBUG: 9 batches created.
138
- 2025-12-03 23:25:11 INFO: Training parser...
139
- 2025-12-03 23:25:11 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
140
- 2025-12-03 23:25:11 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
141
- 2025-12-03 23:25:11 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
142
- 2025-12-03 23:25:11 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
143
- 2025-12-03 23:25:14 INFO: Finished STEP 20/50000, loss = 3.699069 (0.033 sec/batch), lr: 0.003000
144
- 2025-12-03 23:25:15 INFO: Finished STEP 40/50000, loss = 3.841336 (0.035 sec/batch), lr: 0.003000
145
- 2025-12-03 23:25:16 INFO: Finished STEP 60/50000, loss = 2.719965 (0.034 sec/batch), lr: 0.003000
146
- 2025-12-03 23:25:16 INFO: Finished STEP 80/50000, loss = 3.832992 (0.035 sec/batch), lr: 0.003000
147
- 2025-12-03 23:25:17 INFO: Finished STEP 100/50000, loss = 3.050168 (0.035 sec/batch), lr: 0.003000
148
- 2025-12-03 23:25:17 INFO: Evaluating on dev set...
149
- 2025-12-03 23:25:18 INFO: LAS MLAS BLEX
150
- 2025-12-03 23:25:18 INFO: 3.22 0.82 1.64
151
- 2025-12-03 23:25:18 INFO: step 100: train_loss = 1815.927894, dev_score = 0.0322
152
- 2025-12-03 23:25:18 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
153
- 2025-12-03 23:25:18 INFO: new best model saved.
154
- 2025-12-03 23:25:19 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
155
- 2025-12-03 23:25:19 INFO: new model checkpoint saved.
156
- 2025-12-03 23:25:19 INFO: Finished STEP 120/50000, loss = 3.082294 (0.035 sec/batch), lr: 0.003000
157
- 2025-12-03 23:25:20 INFO: Finished STEP 140/50000, loss = 2.829382 (0.035 sec/batch), lr: 0.003000
158
- 2025-12-03 23:25:21 INFO: Finished STEP 160/50000, loss = 2.907810 (0.036 sec/batch), lr: 0.003000
159
- 2025-12-03 23:25:22 INFO: Finished STEP 180/50000, loss = 3.411438 (0.035 sec/batch), lr: 0.003000
160
- 2025-12-03 23:25:22 INFO: Finished STEP 200/50000, loss = 2.634063 (0.037 sec/batch), lr: 0.003000
161
- 2025-12-03 23:25:22 INFO: Evaluating on dev set...
162
- 2025-12-03 23:25:23 INFO: LAS MLAS BLEX
163
- 2025-12-03 23:25:23 INFO: 7.92 6.14 7.51
164
- 2025-12-03 23:25:23 INFO: step 200: train_loss = 2.962158, dev_score = 0.0792
165
- 2025-12-03 23:25:23 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
166
- 2025-12-03 23:25:23 INFO: new best model saved.
167
- 2025-12-03 23:25:24 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
168
- 2025-12-03 23:25:24 INFO: new model checkpoint saved.
169
- 2025-12-03 23:25:25 INFO: Finished STEP 220/50000, loss = 2.243211 (0.036 sec/batch), lr: 0.003000
170
- 2025-12-03 23:25:25 INFO: Finished STEP 240/50000, loss = 3.037997 (0.034 sec/batch), lr: 0.003000
171
- 2025-12-03 23:25:26 INFO: Finished STEP 260/50000, loss = 2.535078 (0.034 sec/batch), lr: 0.003000
172
- 2025-12-03 23:25:27 INFO: Finished STEP 280/50000, loss = 3.026233 (0.035 sec/batch), lr: 0.003000
173
- 2025-12-03 23:25:27 INFO: Finished STEP 300/50000, loss = 3.630493 (0.034 sec/batch), lr: 0.003000
174
- 2025-12-03 23:25:27 INFO: Evaluating on dev set...
175
- 2025-12-03 23:25:28 INFO: LAS MLAS BLEX
176
- 2025-12-03 23:25:28 INFO: 13.61 10.54 12.24
177
- 2025-12-03 23:25:28 INFO: step 300: train_loss = 3.006871, dev_score = 0.1361
178
- 2025-12-03 23:25:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
179
- 2025-12-03 23:25:28 INFO: new best model saved.
180
- 2025-12-03 23:25:29 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
181
- 2025-12-03 23:25:29 INFO: new model checkpoint saved.
182
- 2025-12-03 23:25:30 INFO: Finished STEP 320/50000, loss = 2.784619 (0.034 sec/batch), lr: 0.003000
183
- 2025-12-03 23:25:30 INFO: Finished STEP 340/50000, loss = 3.004477 (0.038 sec/batch), lr: 0.003000
184
- 2025-12-03 23:25:31 INFO: Finished STEP 360/50000, loss = 2.542431 (0.035 sec/batch), lr: 0.003000
185
- 2025-12-03 23:25:32 INFO: Finished STEP 380/50000, loss = 3.084781 (0.036 sec/batch), lr: 0.003000
186
- 2025-12-03 23:25:32 INFO: Finished STEP 400/50000, loss = 2.454229 (0.035 sec/batch), lr: 0.003000
187
- 2025-12-03 23:25:32 INFO: Evaluating on dev set...
188
- 2025-12-03 23:25:33 INFO: LAS MLAS BLEX
189
- 2025-12-03 23:25:33 INFO: 18.81 6.64 10.55
190
- 2025-12-03 23:25:33 INFO: step 400: train_loss = 3.048551, dev_score = 0.1881
191
- 2025-12-03 23:25:33 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
192
- 2025-12-03 23:25:33 INFO: new best model saved.
193
- 2025-12-03 23:25:34 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
194
- 2025-12-03 23:25:34 INFO: new model checkpoint saved.
195
- 2025-12-03 23:25:35 INFO: Finished STEP 420/50000, loss = 2.214305 (0.034 sec/batch), lr: 0.003000
196
- 2025-12-03 23:25:35 INFO: Finished STEP 440/50000, loss = 2.953549 (0.034 sec/batch), lr: 0.003000
197
- 2025-12-03 23:25:36 INFO: Finished STEP 460/50000, loss = 2.711811 (0.037 sec/batch), lr: 0.003000
198
- 2025-12-03 23:25:37 INFO: Finished STEP 480/50000, loss = 2.823795 (0.036 sec/batch), lr: 0.003000
199
- 2025-12-03 23:25:38 INFO: Finished STEP 500/50000, loss = 4.227708 (0.035 sec/batch), lr: 0.003000
200
- 2025-12-03 23:25:38 INFO: Evaluating on dev set...
201
- 2025-12-03 23:25:38 INFO: LAS MLAS BLEX
202
- 2025-12-03 23:25:38 INFO: 18.07 7.92 9.43
203
- 2025-12-03 23:25:38 INFO: step 500: train_loss = 3.085661, dev_score = 0.1807
204
- 2025-12-03 23:25:39 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
205
- 2025-12-03 23:25:39 INFO: new model checkpoint saved.
206
- 2025-12-03 23:25:39 INFO: Finished STEP 520/50000, loss = 3.152939 (0.039 sec/batch), lr: 0.003000
207
- 2025-12-03 23:25:40 INFO: Finished STEP 540/50000, loss = 2.722816 (0.037 sec/batch), lr: 0.003000
208
- 2025-12-03 23:25:41 INFO: Finished STEP 560/50000, loss = 1.797972 (0.034 sec/batch), lr: 0.003000
209
- 2025-12-03 23:25:42 INFO: Finished STEP 580/50000, loss = 1.902476 (0.037 sec/batch), lr: 0.003000
210
- 2025-12-03 23:25:42 INFO: Finished STEP 600/50000, loss = 2.532953 (0.037 sec/batch), lr: 0.003000
211
- 2025-12-03 23:25:42 INFO: Evaluating on dev set...
212
- 2025-12-03 23:25:43 INFO: LAS MLAS BLEX
213
- 2025-12-03 23:25:43 INFO: 24.75 12.64 15.33
214
- 2025-12-03 23:25:43 INFO: step 600: train_loss = 3.006814, dev_score = 0.2475
215
- 2025-12-03 23:25:43 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
216
- 2025-12-03 23:25:43 INFO: new best model saved.
217
- 2025-12-03 23:25:44 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
218
- 2025-12-03 23:25:44 INFO: new model checkpoint saved.
219
- 2025-12-03 23:25:44 INFO: Finished STEP 620/50000, loss = 3.911072 (0.037 sec/batch), lr: 0.003000
220
- 2025-12-03 23:25:45 INFO: Finished STEP 640/50000, loss = 3.704555 (0.039 sec/batch), lr: 0.003000
221
- 2025-12-03 23:25:46 INFO: Finished STEP 660/50000, loss = 2.692690 (0.037 sec/batch), lr: 0.003000
222
- 2025-12-03 23:25:47 INFO: Finished STEP 680/50000, loss = 2.771069 (0.038 sec/batch), lr: 0.003000
223
- 2025-12-03 23:25:47 INFO: Finished STEP 700/50000, loss = 4.281591 (0.036 sec/batch), lr: 0.003000
224
- 2025-12-03 23:25:47 INFO: Evaluating on dev set...
225
- 2025-12-03 23:25:48 INFO: LAS MLAS BLEX
226
- 2025-12-03 23:25:48 INFO: 34.65 21.99 25.31
227
- 2025-12-03 23:25:48 INFO: step 700: train_loss = 3.216450, dev_score = 0.3465
228
- 2025-12-03 23:25:48 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
229
- 2025-12-03 23:25:48 INFO: new best model saved.
230
- 2025-12-03 23:25:49 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
231
- 2025-12-03 23:25:49 INFO: new model checkpoint saved.
232
- 2025-12-03 23:25:50 INFO: Finished STEP 720/50000, loss = 3.350647 (0.040 sec/batch), lr: 0.003000
233
- 2025-12-03 23:25:50 INFO: Finished STEP 740/50000, loss = 2.873540 (0.038 sec/batch), lr: 0.003000
234
- 2025-12-03 23:25:51 INFO: Finished STEP 760/50000, loss = 3.564713 (0.038 sec/batch), lr: 0.003000
235
- 2025-12-03 23:25:52 INFO: Finished STEP 780/50000, loss = 3.640228 (0.041 sec/batch), lr: 0.003000
236
- 2025-12-03 23:25:53 INFO: Finished STEP 800/50000, loss = 2.930106 (0.036 sec/batch), lr: 0.003000
237
- 2025-12-03 23:25:53 INFO: Evaluating on dev set...
238
- 2025-12-03 23:25:53 INFO: LAS MLAS BLEX
239
- 2025-12-03 23:25:53 INFO: 24.75 8.60 12.90
240
- 2025-12-03 23:25:53 INFO: step 800: train_loss = 3.300487, dev_score = 0.2475
241
- 2025-12-03 23:25:54 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
242
- 2025-12-03 23:25:54 INFO: new model checkpoint saved.
243
- 2025-12-03 23:25:54 INFO: Finished STEP 820/50000, loss = 2.968561 (0.035 sec/batch), lr: 0.003000
244
- 2025-12-03 23:25:55 INFO: Finished STEP 840/50000, loss = 2.491823 (0.037 sec/batch), lr: 0.003000
245
- 2025-12-03 23:25:56 INFO: Finished STEP 860/50000, loss = 3.991972 (0.035 sec/batch), lr: 0.003000
246
- 2025-12-03 23:25:57 INFO: Finished STEP 880/50000, loss = 2.541115 (0.037 sec/batch), lr: 0.003000
247
- 2025-12-03 23:25:58 INFO: Finished STEP 900/50000, loss = 3.015432 (0.038 sec/batch), lr: 0.003000
248
- 2025-12-03 23:25:58 INFO: Evaluating on dev set...
249
- 2025-12-03 23:25:58 INFO: LAS MLAS BLEX
250
- 2025-12-03 23:25:58 INFO: 25.50 11.48 13.52
251
- 2025-12-03 23:25:58 INFO: step 900: train_loss = 3.187365, dev_score = 0.2550
252
- 2025-12-03 23:25:59 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
253
- 2025-12-03 23:25:59 INFO: new model checkpoint saved.
254
- 2025-12-03 23:25:59 INFO: Finished STEP 920/50000, loss = 2.031316 (0.038 sec/batch), lr: 0.003000
255
- 2025-12-03 23:26:00 INFO: Finished STEP 940/50000, loss = 2.938839 (0.036 sec/batch), lr: 0.003000
256
- 2025-12-03 23:26:01 INFO: Finished STEP 960/50000, loss = 3.606135 (0.036 sec/batch), lr: 0.003000
257
- 2025-12-03 23:26:02 INFO: Finished STEP 980/50000, loss = 5.427132 (0.038 sec/batch), lr: 0.003000
258
- 2025-12-03 23:26:02 INFO: Finished STEP 1000/50000, loss = 2.710342 (0.039 sec/batch), lr: 0.003000
259
- 2025-12-03 23:26:02 INFO: Evaluating on dev set...
260
- 2025-12-03 23:26:03 INFO: LAS MLAS BLEX
261
- 2025-12-03 23:26:03 INFO: 37.87 21.71 24.22
262
- 2025-12-03 23:26:03 INFO: step 1000: train_loss = 3.299336, dev_score = 0.3787
263
- 2025-12-03 23:26:03 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
264
- 2025-12-03 23:26:03 INFO: new best model saved.
265
- 2025-12-03 23:26:04 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
266
- 2025-12-03 23:26:04 INFO: new model checkpoint saved.
267
- 2025-12-03 23:26:05 INFO: Finished STEP 1020/50000, loss = 3.274420 (0.040 sec/batch), lr: 0.003000
268
- 2025-12-03 23:26:05 INFO: Finished STEP 1040/50000, loss = 3.172289 (0.037 sec/batch), lr: 0.003000
269
- 2025-12-03 23:26:06 INFO: Finished STEP 1060/50000, loss = 2.884028 (0.040 sec/batch), lr: 0.003000
270
- 2025-12-03 23:26:07 INFO: Finished STEP 1080/50000, loss = 4.205043 (0.038 sec/batch), lr: 0.003000
271
- 2025-12-03 23:26:08 INFO: Finished STEP 1100/50000, loss = 3.608851 (0.039 sec/batch), lr: 0.003000
272
- 2025-12-03 23:26:08 INFO: Evaluating on dev set...
273
- 2025-12-03 23:26:08 INFO: LAS MLAS BLEX
274
- 2025-12-03 23:26:08 INFO: 38.61 24.19 26.78
275
- 2025-12-03 23:26:08 INFO: step 1100: train_loss = 3.198060, dev_score = 0.3861
276
- 2025-12-03 23:26:09 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
277
- 2025-12-03 23:26:09 INFO: new best model saved.
278
- 2025-12-03 23:26:09 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
279
- 2025-12-03 23:26:09 INFO: new model checkpoint saved.
280
- 2025-12-03 23:26:10 INFO: Finished STEP 1120/50000, loss = 2.438182 (0.039 sec/batch), lr: 0.003000
281
- 2025-12-03 23:26:11 INFO: Finished STEP 1140/50000, loss = 4.315670 (0.038 sec/batch), lr: 0.003000
282
- 2025-12-03 23:26:11 INFO: Finished STEP 1160/50000, loss = 3.651882 (0.038 sec/batch), lr: 0.003000
283
- 2025-12-03 23:26:12 INFO: Finished STEP 1180/50000, loss = 3.139420 (0.037 sec/batch), lr: 0.003000
284
- 2025-12-03 23:26:13 INFO: Finished STEP 1200/50000, loss = 2.040530 (0.038 sec/batch), lr: 0.003000
285
- 2025-12-03 23:26:13 INFO: Evaluating on dev set...
286
- 2025-12-03 23:26:14 INFO: LAS MLAS BLEX
287
- 2025-12-03 23:26:14 INFO: 39.85 23.98 26.98
288
- 2025-12-03 23:26:14 INFO: step 1200: train_loss = 3.227944, dev_score = 0.3985
289
- 2025-12-03 23:26:14 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
290
- 2025-12-03 23:26:14 INFO: new best model saved.
291
- 2025-12-03 23:26:14 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
292
- 2025-12-03 23:26:14 INFO: new model checkpoint saved.
293
- 2025-12-03 23:26:15 INFO: Finished STEP 1220/50000, loss = 2.697804 (0.041 sec/batch), lr: 0.003000
294
- 2025-12-03 23:26:16 INFO: Finished STEP 1240/50000, loss = 2.911831 (0.038 sec/batch), lr: 0.003000
295
- 2025-12-03 23:26:17 INFO: Finished STEP 1260/50000, loss = 1.945185 (0.040 sec/batch), lr: 0.003000
296
- 2025-12-03 23:26:18 INFO: Finished STEP 1280/50000, loss = 3.025609 (0.038 sec/batch), lr: 0.003000
297
- 2025-12-03 23:26:18 INFO: Finished STEP 1300/50000, loss = 4.230402 (0.037 sec/batch), lr: 0.003000
298
- 2025-12-03 23:26:18 INFO: Evaluating on dev set...
299
- 2025-12-03 23:26:19 INFO: LAS MLAS BLEX
300
- 2025-12-03 23:26:19 INFO: 45.30 30.97 33.55
301
- 2025-12-03 23:26:19 INFO: step 1300: train_loss = 3.111315, dev_score = 0.4530
302
- 2025-12-03 23:26:19 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
303
- 2025-12-03 23:26:19 INFO: new best model saved.
304
- 2025-12-03 23:26:20 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
305
- 2025-12-03 23:26:20 INFO: new model checkpoint saved.
306
- 2025-12-03 23:26:21 INFO: Finished STEP 1320/50000, loss = 1.966291 (0.038 sec/batch), lr: 0.003000
307
- 2025-12-03 23:26:21 INFO: Finished STEP 1340/50000, loss = 3.216881 (0.040 sec/batch), lr: 0.003000
308
- 2025-12-03 23:26:22 INFO: Finished STEP 1360/50000, loss = 2.379959 (0.038 sec/batch), lr: 0.003000
309
- 2025-12-03 23:26:23 INFO: Finished STEP 1380/50000, loss = 4.992296 (0.037 sec/batch), lr: 0.003000
310
- 2025-12-03 23:26:24 INFO: Finished STEP 1400/50000, loss = 3.349003 (0.042 sec/batch), lr: 0.003000
311
- 2025-12-03 23:26:24 INFO: Evaluating on dev set...
312
- 2025-12-03 23:26:24 INFO: LAS MLAS BLEX
313
- 2025-12-03 23:26:24 INFO: 36.88 23.21 24.89
314
- 2025-12-03 23:26:24 INFO: step 1400: train_loss = 3.364135, dev_score = 0.3688
315
- 2025-12-03 23:26:25 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
316
- 2025-12-03 23:26:25 INFO: new model checkpoint saved.
317
- 2025-12-03 23:26:26 INFO: Finished STEP 1420/50000, loss = 4.139513 (0.036 sec/batch), lr: 0.003000
318
- 2025-12-03 23:26:26 INFO: Finished STEP 1440/50000, loss = 2.905265 (0.037 sec/batch), lr: 0.003000
319
- 2025-12-03 23:26:27 INFO: Finished STEP 1460/50000, loss = 3.610150 (0.038 sec/batch), lr: 0.003000
320
- 2025-12-03 23:26:28 INFO: Finished STEP 1480/50000, loss = 4.759534 (0.038 sec/batch), lr: 0.003000
321
- 2025-12-03 23:26:29 INFO: Finished STEP 1500/50000, loss = 2.334270 (0.040 sec/batch), lr: 0.003000
322
- 2025-12-03 23:26:29 INFO: Evaluating on dev set...
323
- 2025-12-03 23:26:29 INFO: LAS MLAS BLEX
324
- 2025-12-03 23:26:29 INFO: 33.66 20.92 32.22
325
- 2025-12-03 23:26:29 INFO: step 1500: train_loss = 3.377781, dev_score = 0.3366
326
- 2025-12-03 23:26:30 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
327
- 2025-12-03 23:26:30 INFO: new model checkpoint saved.
328
- 2025-12-03 23:26:31 INFO: Finished STEP 1520/50000, loss = 2.824460 (0.037 sec/batch), lr: 0.003000
329
- 2025-12-03 23:26:32 INFO: Finished STEP 1540/50000, loss = 4.636815 (0.039 sec/batch), lr: 0.003000
330
- 2025-12-03 23:26:32 INFO: Finished STEP 1560/50000, loss = 3.666232 (0.039 sec/batch), lr: 0.003000
331
- 2025-12-03 23:26:33 INFO: Finished STEP 1580/50000, loss = 2.960546 (0.039 sec/batch), lr: 0.003000
332
- 2025-12-03 23:26:34 INFO: Finished STEP 1600/50000, loss = 2.503325 (0.039 sec/batch), lr: 0.003000
333
- 2025-12-03 23:26:34 INFO: Evaluating on dev set...
334
- 2025-12-03 23:26:34 INFO: LAS MLAS BLEX
335
- 2025-12-03 23:26:34 INFO: 42.57 27.91 31.29
336
- 2025-12-03 23:26:34 INFO: step 1600: train_loss = 3.303970, dev_score = 0.4257
337
- 2025-12-03 23:26:35 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
338
- 2025-12-03 23:26:35 INFO: new model checkpoint saved.
339
- 2025-12-03 23:26:36 INFO: Finished STEP 1620/50000, loss = 3.455626 (0.042 sec/batch), lr: 0.003000
340
- 2025-12-03 23:26:37 INFO: Finished STEP 1640/50000, loss = 3.333741 (0.038 sec/batch), lr: 0.003000
341
- 2025-12-03 23:26:37 INFO: Finished STEP 1660/50000, loss = 3.761724 (0.041 sec/batch), lr: 0.003000
342
- 2025-12-03 23:26:38 INFO: Finished STEP 1680/50000, loss = 3.171466 (0.039 sec/batch), lr: 0.003000
343
- 2025-12-03 23:26:39 INFO: Finished STEP 1700/50000, loss = 3.169466 (0.040 sec/batch), lr: 0.003000
344
- 2025-12-03 23:26:39 INFO: Evaluating on dev set...
345
- 2025-12-03 23:26:39 INFO: LAS MLAS BLEX
346
- 2025-12-03 23:26:39 INFO: 45.05 33.96 39.41
347
- 2025-12-03 23:26:39 INFO: step 1700: train_loss = 3.350673, dev_score = 0.4505
348
- 2025-12-03 23:26:40 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
349
- 2025-12-03 23:26:40 INFO: new model checkpoint saved.
350
- 2025-12-03 23:26:41 INFO: Finished STEP 1720/50000, loss = 2.898433 (0.037 sec/batch), lr: 0.003000
351
- 2025-12-03 23:26:42 INFO: Finished STEP 1740/50000, loss = 4.455222 (0.039 sec/batch), lr: 0.003000
352
- 2025-12-03 23:26:42 INFO: Finished STEP 1760/50000, loss = 2.290793 (0.041 sec/batch), lr: 0.003000
353
- 2025-12-03 23:26:43 INFO: Finished STEP 1780/50000, loss = 3.614108 (0.041 sec/batch), lr: 0.003000
354
- 2025-12-03 23:26:44 INFO: Finished STEP 1800/50000, loss = 2.709010 (0.039 sec/batch), lr: 0.003000
355
- 2025-12-03 23:26:44 INFO: Evaluating on dev set...
356
- 2025-12-03 23:26:45 INFO: LAS MLAS BLEX
357
- 2025-12-03 23:26:45 INFO: 46.04 29.96 33.76
358
- 2025-12-03 23:26:45 INFO: step 1800: train_loss = 3.388469, dev_score = 0.4604
359
- 2025-12-03 23:26:45 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
360
- 2025-12-03 23:26:45 INFO: new best model saved.
361
- 2025-12-03 23:26:45 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
362
- 2025-12-03 23:26:45 INFO: new model checkpoint saved.
363
- 2025-12-03 23:26:46 INFO: Finished STEP 1820/50000, loss = 3.027663 (0.039 sec/batch), lr: 0.003000
364
- 2025-12-03 23:26:47 INFO: Finished STEP 1840/50000, loss = 4.950170 (0.041 sec/batch), lr: 0.003000
365
- 2025-12-03 23:26:48 INFO: Finished STEP 1860/50000, loss = 3.627461 (0.037 sec/batch), lr: 0.003000
366
- 2025-12-03 23:26:49 INFO: Finished STEP 1880/50000, loss = 3.900440 (0.041 sec/batch), lr: 0.003000
367
- 2025-12-03 23:26:49 INFO: Finished STEP 1900/50000, loss = 3.206836 (0.037 sec/batch), lr: 0.003000
368
- 2025-12-03 23:26:49 INFO: Evaluating on dev set...
369
- 2025-12-03 23:26:50 INFO: LAS MLAS BLEX
370
- 2025-12-03 23:26:50 INFO: 41.09 28.63 34.11
371
- 2025-12-03 23:26:50 INFO: step 1900: train_loss = 3.372791, dev_score = 0.4109
372
- 2025-12-03 23:26:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
373
- 2025-12-03 23:26:51 INFO: new model checkpoint saved.
374
- 2025-12-03 23:26:51 INFO: Finished STEP 1920/50000, loss = 4.990382 (0.038 sec/batch), lr: 0.003000
375
- 2025-12-03 23:26:52 INFO: Finished STEP 1940/50000, loss = 3.309713 (0.041 sec/batch), lr: 0.003000
376
- 2025-12-03 23:26:53 INFO: Finished STEP 1960/50000, loss = 3.610244 (0.039 sec/batch), lr: 0.003000
377
- 2025-12-03 23:26:54 INFO: Finished STEP 1980/50000, loss = 3.427943 (0.039 sec/batch), lr: 0.003000
378
- 2025-12-03 23:26:55 INFO: Finished STEP 2000/50000, loss = 3.286536 (0.039 sec/batch), lr: 0.003000
379
- 2025-12-03 23:26:55 INFO: Evaluating on dev set...
380
- 2025-12-03 23:26:55 INFO: LAS MLAS BLEX
381
- 2025-12-03 23:26:55 INFO: 46.04 34.82 37.79
382
- 2025-12-03 23:26:55 INFO: step 2000: train_loss = 3.315330, dev_score = 0.4604
383
- 2025-12-03 23:26:55 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
384
- 2025-12-03 23:26:55 INFO: new best model saved.
385
- 2025-12-03 23:26:56 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
386
- 2025-12-03 23:26:56 INFO: new model checkpoint saved.
387
- 2025-12-03 23:26:57 INFO: Finished STEP 2020/50000, loss = 2.214467 (0.041 sec/batch), lr: 0.003000
388
- 2025-12-03 23:26:58 INFO: Finished STEP 2040/50000, loss = 3.128998 (0.041 sec/batch), lr: 0.003000
389
- 2025-12-03 23:26:58 INFO: Finished STEP 2060/50000, loss = 3.400111 (0.041 sec/batch), lr: 0.003000
390
- 2025-12-03 23:26:59 INFO: Finished STEP 2080/50000, loss = 5.836899 (0.042 sec/batch), lr: 0.003000
391
- 2025-12-03 23:27:00 INFO: Finished STEP 2100/50000, loss = 2.544196 (0.042 sec/batch), lr: 0.003000
392
- 2025-12-03 23:27:00 INFO: Evaluating on dev set...
393
- 2025-12-03 23:27:01 INFO: LAS MLAS BLEX
394
- 2025-12-03 23:27:01 INFO: 41.34 27.60 36.09
395
- 2025-12-03 23:27:01 INFO: step 2100: train_loss = 3.418321, dev_score = 0.4134
396
- 2025-12-03 23:27:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
397
- 2025-12-03 23:27:01 INFO: new model checkpoint saved.
398
- 2025-12-03 23:27:02 INFO: Finished STEP 2120/50000, loss = 4.017645 (0.041 sec/batch), lr: 0.003000
399
- 2025-12-03 23:27:03 INFO: Finished STEP 2140/50000, loss = 4.332951 (0.040 sec/batch), lr: 0.003000
400
- 2025-12-03 23:27:04 INFO: Finished STEP 2160/50000, loss = 2.522452 (0.038 sec/batch), lr: 0.003000
401
- 2025-12-03 23:27:04 INFO: Finished STEP 2180/50000, loss = 3.793148 (0.040 sec/batch), lr: 0.003000
402
- 2025-12-03 23:27:05 INFO: Finished STEP 2200/50000, loss = 2.889349 (0.041 sec/batch), lr: 0.003000
403
- 2025-12-03 23:27:05 INFO: Evaluating on dev set...
404
- 2025-12-03 23:27:06 INFO: LAS MLAS BLEX
405
- 2025-12-03 23:27:06 INFO: 39.60 26.64 35.10
406
- 2025-12-03 23:27:06 INFO: step 2200: train_loss = 3.486956, dev_score = 0.3960
407
- 2025-12-03 23:27:06 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
408
- 2025-12-03 23:27:06 INFO: new model checkpoint saved.
409
- 2025-12-03 23:27:07 INFO: Finished STEP 2220/50000, loss = 3.812532 (0.039 sec/batch), lr: 0.003000
410
- 2025-12-03 23:27:08 INFO: Finished STEP 2240/50000, loss = 4.048740 (0.039 sec/batch), lr: 0.003000
411
- 2025-12-03 23:27:09 INFO: Finished STEP 2260/50000, loss = 3.259020 (0.038 sec/batch), lr: 0.003000
412
- 2025-12-03 23:27:10 INFO: Finished STEP 2280/50000, loss = 3.957003 (0.038 sec/batch), lr: 0.003000
413
- 2025-12-03 23:27:10 INFO: Finished STEP 2300/50000, loss = 3.192188 (0.044 sec/batch), lr: 0.003000
414
- 2025-12-03 23:27:10 INFO: Evaluating on dev set...
415
- 2025-12-03 23:27:11 INFO: LAS MLAS BLEX
416
- 2025-12-03 23:27:11 INFO: 43.81 27.85 33.33
417
- 2025-12-03 23:27:11 INFO: step 2300: train_loss = 3.390584, dev_score = 0.4381
418
- 2025-12-03 23:27:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
419
- 2025-12-03 23:27:12 INFO: new model checkpoint saved.
420
- 2025-12-03 23:27:12 INFO: Finished STEP 2320/50000, loss = 4.818910 (0.040 sec/batch), lr: 0.003000
421
- 2025-12-03 23:27:13 INFO: Finished STEP 2340/50000, loss = 2.406329 (0.042 sec/batch), lr: 0.003000
422
- 2025-12-03 23:27:14 INFO: Finished STEP 2360/50000, loss = 2.298516 (0.040 sec/batch), lr: 0.003000
423
- 2025-12-03 23:27:15 INFO: Finished STEP 2380/50000, loss = 2.806631 (0.041 sec/batch), lr: 0.003000
424
- 2025-12-03 23:27:16 INFO: Finished STEP 2400/50000, loss = 3.427664 (0.040 sec/batch), lr: 0.003000
425
- 2025-12-03 23:27:16 INFO: Evaluating on dev set...
426
- 2025-12-03 23:27:16 INFO: LAS MLAS BLEX
427
- 2025-12-03 23:27:16 INFO: 43.07 31.09 37.39
428
- 2025-12-03 23:27:16 INFO: step 2400: train_loss = 3.488748, dev_score = 0.4307
429
- 2025-12-03 23:27:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
430
- 2025-12-03 23:27:17 INFO: new model checkpoint saved.
431
- 2025-12-03 23:27:18 INFO: Finished STEP 2420/50000, loss = 2.539723 (0.041 sec/batch), lr: 0.003000
432
- 2025-12-03 23:27:19 INFO: Finished STEP 2440/50000, loss = 4.533563 (0.043 sec/batch), lr: 0.003000
433
- 2025-12-03 23:27:19 INFO: Finished STEP 2460/50000, loss = 4.236969 (0.043 sec/batch), lr: 0.003000
434
- 2025-12-03 23:27:20 INFO: Finished STEP 2480/50000, loss = 2.358161 (0.044 sec/batch), lr: 0.003000
435
- 2025-12-03 23:27:21 INFO: Finished STEP 2500/50000, loss = 3.165836 (0.040 sec/batch), lr: 0.003000
436
- 2025-12-03 23:27:21 INFO: Evaluating on dev set...
437
- 2025-12-03 23:27:22 INFO: LAS MLAS BLEX
438
- 2025-12-03 23:27:22 INFO: 46.53 30.44 35.52
439
- 2025-12-03 23:27:22 INFO: step 2500: train_loss = 3.440427, dev_score = 0.4653
440
- 2025-12-03 23:27:22 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
441
- 2025-12-03 23:27:22 INFO: new best model saved.
442
- 2025-12-03 23:27:22 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
443
- 2025-12-03 23:27:22 INFO: new model checkpoint saved.
444
- 2025-12-03 23:27:23 INFO: Finished STEP 2520/50000, loss = 3.732689 (0.039 sec/batch), lr: 0.003000
445
- 2025-12-03 23:27:24 INFO: Finished STEP 2540/50000, loss = 5.190745 (0.040 sec/batch), lr: 0.003000
446
- 2025-12-03 23:27:25 INFO: Finished STEP 2560/50000, loss = 4.239511 (0.041 sec/batch), lr: 0.003000
447
- 2025-12-03 23:27:26 INFO: Finished STEP 2580/50000, loss = 4.053186 (0.040 sec/batch), lr: 0.003000
448
- 2025-12-03 23:27:27 INFO: Finished STEP 2600/50000, loss = 2.809570 (0.042 sec/batch), lr: 0.003000
449
- 2025-12-03 23:27:27 INFO: Evaluating on dev set...
450
- 2025-12-03 23:27:27 INFO: LAS MLAS BLEX
451
- 2025-12-03 23:27:27 INFO: 49.50 37.55 41.77
452
- 2025-12-03 23:27:27 INFO: step 2600: train_loss = 3.507095, dev_score = 0.4950
453
- 2025-12-03 23:27:27 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
454
- 2025-12-03 23:27:27 INFO: new best model saved.
455
- 2025-12-03 23:27:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
456
- 2025-12-03 23:27:28 INFO: new model checkpoint saved.
457
- 2025-12-03 23:27:29 INFO: Finished STEP 2620/50000, loss = 4.161043 (0.040 sec/batch), lr: 0.003000
458
- 2025-12-03 23:27:30 INFO: Finished STEP 2640/50000, loss = 3.440454 (0.041 sec/batch), lr: 0.003000
459
- 2025-12-03 23:27:31 INFO: Finished STEP 2660/50000, loss = 3.371374 (0.039 sec/batch), lr: 0.003000
460
- 2025-12-03 23:27:31 INFO: Finished STEP 2680/50000, loss = 2.921980 (0.039 sec/batch), lr: 0.003000
461
- 2025-12-03 23:27:32 INFO: Finished STEP 2700/50000, loss = 2.089964 (0.040 sec/batch), lr: 0.003000
462
- 2025-12-03 23:27:32 INFO: Evaluating on dev set...
463
- 2025-12-03 23:27:33 INFO: LAS MLAS BLEX
464
- 2025-12-03 23:27:33 INFO: 45.54 33.12 36.94
465
- 2025-12-03 23:27:33 INFO: step 2700: train_loss = 3.381599, dev_score = 0.4554
466
- 2025-12-03 23:27:33 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
467
- 2025-12-03 23:27:33 INFO: new model checkpoint saved.
468
- 2025-12-03 23:27:34 INFO: Finished STEP 2720/50000, loss = 5.878531 (0.041 sec/batch), lr: 0.003000
469
- 2025-12-03 23:27:35 INFO: Finished STEP 2740/50000, loss = 3.237665 (0.040 sec/batch), lr: 0.003000
470
- 2025-12-03 23:27:36 INFO: Finished STEP 2760/50000, loss = 2.492691 (0.042 sec/batch), lr: 0.003000
471
- 2025-12-03 23:27:37 INFO: Finished STEP 2780/50000, loss = 4.720194 (0.039 sec/batch), lr: 0.003000
472
- 2025-12-03 23:27:38 INFO: Finished STEP 2800/50000, loss = 3.760880 (0.042 sec/batch), lr: 0.003000
473
- 2025-12-03 23:27:38 INFO: Evaluating on dev set...
474
- 2025-12-03 23:27:38 INFO: LAS MLAS BLEX
475
- 2025-12-03 23:27:38 INFO: 51.24 38.24 43.28
476
- 2025-12-03 23:27:38 INFO: step 2800: train_loss = 3.545646, dev_score = 0.5124
477
- 2025-12-03 23:27:38 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
478
- 2025-12-03 23:27:38 INFO: new best model saved.
479
- 2025-12-03 23:27:39 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
480
- 2025-12-03 23:27:39 INFO: new model checkpoint saved.
481
- 2025-12-03 23:27:40 INFO: Finished STEP 2820/50000, loss = 3.466887 (0.040 sec/batch), lr: 0.003000
482
- 2025-12-03 23:27:41 INFO: Finished STEP 2840/50000, loss = 2.830301 (0.040 sec/batch), lr: 0.003000
483
- 2025-12-03 23:27:42 INFO: Finished STEP 2860/50000, loss = 3.183891 (0.043 sec/batch), lr: 0.003000
484
- 2025-12-03 23:27:43 INFO: Finished STEP 2880/50000, loss = 3.444857 (0.043 sec/batch), lr: 0.003000
485
- 2025-12-03 23:27:43 INFO: Finished STEP 2900/50000, loss = 3.642260 (0.044 sec/batch), lr: 0.003000
486
- 2025-12-03 23:27:43 INFO: Evaluating on dev set...
487
- 2025-12-03 23:27:44 INFO: LAS MLAS BLEX
488
- 2025-12-03 23:27:44 INFO: 53.71 40.59 44.82
489
- 2025-12-03 23:27:44 INFO: step 2900: train_loss = 3.663115, dev_score = 0.5371
490
- 2025-12-03 23:27:44 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
491
- 2025-12-03 23:27:44 INFO: new best model saved.
492
- 2025-12-03 23:27:45 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
493
- 2025-12-03 23:27:45 INFO: new model checkpoint saved.
494
- 2025-12-03 23:27:46 INFO: Finished STEP 2920/50000, loss = 2.488231 (0.043 sec/batch), lr: 0.003000
495
- 2025-12-03 23:27:46 INFO: Finished STEP 2940/50000, loss = 3.432896 (0.044 sec/batch), lr: 0.003000
496
- 2025-12-03 23:27:47 INFO: Finished STEP 2960/50000, loss = 2.781652 (0.041 sec/batch), lr: 0.003000
497
- 2025-12-03 23:27:48 INFO: Finished STEP 2980/50000, loss = 2.840132 (0.042 sec/batch), lr: 0.003000
498
- 2025-12-03 23:27:49 INFO: Finished STEP 3000/50000, loss = 4.171174 (0.039 sec/batch), lr: 0.003000
499
- 2025-12-03 23:27:49 INFO: Evaluating on dev set...
500
- 2025-12-03 23:27:49 INFO: LAS MLAS BLEX
501
- 2025-12-03 23:27:49 INFO: 51.98 39.50 45.80
502
- 2025-12-03 23:27:49 INFO: step 3000: train_loss = 3.476293, dev_score = 0.5198
503
- 2025-12-03 23:27:50 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
504
- 2025-12-03 23:27:50 INFO: new model checkpoint saved.
505
- 2025-12-03 23:27:51 INFO: Finished STEP 3020/50000, loss = 3.225040 (0.043 sec/batch), lr: 0.003000
506
- 2025-12-03 23:27:52 INFO: Finished STEP 3040/50000, loss = 3.064417 (0.042 sec/batch), lr: 0.003000
507
- 2025-12-03 23:27:53 INFO: Finished STEP 3060/50000, loss = 3.886524 (0.041 sec/batch), lr: 0.003000
508
- 2025-12-03 23:27:53 INFO: Finished STEP 3080/50000, loss = 3.704923 (0.041 sec/batch), lr: 0.003000
509
- 2025-12-03 23:27:54 INFO: Finished STEP 3100/50000, loss = 4.191244 (0.042 sec/batch), lr: 0.003000
510
- 2025-12-03 23:27:54 INFO: Evaluating on dev set...
511
- 2025-12-03 23:27:55 INFO: LAS MLAS BLEX
512
- 2025-12-03 23:27:55 INFO: 49.75 37.32 40.67
513
- 2025-12-03 23:27:55 INFO: step 3100: train_loss = 3.466918, dev_score = 0.4975
514
- 2025-12-03 23:27:55 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
515
- 2025-12-03 23:27:55 INFO: new model checkpoint saved.
516
- 2025-12-03 23:27:56 INFO: Finished STEP 3120/50000, loss = 3.950969 (0.042 sec/batch), lr: 0.003000
517
- 2025-12-03 23:27:57 INFO: Finished STEP 3140/50000, loss = 2.399348 (0.039 sec/batch), lr: 0.003000
518
- 2025-12-03 23:27:58 INFO: Finished STEP 3160/50000, loss = 3.456824 (0.042 sec/batch), lr: 0.003000
519
- 2025-12-03 23:27:59 INFO: Finished STEP 3180/50000, loss = 3.288764 (0.041 sec/batch), lr: 0.003000
520
- 2025-12-03 23:28:00 INFO: Finished STEP 3200/50000, loss = 3.057923 (0.041 sec/batch), lr: 0.003000
521
- 2025-12-03 23:28:00 INFO: Evaluating on dev set...
522
- 2025-12-03 23:28:00 INFO: LAS MLAS BLEX
523
- 2025-12-03 23:28:00 INFO: 47.77 36.55 40.76
524
- 2025-12-03 23:28:00 INFO: step 3200: train_loss = 3.563295, dev_score = 0.4777
525
- 2025-12-03 23:28:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
526
- 2025-12-03 23:28:01 INFO: new model checkpoint saved.
527
- 2025-12-03 23:28:02 INFO: Finished STEP 3220/50000, loss = 4.334808 (0.042 sec/batch), lr: 0.003000
528
- 2025-12-03 23:28:03 INFO: Finished STEP 3240/50000, loss = 2.773743 (0.040 sec/batch), lr: 0.003000
529
- 2025-12-03 23:28:03 INFO: Finished STEP 3260/50000, loss = 3.285001 (0.042 sec/batch), lr: 0.003000
530
- 2025-12-03 23:28:04 INFO: Finished STEP 3280/50000, loss = 3.142590 (0.046 sec/batch), lr: 0.003000
531
- 2025-12-03 23:28:05 INFO: Finished STEP 3300/50000, loss = 2.988616 (0.044 sec/batch), lr: 0.003000
532
- 2025-12-03 23:28:05 INFO: Evaluating on dev set...
533
- 2025-12-03 23:28:06 INFO: LAS MLAS BLEX
534
- 2025-12-03 23:28:06 INFO: 51.98 39.08 41.60
535
- 2025-12-03 23:28:06 INFO: step 3300: train_loss = 3.589203, dev_score = 0.5198
536
- 2025-12-03 23:28:06 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
537
- 2025-12-03 23:28:06 INFO: new model checkpoint saved.
538
- 2025-12-03 23:28:07 INFO: Finished STEP 3320/50000, loss = 5.763182 (0.045 sec/batch), lr: 0.003000
539
- 2025-12-03 23:28:08 INFO: Finished STEP 3340/50000, loss = 3.421094 (0.042 sec/batch), lr: 0.003000
540
- 2025-12-03 23:28:09 INFO: Finished STEP 3360/50000, loss = 5.406409 (0.043 sec/batch), lr: 0.003000
541
- 2025-12-03 23:28:10 INFO: Finished STEP 3380/50000, loss = 3.032209 (0.045 sec/batch), lr: 0.003000
542
- 2025-12-03 23:28:11 INFO: Finished STEP 3400/50000, loss = 3.139112 (0.042 sec/batch), lr: 0.003000
543
- 2025-12-03 23:28:11 INFO: Evaluating on dev set...
544
- 2025-12-03 23:28:11 INFO: LAS MLAS BLEX
545
- 2025-12-03 23:28:11 INFO: 49.75 36.29 40.08
546
- 2025-12-03 23:28:11 INFO: step 3400: train_loss = 3.722110, dev_score = 0.4975
547
- 2025-12-03 23:28:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
548
- 2025-12-03 23:28:12 INFO: new model checkpoint saved.
549
- 2025-12-03 23:28:13 INFO: Finished STEP 3420/50000, loss = 3.515473 (0.044 sec/batch), lr: 0.003000
550
- 2025-12-03 23:28:14 INFO: Finished STEP 3440/50000, loss = 4.545405 (0.042 sec/batch), lr: 0.003000
551
- 2025-12-03 23:28:15 INFO: Finished STEP 3460/50000, loss = 3.489767 (0.043 sec/batch), lr: 0.003000
552
- 2025-12-03 23:28:15 INFO: Finished STEP 3480/50000, loss = 4.931797 (0.042 sec/batch), lr: 0.003000
553
- 2025-12-03 23:28:16 INFO: Finished STEP 3500/50000, loss = 3.305768 (0.042 sec/batch), lr: 0.003000
554
- 2025-12-03 23:28:16 INFO: Evaluating on dev set...
555
- 2025-12-03 23:28:17 INFO: LAS MLAS BLEX
556
- 2025-12-03 23:28:17 INFO: 48.76 37.05 41.68
557
- 2025-12-03 23:28:17 INFO: step 3500: train_loss = 3.662503, dev_score = 0.4876
558
- 2025-12-03 23:28:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
559
- 2025-12-03 23:28:17 INFO: new model checkpoint saved.
560
- 2025-12-03 23:28:18 INFO: Finished STEP 3520/50000, loss = 4.402299 (0.043 sec/batch), lr: 0.003000
561
- 2025-12-03 23:28:19 INFO: Finished STEP 3540/50000, loss = 2.635880 (0.042 sec/batch), lr: 0.003000
562
- 2025-12-03 23:28:20 INFO: Finished STEP 3560/50000, loss = 2.857255 (0.043 sec/batch), lr: 0.003000
563
- 2025-12-03 23:28:21 INFO: Finished STEP 3580/50000, loss = 3.507267 (0.043 sec/batch), lr: 0.003000
564
- 2025-12-03 23:28:22 INFO: Finished STEP 3600/50000, loss = 5.196735 (0.042 sec/batch), lr: 0.003000
565
- 2025-12-03 23:28:22 INFO: Evaluating on dev set...
566
- 2025-12-03 23:28:22 INFO: LAS MLAS BLEX
567
- 2025-12-03 23:28:22 INFO: 51.24 40.84 44.21
568
- 2025-12-03 23:28:22 INFO: step 3600: train_loss = 3.681615, dev_score = 0.5124
569
- 2025-12-03 23:28:23 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
570
- 2025-12-03 23:28:23 INFO: new model checkpoint saved.
571
- 2025-12-03 23:28:24 INFO: Finished STEP 3620/50000, loss = 4.539114 (0.042 sec/batch), lr: 0.003000
572
- 2025-12-03 23:28:25 INFO: Finished STEP 3640/50000, loss = 6.413163 (0.042 sec/batch), lr: 0.003000
573
- 2025-12-03 23:28:25 INFO: Finished STEP 3660/50000, loss = 4.375307 (0.041 sec/batch), lr: 0.003000
574
- 2025-12-03 23:28:26 INFO: Finished STEP 3680/50000, loss = 5.120213 (0.041 sec/batch), lr: 0.003000
575
- 2025-12-03 23:28:27 INFO: Finished STEP 3700/50000, loss = 3.109243 (0.044 sec/batch), lr: 0.003000
576
- 2025-12-03 23:28:27 INFO: Evaluating on dev set...
577
- 2025-12-03 23:28:28 INFO: LAS MLAS BLEX
578
- 2025-12-03 23:28:28 INFO: 49.26 38.99 44.03
579
- 2025-12-03 23:28:28 INFO: step 3700: train_loss = 3.778901, dev_score = 0.4926
580
- 2025-12-03 23:28:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
581
- 2025-12-03 23:28:28 INFO: new model checkpoint saved.
582
- 2025-12-03 23:28:29 INFO: Finished STEP 3720/50000, loss = 3.468835 (0.045 sec/batch), lr: 0.003000
583
- 2025-12-03 23:28:30 INFO: Finished STEP 3740/50000, loss = 3.420460 (0.044 sec/batch), lr: 0.003000
584
- 2025-12-03 23:28:31 INFO: Finished STEP 3760/50000, loss = 2.856275 (0.042 sec/batch), lr: 0.003000
585
- 2025-12-03 23:28:32 INFO: Finished STEP 3780/50000, loss = 2.668692 (0.041 sec/batch), lr: 0.003000
586
- 2025-12-03 23:28:33 INFO: Finished STEP 3800/50000, loss = 4.511623 (0.044 sec/batch), lr: 0.003000
587
- 2025-12-03 23:28:33 INFO: Evaluating on dev set...
588
- 2025-12-03 23:28:33 INFO: LAS MLAS BLEX
589
- 2025-12-03 23:28:33 INFO: 45.30 35.79 39.16
590
- 2025-12-03 23:28:33 INFO: step 3800: train_loss = 3.920723, dev_score = 0.4530
591
- 2025-12-03 23:28:34 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
592
- 2025-12-03 23:28:34 INFO: new model checkpoint saved.
593
- 2025-12-03 23:28:35 INFO: Finished STEP 3820/50000, loss = 4.449274 (0.042 sec/batch), lr: 0.003000
594
- 2025-12-03 23:28:36 INFO: Finished STEP 3840/50000, loss = 2.533618 (0.042 sec/batch), lr: 0.003000
595
- 2025-12-03 23:28:36 INFO: Finished STEP 3860/50000, loss = 3.834488 (0.043 sec/batch), lr: 0.003000
596
- 2025-12-03 23:28:37 INFO: Finished STEP 3880/50000, loss = 3.449773 (0.043 sec/batch), lr: 0.003000
597
- 2025-12-03 23:28:38 INFO: Finished STEP 3900/50000, loss = 3.354862 (0.047 sec/batch), lr: 0.003000
598
- 2025-12-03 23:28:38 INFO: Evaluating on dev set...
599
- 2025-12-03 23:28:39 INFO: LAS MLAS BLEX
600
- 2025-12-03 23:28:39 INFO: 52.48 39.58 44.63
601
- 2025-12-03 23:28:39 INFO: step 3900: train_loss = 3.766196, dev_score = 0.5248
602
- 2025-12-03 23:28:39 INFO: Training ended with 3900 steps.
603
- 2025-12-03 23:28:39 INFO: Best dev F1 = 53.71, at iteration = 2900
604
- 2025-12-03 23:28:39 INFO: Running dev depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
605
- 2025-12-03 23:28:39 INFO: Running parser in predict mode
606
- 2025-12-03 23:28:39 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
607
- 2025-12-03 23:28:41 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
608
- 2025-12-03 23:28:41 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
609
- 2025-12-03 23:28:41 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
610
- 2025-12-03 23:28:41 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
611
- 2025-12-03 23:28:41 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
612
- 2025-12-03 23:28:41 INFO: Loading data with batch size 32...
613
- 2025-12-03 23:28:41 DEBUG: 9 batches created.
614
- 2025-12-03 23:28:42 INFO: F1 scores for each dependency:
615
- Note that unlabeled attachment errors hurt the labeled attachment scores
616
- acl: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
617
- acl:relcl: p 0.1667 r 0.2857 f1 0.2105 (7 actual)
618
- advcl: p 0.0526 r 0.2000 f1 0.0833 (5 actual)
619
- advmod: p 0.4839 r 0.6000 f1 0.5357 (25 actual)
620
- amod: p 0.8148 r 0.7097 f1 0.7586 (31 actual)
621
- appos: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
622
- aux: p 0.8889 r 0.7273 f1 0.8000 (11 actual)
623
- case: p 0.7544 r 0.7679 f1 0.7611 (56 actual)
624
- cc: p 0.7692 r 0.7692 f1 0.7692 (13 actual)
625
- ccomp: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
626
- conj: p 0.5000 r 0.0833 f1 0.1429 (12 actual)
627
- cop: p 0.2500 r 0.3333 f1 0.2857 (3 actual)
628
- csubj: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
629
- det: p 0.8696 r 0.9091 f1 0.8889 (22 actual)
630
- expl: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
631
- iobj: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
632
- mark: p 0.5000 r 0.3333 f1 0.4000 (12 actual)
633
- nmod: p 0.2609 r 0.4000 f1 0.3158 (15 actual)
634
- nmod:poss: p 1.0000 r 0.8947 f1 0.9444 (19 actual)
635
- nsubj: p 0.3421 r 0.7647 f1 0.4727 (17 actual)
636
- nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
637
- obj: p 0.6667 r 0.2727 f1 0.3871 (22 actual)
638
- obl: p 0.3333 r 0.4146 f1 0.3696 (41 actual)
639
- obl:agent: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
640
- orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
641
- parataxis: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
642
- punct: p 0.4808 r 0.4808 f1 0.4808 (52 actual)
643
- root: p 0.4444 r 0.4444 f1 0.4444 (9 actual)
644
- xcomp: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
645
- 2025-12-03 23:28:42 INFO: LAS MLAS BLEX
646
- 2025-12-03 23:28:42 INFO: 53.71 40.59 44.82
647
- 2025-12-03 23:28:42 INFO: Parser score:
648
- 2025-12-03 23:28:42 INFO: sv_diachronic 53.71
649
- 2025-12-03 23:28:42 INFO: Finished running dev set on
650
- UD_Swedish-diachronic
651
- UAS LAS CLAS MLAS BLEX
652
- 66.58 53.71 44.82 40.59 44.82
653
- 2025-12-03 23:28:42 INFO: Running test depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
654
- 2025-12-03 23:28:42 INFO: Running parser in predict mode
655
- 2025-12-03 23:28:42 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
656
- 2025-12-03 23:28:44 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
657
- 2025-12-03 23:28:44 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
658
- 2025-12-03 23:28:44 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
659
- 2025-12-03 23:28:44 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
660
- 2025-12-03 23:28:44 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
661
- 2025-12-03 23:28:44 INFO: Loading data with batch size 32...
662
- 2025-12-03 23:28:44 DEBUG: 93 batches created.
663
- 2025-12-03 23:28:49 INFO: F1 scores for each dependency:
664
- Note that unlabeled attachment errors hurt the labeled attachment scores
665
- acl: p 0.0000 r 0.0000 f1 0.0000 (32 actual)
666
- acl:cleft: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
667
- acl:relcl: p 0.1167 r 0.0933 f1 0.1037 (75 actual)
668
- advcl: p 0.0544 r 0.2667 f1 0.0904 (60 actual)
669
- advcl:relcl: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
670
- advmod: p 0.5083 r 0.5746 f1 0.5394 (268 actual)
671
- amod: p 0.6795 r 0.6913 f1 0.6853 (230 actual)
672
- appos: p 0.0000 r 0.0000 f1 0.0000 (13 actual)
673
- aux: p 0.7556 r 0.8095 f1 0.7816 (84 actual)
674
- aux:pass: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
675
- case: p 0.7245 r 0.7051 f1 0.7147 (373 actual)
676
- cc: p 0.5759 r 0.5871 f1 0.5815 (155 actual)
677
- ccomp: p 0.0000 r 0.0000 f1 0.0000 (35 actual)
678
- compound:prt: p 0.0000 r 0.0000 f1 0.0000 (21 actual)
679
- conj: p 0.1053 r 0.0253 f1 0.0408 (158 actual)
680
- cop: p 0.7619 r 0.3478 f1 0.4776 (46 actual)
681
- csubj: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
682
- dep: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
683
- det: p 0.7833 r 0.7644 f1 0.7737 (208 actual)
684
- discourse: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
685
- dislocated: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
686
- expl: p 0.0000 r 0.0000 f1 0.0000 (11 actual)
687
- expl:pv: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
688
- fixed: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
689
- flat: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
690
- flat:name: p 0.0000 r 0.0000 f1 0.0000 (12 actual)
691
- goeswith: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
692
- iobj: p 0.0000 r 0.0000 f1 0.0000 (14 actual)
693
- mark: p 0.6569 r 0.5882 f1 0.6207 (153 actual)
694
- nmod: p 0.2320 r 0.2843 f1 0.2555 (102 actual)
695
- nmod:poss: p 0.8102 r 0.7817 f1 0.7957 (142 actual)
696
- nsubj: p 0.3947 r 0.6429 f1 0.4891 (280 actual)
697
- nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (25 actual)
698
- nummod: p 0.0000 r 0.0000 f1 0.0000 (10 actual)
699
- obj: p 0.6111 r 0.1803 f1 0.2785 (183 actual)
700
- obl: p 0.2623 r 0.4029 f1 0.3177 (278 actual)
701
- obl:agent: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
702
- orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
703
- parataxis: p 0.0000 r 0.0000 f1 0.0000 (18 actual)
704
- punct: p 0.3852 r 0.3906 f1 0.3879 (425 actual)
705
- reparandum: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
706
- root: p 0.5253 r 0.5253 f1 0.5253 (99 actual)
707
- vocative: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
708
- xcomp: p 0.0000 r 0.0000 f1 0.0000 (75 actual)
709
- 2025-12-03 23:28:49 INFO: LAS MLAS BLEX
710
- 2025-12-03 23:28:49 INFO: 47.60 36.14 39.67
711
- 2025-12-03 23:28:49 INFO: Parser score:
712
- 2025-12-03 23:28:49 INFO: sv_diachronic 47.60
713
- 2025-12-03 23:28:49 INFO: Finished running test set on
714
- UD_Swedish-diachronic
715
- UAS LAS CLAS MLAS BLEX
716
- 62.40 47.60 39.67 36.14 39.67
717
- DONE.
718
- Full log saved to: logs/log_conll17.pt_sv_20251203_232257.txt
719
- Symlink updated: logs/latest.txt → log_conll17.pt_sv_20251203_232257.txt
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_conll17.pt_sv_20260218_190034.txt DELETED
@@ -1,120 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_20260218_190034.txt ===
2
- Language codes: sv
3
- Using pretrained model: conll17.pt
4
-
5
- Running: python prepare-train-val-SILVER.py sv
6
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
7
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
8
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
9
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
10
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
11
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
12
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
13
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
14
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
15
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
16
- Loading DigPhil MACHINE for SILVER (minus gold overlap)…
17
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
18
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
19
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
20
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
21
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
22
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
23
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
24
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
25
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
26
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
27
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
28
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
29
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
30
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
31
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
32
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
33
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
34
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
35
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
36
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
37
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
38
- Adding 75 remaining gold sentences to TRAIN
39
- Cleaning TRAIN...
40
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
41
- Cleaning SILVER...
42
- [REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
43
- [REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
44
- [REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
45
- [REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
46
- [REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
47
- [REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
48
- [REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
49
- [REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
50
- [REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
51
- [REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
52
- [REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
53
- [REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
54
- [REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
55
- [REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
56
- [REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
57
- [REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
58
- [REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
59
- [REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
60
- [REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
61
- [REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
62
- [REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
63
- [REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
64
- [REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
65
- [REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
66
- [REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
67
- [REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
68
- [REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
69
- [REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
70
- [REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
71
- [REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
72
- [REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
73
- [REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
74
- [REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
75
- [REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
76
- [REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
77
- [REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
78
- [REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
79
- [REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
80
- [REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
81
- [REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
82
- [REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
83
- [REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
84
- [REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
85
- [REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
86
- [REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
87
- [REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
88
- [REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
89
- [REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
90
- [REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
91
- [REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
92
- [REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
93
- [REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
94
- [REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
95
- [REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
96
- [REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
97
- [REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
98
- [REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
99
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
100
- [REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
101
- [REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
102
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
103
- [REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
104
- [REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
105
- Cleaning DEV...
106
- Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13868 valid sentences)
107
- Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (46432 valid sentences)
108
- Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (34 valid sentences)
109
- Done.
110
- Sourcing scripts/config.sh
111
- Running stanza dataset preparation…
112
- 2026-02-18 19:00:38 INFO: Datasets program called with:
113
- /home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --use_bert --bert_model vesteinn/ScandiBERT --silver_file ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu
114
- usage: python -m stanza.utils.datasets.prepare_depparse_treebank
115
- [-h] [--gold] [--predicted]
116
- [--wordvec_pretrain_file WORDVEC_PRETRAIN_FILE]
117
- [--tagger_model TAGGER_MODEL] [--save_dir SAVE_DIR]
118
- [--no_download_tagger] [--charlm CHARLM] [--no_charlm]
119
- treebanks [treebanks ...]
120
- python -m stanza.utils.datasets.prepare_depparse_treebank: error: unrecognized arguments: --use_bert --bert_model vesteinn/ScandiBERT --silver_file ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_conll17.pt_sv_20260218_190242.txt DELETED
@@ -1,143 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_20260218_190242.txt ===
2
- Language codes: sv
3
- Using pretrained model: conll17.pt
4
-
5
- Running: python prepare-train-val-SILVER.py sv
6
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
7
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
8
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
9
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
10
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
11
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
12
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
13
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
14
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
15
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
16
- Loading DigPhil MACHINE for SILVER (minus gold overlap)…
17
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
18
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
19
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
20
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
21
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
22
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
23
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
24
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
25
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
26
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
27
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
28
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
29
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
30
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
31
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
32
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
33
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
34
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
35
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
36
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
37
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
38
- Adding 75 remaining gold sentences to TRAIN
39
- Cleaning TRAIN...
40
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
41
- Cleaning SILVER...
42
- [REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
43
- [REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
44
- [REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
45
- [REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
46
- [REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
47
- [REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
48
- [REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
49
- [REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
50
- [REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
51
- [REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
52
- [REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
53
- [REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
54
- [REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
55
- [REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
56
- [REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
57
- [REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
58
- [REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
59
- [REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
60
- [REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
61
- [REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
62
- [REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
63
- [REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
64
- [REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
65
- [REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
66
- [REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
67
- [REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
68
- [REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
69
- [REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
70
- [REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
71
- [REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
72
- [REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
73
- [REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
74
- [REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
75
- [REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
76
- [REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
77
- [REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
78
- [REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
79
- [REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
80
- [REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
81
- [REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
82
- [REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
83
- [REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
84
- [REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
85
- [REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
86
- [REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
87
- [REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
88
- [REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
89
- [REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
90
- [REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
91
- [REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
92
- [REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
93
- [REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
94
- [REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
95
- [REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
96
- [REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
97
- [REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
98
- [REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
99
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
100
- [REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
101
- [REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
102
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
103
- [REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
104
- [REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
105
- Cleaning DEV...
106
- Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13868 valid sentences)
107
- Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (46432 valid sentences)
108
- Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (34 valid sentences)
109
- Done.
110
- Sourcing scripts/config.sh
111
- Running stanza dataset preparation…
112
- 2026-02-18 19:02:45 INFO: Datasets program called with:
113
- /home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic
114
- 2026-02-18 19:02:45 DEBUG: Looking for tagger for lang |sv| dataset |diachronic|. Suggested model |None|. Looking first in |saved_models/pos|.
115
- 2026-02-18 19:02:45 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
116
-
117
- 2026-02-18 19:02:46 INFO: Downloaded file to /home/urdatorn/stanza_resources/resources.json
118
- 2026-02-18 19:02:46 DEBUG: Processing parameter "processors"...
119
- 2026-02-18 19:02:46 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
120
- 2026-02-18 19:02:46 INFO: Downloading these customized packages for language: sv (Swedish)...
121
- =======================
122
- | Processor | Package |
123
- -----------------------
124
- =======================
125
-
126
- 2026-02-18 19:02:46 INFO: Finished downloading models and saved to /home/urdatorn/stanza_resources
127
- Traceback (most recent call last):
128
- File "<frozen runpy>", line 198, in _run_module_as_main
129
- File "<frozen runpy>", line 88, in _run_code
130
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 151, in <module>
131
- main()
132
- ~~~~^^
133
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 147, in main
134
- common.main(process_treebank, common.ModelType.DEPPARSE, add_specific_args)
135
- ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
136
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 335, in main
137
- process_treebank(treebank, model_type, paths, args)
138
- ~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
139
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 109, in process_treebank
140
- tagger_model = choose_tagger_model(short_language, dataset, args.tagger_model, args)
141
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 90, in choose_tagger_model
142
- raise FileNotFoundError("Could not figure out which model file to use for %s. Just tried to download to %s the models %s" % (short_language, args.save_dir, download_list))
143
- FileNotFoundError: Could not figure out which model file to use for sv. Just tried to download to saved_models/pos the models []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_conll17.pt_sv_20260218_190400.txt DELETED
@@ -1,38 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_20260218_190400.txt ===
2
- Language codes: sv
3
- Using pretrained model: conll17.pt
4
-
5
- Sourcing scripts/config.sh
6
- Running stanza dataset preparation…
7
- 2026-02-18 19:04:02 INFO: Datasets program called with:
8
- /home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic
9
- 2026-02-18 19:04:02 DEBUG: Looking for tagger for lang |sv| dataset |diachronic|. Suggested model |None|. Looking first in |saved_models/pos|.
10
- 2026-02-18 19:04:02 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
11
-
12
- 2026-02-18 19:04:02 INFO: Downloaded file to /home/urdatorn/stanza_resources/resources.json
13
- 2026-02-18 19:04:02 DEBUG: Processing parameter "processors"...
14
- 2026-02-18 19:04:02 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
15
- 2026-02-18 19:04:02 INFO: Downloading these customized packages for language: sv (Swedish)...
16
- =======================
17
- | Processor | Package |
18
- -----------------------
19
- =======================
20
-
21
- 2026-02-18 19:04:02 INFO: Finished downloading models and saved to /home/urdatorn/stanza_resources
22
- Traceback (most recent call last):
23
- File "<frozen runpy>", line 198, in _run_module_as_main
24
- File "<frozen runpy>", line 88, in _run_code
25
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 151, in <module>
26
- main()
27
- ~~~~^^
28
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 147, in main
29
- common.main(process_treebank, common.ModelType.DEPPARSE, add_specific_args)
30
- ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
31
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 335, in main
32
- process_treebank(treebank, model_type, paths, args)
33
- ~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
34
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 109, in process_treebank
35
- tagger_model = choose_tagger_model(short_language, dataset, args.tagger_model, args)
36
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 90, in choose_tagger_model
37
- raise FileNotFoundError("Could not figure out which model file to use for %s. Just tried to download to %s the models %s" % (short_language, args.save_dir, download_list))
38
- FileNotFoundError: Could not figure out which model file to use for sv. Just tried to download to saved_models/pos the models []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_conll17.pt_sv_20260218_190808.txt DELETED
@@ -1,38 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_20260218_190808.txt ===
2
- Language codes: sv
3
- Using pretrained model: conll17.pt
4
-
5
- Sourcing scripts/config.sh
6
- Running stanza dataset preparation…
7
- 2026-02-18 19:08:10 INFO: Datasets program called with:
8
- /home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic
9
- 2026-02-18 19:08:10 DEBUG: Looking for tagger for lang |sv| dataset |diachronic|. Suggested model |None|. Looking first in |saved_models/pos|.
10
- 2026-02-18 19:08:10 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
11
-
12
- 2026-02-18 19:08:11 INFO: Downloaded file to /home/urdatorn/stanza_resources/resources.json
13
- 2026-02-18 19:08:11 DEBUG: Processing parameter "processors"...
14
- 2026-02-18 19:08:11 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
15
- 2026-02-18 19:08:11 INFO: Downloading these customized packages for language: sv (Swedish)...
16
- =======================
17
- | Processor | Package |
18
- -----------------------
19
- =======================
20
-
21
- 2026-02-18 19:08:11 INFO: Finished downloading models and saved to /home/urdatorn/stanza_resources
22
- Traceback (most recent call last):
23
- File "<frozen runpy>", line 198, in _run_module_as_main
24
- File "<frozen runpy>", line 88, in _run_code
25
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 151, in <module>
26
- main()
27
- ~~~~^^
28
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 147, in main
29
- common.main(process_treebank, common.ModelType.DEPPARSE, add_specific_args)
30
- ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
31
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 335, in main
32
- process_treebank(treebank, model_type, paths, args)
33
- ~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
34
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 109, in process_treebank
35
- tagger_model = choose_tagger_model(short_language, dataset, args.tagger_model, args)
36
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 90, in choose_tagger_model
37
- raise FileNotFoundError("Could not figure out which model file to use for %s. Just tried to download to %s the models %s" % (short_language, args.save_dir, download_list))
38
- FileNotFoundError: Could not figure out which model file to use for sv. Just tried to download to saved_models/pos the models []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_conll17.pt_sv_20260218_190956.txt DELETED
@@ -1,38 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_20260218_190956.txt ===
2
- Language codes: sv
3
- Using pretrained model: conll17.pt
4
-
5
- Sourcing scripts/config.sh
6
- Running stanza dataset preparation…
7
- 2026-02-18 19:09:58 INFO: Datasets program called with:
8
- /home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
9
- 2026-02-18 19:09:58 DEBUG: Looking for tagger for lang |sv| dataset |diachronic|. Suggested model |None|. Looking first in |saved_models/pos|.
10
- 2026-02-18 19:09:58 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
11
-
12
- 2026-02-18 19:09:58 INFO: Downloaded file to /home/urdatorn/stanza_resources/resources.json
13
- 2026-02-18 19:09:58 DEBUG: Processing parameter "processors"...
14
- 2026-02-18 19:09:58 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
15
- 2026-02-18 19:09:58 INFO: Downloading these customized packages for language: sv (Swedish)...
16
- =======================
17
- | Processor | Package |
18
- -----------------------
19
- =======================
20
-
21
- 2026-02-18 19:09:58 INFO: Finished downloading models and saved to /home/urdatorn/stanza_resources
22
- Traceback (most recent call last):
23
- File "<frozen runpy>", line 198, in _run_module_as_main
24
- File "<frozen runpy>", line 88, in _run_code
25
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 151, in <module>
26
- main()
27
- ~~~~^^
28
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 147, in main
29
- common.main(process_treebank, common.ModelType.DEPPARSE, add_specific_args)
30
- ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
31
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 335, in main
32
- process_treebank(treebank, model_type, paths, args)
33
- ~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
34
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 109, in process_treebank
35
- tagger_model = choose_tagger_model(short_language, dataset, args.tagger_model, args)
36
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 90, in choose_tagger_model
37
- raise FileNotFoundError("Could not figure out which model file to use for %s. Just tried to download to %s the models %s" % (short_language, args.save_dir, download_list))
38
- FileNotFoundError: Could not figure out which model file to use for sv. Just tried to download to saved_models/pos the models []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_conll17.pt_sv_20260218_191305.txt DELETED
@@ -1,38 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_20260218_191305.txt ===
2
- Language codes: sv
3
- Using pretrained model: conll17.pt
4
-
5
- Sourcing scripts/config.sh
6
- Running stanza dataset preparation…
7
- 2026-02-18 19:13:07 INFO: Datasets program called with:
8
- /home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
9
- 2026-02-18 19:13:07 DEBUG: Looking for tagger for lang |sv| dataset |diachronic|. Suggested model |None|. Looking first in |saved_models/pos|.
10
- 2026-02-18 19:13:07 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
11
-
12
- 2026-02-18 19:13:07 INFO: Downloaded file to /home/urdatorn/stanza_resources/resources.json
13
- 2026-02-18 19:13:07 DEBUG: Processing parameter "processors"...
14
- 2026-02-18 19:13:07 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
15
- 2026-02-18 19:13:07 INFO: Downloading these customized packages for language: sv (Swedish)...
16
- =======================
17
- | Processor | Package |
18
- -----------------------
19
- =======================
20
-
21
- 2026-02-18 19:13:07 INFO: Finished downloading models and saved to /home/urdatorn/stanza_resources
22
- Traceback (most recent call last):
23
- File "<frozen runpy>", line 198, in _run_module_as_main
24
- File "<frozen runpy>", line 88, in _run_code
25
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 151, in <module>
26
- main()
27
- ~~~~^^
28
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 147, in main
29
- common.main(process_treebank, common.ModelType.DEPPARSE, add_specific_args)
30
- ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
31
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 335, in main
32
- process_treebank(treebank, model_type, paths, args)
33
- ~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
34
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 109, in process_treebank
35
- tagger_model = choose_tagger_model(short_language, dataset, args.tagger_model, args)
36
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 90, in choose_tagger_model
37
- raise FileNotFoundError("Could not figure out which model file to use for %s. Just tried to download to %s the models %s" % (short_language, args.save_dir, download_list))
38
- FileNotFoundError: Could not figure out which model file to use for sv. Just tried to download to saved_models/pos the models []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_conll17.pt_sv_20260218_191719.txt DELETED
@@ -1,38 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_20260218_191719.txt ===
2
- Language codes: sv
3
- Using pretrained model: conll17.pt
4
-
5
- Sourcing scripts/config.sh
6
- Running stanza dataset preparation…
7
- 2026-02-18 19:17:21 INFO: Datasets program called with:
8
- /home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
9
- 2026-02-18 19:17:21 DEBUG: Looking for tagger for lang |sv| dataset |diachronic|. Suggested model |None|. Looking first in |saved_models/pos|.
10
- 2026-02-18 19:17:21 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
11
-
12
- 2026-02-18 19:17:21 INFO: Downloaded file to /home/urdatorn/stanza_resources/resources.json
13
- 2026-02-18 19:17:21 DEBUG: Processing parameter "processors"...
14
- 2026-02-18 19:17:21 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
15
- 2026-02-18 19:17:21 INFO: Downloading these customized packages for language: sv (Swedish)...
16
- =======================
17
- | Processor | Package |
18
- -----------------------
19
- =======================
20
-
21
- 2026-02-18 19:17:21 INFO: Finished downloading models and saved to /home/urdatorn/stanza_resources
22
- Traceback (most recent call last):
23
- File "<frozen runpy>", line 198, in _run_module_as_main
24
- File "<frozen runpy>", line 88, in _run_code
25
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 151, in <module>
26
- main()
27
- ~~~~^^
28
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 147, in main
29
- common.main(process_treebank, common.ModelType.DEPPARSE, add_specific_args)
30
- ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
31
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 335, in main
32
- process_treebank(treebank, model_type, paths, args)
33
- ~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
34
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 109, in process_treebank
35
- tagger_model = choose_tagger_model(short_language, dataset, args.tagger_model, args)
36
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 90, in choose_tagger_model
37
- raise FileNotFoundError("Could not figure out which model file to use for %s. Just tried to download to %s the models %s" % (short_language, args.save_dir, download_list))
38
- FileNotFoundError: Could not figure out which model file to use for sv. Just tried to download to saved_models/pos the models []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_conll17.pt_sv_20260219_104841.txt DELETED
@@ -1,44 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_20260219_104841.txt ===
2
- Language codes: sv
3
- Using pretrained model: conll17.pt
4
-
5
- Sourcing scripts/config.sh
6
- Running stanza dataset preparation…
7
- 2026-02-19 10:48:43 INFO: Datasets program called with:
8
- /home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt --gold
9
- Augmented 60 quotes: Counter({'《》': 9, '»«': 8, '″″': 8, '““': 7, '„”': 6, '""': 6, '„“': 6, '«»': 4, '””': 4, '「」': 2})
10
- Traceback (most recent call last):
11
- File "<frozen runpy>", line 198, in _run_module_as_main
12
- File "<frozen runpy>", line 88, in _run_code
13
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 151, in <module>
14
- main()
15
- ~~~~^^
16
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 147, in main
17
- common.main(process_treebank, common.ModelType.DEPPARSE, add_specific_args)
18
- ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
19
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 335, in main
20
- process_treebank(treebank, model_type, paths, args)
21
- ~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
22
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 96, in process_treebank
23
- prepare_tokenizer_treebank.copy_conllu_treebank(treebank, model_type, paths, paths["DEPPARSE_DATA_DIR"])
24
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
25
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_tokenizer_treebank.py", line 75, in copy_conllu_treebank
26
- process_treebank(treebank, model_type, paths, args)
27
- ~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
28
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_tokenizer_treebank.py", line 1499, in process_treebank
29
- process_ud_treebank(treebank, udbase_dir, tokenizer_dir, short_name, short_language, args.augment)
30
- ~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
31
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_tokenizer_treebank.py", line 1357, in process_ud_treebank
32
- prepare_ud_dataset(treebank, udbase_dir, tokenizer_dir, short_name, short_language, "test", augment)
33
- ~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
34
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_tokenizer_treebank.py", line 1332, in prepare_ud_dataset
35
- input_conllu = common.find_treebank_dataset_file(treebank, udbase_dir, dataset, "conllu", fail=True)
36
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 196, in find_treebank_dataset_file
37
- raise FileNotFoundError("Could not find any treebank files which matched {}\nIf you have the data elsewhere, you can change the base directory for the search by changing the {} environment variable".format(filename, env_var))
38
- FileNotFoundError: Could not find any treebank files which matched /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/*-ud-test.conllu
39
- If you have the data elsewhere, you can change the base directory for the search by changing the UDBASE environment variable
40
- Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
41
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /tmp/tmp468rp5x_/sv_diachronic.train.gold.conllu
42
- Swapped 'w1, w2' for 'w1 ,w2' 93 times
43
- Added 86 new sentences with asdf, zzzz -> asdf,zzzz
44
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /tmp/tmp468rp5x_/sv_diachronic.dev.gold.conllu
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_conll17.pt_sv_20260219_105228.txt DELETED
@@ -1,315 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_20260219_105228.txt ===
2
- Language codes: sv
3
- Using pretrained model: conll17.pt
4
-
5
- Running: python prepare-train-val-SILVER.py sv
6
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
7
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
8
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
9
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
10
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
11
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
12
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
13
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
14
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
15
- Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
16
- Loading DigPhil MACHINE for SILVER (minus gold overlap)…
17
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
18
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
19
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
20
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
21
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
22
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
23
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
24
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
25
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
26
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
27
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
28
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
29
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
30
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
31
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
32
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
33
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
34
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
35
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
36
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
37
- Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
38
- Adding 74 remaining gold sentences to TRAIN
39
- Cleaning TRAIN...
40
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
41
- Cleaning SILVER...
42
- [REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
43
- [REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
44
- [REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
45
- [REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
46
- [REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
47
- [REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
48
- [REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
49
- [REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
50
- [REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
51
- [REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
52
- [REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
53
- [REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
54
- [REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
55
- [REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
56
- [REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
57
- [REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
58
- [REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
59
- [REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
60
- [REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
61
- [REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
62
- [REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
63
- [REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
64
- [REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
65
- [REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
66
- [REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
67
- [REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
68
- [REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
69
- [REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
70
- [REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
71
- [REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
72
- [REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
73
- [REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
74
- [REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
75
- [REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
76
- [REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
77
- [REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
78
- [REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
79
- [REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
80
- [REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
81
- [REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
82
- [REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
83
- [REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
84
- [REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
85
- [REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
86
- [REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
87
- [REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
88
- [REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
89
- [REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
90
- [REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
91
- [REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
92
- [REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
93
- [REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
94
- [REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
95
- [REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
96
- [REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
97
- [REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
98
- [REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
99
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
100
- [REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
101
- [REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
102
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
103
- [REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
104
- [REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
105
- Cleaning DEV...
106
- Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13867 valid sentences)
107
- Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (46432 valid sentences)
108
- Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (35 valid sentences)
109
- Writing TEST → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (copy of dev, required by Stanza)
110
- Done.
111
- Sourcing scripts/config.sh
112
- Running stanza dataset preparation…
113
- 2026-02-19 10:52:31 INFO: Datasets program called with:
114
- /home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt --gold
115
- Augmented 60 quotes: Counter({'《》': 9, '»«': 8, '″″': 8, '““': 7, '„”': 6, '""': 6, '„“': 6, '«»': 4, '””': 4, '「」': 2})
116
- Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
117
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /tmp/tmpwrb749kh/sv_diachronic.train.gold.conllu
118
- Swapped 'w1, w2' for 'w1 ,w2' 98 times
119
- Added 86 new sentences with asdf, zzzz -> asdf,zzzz
120
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /tmp/tmpwrb749kh/sv_diachronic.dev.gold.conllu
121
- Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /tmp/tmpwrb749kh/sv_diachronic.test.gold.conllu
122
- Copying from /tmp/tmpwrb749kh/sv_diachronic.train.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
123
- Copying from /tmp/tmpwrb749kh/sv_diachronic.dev.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
124
- Copying from /tmp/tmpwrb749kh/sv_diachronic.test.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
125
- Running stanza dependency parser training…
126
- 2026-02-19 10:52:34 INFO: Training program called with:
127
- /home/urdatorn/git/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --batch_size 32 --dropout 0.33 --use_bert --bert_model vesteinn/ScandiBERT --silver_file ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu
128
- 2026-02-19 10:52:34 DEBUG: UD_Swedish-diachronic: sv_diachronic
129
- 2026-02-19 10:52:34 INFO: Using model /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
130
- 2026-02-19 10:52:34 INFO: Using model /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
131
- 2026-02-19 10:52:34 INFO: Using default pretrain for language sv, found in /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt To use a different pretrain, specify --wordvec_pretrain_file
132
- 2026-02-19 10:52:34 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_scandibert_parser.pt does not exist, training new model
133
- 2026-02-19 10:52:34 INFO: Using model /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
134
- 2026-02-19 10:52:34 INFO: Using model /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
135
- 2026-02-19 10:52:34 INFO: Using default pretrain for language sv, found in /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt To use a different pretrain, specify --wordvec_pretrain_file
136
- 2026-02-19 10:52:34 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/home/urdatorn/stanza_resources/sv/pretrain', '--train_file', '/home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--wordvec_pretrain_file', '/home/urdatorn/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt', '--batch_size', '32', '--dropout', '0.33', '--bert_model', 'vesteinn/ScandiBERT', '--silver_file', 'ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu']
137
- 2026-02-19 10:52:34 INFO: Running parser in train mode
138
- 2026-02-19 10:52:35 DEBUG: Loaded pretrain from /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
139
- 2026-02-19 10:52:35 INFO: Using 7 as the word cutoff based on the size of the pretrain (250000)
140
- 2026-02-19 10:52:35 INFO: Using pretrained contextualized char embedding
141
- 2026-02-19 10:52:35 INFO: ARGS USED AT TRAINING TIME:
142
- augment_nopunct: None
143
- batch_size: 32
144
- bert_finetune: False
145
- bert_finetune_layers: None
146
- bert_hidden_layers: 4
147
- bert_learning_rate: 1.0
148
- bert_model: vesteinn/ScandiBERT
149
- bert_start_finetuning: 200
150
- bert_warmup_steps: 200
151
- bert_weight_decay: 0.0
152
- beta2: 0.999
153
- char: True
154
- char_emb_dim: 100
155
- char_hidden_dim: 400
156
- char_num_layers: 1
157
- char_rec_dropout: 0
158
- charlm: True
159
- charlm_backward_file: /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
160
- charlm_forward_file: /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt
161
- charlm_save_dir: saved_models/charlm
162
- charlm_shorthand: sv_conll17
163
- checkpoint: True
164
- checkpoint_interval: 500
165
- checkpoint_save_name: None
166
- continue_from: None
167
- data_dir: data/depparse
168
- deep_biaff_hidden_dim: 400
169
- deep_biaff_output_dim: 160
170
- device: cpu
171
- distance: True
172
- dropout: 0.33
173
- eval_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
174
- eval_interval: 100
175
- gold_labels: True
176
- hidden_dim: 400
177
- lang: sv
178
- linearization: True
179
- log_norms: False
180
- log_step: 20
181
- lora_alpha: 128
182
- lora_dropout: 0.1
183
- lora_modules_to_save: []
184
- lora_rank: 64
185
- lora_target_modules: ['query', 'value', 'output.dense', 'intermediate.dense']
186
- lr: 2.0
187
- max_grad_norm: 1.0
188
- max_steps: 50000
189
- max_steps_before_stop: 2000
190
- mode: train
191
- model_type: graph
192
- num_layers: 3
193
- optim: adadelta
194
- output_file: None
195
- output_latex: False
196
- pretrain: True
197
- pretrain_max_vocab: 250000
198
- rec_dropout: 0
199
- reversed: False
200
- sample_train: 1.0
201
- save_dir: saved_models/depparse
202
- save_name: {shorthand}_{embedding}_parser.pt
203
- second_batch_size: None
204
- second_bert_learning_rate: 0.001
205
- second_lr: 0.0002
206
- second_optim: adam
207
- second_optim_start_step: 10000
208
- second_warmup_steps: 200
209
- seed: 1234
210
- shorthand: sv_diachronic
211
- silver_file: ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu
212
- silver_weight: 0.5
213
- tag_emb_dim: 50
214
- train_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
215
- train_size: None
216
- transformed_dim: 125
217
- transition_embedding_dim: 20
218
- transition_hidden_dim: 20
219
- transition_merge_hidden_dim: 200
220
- transition_subtree_combination: SubtreeCombination.NONE
221
- transition_subtree_nonlinearity: none
222
- use_arc_embedding: False
223
- use_peft: False
224
- use_ufeats: True
225
- use_upos: True
226
- use_xpos: True
227
- wandb: False
228
- wandb_name: None
229
- weight_decay: 1e-05
230
- word_cutoff: 7
231
- word_dropout: 0.33
232
- word_emb_dim: 75
233
- wordvec_dir: /home/urdatorn/stanza_resources/sv/pretrain
234
- wordvec_file: None
235
- wordvec_pretrain_file: /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
236
-
237
- 2026-02-19 10:52:35 INFO: Loading data with batch size 32...
238
- 2026-02-19 10:52:36 INFO: Train File /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 13953
239
- 2026-02-19 10:52:36 INFO: Original data size: 13953
240
- 2026-02-19 10:52:38 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
241
- 2026-02-19 10:52:38 INFO: Original length = 13953
242
- 2026-02-19 10:52:38 INFO: Filtered length = 13953
243
- 2026-02-19 10:52:40 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='-') for the xpos factory for sv_diachronic
244
- 2026-02-19 10:52:42 DEBUG: 9512 batches created.
245
- 2026-02-19 10:52:42 INFO: Augmented data size: 14711
246
- 2026-02-19 10:52:45 DEBUG: 10020 batches created.
247
- 2026-02-19 10:52:45 DEBUG: 34 batches created.
248
- 2026-02-19 10:52:45 DEBUG: 34 batches created.
249
- 2026-02-19 10:52:45 INFO: Training parser...
250
-
251
- XLMRobertaModel LOAD REPORT from: vesteinn/ScandiBERT
252
- Key | Status |
253
- --------------------------------+------------+-
254
- lm_head.bias | UNEXPECTED |
255
- lm_head.decoder.bias | UNEXPECTED |
256
- lm_head.layer_norm.bias | UNEXPECTED |
257
- lm_head.dense.weight | UNEXPECTED |
258
- lm_head.layer_norm.weight | UNEXPECTED |
259
- roberta.embeddings.position_ids | UNEXPECTED |
260
- lm_head.dense.bias | UNEXPECTED |
261
- pooler.dense.bias | MISSING |
262
- pooler.dense.weight | MISSING |
263
-
264
- Notes:
265
- - UNEXPECTED :can be ignored when loading from different task/architecture; not ok if you expect identical arch.
266
- - MISSING :those params were newly initialized because missing from the checkpoint. Consider training on your downstream task.
267
- Warning: You are sending unauthenticated requests to the HF Hub. Please set a HF_TOKEN to enable higher rate limits and faster downloads.
268
- Traceback (most recent call last):
269
- File "<frozen runpy>", line 198, in _run_module_as_main
270
- File "<frozen runpy>", line 88, in _run_code
271
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/training/run_depparse.py", line 145, in <module>
272
- main()
273
- ~~~~^^
274
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/training/run_depparse.py", line 142, in main
275
- common.main(run_treebank, "depparse", "parser", add_depparse_args, sub_argparse=parser.build_argparse(), build_model_filename=build_model_filename, choose_charlm_method=choose_depparse_charlm)
276
- ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
277
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/training/common.py", line 201, in main
278
- run_treebank(mode, paths, treebank, short_name, command_args, extra_args + save_name_args)
279
- ~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
280
- File "/home/urdatorn/git/stanza-digphil/stanza/utils/training/run_depparse.py", line 94, in run_treebank
281
- parser.main(train_args)
282
- ~~~~~~~~~~~^^^^^^^^^^^^
283
- File "/home/urdatorn/git/stanza-digphil/stanza/models/parser.py", line 266, in main
284
- return train(args)
285
- File "/home/urdatorn/git/stanza-digphil/stanza/models/parser.py", line 392, in train
286
- trainer = model_type(args=args, vocab=vocab, pretrain=pretrain, device=args['device'])
287
- File "/home/urdatorn/git/stanza-digphil/stanza/models/depparse/trainer.py", line 65, in __init__
288
- bert_model, bert_tokenizer = load_bert(self.args['bert_model'])
289
- ~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^
290
- File "/home/urdatorn/git/stanza-digphil/stanza/models/common/foundation_cache.py", line 115, in load_bert
291
- return bert_embedding.load_bert(model_name, local_files_only=local_files_only)
292
- ~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
293
- File "/home/urdatorn/git/stanza-digphil/stanza/models/common/bert_embedding.py", line 66, in load_bert
294
- bert_tokenizer = load_tokenizer(model_name, tokenizer_kwargs=tokenizer_kwargs, local_files_only=local_files_only)
295
- File "/home/urdatorn/git/stanza-digphil/stanza/models/common/bert_embedding.py", line 50, in load_tokenizer
296
- bert_tokenizer = AutoTokenizer.from_pretrained(model_name, **bert_args)
297
- File "/home/urdatorn/.pyenv/versions/venv/lib/python3.14/site-packages/transformers/models/auto/tokenization_auto.py", line 712, in from_pretrained
298
- return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs)
299
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
300
- File "/home/urdatorn/.pyenv/versions/venv/lib/python3.14/site-packages/transformers/tokenization_utils_base.py", line 1712, in from_pretrained
301
- return cls._from_pretrained(
302
- ~~~~~~~~~~~~~~~~~~~~^
303
- resolved_vocab_files,
304
- ^^^^^^^^^^^^^^^^^^^^^
305
- ...<9 lines>...
306
- **kwargs,
307
- ^^^^^^^^^
308
- )
309
- ^
310
- File "/home/urdatorn/.pyenv/versions/venv/lib/python3.14/site-packages/transformers/tokenization_utils_base.py", line 1897, in _from_pretrained
311
- init_kwargs = cls.convert_to_native_format(**init_kwargs)
312
- File "/home/urdatorn/.pyenv/versions/venv/lib/python3.14/site-packages/transformers/tokenization_utils_tokenizers.py", line 127, in convert_to_native_format
313
- if vocab and isinstance(vocab[0], (list, tuple)):
314
- ~~~~~^^^
315
- KeyError: 0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_conll17.pt_sv_20260219_105756.txt DELETED
The diff for this file is too large to render. See raw diff
 
logs/log_conll17.pt_sv_20260219_114115.txt DELETED
The diff for this file is too large to render. See raw diff
 
logs/log_conll17.pt_sv_diachron_20251203_223822.txt DELETED
@@ -1,731 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_diachron_20251203_223822.txt ===
2
- Language codes: sv diachron
3
- Using pretrained model: conll17.pt
4
-
5
- Running: python prepare-train-val-test.py sv diachron
6
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
7
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
8
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
9
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
10
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
11
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
12
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/ucxn_ud_swedish-talbanken.conllu
13
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
14
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
15
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
16
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
17
- Including DigPhil MACHINE in TRAIN (minus gold)…
18
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
19
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
20
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
21
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
22
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
23
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
24
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
25
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
26
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
27
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
28
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
29
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
30
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
31
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
32
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
33
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
34
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
35
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
36
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
37
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
38
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
39
- Cleaning TRAIN...
40
- [REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
41
- [REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
42
- [REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
43
- [REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
44
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
45
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
46
- [REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
47
- [REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
48
- [REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
49
- [REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
50
- [REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
51
- [REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
52
- [REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
53
- [REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
54
- [REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
55
- [REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
56
- [REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
57
- [REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
58
- [REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
59
- [REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
60
- [REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
61
- [REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
62
- [REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
63
- [REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
64
- [REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
65
- [REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
66
- [REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
67
- [REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
68
- [REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
69
- [REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
70
- [REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
71
- [REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
72
- [REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
73
- [REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
74
- [REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
75
- [REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
76
- [REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
77
- [REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
78
- [REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
79
- [REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
80
- [REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
81
- [REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
82
- [REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
83
- [REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
84
- [REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
85
- [REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
86
- [REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
87
- [REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
88
- [REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
89
- [REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
90
- [REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
91
- [REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
92
- [REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
93
- [REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
94
- [REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
95
- [REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
96
- [REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
97
- [REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
98
- [REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
99
- [REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
100
- [REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
101
- [REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
102
- [REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
103
- Cleaning DEV...
104
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
105
- Cleaning TEST...
106
- Writing TRAIN → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (66252 valid sentences)
107
- Writing DEV → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (9 valid sentences)
108
- Writing TEST → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (99 valid sentences)
109
- Done.
110
- Sourcing scripts/config_alvis.sh
111
- Running stanza dataset preparation…
112
- 2025-12-03 22:38:31 INFO: Datasets program called with:
113
- /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
114
- 2025-12-03 22:38:31 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
115
-
116
- 2025-12-03 22:38:31 INFO: Downloaded file to /cephyr/users/cleland/Alvis/stanza_resources/resources.json
117
- 2025-12-03 22:38:31 DEBUG: Processing parameter "processors"...
118
- 2025-12-03 22:38:31 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
119
- 2025-12-03 22:38:31 INFO: Downloading these customized packages for language: sv (Swedish)...
120
- =======================
121
- | Processor | Package |
122
- -----------------------
123
- =======================
124
-
125
- 2025-12-03 22:38:31 INFO: Finished downloading models and saved to /cephyr/users/cleland/Alvis/stanza_resources
126
- 2025-12-03 22:38:31 INFO: Using tagger model in /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt for sv_diachronic
127
- 2025-12-03 22:38:31 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
128
- 2025-12-03 22:38:31 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
129
- Augmented 192 quotes: Counter({'„”': 28, '""': 27, '「」': 20, '″″': 20, '»«': 18, '«»': 18, '《》': 17, '””': 17, '„“': 15, '““': 12})
130
- 2025-12-03 22:38:34 INFO: Running tagger to retag /local/tmp.5441282/tmpi3ipyceb/sv_diachronic.train.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
131
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5441282/tmpi3ipyceb/sv_diachronic.train.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu']
132
- 2025-12-03 22:38:34 INFO: Running tagger in predict mode
133
- 2025-12-03 22:38:34 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
134
- 2025-12-03 22:38:36 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
135
- 2025-12-03 22:38:36 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
136
- 2025-12-03 22:38:36 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
137
- 2025-12-03 22:38:36 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
138
- 2025-12-03 22:38:37 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
139
- 2025-12-03 22:38:39 INFO: Loading data with batch size 250...
140
- 2025-12-03 22:39:18 INFO: Start evaluation...
141
- 2025-12-03 22:42:53 INFO: UPOS XPOS UFeats AllTags
142
- 2025-12-03 22:42:53 INFO: 99.37 88.39 98.13 87.63
143
- 2025-12-03 22:42:53 INFO: POS Tagger score: sv_diachronic 87.63
144
- 2025-12-03 22:42:53 INFO: Running tagger to retag /local/tmp.5441282/tmpi3ipyceb/sv_diachronic.dev.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
145
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5441282/tmpi3ipyceb/sv_diachronic.dev.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu']
146
- 2025-12-03 22:42:53 INFO: Running tagger in predict mode
147
- 2025-12-03 22:42:53 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
148
- 2025-12-03 22:42:55 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
149
- 2025-12-03 22:42:55 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
150
- 2025-12-03 22:42:55 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
151
- 2025-12-03 22:42:55 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
152
- 2025-12-03 22:42:55 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
153
- 2025-12-03 22:42:55 INFO: Loading data with batch size 250...
154
- 2025-12-03 22:42:55 INFO: Start evaluation...
155
- 2025-12-03 22:42:56 INFO: UPOS XPOS UFeats AllTags
156
- 2025-12-03 22:42:56 INFO: 93.32 90.84 93.32 85.64
157
- 2025-12-03 22:42:56 INFO: POS Tagger score: sv_diachronic 85.64
158
- 2025-12-03 22:42:56 INFO: Running tagger to retag /local/tmp.5441282/tmpi3ipyceb/sv_diachronic.test.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
159
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5441282/tmpi3ipyceb/sv_diachronic.test.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu']
160
- 2025-12-03 22:42:56 INFO: Running tagger in predict mode
161
- 2025-12-03 22:42:56 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
162
- 2025-12-03 22:42:57 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
163
- 2025-12-03 22:42:57 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
164
- 2025-12-03 22:42:57 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
165
- 2025-12-03 22:42:57 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
166
- 2025-12-03 22:42:58 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
167
- 2025-12-03 22:42:58 INFO: Loading data with batch size 250...
168
- 2025-12-03 22:42:58 INFO: Start evaluation...
169
- 2025-12-03 22:42:58 INFO: UPOS XPOS UFeats AllTags
170
- 2025-12-03 22:42:58 INFO: 93.14 96.78 95.32 90.28
171
- 2025-12-03 22:42:58 INFO: POS Tagger score: sv_diachronic 90.28
172
- Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
173
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /local/tmp.5441282/tmpi3ipyceb/sv_diachronic.train.gold.conllu
174
- Swapped 'w1, w2' for 'w1 ,w2' 132 times
175
- Added 506 new sentences with asdf, zzzz -> asdf,zzzz
176
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /local/tmp.5441282/tmpi3ipyceb/sv_diachronic.dev.gold.conllu
177
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /local/tmp.5441282/tmpi3ipyceb/sv_diachronic.test.gold.conllu
178
- Running stanza dependency parser training…
179
- 2025-12-03 22:43:12 INFO: Training program called with:
180
- /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt --batch_size 32 --dropout 0.33
181
- 2025-12-03 22:43:12 DEBUG: UD_Swedish-diachronic: sv_diachronic
182
- 2025-12-03 22:43:12 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
183
- 2025-12-03 22:43:12 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
184
- 2025-12-03 22:43:12 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_charlm_parser.pt does not exist, training new model
185
- 2025-12-03 22:43:12 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
186
- 2025-12-03 22:43:12 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
187
- 2025-12-03 22:43:12 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--train_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
188
- 2025-12-03 22:43:12 INFO: Running parser in train mode
189
- 2025-12-03 22:43:12 INFO: Using pretrained contextualized char embedding
190
- 2025-12-03 22:43:12 INFO: Loading data with batch size 32...
191
- 2025-12-03 22:43:20 INFO: Train File /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 66758
192
- 2025-12-03 22:43:20 INFO: Original data size: 66758
193
- 2025-12-03 22:43:21 INFO: Augmented data size: 66875
194
- 2025-12-03 22:43:38 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
195
- 2025-12-03 22:43:38 INFO: Original length = 66875
196
- 2025-12-03 22:43:38 INFO: Filtered length = 66875
197
- 2025-12-03 22:43:55 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='|') for the xpos factory for sv_diachronic
198
- 2025-12-03 22:44:01 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
199
- 2025-12-03 22:44:15 DEBUG: 38839 batches created.
200
- 2025-12-03 22:44:15 DEBUG: 9 batches created.
201
- 2025-12-03 22:44:15 INFO: Training parser...
202
- 2025-12-03 22:44:15 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
203
- 2025-12-03 22:44:15 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
204
- 2025-12-03 22:44:15 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
205
- 2025-12-03 22:44:16 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
206
- 2025-12-03 22:44:24 INFO: Finished STEP 20/50000, loss = 6.296603 (0.201 sec/batch), lr: 0.003000
207
- 2025-12-03 22:44:27 INFO: Finished STEP 40/50000, loss = 5.081994 (0.181 sec/batch), lr: 0.003000
208
- 2025-12-03 22:44:31 INFO: Finished STEP 60/50000, loss = 4.866196 (0.144 sec/batch), lr: 0.003000
209
- 2025-12-03 22:44:34 INFO: Finished STEP 80/50000, loss = 4.271401 (0.138 sec/batch), lr: 0.003000
210
- 2025-12-03 22:44:36 INFO: Finished STEP 100/50000, loss = 4.687382 (0.143 sec/batch), lr: 0.003000
211
- 2025-12-03 22:44:36 INFO: Evaluating on dev set...
212
- 2025-12-03 22:44:37 INFO: LAS MLAS BLEX
213
- 2025-12-03 22:44:37 INFO: 43.56 33.47 35.15
214
- 2025-12-03 22:44:37 INFO: step 100: train_loss = 7.396650, dev_score = 0.4356
215
- 2025-12-03 22:44:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
216
- 2025-12-03 22:44:37 INFO: new best model saved.
217
- 2025-12-03 22:44:38 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
218
- 2025-12-03 22:44:38 INFO: new model checkpoint saved.
219
- 2025-12-03 22:44:41 INFO: Finished STEP 120/50000, loss = 3.784610 (0.125 sec/batch), lr: 0.003000
220
- 2025-12-03 22:44:43 INFO: Finished STEP 140/50000, loss = 4.674234 (0.131 sec/batch), lr: 0.003000
221
- 2025-12-03 22:44:46 INFO: Finished STEP 160/50000, loss = 4.012815 (0.137 sec/batch), lr: 0.003000
222
- 2025-12-03 22:44:48 INFO: Finished STEP 180/50000, loss = 4.966728 (0.122 sec/batch), lr: 0.003000
223
- 2025-12-03 22:44:51 INFO: Finished STEP 200/50000, loss = 4.140534 (0.121 sec/batch), lr: 0.003000
224
- 2025-12-03 22:44:51 INFO: Evaluating on dev set...
225
- 2025-12-03 22:44:51 INFO: LAS MLAS BLEX
226
- 2025-12-03 22:44:51 INFO: 51.98 38.30 43.83
227
- 2025-12-03 22:44:51 INFO: step 200: train_loss = 4.683159, dev_score = 0.5198
228
- 2025-12-03 22:44:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
229
- 2025-12-03 22:44:51 INFO: new best model saved.
230
- 2025-12-03 22:44:52 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
231
- 2025-12-03 22:44:52 INFO: new model checkpoint saved.
232
- 2025-12-03 22:44:55 INFO: Finished STEP 220/50000, loss = 3.998342 (0.126 sec/batch), lr: 0.003000
233
- 2025-12-03 22:44:57 INFO: Finished STEP 240/50000, loss = 5.531576 (0.124 sec/batch), lr: 0.003000
234
- 2025-12-03 22:44:59 INFO: Finished STEP 260/50000, loss = 4.806288 (0.116 sec/batch), lr: 0.003000
235
- 2025-12-03 22:45:02 INFO: Finished STEP 280/50000, loss = 4.965835 (0.111 sec/batch), lr: 0.003000
236
- 2025-12-03 22:45:04 INFO: Finished STEP 300/50000, loss = 3.812420 (0.115 sec/batch), lr: 0.003000
237
- 2025-12-03 22:45:04 INFO: Evaluating on dev set...
238
- 2025-12-03 22:45:04 INFO: LAS MLAS BLEX
239
- 2025-12-03 22:45:04 INFO: 55.45 44.40 49.47
240
- 2025-12-03 22:45:04 INFO: step 300: train_loss = 4.276576, dev_score = 0.5545
241
- 2025-12-03 22:45:05 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
242
- 2025-12-03 22:45:05 INFO: new best model saved.
243
- 2025-12-03 22:45:05 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
244
- 2025-12-03 22:45:05 INFO: new model checkpoint saved.
245
- 2025-12-03 22:45:07 INFO: Finished STEP 320/50000, loss = 3.429461 (0.111 sec/batch), lr: 0.003000
246
- 2025-12-03 22:45:10 INFO: Finished STEP 340/50000, loss = 3.642277 (0.118 sec/batch), lr: 0.003000
247
- 2025-12-03 22:45:12 INFO: Finished STEP 360/50000, loss = 3.760803 (0.116 sec/batch), lr: 0.003000
248
- 2025-12-03 22:45:14 INFO: Finished STEP 380/50000, loss = 4.786219 (0.105 sec/batch), lr: 0.003000
249
- 2025-12-03 22:45:16 INFO: Finished STEP 400/50000, loss = 2.972147 (0.108 sec/batch), lr: 0.003000
250
- 2025-12-03 22:45:16 INFO: Evaluating on dev set...
251
- 2025-12-03 22:45:17 INFO: LAS MLAS BLEX
252
- 2025-12-03 22:45:17 INFO: 57.43 46.61 50.85
253
- 2025-12-03 22:45:17 INFO: step 400: train_loss = 3.942294, dev_score = 0.5743
254
- 2025-12-03 22:45:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
255
- 2025-12-03 22:45:17 INFO: new best model saved.
256
- 2025-12-03 22:45:18 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
257
- 2025-12-03 22:45:18 INFO: new model checkpoint saved.
258
- 2025-12-03 22:45:20 INFO: Finished STEP 420/50000, loss = 4.565236 (0.105 sec/batch), lr: 0.003000
259
- 2025-12-03 22:45:22 INFO: Finished STEP 440/50000, loss = 4.531408 (0.098 sec/batch), lr: 0.003000
260
- 2025-12-03 22:45:24 INFO: Finished STEP 460/50000, loss = 5.082150 (0.090 sec/batch), lr: 0.003000
261
- 2025-12-03 22:45:26 INFO: Finished STEP 480/50000, loss = 2.923217 (0.110 sec/batch), lr: 0.003000
262
- 2025-12-03 22:45:28 INFO: Finished STEP 500/50000, loss = 3.408716 (0.100 sec/batch), lr: 0.003000
263
- 2025-12-03 22:45:28 INFO: Evaluating on dev set...
264
- 2025-12-03 22:45:28 INFO: LAS MLAS BLEX
265
- 2025-12-03 22:45:28 INFO: 60.89 51.48 55.27
266
- 2025-12-03 22:45:28 INFO: step 500: train_loss = 4.009553, dev_score = 0.6089
267
- 2025-12-03 22:45:29 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
268
- 2025-12-03 22:45:29 INFO: new best model saved.
269
- 2025-12-03 22:45:29 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
270
- 2025-12-03 22:45:29 INFO: new model checkpoint saved.
271
- 2025-12-03 22:45:31 INFO: Finished STEP 520/50000, loss = 5.306584 (0.101 sec/batch), lr: 0.003000
272
- 2025-12-03 22:45:33 INFO: Finished STEP 540/50000, loss = 3.657916 (0.100 sec/batch), lr: 0.003000
273
- 2025-12-03 22:45:35 INFO: Finished STEP 560/50000, loss = 4.001961 (0.086 sec/batch), lr: 0.003000
274
- 2025-12-03 22:45:37 INFO: Finished STEP 580/50000, loss = 4.052956 (0.085 sec/batch), lr: 0.003000
275
- 2025-12-03 22:45:39 INFO: Finished STEP 600/50000, loss = 3.852856 (0.099 sec/batch), lr: 0.003000
276
- 2025-12-03 22:45:39 INFO: Evaluating on dev set...
277
- 2025-12-03 22:45:40 INFO: LAS MLAS BLEX
278
- 2025-12-03 22:45:40 INFO: 59.90 50.00 55.51
279
- 2025-12-03 22:45:40 INFO: step 600: train_loss = 3.724004, dev_score = 0.5990
280
- 2025-12-03 22:45:40 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
281
- 2025-12-03 22:45:40 INFO: new model checkpoint saved.
282
- 2025-12-03 22:45:42 INFO: Finished STEP 620/50000, loss = 3.484413 (0.101 sec/batch), lr: 0.003000
283
- 2025-12-03 22:45:44 INFO: Finished STEP 640/50000, loss = 3.220206 (0.102 sec/batch), lr: 0.003000
284
- 2025-12-03 22:45:46 INFO: Finished STEP 660/50000, loss = 3.800846 (0.094 sec/batch), lr: 0.003000
285
- 2025-12-03 22:45:48 INFO: Finished STEP 680/50000, loss = 3.769488 (0.093 sec/batch), lr: 0.003000
286
- 2025-12-03 22:45:50 INFO: Finished STEP 700/50000, loss = 3.285978 (0.097 sec/batch), lr: 0.003000
287
- 2025-12-03 22:45:50 INFO: Evaluating on dev set...
288
- 2025-12-03 22:45:50 INFO: LAS MLAS BLEX
289
- 2025-12-03 22:45:50 INFO: 57.92 46.93 53.70
290
- 2025-12-03 22:45:50 INFO: step 700: train_loss = 3.829402, dev_score = 0.5792
291
- 2025-12-03 22:45:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
292
- 2025-12-03 22:45:51 INFO: new model checkpoint saved.
293
- 2025-12-03 22:45:53 INFO: Finished STEP 720/50000, loss = 4.905636 (0.100 sec/batch), lr: 0.003000
294
- 2025-12-03 22:45:55 INFO: Finished STEP 740/50000, loss = 3.337056 (0.102 sec/batch), lr: 0.003000
295
- 2025-12-03 22:45:57 INFO: Finished STEP 760/50000, loss = 5.074053 (0.092 sec/batch), lr: 0.003000
296
- 2025-12-03 22:45:59 INFO: Finished STEP 780/50000, loss = 4.024767 (0.089 sec/batch), lr: 0.003000
297
- 2025-12-03 22:46:00 INFO: Finished STEP 800/50000, loss = 4.552472 (0.091 sec/batch), lr: 0.003000
298
- 2025-12-03 22:46:00 INFO: Evaluating on dev set...
299
- 2025-12-03 22:46:01 INFO: LAS MLAS BLEX
300
- 2025-12-03 22:46:01 INFO: 61.39 49.05 53.70
301
- 2025-12-03 22:46:01 INFO: step 800: train_loss = 3.612613, dev_score = 0.6139
302
- 2025-12-03 22:46:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
303
- 2025-12-03 22:46:01 INFO: new best model saved.
304
- 2025-12-03 22:46:02 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
305
- 2025-12-03 22:46:02 INFO: new model checkpoint saved.
306
- 2025-12-03 22:46:04 INFO: Finished STEP 820/50000, loss = 6.260066 (0.090 sec/batch), lr: 0.003000
307
- 2025-12-03 22:46:05 INFO: Finished STEP 840/50000, loss = 4.288107 (0.092 sec/batch), lr: 0.003000
308
- 2025-12-03 22:46:07 INFO: Finished STEP 860/50000, loss = 3.849105 (0.097 sec/batch), lr: 0.003000
309
- 2025-12-03 22:46:09 INFO: Finished STEP 880/50000, loss = 2.492743 (0.088 sec/batch), lr: 0.003000
310
- 2025-12-03 22:46:11 INFO: Finished STEP 900/50000, loss = 4.095746 (0.090 sec/batch), lr: 0.003000
311
- 2025-12-03 22:46:11 INFO: Evaluating on dev set...
312
- 2025-12-03 22:46:11 INFO: LAS MLAS BLEX
313
- 2025-12-03 22:46:11 INFO: 60.40 51.68 55.04
314
- 2025-12-03 22:46:11 INFO: step 900: train_loss = 3.946219, dev_score = 0.6040
315
- 2025-12-03 22:46:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
316
- 2025-12-03 22:46:12 INFO: new model checkpoint saved.
317
- 2025-12-03 22:46:14 INFO: Finished STEP 920/50000, loss = 2.770604 (0.084 sec/batch), lr: 0.003000
318
- 2025-12-03 22:46:16 INFO: Finished STEP 940/50000, loss = 3.612128 (0.100 sec/batch), lr: 0.003000
319
- 2025-12-03 22:46:17 INFO: Finished STEP 960/50000, loss = 4.249920 (0.089 sec/batch), lr: 0.003000
320
- 2025-12-03 22:46:19 INFO: Finished STEP 980/50000, loss = 2.418294 (0.091 sec/batch), lr: 0.003000
321
- 2025-12-03 22:46:21 INFO: Finished STEP 1000/50000, loss = 4.842584 (0.094 sec/batch), lr: 0.003000
322
- 2025-12-03 22:46:21 INFO: Evaluating on dev set...
323
- 2025-12-03 22:46:21 INFO: LAS MLAS BLEX
324
- 2025-12-03 22:46:21 INFO: 61.39 53.70 56.66
325
- 2025-12-03 22:46:21 INFO: step 1000: train_loss = 3.696227, dev_score = 0.6139
326
- 2025-12-03 22:46:22 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
327
- 2025-12-03 22:46:22 INFO: new best model saved.
328
- 2025-12-03 22:46:22 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
329
- 2025-12-03 22:46:22 INFO: new model checkpoint saved.
330
- 2025-12-03 22:46:24 INFO: Finished STEP 1020/50000, loss = 3.346077 (0.088 sec/batch), lr: 0.003000
331
- 2025-12-03 22:46:26 INFO: Finished STEP 1040/50000, loss = 3.296835 (0.087 sec/batch), lr: 0.003000
332
- 2025-12-03 22:46:28 INFO: Finished STEP 1060/50000, loss = 2.515590 (0.088 sec/batch), lr: 0.003000
333
- 2025-12-03 22:46:29 INFO: Finished STEP 1080/50000, loss = 3.913376 (0.075 sec/batch), lr: 0.003000
334
- 2025-12-03 22:46:31 INFO: Finished STEP 1100/50000, loss = 5.241524 (0.077 sec/batch), lr: 0.003000
335
- 2025-12-03 22:46:31 INFO: Evaluating on dev set...
336
- 2025-12-03 22:46:31 INFO: LAS MLAS BLEX
337
- 2025-12-03 22:46:31 INFO: 59.65 49.47 53.28
338
- 2025-12-03 22:46:31 INFO: step 1100: train_loss = 3.721023, dev_score = 0.5965
339
- 2025-12-03 22:46:32 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
340
- 2025-12-03 22:46:32 INFO: new model checkpoint saved.
341
- 2025-12-03 22:46:34 INFO: Finished STEP 1120/50000, loss = 2.731287 (0.086 sec/batch), lr: 0.003000
342
- 2025-12-03 22:46:36 INFO: Finished STEP 1140/50000, loss = 3.536034 (0.080 sec/batch), lr: 0.003000
343
- 2025-12-03 22:46:37 INFO: Finished STEP 1160/50000, loss = 3.398331 (0.079 sec/batch), lr: 0.003000
344
- 2025-12-03 22:46:39 INFO: Finished STEP 1180/50000, loss = 5.028436 (0.078 sec/batch), lr: 0.003000
345
- 2025-12-03 22:46:41 INFO: Finished STEP 1200/50000, loss = 3.061586 (0.077 sec/batch), lr: 0.003000
346
- 2025-12-03 22:46:41 INFO: Evaluating on dev set...
347
- 2025-12-03 22:46:41 INFO: LAS MLAS BLEX
348
- 2025-12-03 22:46:41 INFO: 61.14 53.28 56.66
349
- 2025-12-03 22:46:41 INFO: step 1200: train_loss = 3.522418, dev_score = 0.6114
350
- 2025-12-03 22:46:42 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
351
- 2025-12-03 22:46:42 INFO: new model checkpoint saved.
352
- 2025-12-03 22:46:43 INFO: Finished STEP 1220/50000, loss = 2.513215 (0.089 sec/batch), lr: 0.003000
353
- 2025-12-03 22:46:45 INFO: Finished STEP 1240/50000, loss = 4.864305 (0.085 sec/batch), lr: 0.003000
354
- 2025-12-03 22:46:47 INFO: Finished STEP 1260/50000, loss = 7.466803 (0.082 sec/batch), lr: 0.003000
355
- 2025-12-03 22:46:49 INFO: Finished STEP 1280/50000, loss = 3.586653 (0.086 sec/batch), lr: 0.003000
356
- 2025-12-03 22:46:50 INFO: Finished STEP 1300/50000, loss = 3.791464 (0.078 sec/batch), lr: 0.003000
357
- 2025-12-03 22:46:50 INFO: Evaluating on dev set...
358
- 2025-12-03 22:46:51 INFO: LAS MLAS BLEX
359
- 2025-12-03 22:46:51 INFO: 61.63 52.01 55.81
360
- 2025-12-03 22:46:51 INFO: step 1300: train_loss = 3.743181, dev_score = 0.6163
361
- 2025-12-03 22:46:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
362
- 2025-12-03 22:46:51 INFO: new best model saved.
363
- 2025-12-03 22:46:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
364
- 2025-12-03 22:46:51 INFO: new model checkpoint saved.
365
- 2025-12-03 22:46:53 INFO: Finished STEP 1320/50000, loss = 2.740164 (0.083 sec/batch), lr: 0.003000
366
- 2025-12-03 22:46:55 INFO: Finished STEP 1340/50000, loss = 4.084477 (0.088 sec/batch), lr: 0.003000
367
- 2025-12-03 22:46:56 INFO: Finished STEP 1360/50000, loss = 4.818987 (0.070 sec/batch), lr: 0.003000
368
- 2025-12-03 22:46:58 INFO: Finished STEP 1380/50000, loss = 2.790164 (0.076 sec/batch), lr: 0.003000
369
- 2025-12-03 22:47:00 INFO: Finished STEP 1400/50000, loss = 4.602959 (0.081 sec/batch), lr: 0.003000
370
- 2025-12-03 22:47:00 INFO: Evaluating on dev set...
371
- 2025-12-03 22:47:00 INFO: LAS MLAS BLEX
372
- 2025-12-03 22:47:00 INFO: 62.13 55.08 58.05
373
- 2025-12-03 22:47:00 INFO: step 1400: train_loss = 3.724900, dev_score = 0.6213
374
- 2025-12-03 22:47:00 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
375
- 2025-12-03 22:47:00 INFO: new best model saved.
376
- 2025-12-03 22:47:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
377
- 2025-12-03 22:47:01 INFO: new model checkpoint saved.
378
- 2025-12-03 22:47:03 INFO: Finished STEP 1420/50000, loss = 4.015442 (0.079 sec/batch), lr: 0.003000
379
- 2025-12-03 22:47:04 INFO: Finished STEP 1440/50000, loss = 3.268615 (0.082 sec/batch), lr: 0.003000
380
- 2025-12-03 22:47:06 INFO: Finished STEP 1460/50000, loss = 3.735999 (0.080 sec/batch), lr: 0.003000
381
- 2025-12-03 22:47:07 INFO: Finished STEP 1480/50000, loss = 3.804015 (0.073 sec/batch), lr: 0.003000
382
- 2025-12-03 22:47:09 INFO: Finished STEP 1500/50000, loss = 4.517982 (0.087 sec/batch), lr: 0.003000
383
- 2025-12-03 22:47:09 INFO: Evaluating on dev set...
384
- 2025-12-03 22:47:10 INFO: LAS MLAS BLEX
385
- 2025-12-03 22:47:10 INFO: 61.63 52.54 56.36
386
- 2025-12-03 22:47:10 INFO: step 1500: train_loss = 3.783271, dev_score = 0.6163
387
- 2025-12-03 22:47:10 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
388
- 2025-12-03 22:47:10 INFO: new model checkpoint saved.
389
- 2025-12-03 22:47:12 INFO: Finished STEP 1520/50000, loss = 3.094217 (0.085 sec/batch), lr: 0.003000
390
- 2025-12-03 22:47:13 INFO: Finished STEP 1540/50000, loss = 3.673062 (0.073 sec/batch), lr: 0.003000
391
- 2025-12-03 22:47:15 INFO: Finished STEP 1560/50000, loss = 3.584330 (0.078 sec/batch), lr: 0.003000
392
- 2025-12-03 22:47:16 INFO: Finished STEP 1580/50000, loss = 3.714322 (0.078 sec/batch), lr: 0.003000
393
- 2025-12-03 22:47:18 INFO: Finished STEP 1600/50000, loss = 3.680001 (0.070 sec/batch), lr: 0.003000
394
- 2025-12-03 22:47:18 INFO: Evaluating on dev set...
395
- 2025-12-03 22:47:18 INFO: LAS MLAS BLEX
396
- 2025-12-03 22:47:18 INFO: 60.89 49.47 54.55
397
- 2025-12-03 22:47:18 INFO: step 1600: train_loss = 3.629141, dev_score = 0.6089
398
- 2025-12-03 22:47:19 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
399
- 2025-12-03 22:47:19 INFO: new model checkpoint saved.
400
- 2025-12-03 22:47:21 INFO: Finished STEP 1620/50000, loss = 4.405625 (0.076 sec/batch), lr: 0.003000
401
- 2025-12-03 22:47:22 INFO: Finished STEP 1640/50000, loss = 6.583941 (0.084 sec/batch), lr: 0.003000
402
- 2025-12-03 22:47:24 INFO: Finished STEP 1660/50000, loss = 2.624713 (0.082 sec/batch), lr: 0.003000
403
- 2025-12-03 22:47:25 INFO: Finished STEP 1680/50000, loss = 3.746646 (0.081 sec/batch), lr: 0.003000
404
- 2025-12-03 22:47:27 INFO: Finished STEP 1700/50000, loss = 3.647429 (0.078 sec/batch), lr: 0.003000
405
- 2025-12-03 22:47:27 INFO: Evaluating on dev set...
406
- 2025-12-03 22:47:27 INFO: LAS MLAS BLEX
407
- 2025-12-03 22:47:27 INFO: 62.38 53.59 56.96
408
- 2025-12-03 22:47:27 INFO: step 1700: train_loss = 3.711031, dev_score = 0.6238
409
- 2025-12-03 22:47:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
410
- 2025-12-03 22:47:28 INFO: new best model saved.
411
- 2025-12-03 22:47:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
412
- 2025-12-03 22:47:28 INFO: new model checkpoint saved.
413
- 2025-12-03 22:47:30 INFO: Finished STEP 1720/50000, loss = 3.788727 (0.082 sec/batch), lr: 0.003000
414
- 2025-12-03 22:47:31 INFO: Finished STEP 1740/50000, loss = 4.010241 (0.075 sec/batch), lr: 0.003000
415
- 2025-12-03 22:47:33 INFO: Finished STEP 1760/50000, loss = 3.686716 (0.081 sec/batch), lr: 0.003000
416
- 2025-12-03 22:47:35 INFO: Finished STEP 1780/50000, loss = 2.884626 (0.079 sec/batch), lr: 0.003000
417
- 2025-12-03 22:47:36 INFO: Finished STEP 1800/50000, loss = 3.530944 (0.073 sec/batch), lr: 0.003000
418
- 2025-12-03 22:47:36 INFO: Evaluating on dev set...
419
- 2025-12-03 22:47:37 INFO: LAS MLAS BLEX
420
- 2025-12-03 22:47:37 INFO: 63.37 53.81 58.47
421
- 2025-12-03 22:47:37 INFO: step 1800: train_loss = 3.627865, dev_score = 0.6337
422
- 2025-12-03 22:47:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
423
- 2025-12-03 22:47:37 INFO: new best model saved.
424
- 2025-12-03 22:47:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
425
- 2025-12-03 22:47:37 INFO: new model checkpoint saved.
426
- 2025-12-03 22:47:39 INFO: Finished STEP 1820/50000, loss = 3.387048 (0.078 sec/batch), lr: 0.003000
427
- 2025-12-03 22:47:40 INFO: Finished STEP 1840/50000, loss = 3.075635 (0.081 sec/batch), lr: 0.003000
428
- 2025-12-03 22:47:42 INFO: Finished STEP 1860/50000, loss = 3.544627 (0.074 sec/batch), lr: 0.003000
429
- 2025-12-03 22:47:44 INFO: Finished STEP 1880/50000, loss = 2.697122 (0.073 sec/batch), lr: 0.003000
430
- 2025-12-03 22:47:45 INFO: Finished STEP 1900/50000, loss = 4.582170 (0.074 sec/batch), lr: 0.003000
431
- 2025-12-03 22:47:45 INFO: Evaluating on dev set...
432
- 2025-12-03 22:47:46 INFO: LAS MLAS BLEX
433
- 2025-12-03 22:47:46 INFO: 63.12 55.81 57.93
434
- 2025-12-03 22:47:46 INFO: step 1900: train_loss = 3.833838, dev_score = 0.6312
435
- 2025-12-03 22:47:46 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
436
- 2025-12-03 22:47:46 INFO: new model checkpoint saved.
437
- 2025-12-03 22:47:48 INFO: Finished STEP 1920/50000, loss = 4.797754 (0.071 sec/batch), lr: 0.003000
438
- 2025-12-03 22:47:49 INFO: Finished STEP 1940/50000, loss = 3.999227 (0.076 sec/batch), lr: 0.003000
439
- 2025-12-03 22:47:51 INFO: Finished STEP 1960/50000, loss = 2.863396 (0.081 sec/batch), lr: 0.003000
440
- 2025-12-03 22:47:52 INFO: Finished STEP 1980/50000, loss = 3.603798 (0.074 sec/batch), lr: 0.003000
441
- 2025-12-03 22:47:54 INFO: Finished STEP 2000/50000, loss = 2.377973 (0.075 sec/batch), lr: 0.003000
442
- 2025-12-03 22:47:54 INFO: Evaluating on dev set...
443
- 2025-12-03 22:47:54 INFO: LAS MLAS BLEX
444
- 2025-12-03 22:47:54 INFO: 58.91 50.63 54.85
445
- 2025-12-03 22:47:54 INFO: step 2000: train_loss = 3.578752, dev_score = 0.5891
446
- 2025-12-03 22:47:55 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
447
- 2025-12-03 22:47:55 INFO: new model checkpoint saved.
448
- 2025-12-03 22:47:56 INFO: Finished STEP 2020/50000, loss = 3.455811 (0.069 sec/batch), lr: 0.003000
449
- 2025-12-03 22:47:58 INFO: Finished STEP 2040/50000, loss = 4.113457 (0.067 sec/batch), lr: 0.003000
450
- 2025-12-03 22:47:59 INFO: Finished STEP 2060/50000, loss = 3.953318 (0.070 sec/batch), lr: 0.003000
451
- 2025-12-03 22:48:01 INFO: Finished STEP 2080/50000, loss = 2.824056 (0.075 sec/batch), lr: 0.003000
452
- 2025-12-03 22:48:02 INFO: Finished STEP 2100/50000, loss = 4.579782 (0.079 sec/batch), lr: 0.003000
453
- 2025-12-03 22:48:02 INFO: Evaluating on dev set...
454
- 2025-12-03 22:48:03 INFO: LAS MLAS BLEX
455
- 2025-12-03 22:48:03 INFO: 61.63 52.85 57.51
456
- 2025-12-03 22:48:03 INFO: step 2100: train_loss = 3.729740, dev_score = 0.6163
457
- 2025-12-03 22:48:03 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
458
- 2025-12-03 22:48:03 INFO: new model checkpoint saved.
459
- 2025-12-03 22:48:05 INFO: Finished STEP 2120/50000, loss = 3.769734 (0.077 sec/batch), lr: 0.003000
460
- 2025-12-03 22:48:06 INFO: Finished STEP 2140/50000, loss = 3.904819 (0.071 sec/batch), lr: 0.003000
461
- 2025-12-03 22:48:08 INFO: Finished STEP 2160/50000, loss = 3.850474 (0.068 sec/batch), lr: 0.003000
462
- 2025-12-03 22:48:09 INFO: Finished STEP 2180/50000, loss = 3.001653 (0.072 sec/batch), lr: 0.003000
463
- 2025-12-03 22:48:11 INFO: Finished STEP 2200/50000, loss = 1.957051 (0.074 sec/batch), lr: 0.003000
464
- 2025-12-03 22:48:11 INFO: Evaluating on dev set...
465
- 2025-12-03 22:48:11 INFO: LAS MLAS BLEX
466
- 2025-12-03 22:48:11 INFO: 60.64 51.16 55.39
467
- 2025-12-03 22:48:11 INFO: step 2200: train_loss = 3.531071, dev_score = 0.6064
468
- 2025-12-03 22:48:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
469
- 2025-12-03 22:48:12 INFO: new model checkpoint saved.
470
- 2025-12-03 22:48:13 INFO: Finished STEP 2220/50000, loss = 3.429658 (0.077 sec/batch), lr: 0.003000
471
- 2025-12-03 22:48:15 INFO: Finished STEP 2240/50000, loss = 3.791333 (0.073 sec/batch), lr: 0.003000
472
- 2025-12-03 22:48:16 INFO: Finished STEP 2260/50000, loss = 2.263699 (0.072 sec/batch), lr: 0.003000
473
- 2025-12-03 22:48:18 INFO: Finished STEP 2280/50000, loss = 3.650295 (0.067 sec/batch), lr: 0.003000
474
- 2025-12-03 22:48:19 INFO: Finished STEP 2300/50000, loss = 4.082314 (0.071 sec/batch), lr: 0.003000
475
- 2025-12-03 22:48:19 INFO: Evaluating on dev set...
476
- 2025-12-03 22:48:20 INFO: LAS MLAS BLEX
477
- 2025-12-03 22:48:20 INFO: 62.13 54.55 57.93
478
- 2025-12-03 22:48:20 INFO: step 2300: train_loss = 3.777273, dev_score = 0.6213
479
- 2025-12-03 22:48:20 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
480
- 2025-12-03 22:48:20 INFO: new model checkpoint saved.
481
- 2025-12-03 22:48:22 INFO: Finished STEP 2320/50000, loss = 3.704285 (0.071 sec/batch), lr: 0.003000
482
- 2025-12-03 22:48:23 INFO: Finished STEP 2340/50000, loss = 2.753342 (0.069 sec/batch), lr: 0.003000
483
- 2025-12-03 22:48:25 INFO: Finished STEP 2360/50000, loss = 3.819938 (0.072 sec/batch), lr: 0.003000
484
- 2025-12-03 22:48:26 INFO: Finished STEP 2380/50000, loss = 4.015243 (0.075 sec/batch), lr: 0.003000
485
- 2025-12-03 22:48:27 INFO: Finished STEP 2400/50000, loss = 4.291789 (0.077 sec/batch), lr: 0.003000
486
- 2025-12-03 22:48:27 INFO: Evaluating on dev set...
487
- 2025-12-03 22:48:28 INFO: LAS MLAS BLEX
488
- 2025-12-03 22:48:28 INFO: 60.64 53.28 57.08
489
- 2025-12-03 22:48:28 INFO: step 2400: train_loss = 3.659402, dev_score = 0.6064
490
- 2025-12-03 22:48:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
491
- 2025-12-03 22:48:28 INFO: new model checkpoint saved.
492
- 2025-12-03 22:48:30 INFO: Finished STEP 2420/50000, loss = 3.808647 (0.070 sec/batch), lr: 0.003000
493
- 2025-12-03 22:48:31 INFO: Finished STEP 2440/50000, loss = 3.675776 (0.070 sec/batch), lr: 0.003000
494
- 2025-12-03 22:48:33 INFO: Finished STEP 2460/50000, loss = 4.045568 (0.071 sec/batch), lr: 0.003000
495
- 2025-12-03 22:48:34 INFO: Finished STEP 2480/50000, loss = 2.708247 (0.071 sec/batch), lr: 0.003000
496
- 2025-12-03 22:48:36 INFO: Finished STEP 2500/50000, loss = 3.840647 (0.072 sec/batch), lr: 0.003000
497
- 2025-12-03 22:48:36 INFO: Evaluating on dev set...
498
- 2025-12-03 22:48:36 INFO: LAS MLAS BLEX
499
- 2025-12-03 22:48:36 INFO: 63.61 54.97 57.93
500
- 2025-12-03 22:48:36 INFO: step 2500: train_loss = 3.616523, dev_score = 0.6361
501
- 2025-12-03 22:48:36 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
502
- 2025-12-03 22:48:36 INFO: new best model saved.
503
- 2025-12-03 22:48:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
504
- 2025-12-03 22:48:37 INFO: new model checkpoint saved.
505
- 2025-12-03 22:48:38 INFO: Finished STEP 2520/50000, loss = 3.329048 (0.075 sec/batch), lr: 0.003000
506
- 2025-12-03 22:48:40 INFO: Finished STEP 2540/50000, loss = 3.042371 (0.066 sec/batch), lr: 0.003000
507
- 2025-12-03 22:48:41 INFO: Finished STEP 2560/50000, loss = 3.319040 (0.070 sec/batch), lr: 0.003000
508
- 2025-12-03 22:48:43 INFO: Finished STEP 2580/50000, loss = 4.341519 (0.056 sec/batch), lr: 0.003000
509
- 2025-12-03 22:48:44 INFO: Finished STEP 2600/50000, loss = 5.309865 (0.070 sec/batch), lr: 0.003000
510
- 2025-12-03 22:48:44 INFO: Evaluating on dev set...
511
- 2025-12-03 22:48:44 INFO: LAS MLAS BLEX
512
- 2025-12-03 22:48:44 INFO: 61.14 52.01 56.24
513
- 2025-12-03 22:48:44 INFO: step 2600: train_loss = 3.623874, dev_score = 0.6114
514
- 2025-12-03 22:48:45 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
515
- 2025-12-03 22:48:45 INFO: new model checkpoint saved.
516
- 2025-12-03 22:48:46 INFO: Finished STEP 2620/50000, loss = 2.560167 (0.066 sec/batch), lr: 0.003000
517
- 2025-12-03 22:48:48 INFO: Finished STEP 2640/50000, loss = 3.262659 (0.066 sec/batch), lr: 0.003000
518
- 2025-12-03 22:48:49 INFO: Finished STEP 2660/50000, loss = 2.611564 (0.072 sec/batch), lr: 0.003000
519
- 2025-12-03 22:48:51 INFO: Finished STEP 2680/50000, loss = 3.953672 (0.070 sec/batch), lr: 0.003000
520
- 2025-12-03 22:48:52 INFO: Finished STEP 2700/50000, loss = 3.726388 (0.072 sec/batch), lr: 0.003000
521
- 2025-12-03 22:48:52 INFO: Evaluating on dev set...
522
- 2025-12-03 22:48:53 INFO: LAS MLAS BLEX
523
- 2025-12-03 22:48:53 INFO: 61.88 53.39 57.20
524
- 2025-12-03 22:48:53 INFO: step 2700: train_loss = 3.740555, dev_score = 0.6188
525
- 2025-12-03 22:48:53 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
526
- 2025-12-03 22:48:53 INFO: new model checkpoint saved.
527
- 2025-12-03 22:48:54 INFO: Finished STEP 2720/50000, loss = 3.536426 (0.066 sec/batch), lr: 0.003000
528
- 2025-12-03 22:48:56 INFO: Finished STEP 2740/50000, loss = 4.492881 (0.072 sec/batch), lr: 0.003000
529
- 2025-12-03 22:48:57 INFO: Finished STEP 2760/50000, loss = 3.437390 (0.073 sec/batch), lr: 0.003000
530
- 2025-12-03 22:48:59 INFO: Finished STEP 2780/50000, loss = 3.811538 (0.065 sec/batch), lr: 0.003000
531
- 2025-12-03 22:49:00 INFO: Finished STEP 2800/50000, loss = 2.615445 (0.069 sec/batch), lr: 0.003000
532
- 2025-12-03 22:49:00 INFO: Evaluating on dev set...
533
- 2025-12-03 22:49:01 INFO: LAS MLAS BLEX
534
- 2025-12-03 22:49:01 INFO: 61.14 51.59 56.66
535
- 2025-12-03 22:49:01 INFO: step 2800: train_loss = 3.751917, dev_score = 0.6114
536
- 2025-12-03 22:49:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
537
- 2025-12-03 22:49:01 INFO: new model checkpoint saved.
538
- 2025-12-03 22:49:03 INFO: Finished STEP 2820/50000, loss = 3.525189 (0.076 sec/batch), lr: 0.003000
539
- 2025-12-03 22:49:04 INFO: Finished STEP 2840/50000, loss = 3.113450 (0.066 sec/batch), lr: 0.003000
540
- 2025-12-03 22:49:05 INFO: Finished STEP 2860/50000, loss = 2.488067 (0.066 sec/batch), lr: 0.003000
541
- 2025-12-03 22:49:07 INFO: Finished STEP 2880/50000, loss = 2.702721 (0.071 sec/batch), lr: 0.003000
542
- 2025-12-03 22:49:08 INFO: Finished STEP 2900/50000, loss = 3.248651 (0.068 sec/batch), lr: 0.003000
543
- 2025-12-03 22:49:08 INFO: Evaluating on dev set...
544
- 2025-12-03 22:49:09 INFO: LAS MLAS BLEX
545
- 2025-12-03 22:49:09 INFO: 62.13 52.43 56.24
546
- 2025-12-03 22:49:09 INFO: step 2900: train_loss = 3.761651, dev_score = 0.6213
547
- 2025-12-03 22:49:09 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
548
- 2025-12-03 22:49:09 INFO: new model checkpoint saved.
549
- 2025-12-03 22:49:11 INFO: Finished STEP 2920/50000, loss = 3.129019 (0.072 sec/batch), lr: 0.003000
550
- 2025-12-03 22:49:12 INFO: Finished STEP 2940/50000, loss = 3.579517 (0.067 sec/batch), lr: 0.003000
551
- 2025-12-03 22:49:13 INFO: Finished STEP 2960/50000, loss = 3.505895 (0.067 sec/batch), lr: 0.003000
552
- 2025-12-03 22:49:15 INFO: Finished STEP 2980/50000, loss = 2.193599 (0.073 sec/batch), lr: 0.003000
553
- 2025-12-03 22:49:16 INFO: Finished STEP 3000/50000, loss = 1.882619 (0.064 sec/batch), lr: 0.003000
554
- 2025-12-03 22:49:16 INFO: Evaluating on dev set...
555
- 2025-12-03 22:49:17 INFO: LAS MLAS BLEX
556
- 2025-12-03 22:49:17 INFO: 61.63 51.16 55.81
557
- 2025-12-03 22:49:17 INFO: step 3000: train_loss = 4.014063, dev_score = 0.6163
558
- 2025-12-03 22:49:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
559
- 2025-12-03 22:49:17 INFO: new model checkpoint saved.
560
- 2025-12-03 22:49:19 INFO: Finished STEP 3020/50000, loss = 5.977424 (0.058 sec/batch), lr: 0.003000
561
- 2025-12-03 22:49:20 INFO: Finished STEP 3040/50000, loss = 3.319470 (0.063 sec/batch), lr: 0.003000
562
- 2025-12-03 22:49:21 INFO: Finished STEP 3060/50000, loss = 3.211635 (0.069 sec/batch), lr: 0.003000
563
- 2025-12-03 22:49:23 INFO: Finished STEP 3080/50000, loss = 3.183325 (0.065 sec/batch), lr: 0.003000
564
- 2025-12-03 22:49:24 INFO: Finished STEP 3100/50000, loss = 2.788619 (0.069 sec/batch), lr: 0.003000
565
- 2025-12-03 22:49:24 INFO: Evaluating on dev set...
566
- 2025-12-03 22:49:25 INFO: LAS MLAS BLEX
567
- 2025-12-03 22:49:25 INFO: 60.89 51.05 56.54
568
- 2025-12-03 22:49:25 INFO: step 3100: train_loss = 3.763229, dev_score = 0.6089
569
- 2025-12-03 22:49:25 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
570
- 2025-12-03 22:49:25 INFO: new model checkpoint saved.
571
- 2025-12-03 22:49:27 INFO: Finished STEP 3120/50000, loss = 4.927651 (0.063 sec/batch), lr: 0.003000
572
- 2025-12-03 22:49:28 INFO: Finished STEP 3140/50000, loss = 3.578332 (0.068 sec/batch), lr: 0.003000
573
- 2025-12-03 22:49:29 INFO: Finished STEP 3160/50000, loss = 3.173575 (0.076 sec/batch), lr: 0.003000
574
- 2025-12-03 22:49:31 INFO: Finished STEP 3180/50000, loss = 4.860454 (0.075 sec/batch), lr: 0.003000
575
- 2025-12-03 22:49:32 INFO: Finished STEP 3200/50000, loss = 1.778136 (0.065 sec/batch), lr: 0.003000
576
- 2025-12-03 22:49:32 INFO: Evaluating on dev set...
577
- 2025-12-03 22:49:33 INFO: LAS MLAS BLEX
578
- 2025-12-03 22:49:33 INFO: 60.15 50.74 54.12
579
- 2025-12-03 22:49:33 INFO: step 3200: train_loss = 3.791774, dev_score = 0.6015
580
- 2025-12-03 22:49:33 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
581
- 2025-12-03 22:49:33 INFO: new model checkpoint saved.
582
- 2025-12-03 22:49:35 INFO: Finished STEP 3220/50000, loss = 8.492014 (0.070 sec/batch), lr: 0.003000
583
- 2025-12-03 22:49:36 INFO: Finished STEP 3240/50000, loss = 3.045755 (0.064 sec/batch), lr: 0.003000
584
- 2025-12-03 22:49:37 INFO: Finished STEP 3260/50000, loss = 4.122291 (0.076 sec/batch), lr: 0.003000
585
- 2025-12-03 22:49:39 INFO: Finished STEP 3280/50000, loss = 3.458145 (0.063 sec/batch), lr: 0.003000
586
- 2025-12-03 22:49:40 INFO: Finished STEP 3300/50000, loss = 2.681028 (0.064 sec/batch), lr: 0.003000
587
- 2025-12-03 22:49:40 INFO: Evaluating on dev set...
588
- 2025-12-03 22:49:41 INFO: LAS MLAS BLEX
589
- 2025-12-03 22:49:41 INFO: 61.88 51.59 55.39
590
- 2025-12-03 22:49:41 INFO: step 3300: train_loss = 3.750064, dev_score = 0.6188
591
- 2025-12-03 22:49:41 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
592
- 2025-12-03 22:49:41 INFO: new model checkpoint saved.
593
- 2025-12-03 22:49:42 INFO: Finished STEP 3320/50000, loss = 3.499715 (0.066 sec/batch), lr: 0.003000
594
- 2025-12-03 22:49:44 INFO: Finished STEP 3340/50000, loss = 3.765631 (0.064 sec/batch), lr: 0.003000
595
- 2025-12-03 22:49:45 INFO: Finished STEP 3360/50000, loss = 2.908885 (0.064 sec/batch), lr: 0.003000
596
- 2025-12-03 22:49:47 INFO: Finished STEP 3380/50000, loss = 4.147782 (0.067 sec/batch), lr: 0.003000
597
- 2025-12-03 22:49:48 INFO: Finished STEP 3400/50000, loss = 4.854831 (0.073 sec/batch), lr: 0.003000
598
- 2025-12-03 22:49:48 INFO: Evaluating on dev set...
599
- 2025-12-03 22:49:48 INFO: LAS MLAS BLEX
600
- 2025-12-03 22:49:48 INFO: 61.39 51.59 54.55
601
- 2025-12-03 22:49:48 INFO: step 3400: train_loss = 3.827870, dev_score = 0.6139
602
- 2025-12-03 22:49:49 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
603
- 2025-12-03 22:49:49 INFO: new model checkpoint saved.
604
- 2025-12-03 22:49:50 INFO: Finished STEP 3420/50000, loss = 3.672020 (0.066 sec/batch), lr: 0.003000
605
- 2025-12-03 22:49:52 INFO: Finished STEP 3440/50000, loss = 3.315593 (0.064 sec/batch), lr: 0.003000
606
- 2025-12-03 22:49:53 INFO: Finished STEP 3460/50000, loss = 2.334443 (0.064 sec/batch), lr: 0.003000
607
- 2025-12-03 22:49:54 INFO: Finished STEP 3480/50000, loss = 4.692723 (0.061 sec/batch), lr: 0.003000
608
- 2025-12-03 22:49:56 INFO: Finished STEP 3500/50000, loss = 2.920178 (0.064 sec/batch), lr: 0.003000
609
- 2025-12-03 22:49:56 INFO: Evaluating on dev set...
610
- 2025-12-03 22:49:56 INFO: LAS MLAS BLEX
611
- 2025-12-03 22:49:56 INFO: 60.40 50.74 53.70
612
- 2025-12-03 22:49:56 INFO: step 3500: train_loss = 3.750002, dev_score = 0.6040
613
- 2025-12-03 22:49:56 INFO: Training ended with 3500 steps.
614
- 2025-12-03 22:49:56 INFO: Best dev F1 = 63.61, at iteration = 2500
615
- 2025-12-03 22:49:57 INFO: Running dev depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
616
- 2025-12-03 22:49:57 INFO: Running parser in predict mode
617
- 2025-12-03 22:49:57 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
618
- 2025-12-03 22:49:59 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
619
- 2025-12-03 22:49:59 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
620
- 2025-12-03 22:49:59 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
621
- 2025-12-03 22:49:59 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
622
- 2025-12-03 22:50:00 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
623
- 2025-12-03 22:50:00 INFO: Loading data with batch size 32...
624
- 2025-12-03 22:50:00 DEBUG: 9 batches created.
625
- 2025-12-03 22:50:00 INFO: F1 scores for each dependency:
626
- Note that unlabeled attachment errors hurt the labeled attachment scores
627
- acl: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
628
- acl:relcl: p 0.2222 r 0.2857 f1 0.2500 (7 actual)
629
- advcl: p 0.1667 r 0.2000 f1 0.1818 (5 actual)
630
- advmod: p 0.4828 r 0.5600 f1 0.5185 (25 actual)
631
- amod: p 0.8889 r 0.7742 f1 0.8276 (31 actual)
632
- appos: p 1.0000 r 0.5000 f1 0.6667 (4 actual)
633
- aux: p 0.8182 r 0.8182 f1 0.8182 (11 actual)
634
- case: p 0.9444 r 0.9107 f1 0.9273 (56 actual)
635
- cc: p 0.6923 r 0.6923 f1 0.6923 (13 actual)
636
- ccomp: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
637
- compound:prt: p 0.0000 r 0.0000 f1 0.0000 (0 actual)
638
- conj: p 0.3750 r 0.5000 f1 0.4286 (12 actual)
639
- cop: p 0.5000 r 0.3333 f1 0.4000 (3 actual)
640
- csubj: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
641
- det: p 0.8696 r 0.9091 f1 0.8889 (22 actual)
642
- expl: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
643
- iobj: p 0.3333 r 0.5000 f1 0.4000 (2 actual)
644
- mark: p 0.6364 r 0.5833 f1 0.6087 (12 actual)
645
- nmod: p 0.4118 r 0.4667 f1 0.4375 (15 actual)
646
- nmod:poss: p 1.0000 r 0.8947 f1 0.9444 (19 actual)
647
- nsubj: p 0.5200 r 0.7647 f1 0.6190 (17 actual)
648
- nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
649
- obj: p 0.5769 r 0.6818 f1 0.6250 (22 actual)
650
- obl: p 0.5714 r 0.5854 f1 0.5783 (41 actual)
651
- obl:agent: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
652
- orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
653
- parataxis: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
654
- punct: p 0.4423 r 0.4423 f1 0.4423 (52 actual)
655
- root: p 0.5556 r 0.5556 f1 0.5556 (9 actual)
656
- xcomp: p 0.6667 r 0.2500 f1 0.3636 (8 actual)
657
- 2025-12-03 22:50:00 INFO: LAS MLAS BLEX
658
- 2025-12-03 22:50:00 INFO: 63.61 54.97 57.93
659
- 2025-12-03 22:50:00 INFO: Parser score:
660
- 2025-12-03 22:50:00 INFO: sv_diachronic 63.61
661
- 2025-12-03 22:50:00 INFO: Finished running dev set on
662
- UD_Swedish-diachronic
663
- UAS LAS CLAS MLAS BLEX
664
- 70.54 63.61 57.93 54.97 57.93
665
- 2025-12-03 22:50:00 INFO: Running test depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
666
- 2025-12-03 22:50:00 INFO: Running parser in predict mode
667
- 2025-12-03 22:50:00 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
668
- 2025-12-03 22:50:02 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
669
- 2025-12-03 22:50:02 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
670
- 2025-12-03 22:50:02 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
671
- 2025-12-03 22:50:02 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
672
- 2025-12-03 22:50:02 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
673
- 2025-12-03 22:50:02 INFO: Loading data with batch size 32...
674
- 2025-12-03 22:50:02 DEBUG: 93 batches created.
675
- 2025-12-03 22:50:07 INFO: F1 scores for each dependency:
676
- Note that unlabeled attachment errors hurt the labeled attachment scores
677
- acl: p 0.3333 r 0.0312 f1 0.0571 (32 actual)
678
- acl:cleft: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
679
- acl:relcl: p 0.3061 r 0.2000 f1 0.2419 (75 actual)
680
- advcl: p 0.0893 r 0.1667 f1 0.1163 (60 actual)
681
- advcl:relcl: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
682
- advmod: p 0.5745 r 0.5896 f1 0.5820 (268 actual)
683
- amod: p 0.8139 r 0.8174 f1 0.8156 (230 actual)
684
- appos: p 0.0000 r 0.0000 f1 0.0000 (13 actual)
685
- aux: p 0.8554 r 0.8452 f1 0.8503 (84 actual)
686
- aux:pass: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
687
- case: p 0.8661 r 0.8150 f1 0.8398 (373 actual)
688
- cc: p 0.6474 r 0.6516 f1 0.6495 (155 actual)
689
- ccomp: p 0.0000 r 0.0000 f1 0.0000 (35 actual)
690
- compound:prt: p 0.6800 r 0.8095 f1 0.7391 (21 actual)
691
- conj: p 0.2938 r 0.2975 f1 0.2956 (158 actual)
692
- cop: p 0.7188 r 0.5000 f1 0.5897 (46 actual)
693
- csubj: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
694
- dep: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
695
- det: p 0.8308 r 0.8029 f1 0.8166 (208 actual)
696
- discourse: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
697
- dislocated: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
698
- expl: p 0.5000 r 0.0909 f1 0.1538 (11 actual)
699
- expl:pv: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
700
- fixed: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
701
- flat: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
702
- flat:name: p 0.0000 r 0.0000 f1 0.0000 (12 actual)
703
- goeswith: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
704
- iobj: p 0.1935 r 0.4286 f1 0.2667 (14 actual)
705
- mark: p 0.6624 r 0.6797 f1 0.6710 (153 actual)
706
- nmod: p 0.2661 r 0.2843 f1 0.2749 (102 actual)
707
- nmod:poss: p 0.8865 r 0.8803 f1 0.8834 (142 actual)
708
- nsubj: p 0.5385 r 0.6750 f1 0.5990 (280 actual)
709
- nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (25 actual)
710
- nummod: p 0.8000 r 0.8000 f1 0.8000 (10 actual)
711
- obj: p 0.5556 r 0.6011 f1 0.5774 (183 actual)
712
- obl: p 0.5030 r 0.5935 f1 0.5446 (278 actual)
713
- obl:agent: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
714
- orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
715
- parataxis: p 0.1000 r 0.2222 f1 0.1379 (18 actual)
716
- punct: p 0.4685 r 0.4729 f1 0.4707 (425 actual)
717
- reparandum: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
718
- root: p 0.5657 r 0.5657 f1 0.5657 (99 actual)
719
- vocative: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
720
- xcomp: p 0.5000 r 0.3467 f1 0.4094 (75 actual)
721
- 2025-12-03 22:50:07 INFO: LAS MLAS BLEX
722
- 2025-12-03 22:50:07 INFO: 59.06 49.85 53.26
723
- 2025-12-03 22:50:07 INFO: Parser score:
724
- 2025-12-03 22:50:07 INFO: sv_diachronic 59.06
725
- 2025-12-03 22:50:07 INFO: Finished running test set on
726
- UD_Swedish-diachronic
727
- UAS LAS CLAS MLAS BLEX
728
- 68.07 59.06 53.26 49.85 53.26
729
- DONE.
730
- Full log saved to: logs/log_conll17.pt_sv_diachron_20251203_223822.txt
731
- Symlink updated: logs/latest.txt → log_conll17.pt_sv_diachron_20251203_223822.txt
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_conll17.pt_sv_diachron_20251212_145854.txt DELETED
@@ -1,161 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_diachron_20251212_145854.txt ===
2
- Language codes: sv diachron
3
- Using pretrained model: conll17.pt
4
-
5
- Running: python prepare-train-val-test.py sv diachron
6
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec991-spf148.conllu
7
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec988-spf145.conllu
8
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
9
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
10
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
11
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec252-BremerF_Teckningar1.conllu
12
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
13
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
14
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
15
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
16
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec452-NyblomH_FantasierFyra.conllu
17
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1102-spf259.conllu
18
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-letter141673-Stalhammar.conllu
19
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1033-spf190.conllu
20
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
21
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
22
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
23
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
24
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
25
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
26
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
27
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
28
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
29
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
30
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
31
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec25-Runius.conllu
32
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
33
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
34
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec987-spf144.conllu
35
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1063-spf220.conllu
36
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
37
- Including DigPhil MACHINE in TRAIN (minus gold)…
38
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
39
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
40
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
41
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
42
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
43
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
44
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
45
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
46
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
47
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
48
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
49
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
50
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
51
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
52
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
53
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
54
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
55
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
56
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
57
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
58
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
59
- Cleaning TRAIN...
60
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
61
- [REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
62
- [REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
63
- [REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
64
- [REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
65
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
66
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
67
- [REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
68
- [REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
69
- [REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
70
- [REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
71
- [REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
72
- [REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
73
- [REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
74
- [REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
75
- [REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
76
- [REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
77
- [REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
78
- [REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
79
- [REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
80
- [REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
81
- [REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
82
- [REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
83
- [REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
84
- [REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
85
- [REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
86
- [REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
87
- [REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
88
- [REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
89
- [REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
90
- [REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
91
- [REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
92
- [REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
93
- [REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
94
- [REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
95
- [REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
96
- [REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
97
- [REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
98
- [REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
99
- [REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
100
- [REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
101
- [REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
102
- [REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
103
- [REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
104
- [REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
105
- [REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
106
- [REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
107
- [REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
108
- [REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
109
- [REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
110
- [REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
111
- [REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
112
- [REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
113
- [REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
114
- [REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
115
- [REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
116
- [REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
117
- [REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
118
- [REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
119
- [REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
120
- [REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
121
- [REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
122
- [REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
123
- [REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
124
- Cleaning DEV...
125
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
126
- Cleaning TEST...
127
- Writing TRAIN → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (60334 valid sentences)
128
- Writing DEV → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (9 valid sentences)
129
- Writing TEST → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (99 valid sentences)
130
- Done.
131
- Sourcing scripts/config_alvis.sh
132
- Running stanza dataset preparation…
133
- 2025-12-12 14:59:02 INFO: Datasets program called with:
134
- /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
135
- 2025-12-12 14:59:02 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
136
-
137
- 2025-12-12 14:59:02 INFO: Downloaded file to /cephyr/users/cleland/Alvis/stanza_resources/resources.json
138
- 2025-12-12 14:59:02 DEBUG: Processing parameter "processors"...
139
- 2025-12-12 14:59:02 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
140
- 2025-12-12 14:59:02 INFO: Downloading these customized packages for language: sv (Swedish)...
141
- =======================
142
- | Processor | Package |
143
- -----------------------
144
- =======================
145
-
146
- 2025-12-12 14:59:02 INFO: Finished downloading models and saved to /cephyr/users/cleland/Alvis/stanza_resources
147
- 2025-12-12 14:59:02 INFO: Using tagger model in /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt for sv_diachronic
148
- 2025-12-12 14:59:02 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
149
- 2025-12-12 14:59:02 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
150
- Augmented 189 quotes: Counter({'""': 26, '「」': 23, '„”': 23, '»«': 22, '″″': 18, '””': 18, '““': 17, '《》': 17, '„“': 13, '«»': 12})
151
- 2025-12-12 14:59:05 INFO: Running tagger to retag /local/tmp.5491708/tmpc2soyxjt/sv_diachronic.train.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
152
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5491708/tmpc2soyxjt/sv_diachronic.train.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu']
153
- 2025-12-12 14:59:05 INFO: Running tagger in predict mode
154
- 2025-12-12 14:59:05 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
155
- 2025-12-12 14:59:07 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
156
- 2025-12-12 14:59:07 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
157
- 2025-12-12 14:59:07 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
158
- 2025-12-12 14:59:07 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
159
- 2025-12-12 14:59:08 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
160
- 2025-12-12 14:59:11 INFO: Loading data with batch size 250...
161
- ./make_new_model.sh: line 58: 3492354 Terminated python -m stanza.utils.datasets.prepare_depparse_treebank UD_Swedish-diachronic --wordvec_pretrain_file "/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/${PRETRAINED_MODEL}"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_conll17.pt_sv_diachron_20251212_150001.txt DELETED
The diff for this file is too large to render. See raw diff
 
logs/log_conll17.pt_sv_diachron_20260116_065352.txt DELETED
@@ -1,830 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_diachron_20260116_065352.txt ===
2
- Language codes: sv diachron
3
- Using pretrained model: conll17.pt
4
-
5
- Running: python prepare-train-val-test.py sv diachron
6
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec991-spf148.conllu
7
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec988-spf145.conllu
8
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
9
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
10
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
11
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec252-BremerF_Teckningar1.conllu
12
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
13
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
14
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
15
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
16
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec452-NyblomH_FantasierFyra.conllu
17
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1102-spf259.conllu
18
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-letter141673-Stalhammar.conllu
19
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1033-spf190.conllu
20
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
21
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
22
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
23
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
24
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
25
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
26
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
27
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
28
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
29
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
30
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
31
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec25-Runius.conllu
32
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
33
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
34
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec987-spf144.conllu
35
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1063-spf220.conllu
36
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
37
- Including DigPhil MACHINE in TRAIN (minus gold)…
38
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
39
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
40
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
41
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
42
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
43
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
44
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
45
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
46
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
47
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
48
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
49
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
50
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
51
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
52
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
53
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
54
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
55
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
56
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
57
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
58
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
59
- Cleaning TRAIN...
60
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
61
- [REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
62
- [REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
63
- [REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
64
- [REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
65
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
66
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
67
- [REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
68
- [REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
69
- [REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
70
- [REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
71
- [REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
72
- [REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
73
- [REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
74
- [REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
75
- [REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
76
- [REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
77
- [REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
78
- [REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
79
- [REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
80
- [REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
81
- [REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
82
- [REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
83
- [REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
84
- [REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
85
- [REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
86
- [REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
87
- [REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
88
- [REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
89
- [REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
90
- [REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
91
- [REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
92
- [REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
93
- [REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
94
- [REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
95
- [REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
96
- [REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
97
- [REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
98
- [REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
99
- [REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
100
- [REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
101
- [REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
102
- [REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
103
- [REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
104
- [REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
105
- [REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
106
- [REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
107
- [REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
108
- [REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
109
- [REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
110
- [REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
111
- [REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
112
- [REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
113
- [REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
114
- [REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
115
- [REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
116
- [REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
117
- [REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
118
- [REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
119
- [REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
120
- [REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
121
- [REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
122
- [REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
123
- [REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
124
- Cleaning DEV...
125
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
126
- Cleaning TEST...
127
- Writing TRAIN → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (60334 valid sentences)
128
- Writing DEV → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (9 valid sentences)
129
- Writing TEST → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (99 valid sentences)
130
- Done.
131
- Sourcing scripts/config_alvis.sh
132
- Running stanza dataset preparation…
133
- 2026-01-16 06:54:03 INFO: Datasets program called with:
134
- /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
135
- 2026-01-16 06:54:03 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
136
-
137
- 2026-01-16 06:54:03 INFO: Downloaded file to /cephyr/users/cleland/Alvis/stanza_resources/resources.json
138
- 2026-01-16 06:54:03 DEBUG: Processing parameter "processors"...
139
- 2026-01-16 06:54:03 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
140
- 2026-01-16 06:54:03 INFO: Downloading these customized packages for language: sv (Swedish)...
141
- =======================
142
- | Processor | Package |
143
- -----------------------
144
- =======================
145
-
146
- 2026-01-16 06:54:03 INFO: Finished downloading models and saved to /cephyr/users/cleland/Alvis/stanza_resources
147
- 2026-01-16 06:54:03 INFO: Using tagger model in /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt for sv_diachronic
148
- 2026-01-16 06:54:03 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
149
- 2026-01-16 06:54:03 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
150
- Augmented 189 quotes: Counter({'""': 26, '「」': 23, '„”': 23, '»«': 22, '″″': 18, '””': 18, '““': 17, '《》': 17, '„“': 13, '«»': 12})
151
- 2026-01-16 06:54:06 INFO: Running tagger to retag /local/tmp.5658622/tmp9_y5_489/sv_diachronic.train.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
152
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5658622/tmp9_y5_489/sv_diachronic.train.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu']
153
- 2026-01-16 06:54:06 INFO: Running tagger in predict mode
154
- 2026-01-16 06:54:06 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
155
- 2026-01-16 06:54:08 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
156
- 2026-01-16 06:54:08 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
157
- 2026-01-16 06:54:08 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
158
- 2026-01-16 06:54:08 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
159
- 2026-01-16 06:54:08 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
160
- 2026-01-16 06:54:12 INFO: Loading data with batch size 250...
161
- 2026-01-16 06:54:44 INFO: Start evaluation...
162
- 2026-01-16 06:57:49 INFO: UPOS XPOS UFeats AllTags
163
- 2026-01-16 06:57:49 INFO: 99.37 87.38 98.56 87.09
164
- 2026-01-16 06:57:49 INFO: POS Tagger score: sv_diachronic 87.09
165
- 2026-01-16 06:57:50 INFO: Running tagger to retag /local/tmp.5658622/tmp9_y5_489/sv_diachronic.dev.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
166
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5658622/tmp9_y5_489/sv_diachronic.dev.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu']
167
- 2026-01-16 06:57:50 INFO: Running tagger in predict mode
168
- 2026-01-16 06:57:50 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
169
- 2026-01-16 06:57:51 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
170
- 2026-01-16 06:57:51 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
171
- 2026-01-16 06:57:51 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
172
- 2026-01-16 06:57:52 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
173
- 2026-01-16 06:57:52 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
174
- 2026-01-16 06:57:52 INFO: Loading data with batch size 250...
175
- 2026-01-16 06:57:52 INFO: Start evaluation...
176
- 2026-01-16 06:57:52 INFO: UPOS XPOS UFeats AllTags
177
- 2026-01-16 06:57:52 INFO: 93.32 90.84 93.32 85.64
178
- 2026-01-16 06:57:52 INFO: POS Tagger score: sv_diachronic 85.64
179
- 2026-01-16 06:57:52 INFO: Running tagger to retag /local/tmp.5658622/tmp9_y5_489/sv_diachronic.test.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
180
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5658622/tmp9_y5_489/sv_diachronic.test.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu']
181
- 2026-01-16 06:57:52 INFO: Running tagger in predict mode
182
- 2026-01-16 06:57:52 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
183
- 2026-01-16 06:57:54 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
184
- 2026-01-16 06:57:54 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
185
- 2026-01-16 06:57:54 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
186
- 2026-01-16 06:57:54 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
187
- 2026-01-16 06:57:54 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
188
- 2026-01-16 06:57:54 INFO: Loading data with batch size 250...
189
- 2026-01-16 06:57:54 INFO: Start evaluation...
190
- 2026-01-16 06:57:55 INFO: UPOS XPOS UFeats AllTags
191
- 2026-01-16 06:57:55 INFO: 93.14 96.78 95.32 90.28
192
- 2026-01-16 06:57:55 INFO: POS Tagger score: sv_diachronic 90.28
193
- Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
194
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /local/tmp.5658622/tmp9_y5_489/sv_diachronic.train.gold.conllu
195
- Swapped 'w1, w2' for 'w1 ,w2' 106 times
196
- Added 520 new sentences with asdf, zzzz -> asdf,zzzz
197
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /local/tmp.5658622/tmp9_y5_489/sv_diachronic.dev.gold.conllu
198
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /local/tmp.5658622/tmp9_y5_489/sv_diachronic.test.gold.conllu
199
- Running stanza dependency parser training…
200
- 2026-01-16 06:58:08 INFO: Training program called with:
201
- /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt --batch_size 8 --dropout 0.33
202
- 2026-01-16 06:58:08 DEBUG: UD_Swedish-diachronic: sv_diachronic
203
- 2026-01-16 06:58:08 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
204
- 2026-01-16 06:58:08 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
205
- 2026-01-16 06:58:08 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_charlm_parser.pt does not exist, training new model
206
- 2026-01-16 06:58:08 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
207
- 2026-01-16 06:58:08 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
208
- 2026-01-16 06:58:08 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--train_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '8', '--dropout', '0.33']
209
- 2026-01-16 06:58:08 INFO: Running parser in train mode
210
- 2026-01-16 06:58:08 INFO: Using pretrained contextualized char embedding
211
- 2026-01-16 06:58:08 INFO: Loading data with batch size 8...
212
- 2026-01-16 06:58:14 INFO: Train File /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 60854
213
- 2026-01-16 06:58:14 INFO: Original data size: 60854
214
- 2026-01-16 06:58:14 INFO: Augmented data size: 60854
215
- 2026-01-16 06:58:29 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
216
- 2026-01-16 06:58:29 INFO: Original length = 60854
217
- 2026-01-16 06:58:29 INFO: Filtered length = 60854
218
- 2026-01-16 06:58:43 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='|') for the xpos factory for sv_diachronic
219
- 2026-01-16 06:58:48 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
220
- 2026-01-16 06:58:59 DEBUG: 58067 batches created.
221
- 2026-01-16 06:58:59 DEBUG: 9 batches created.
222
- 2026-01-16 06:58:59 INFO: Training parser...
223
- 2026-01-16 06:58:59 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
224
- 2026-01-16 06:58:59 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
225
- 2026-01-16 06:59:00 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
226
- 2026-01-16 06:59:00 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
227
- 2026-01-16 06:59:04 INFO: Finished STEP 20/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
228
- 2026-01-16 06:59:05 INFO: Finished STEP 40/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
229
- 2026-01-16 06:59:05 INFO: Finished STEP 60/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
230
- 2026-01-16 06:59:06 INFO: Finished STEP 80/50000, loss = 0.000000 (0.032 sec/batch), lr: 0.003000
231
- 2026-01-16 06:59:07 INFO: Finished STEP 100/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
232
- 2026-01-16 06:59:07 INFO: Evaluating on dev set...
233
- 2026-01-16 06:59:07 INFO: LAS MLAS BLEX
234
- 2026-01-16 06:59:07 INFO: 0.50 0.31 0.63
235
- 2026-01-16 06:59:07 INFO: step 100: train_loss = 192.073832, dev_score = 0.0050
236
- 2026-01-16 06:59:07 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
237
- 2026-01-16 06:59:07 INFO: new best model saved.
238
- 2026-01-16 06:59:08 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
239
- 2026-01-16 06:59:08 INFO: new model checkpoint saved.
240
- 2026-01-16 06:59:08 INFO: Finished STEP 120/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
241
- 2026-01-16 06:59:09 INFO: Finished STEP 140/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
242
- 2026-01-16 06:59:10 INFO: Finished STEP 160/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
243
- 2026-01-16 06:59:10 INFO: Finished STEP 180/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
244
- 2026-01-16 06:59:11 INFO: Finished STEP 200/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
245
- 2026-01-16 06:59:11 INFO: Evaluating on dev set...
246
- 2026-01-16 06:59:11 INFO: LAS MLAS BLEX
247
- 2026-01-16 06:59:11 INFO: 0.50 0.31 0.63
248
- 2026-01-16 06:59:11 INFO: step 200: train_loss = 0.000000, dev_score = 0.0050
249
- 2026-01-16 06:59:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
250
- 2026-01-16 06:59:12 INFO: new best model saved.
251
- 2026-01-16 06:59:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
252
- 2026-01-16 06:59:12 INFO: new model checkpoint saved.
253
- 2026-01-16 06:59:13 INFO: Finished STEP 220/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
254
- 2026-01-16 06:59:13 INFO: Finished STEP 240/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
255
- 2026-01-16 06:59:14 INFO: Finished STEP 260/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
256
- 2026-01-16 06:59:15 INFO: Finished STEP 280/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
257
- 2026-01-16 06:59:15 INFO: Finished STEP 300/50000, loss = 2.877054 (0.032 sec/batch), lr: 0.003000
258
- 2026-01-16 06:59:15 INFO: Evaluating on dev set...
259
- 2026-01-16 06:59:16 INFO: LAS MLAS BLEX
260
- 2026-01-16 06:59:16 INFO: 1.49 0.00 0.00
261
- 2026-01-16 06:59:16 INFO: step 300: train_loss = 32.398629, dev_score = 0.0149
262
- 2026-01-16 06:59:16 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
263
- 2026-01-16 06:59:16 INFO: new best model saved.
264
- 2026-01-16 06:59:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
265
- 2026-01-16 06:59:17 INFO: new model checkpoint saved.
266
- 2026-01-16 06:59:17 INFO: Finished STEP 320/50000, loss = 4.388504 (0.032 sec/batch), lr: 0.003000
267
- 2026-01-16 06:59:18 INFO: Finished STEP 340/50000, loss = 1.865119 (0.031 sec/batch), lr: 0.003000
268
- 2026-01-16 06:59:18 INFO: Finished STEP 360/50000, loss = 2.029490 (0.031 sec/batch), lr: 0.003000
269
- 2026-01-16 06:59:19 INFO: Finished STEP 380/50000, loss = 3.120337 (0.033 sec/batch), lr: 0.003000
270
- 2026-01-16 06:59:20 INFO: Finished STEP 400/50000, loss = 0.527415 (0.031 sec/batch), lr: 0.003000
271
- 2026-01-16 06:59:20 INFO: Evaluating on dev set...
272
- 2026-01-16 06:59:20 INFO: LAS MLAS BLEX
273
- 2026-01-16 06:59:20 INFO: 2.97 2.74 2.74
274
- 2026-01-16 06:59:20 INFO: step 400: train_loss = 2.460383, dev_score = 0.0297
275
- 2026-01-16 06:59:21 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
276
- 2026-01-16 06:59:21 INFO: new best model saved.
277
- 2026-01-16 06:59:21 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
278
- 2026-01-16 06:59:21 INFO: new model checkpoint saved.
279
- 2026-01-16 06:59:22 INFO: Finished STEP 420/50000, loss = 0.000224 (0.030 sec/batch), lr: 0.003000
280
- 2026-01-16 06:59:22 INFO: Finished STEP 440/50000, loss = 1.204638 (0.031 sec/batch), lr: 0.003000
281
- 2026-01-16 06:59:23 INFO: Finished STEP 460/50000, loss = 1.052635 (0.031 sec/batch), lr: 0.003000
282
- 2026-01-16 06:59:24 INFO: Finished STEP 480/50000, loss = 2.311400 (0.032 sec/batch), lr: 0.003000
283
- 2026-01-16 06:59:24 INFO: Finished STEP 500/50000, loss = 1.092946 (0.031 sec/batch), lr: 0.003000
284
- 2026-01-16 06:59:24 INFO: Evaluating on dev set...
285
- 2026-01-16 06:59:25 INFO: LAS MLAS BLEX
286
- 2026-01-16 06:59:25 INFO: 2.48 0.70 1.41
287
- 2026-01-16 06:59:25 INFO: step 500: train_loss = 1.338739, dev_score = 0.0248
288
- 2026-01-16 06:59:25 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
289
- 2026-01-16 06:59:25 INFO: new model checkpoint saved.
290
- 2026-01-16 06:59:26 INFO: Finished STEP 520/50000, loss = 0.652456 (0.030 sec/batch), lr: 0.003000
291
- 2026-01-16 06:59:27 INFO: Finished STEP 540/50000, loss = 0.147004 (0.031 sec/batch), lr: 0.003000
292
- 2026-01-16 06:59:27 INFO: Finished STEP 560/50000, loss = 0.782623 (0.030 sec/batch), lr: 0.003000
293
- 2026-01-16 06:59:28 INFO: Finished STEP 580/50000, loss = 0.076848 (0.030 sec/batch), lr: 0.003000
294
- 2026-01-16 06:59:28 INFO: Finished STEP 600/50000, loss = 0.429073 (0.031 sec/batch), lr: 0.003000
295
- 2026-01-16 06:59:28 INFO: Evaluating on dev set...
296
- 2026-01-16 06:59:29 INFO: LAS MLAS BLEX
297
- 2026-01-16 06:59:29 INFO: 3.71 2.21 2.65
298
- 2026-01-16 06:59:29 INFO: step 600: train_loss = 1.109608, dev_score = 0.0371
299
- 2026-01-16 06:59:29 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
300
- 2026-01-16 06:59:29 INFO: new best model saved.
301
- 2026-01-16 06:59:30 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
302
- 2026-01-16 06:59:30 INFO: new model checkpoint saved.
303
- 2026-01-16 06:59:30 INFO: Finished STEP 620/50000, loss = 0.636850 (0.033 sec/batch), lr: 0.003000
304
- 2026-01-16 06:59:31 INFO: Finished STEP 640/50000, loss = 0.097528 (0.032 sec/batch), lr: 0.003000
305
- 2026-01-16 06:59:32 INFO: Finished STEP 660/50000, loss = 0.328014 (0.031 sec/batch), lr: 0.003000
306
- 2026-01-16 06:59:32 INFO: Finished STEP 680/50000, loss = 1.418195 (0.031 sec/batch), lr: 0.003000
307
- 2026-01-16 06:59:33 INFO: Finished STEP 700/50000, loss = 0.000002 (0.031 sec/batch), lr: 0.003000
308
- 2026-01-16 06:59:33 INFO: Evaluating on dev set...
309
- 2026-01-16 06:59:34 INFO: LAS MLAS BLEX
310
- 2026-01-16 06:59:34 INFO: 3.47 2.50 2.50
311
- 2026-01-16 06:59:34 INFO: step 700: train_loss = 1.152763, dev_score = 0.0347
312
- 2026-01-16 06:59:34 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
313
- 2026-01-16 06:59:34 INFO: new model checkpoint saved.
314
- 2026-01-16 06:59:35 INFO: Finished STEP 720/50000, loss = 3.007647 (0.031 sec/batch), lr: 0.003000
315
- 2026-01-16 06:59:35 INFO: Finished STEP 740/50000, loss = 1.066953 (0.031 sec/batch), lr: 0.003000
316
- 2026-01-16 06:59:36 INFO: Finished STEP 760/50000, loss = 1.151768 (0.030 sec/batch), lr: 0.003000
317
- 2026-01-16 06:59:37 INFO: Finished STEP 780/50000, loss = 1.043154 (0.032 sec/batch), lr: 0.003000
318
- 2026-01-16 06:59:37 INFO: Finished STEP 800/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
319
- 2026-01-16 06:59:37 INFO: Evaluating on dev set...
320
- 2026-01-16 06:59:38 INFO: LAS MLAS BLEX
321
- 2026-01-16 06:59:38 INFO: 3.71 2.21 2.58
322
- 2026-01-16 06:59:38 INFO: step 800: train_loss = 1.042035, dev_score = 0.0371
323
- 2026-01-16 06:59:38 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
324
- 2026-01-16 06:59:38 INFO: new best model saved.
325
- 2026-01-16 06:59:38 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
326
- 2026-01-16 06:59:38 INFO: new model checkpoint saved.
327
- 2026-01-16 06:59:39 INFO: Finished STEP 820/50000, loss = 0.309987 (0.031 sec/batch), lr: 0.003000
328
- 2026-01-16 06:59:40 INFO: Finished STEP 840/50000, loss = 0.046532 (0.030 sec/batch), lr: 0.003000
329
- 2026-01-16 06:59:40 INFO: Finished STEP 860/50000, loss = 2.719592 (0.031 sec/batch), lr: 0.003000
330
- 2026-01-16 06:59:41 INFO: Finished STEP 880/50000, loss = 1.216663 (0.030 sec/batch), lr: 0.003000
331
- 2026-01-16 06:59:42 INFO: Finished STEP 900/50000, loss = 1.323724 (0.032 sec/batch), lr: 0.003000
332
- 2026-01-16 06:59:42 INFO: Evaluating on dev set...
333
- 2026-01-16 06:59:42 INFO: LAS MLAS BLEX
334
- 2026-01-16 06:59:42 INFO: 3.71 2.71 3.05
335
- 2026-01-16 06:59:42 INFO: step 900: train_loss = 1.123192, dev_score = 0.0371
336
- 2026-01-16 06:59:42 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
337
- 2026-01-16 06:59:42 INFO: new best model saved.
338
- 2026-01-16 06:59:43 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
339
- 2026-01-16 06:59:43 INFO: new model checkpoint saved.
340
- 2026-01-16 06:59:43 INFO: Finished STEP 920/50000, loss = 0.152121 (0.031 sec/batch), lr: 0.003000
341
- 2026-01-16 06:59:44 INFO: Finished STEP 940/50000, loss = 0.897826 (0.031 sec/batch), lr: 0.003000
342
- 2026-01-16 06:59:45 INFO: Finished STEP 960/50000, loss = 0.000258 (0.030 sec/batch), lr: 0.003000
343
- 2026-01-16 06:59:45 INFO: Finished STEP 980/50000, loss = 2.134546 (0.030 sec/batch), lr: 0.003000
344
- 2026-01-16 06:59:46 INFO: Finished STEP 1000/50000, loss = 0.035686 (0.030 sec/batch), lr: 0.003000
345
- 2026-01-16 06:59:46 INFO: Evaluating on dev set...
346
- 2026-01-16 06:59:46 INFO: LAS MLAS BLEX
347
- 2026-01-16 06:59:46 INFO: 3.71 2.49 2.85
348
- 2026-01-16 06:59:46 INFO: step 1000: train_loss = 0.984753, dev_score = 0.0371
349
- 2026-01-16 06:59:47 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
350
- 2026-01-16 06:59:47 INFO: new best model saved.
351
- 2026-01-16 06:59:47 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
352
- 2026-01-16 06:59:47 INFO: new model checkpoint saved.
353
- 2026-01-16 06:59:48 INFO: Finished STEP 1020/50000, loss = 1.837077 (0.031 sec/batch), lr: 0.003000
354
- 2026-01-16 06:59:48 INFO: Finished STEP 1040/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
355
- 2026-01-16 06:59:49 INFO: Finished STEP 1060/50000, loss = 2.368280 (0.030 sec/batch), lr: 0.003000
356
- 2026-01-16 06:59:50 INFO: Finished STEP 1080/50000, loss = 0.469828 (0.031 sec/batch), lr: 0.003000
357
- 2026-01-16 06:59:50 INFO: Finished STEP 1100/50000, loss = 0.000491 (0.031 sec/batch), lr: 0.003000
358
- 2026-01-16 06:59:50 INFO: Evaluating on dev set...
359
- 2026-01-16 06:59:51 INFO: LAS MLAS BLEX
360
- 2026-01-16 06:59:51 INFO: 4.95 4.11 5.48
361
- 2026-01-16 06:59:51 INFO: step 1100: train_loss = 1.174486, dev_score = 0.0495
362
- 2026-01-16 06:59:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
363
- 2026-01-16 06:59:51 INFO: new best model saved.
364
- 2026-01-16 06:59:52 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
365
- 2026-01-16 06:59:52 INFO: new model checkpoint saved.
366
- 2026-01-16 06:59:52 INFO: Finished STEP 1120/50000, loss = 5.331606 (0.033 sec/batch), lr: 0.003000
367
- 2026-01-16 06:59:53 INFO: Finished STEP 1140/50000, loss = 4.487067 (0.032 sec/batch), lr: 0.003000
368
- 2026-01-16 06:59:54 INFO: Finished STEP 1160/50000, loss = 1.940725 (0.031 sec/batch), lr: 0.003000
369
- 2026-01-16 06:59:54 INFO: Finished STEP 1180/50000, loss = 2.472967 (0.032 sec/batch), lr: 0.003000
370
- 2026-01-16 06:59:55 INFO: Finished STEP 1200/50000, loss = 0.686162 (0.032 sec/batch), lr: 0.003000
371
- 2026-01-16 06:59:55 INFO: Evaluating on dev set...
372
- 2026-01-16 06:59:56 INFO: LAS MLAS BLEX
373
- 2026-01-16 06:59:56 INFO: 6.68 4.08 5.10
374
- 2026-01-16 06:59:56 INFO: step 1200: train_loss = 2.657063, dev_score = 0.0668
375
- 2026-01-16 06:59:56 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
376
- 2026-01-16 06:59:56 INFO: new best model saved.
377
- 2026-01-16 06:59:56 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
378
- 2026-01-16 06:59:56 INFO: new model checkpoint saved.
379
- 2026-01-16 06:59:57 INFO: Finished STEP 1220/50000, loss = 0.828940 (0.031 sec/batch), lr: 0.003000
380
- 2026-01-16 06:59:58 INFO: Finished STEP 1240/50000, loss = 1.596623 (0.032 sec/batch), lr: 0.003000
381
- 2026-01-16 06:59:58 INFO: Finished STEP 1260/50000, loss = 6.153142 (0.032 sec/batch), lr: 0.003000
382
- 2026-01-16 06:59:59 INFO: Finished STEP 1280/50000, loss = 1.988011 (0.031 sec/batch), lr: 0.003000
383
- 2026-01-16 07:00:00 INFO: Finished STEP 1300/50000, loss = 3.000342 (0.032 sec/batch), lr: 0.003000
384
- 2026-01-16 07:00:00 INFO: Evaluating on dev set...
385
- 2026-01-16 07:00:00 INFO: LAS MLAS BLEX
386
- 2026-01-16 07:00:00 INFO: 7.67 5.80 6.48
387
- 2026-01-16 07:00:00 INFO: step 1300: train_loss = 1.683935, dev_score = 0.0767
388
- 2026-01-16 07:00:00 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
389
- 2026-01-16 07:00:00 INFO: new best model saved.
390
- 2026-01-16 07:00:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
391
- 2026-01-16 07:00:01 INFO: new model checkpoint saved.
392
- 2026-01-16 07:00:02 INFO: Finished STEP 1320/50000, loss = 2.859746 (0.032 sec/batch), lr: 0.003000
393
- 2026-01-16 07:00:02 INFO: Finished STEP 1340/50000, loss = 1.214685 (0.032 sec/batch), lr: 0.003000
394
- 2026-01-16 07:00:03 INFO: Finished STEP 1360/50000, loss = 0.369348 (0.032 sec/batch), lr: 0.003000
395
- 2026-01-16 07:00:04 INFO: Finished STEP 1380/50000, loss = 1.401275 (0.032 sec/batch), lr: 0.003000
396
- 2026-01-16 07:00:04 INFO: Finished STEP 1400/50000, loss = 1.433550 (0.032 sec/batch), lr: 0.003000
397
- 2026-01-16 07:00:04 INFO: Evaluating on dev set...
398
- 2026-01-16 07:00:05 INFO: LAS MLAS BLEX
399
- 2026-01-16 07:00:05 INFO: 9.65 4.51 5.64
400
- 2026-01-16 07:00:05 INFO: step 1400: train_loss = 1.665584, dev_score = 0.0965
401
- 2026-01-16 07:00:05 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
402
- 2026-01-16 07:00:05 INFO: new best model saved.
403
- 2026-01-16 07:00:05 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
404
- 2026-01-16 07:00:05 INFO: new model checkpoint saved.
405
- 2026-01-16 07:00:06 INFO: Finished STEP 1420/50000, loss = 1.668130 (0.032 sec/batch), lr: 0.003000
406
- 2026-01-16 07:00:07 INFO: Finished STEP 1440/50000, loss = 0.647845 (0.031 sec/batch), lr: 0.003000
407
- 2026-01-16 07:00:07 INFO: Finished STEP 1460/50000, loss = 1.816998 (0.034 sec/batch), lr: 0.003000
408
- 2026-01-16 07:00:08 INFO: Finished STEP 1480/50000, loss = 1.645519 (0.032 sec/batch), lr: 0.003000
409
- 2026-01-16 07:00:09 INFO: Finished STEP 1500/50000, loss = 3.368581 (0.032 sec/batch), lr: 0.003000
410
- 2026-01-16 07:00:09 INFO: Evaluating on dev set...
411
- 2026-01-16 07:00:10 INFO: LAS MLAS BLEX
412
- 2026-01-16 07:00:10 INFO: 11.14 6.09 6.81
413
- 2026-01-16 07:00:10 INFO: step 1500: train_loss = 1.689450, dev_score = 0.1114
414
- 2026-01-16 07:00:10 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
415
- 2026-01-16 07:00:10 INFO: new best model saved.
416
- 2026-01-16 07:00:10 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
417
- 2026-01-16 07:00:10 INFO: new model checkpoint saved.
418
- 2026-01-16 07:00:11 INFO: Finished STEP 1520/50000, loss = 0.696790 (0.033 sec/batch), lr: 0.003000
419
- 2026-01-16 07:00:12 INFO: Finished STEP 1540/50000, loss = 0.979858 (0.032 sec/batch), lr: 0.003000
420
- 2026-01-16 07:00:12 INFO: Finished STEP 1560/50000, loss = 1.371460 (0.032 sec/batch), lr: 0.003000
421
- 2026-01-16 07:00:13 INFO: Finished STEP 1580/50000, loss = 1.076014 (0.031 sec/batch), lr: 0.003000
422
- 2026-01-16 07:00:14 INFO: Finished STEP 1600/50000, loss = 0.347749 (0.031 sec/batch), lr: 0.003000
423
- 2026-01-16 07:00:14 INFO: Evaluating on dev set...
424
- 2026-01-16 07:00:14 INFO: LAS MLAS BLEX
425
- 2026-01-16 07:00:14 INFO: 18.07 4.21 5.35
426
- 2026-01-16 07:00:14 INFO: step 1600: train_loss = 1.549157, dev_score = 0.1807
427
- 2026-01-16 07:00:14 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
428
- 2026-01-16 07:00:14 INFO: new best model saved.
429
- 2026-01-16 07:00:15 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
430
- 2026-01-16 07:00:15 INFO: new model checkpoint saved.
431
- 2026-01-16 07:00:16 INFO: Finished STEP 1620/50000, loss = 0.799195 (0.032 sec/batch), lr: 0.003000
432
- 2026-01-16 07:00:16 INFO: Finished STEP 1640/50000, loss = 1.727682 (0.032 sec/batch), lr: 0.003000
433
- 2026-01-16 07:00:17 INFO: Finished STEP 1660/50000, loss = 2.875053 (0.033 sec/batch), lr: 0.003000
434
- 2026-01-16 07:00:18 INFO: Finished STEP 1680/50000, loss = 2.254138 (0.031 sec/batch), lr: 0.003000
435
- 2026-01-16 07:00:18 INFO: Finished STEP 1700/50000, loss = 0.754430 (0.031 sec/batch), lr: 0.003000
436
- 2026-01-16 07:00:18 INFO: Evaluating on dev set...
437
- 2026-01-16 07:00:19 INFO: LAS MLAS BLEX
438
- 2026-01-16 07:00:19 INFO: 11.39 4.62 5.01
439
- 2026-01-16 07:00:19 INFO: step 1700: train_loss = 1.452840, dev_score = 0.1139
440
- 2026-01-16 07:00:19 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
441
- 2026-01-16 07:00:19 INFO: new model checkpoint saved.
442
- 2026-01-16 07:00:20 INFO: Finished STEP 1720/50000, loss = 0.672960 (0.032 sec/batch), lr: 0.003000
443
- 2026-01-16 07:00:21 INFO: Finished STEP 1740/50000, loss = 1.758359 (0.031 sec/batch), lr: 0.003000
444
- 2026-01-16 07:00:21 INFO: Finished STEP 1760/50000, loss = 0.112267 (0.032 sec/batch), lr: 0.003000
445
- 2026-01-16 07:00:22 INFO: Finished STEP 1780/50000, loss = 2.130425 (0.035 sec/batch), lr: 0.003000
446
- 2026-01-16 07:00:23 INFO: Finished STEP 1800/50000, loss = 0.867664 (0.032 sec/batch), lr: 0.003000
447
- 2026-01-16 07:00:23 INFO: Evaluating on dev set...
448
- 2026-01-16 07:00:23 INFO: LAS MLAS BLEX
449
- 2026-01-16 07:00:23 INFO: 8.17 5.17 5.52
450
- 2026-01-16 07:00:23 INFO: step 1800: train_loss = 1.566285, dev_score = 0.0817
451
- 2026-01-16 07:00:24 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
452
- 2026-01-16 07:00:24 INFO: new model checkpoint saved.
453
- 2026-01-16 07:00:24 INFO: Finished STEP 1820/50000, loss = 0.918402 (0.032 sec/batch), lr: 0.003000
454
- 2026-01-16 07:00:25 INFO: Finished STEP 1840/50000, loss = 0.602067 (0.031 sec/batch), lr: 0.003000
455
- 2026-01-16 07:00:26 INFO: Finished STEP 1860/50000, loss = 4.889768 (0.031 sec/batch), lr: 0.003000
456
- 2026-01-16 07:00:26 INFO: Finished STEP 1880/50000, loss = 3.249488 (0.037 sec/batch), lr: 0.003000
457
- 2026-01-16 07:00:27 INFO: Finished STEP 1900/50000, loss = 1.829388 (0.032 sec/batch), lr: 0.003000
458
- 2026-01-16 07:00:27 INFO: Evaluating on dev set...
459
- 2026-01-16 07:00:28 INFO: LAS MLAS BLEX
460
- 2026-01-16 07:00:28 INFO: 9.16 3.77 5.28
461
- 2026-01-16 07:00:28 INFO: step 1900: train_loss = 1.531276, dev_score = 0.0916
462
- 2026-01-16 07:00:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
463
- 2026-01-16 07:00:28 INFO: new model checkpoint saved.
464
- 2026-01-16 07:00:29 INFO: Finished STEP 1920/50000, loss = 1.377406 (0.033 sec/batch), lr: 0.003000
465
- 2026-01-16 07:00:29 INFO: Finished STEP 1940/50000, loss = 0.690662 (0.032 sec/batch), lr: 0.003000
466
- 2026-01-16 07:00:30 INFO: Finished STEP 1960/50000, loss = 0.488795 (0.032 sec/batch), lr: 0.003000
467
- 2026-01-16 07:00:31 INFO: Finished STEP 1980/50000, loss = 1.489328 (0.031 sec/batch), lr: 0.003000
468
- 2026-01-16 07:00:31 INFO: Finished STEP 2000/50000, loss = 0.265036 (0.032 sec/batch), lr: 0.003000
469
- 2026-01-16 07:00:31 INFO: Evaluating on dev set...
470
- 2026-01-16 07:00:32 INFO: LAS MLAS BLEX
471
- 2026-01-16 07:00:32 INFO: 3.71 2.34 3.90
472
- 2026-01-16 07:00:32 INFO: step 2000: train_loss = 1.577713, dev_score = 0.0371
473
- 2026-01-16 07:00:32 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
474
- 2026-01-16 07:00:32 INFO: new model checkpoint saved.
475
- 2026-01-16 07:00:33 INFO: Finished STEP 2020/50000, loss = 1.118794 (0.031 sec/batch), lr: 0.003000
476
- 2026-01-16 07:00:34 INFO: Finished STEP 2040/50000, loss = 1.678119 (0.032 sec/batch), lr: 0.003000
477
- 2026-01-16 07:00:34 INFO: Finished STEP 2060/50000, loss = 0.566802 (0.032 sec/batch), lr: 0.003000
478
- 2026-01-16 07:00:35 INFO: Finished STEP 2080/50000, loss = 1.566789 (0.032 sec/batch), lr: 0.003000
479
- 2026-01-16 07:00:36 INFO: Finished STEP 2100/50000, loss = 1.108176 (0.031 sec/batch), lr: 0.003000
480
- 2026-01-16 07:00:36 INFO: Evaluating on dev set...
481
- 2026-01-16 07:00:36 INFO: LAS MLAS BLEX
482
- 2026-01-16 07:00:36 INFO: 8.17 5.37 6.91
483
- 2026-01-16 07:00:36 INFO: step 2100: train_loss = 1.353945, dev_score = 0.0817
484
- 2026-01-16 07:00:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
485
- 2026-01-16 07:00:37 INFO: new model checkpoint saved.
486
- 2026-01-16 07:00:37 INFO: Finished STEP 2120/50000, loss = 0.896808 (0.033 sec/batch), lr: 0.003000
487
- 2026-01-16 07:00:38 INFO: Finished STEP 2140/50000, loss = 3.207397 (0.032 sec/batch), lr: 0.003000
488
- 2026-01-16 07:00:39 INFO: Finished STEP 2160/50000, loss = 0.004755 (0.037 sec/batch), lr: 0.003000
489
- 2026-01-16 07:00:39 INFO: Finished STEP 2180/50000, loss = 0.760959 (0.031 sec/batch), lr: 0.003000
490
- 2026-01-16 07:00:40 INFO: Finished STEP 2200/50000, loss = 2.696491 (0.032 sec/batch), lr: 0.003000
491
- 2026-01-16 07:00:40 INFO: Evaluating on dev set...
492
- 2026-01-16 07:00:41 INFO: LAS MLAS BLEX
493
- 2026-01-16 07:00:41 INFO: 7.67 1.83 2.19
494
- 2026-01-16 07:00:41 INFO: step 2200: train_loss = 1.574935, dev_score = 0.0767
495
- 2026-01-16 07:00:41 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
496
- 2026-01-16 07:00:41 INFO: new model checkpoint saved.
497
- 2026-01-16 07:00:42 INFO: Finished STEP 2220/50000, loss = 2.683285 (0.031 sec/batch), lr: 0.003000
498
- 2026-01-16 07:00:42 INFO: Finished STEP 2240/50000, loss = 2.951740 (0.031 sec/batch), lr: 0.003000
499
- 2026-01-16 07:00:43 INFO: Finished STEP 2260/50000, loss = 4.218589 (0.031 sec/batch), lr: 0.003000
500
- 2026-01-16 07:00:44 INFO: Finished STEP 2280/50000, loss = 1.695766 (0.031 sec/batch), lr: 0.003000
501
- 2026-01-16 07:00:44 INFO: Finished STEP 2300/50000, loss = 2.362932 (0.031 sec/batch), lr: 0.003000
502
- 2026-01-16 07:00:44 INFO: Evaluating on dev set...
503
- 2026-01-16 07:00:45 INFO: LAS MLAS BLEX
504
- 2026-01-16 07:00:45 INFO: 12.62 3.67 4.48
505
- 2026-01-16 07:00:45 INFO: step 2300: train_loss = 3.007224, dev_score = 0.1262
506
- 2026-01-16 07:00:46 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
507
- 2026-01-16 07:00:46 INFO: new model checkpoint saved.
508
- 2026-01-16 07:00:46 INFO: Finished STEP 2320/50000, loss = 2.829835 (0.031 sec/batch), lr: 0.003000
509
- 2026-01-16 07:00:47 INFO: Finished STEP 2340/50000, loss = 1.246457 (0.031 sec/batch), lr: 0.003000
510
- 2026-01-16 07:00:47 INFO: Finished STEP 2360/50000, loss = 1.362900 (0.032 sec/batch), lr: 0.003000
511
- 2026-01-16 07:00:48 INFO: Finished STEP 2380/50000, loss = 1.726964 (0.031 sec/batch), lr: 0.003000
512
- 2026-01-16 07:00:49 INFO: Finished STEP 2400/50000, loss = 1.644795 (0.031 sec/batch), lr: 0.003000
513
- 2026-01-16 07:00:49 INFO: Evaluating on dev set...
514
- 2026-01-16 07:00:49 INFO: LAS MLAS BLEX
515
- 2026-01-16 07:00:49 INFO: 11.14 6.71 7.77
516
- 2026-01-16 07:00:49 INFO: step 2400: train_loss = 2.391629, dev_score = 0.1114
517
- 2026-01-16 07:00:50 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
518
- 2026-01-16 07:00:50 INFO: new model checkpoint saved.
519
- 2026-01-16 07:00:50 INFO: Finished STEP 2420/50000, loss = 1.448694 (0.030 sec/batch), lr: 0.003000
520
- 2026-01-16 07:00:51 INFO: Finished STEP 2440/50000, loss = 0.224588 (0.030 sec/batch), lr: 0.003000
521
- 2026-01-16 07:00:52 INFO: Finished STEP 2460/50000, loss = 1.348104 (0.031 sec/batch), lr: 0.003000
522
- 2026-01-16 07:00:52 INFO: Finished STEP 2480/50000, loss = 1.120497 (0.031 sec/batch), lr: 0.003000
523
- 2026-01-16 07:00:53 INFO: Finished STEP 2500/50000, loss = 2.224450 (0.032 sec/batch), lr: 0.003000
524
- 2026-01-16 07:00:53 INFO: Evaluating on dev set...
525
- 2026-01-16 07:00:53 INFO: LAS MLAS BLEX
526
- 2026-01-16 07:00:53 INFO: 18.56 6.11 7.63
527
- 2026-01-16 07:00:53 INFO: step 2500: train_loss = 2.279046, dev_score = 0.1856
528
- 2026-01-16 07:00:54 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
529
- 2026-01-16 07:00:54 INFO: new best model saved.
530
- 2026-01-16 07:00:54 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
531
- 2026-01-16 07:00:54 INFO: new model checkpoint saved.
532
- 2026-01-16 07:00:55 INFO: Finished STEP 2520/50000, loss = 4.519064 (0.031 sec/batch), lr: 0.003000
533
- 2026-01-16 07:00:55 INFO: Finished STEP 2540/50000, loss = 3.034460 (0.034 sec/batch), lr: 0.003000
534
- 2026-01-16 07:00:56 INFO: Finished STEP 2560/50000, loss = 2.673581 (0.032 sec/batch), lr: 0.003000
535
- 2026-01-16 07:00:57 INFO: Finished STEP 2580/50000, loss = 0.801564 (0.030 sec/batch), lr: 0.003000
536
- 2026-01-16 07:00:57 INFO: Finished STEP 2600/50000, loss = 0.521267 (0.032 sec/batch), lr: 0.003000
537
- 2026-01-16 07:00:57 INFO: Evaluating on dev set...
538
- 2026-01-16 07:00:58 INFO: LAS MLAS BLEX
539
- 2026-01-16 07:00:58 INFO: 18.56 6.01 7.21
540
- 2026-01-16 07:00:58 INFO: step 2600: train_loss = 2.528788, dev_score = 0.1856
541
- 2026-01-16 07:00:58 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
542
- 2026-01-16 07:00:58 INFO: new best model saved.
543
- 2026-01-16 07:00:59 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
544
- 2026-01-16 07:00:59 INFO: new model checkpoint saved.
545
- 2026-01-16 07:00:59 INFO: Finished STEP 2620/50000, loss = 1.897941 (0.031 sec/batch), lr: 0.003000
546
- 2026-01-16 07:01:00 INFO: Finished STEP 2640/50000, loss = 2.104889 (0.031 sec/batch), lr: 0.003000
547
- 2026-01-16 07:01:00 INFO: Finished STEP 2660/50000, loss = 2.285084 (0.031 sec/batch), lr: 0.003000
548
- 2026-01-16 07:01:01 INFO: Finished STEP 2680/50000, loss = 2.218475 (0.031 sec/batch), lr: 0.003000
549
- 2026-01-16 07:01:02 INFO: Finished STEP 2700/50000, loss = 6.859745 (0.031 sec/batch), lr: 0.003000
550
- 2026-01-16 07:01:02 INFO: Evaluating on dev set...
551
- 2026-01-16 07:01:02 INFO: LAS MLAS BLEX
552
- 2026-01-16 07:01:02 INFO: 19.55 6.19 8.12
553
- 2026-01-16 07:01:02 INFO: step 2700: train_loss = 2.664627, dev_score = 0.1955
554
- 2026-01-16 07:01:03 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
555
- 2026-01-16 07:01:03 INFO: new best model saved.
556
- 2026-01-16 07:01:03 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
557
- 2026-01-16 07:01:03 INFO: new model checkpoint saved.
558
- 2026-01-16 07:01:04 INFO: Finished STEP 2720/50000, loss = 2.067003 (0.031 sec/batch), lr: 0.003000
559
- 2026-01-16 07:01:04 INFO: Finished STEP 2740/50000, loss = 1.080532 (0.031 sec/batch), lr: 0.003000
560
- 2026-01-16 07:01:05 INFO: Finished STEP 2760/50000, loss = 4.616343 (0.032 sec/batch), lr: 0.003000
561
- 2026-01-16 07:01:06 INFO: Finished STEP 2780/50000, loss = 1.479079 (0.032 sec/batch), lr: 0.003000
562
- 2026-01-16 07:01:06 INFO: Finished STEP 2800/50000, loss = 1.063658 (0.032 sec/batch), lr: 0.003000
563
- 2026-01-16 07:01:06 INFO: Evaluating on dev set...
564
- 2026-01-16 07:01:07 INFO: LAS MLAS BLEX
565
- 2026-01-16 07:01:07 INFO: 17.82 4.66 6.21
566
- 2026-01-16 07:01:07 INFO: step 2800: train_loss = 2.274723, dev_score = 0.1782
567
- 2026-01-16 07:01:08 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
568
- 2026-01-16 07:01:08 INFO: new model checkpoint saved.
569
- 2026-01-16 07:01:08 INFO: Finished STEP 2820/50000, loss = 2.248970 (0.032 sec/batch), lr: 0.003000
570
- 2026-01-16 07:01:09 INFO: Finished STEP 2840/50000, loss = 3.309697 (0.031 sec/batch), lr: 0.003000
571
- 2026-01-16 07:01:10 INFO: Finished STEP 2860/50000, loss = 0.586270 (0.032 sec/batch), lr: 0.003000
572
- 2026-01-16 07:01:10 INFO: Finished STEP 2880/50000, loss = 1.731944 (0.031 sec/batch), lr: 0.003000
573
- 2026-01-16 07:01:11 INFO: Finished STEP 2900/50000, loss = 2.552610 (0.030 sec/batch), lr: 0.003000
574
- 2026-01-16 07:01:11 INFO: Evaluating on dev set...
575
- 2026-01-16 07:01:11 INFO: LAS MLAS BLEX
576
- 2026-01-16 07:01:11 INFO: 17.33 5.02 6.95
577
- 2026-01-16 07:01:11 INFO: step 2900: train_loss = 2.368524, dev_score = 0.1733
578
- 2026-01-16 07:01:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
579
- 2026-01-16 07:01:12 INFO: new model checkpoint saved.
580
- 2026-01-16 07:01:13 INFO: Finished STEP 2920/50000, loss = 3.126576 (0.031 sec/batch), lr: 0.003000
581
- 2026-01-16 07:01:13 INFO: Finished STEP 2940/50000, loss = 1.741402 (0.031 sec/batch), lr: 0.003000
582
- 2026-01-16 07:01:14 INFO: Finished STEP 2960/50000, loss = 3.777675 (0.031 sec/batch), lr: 0.003000
583
- 2026-01-16 07:01:15 INFO: Finished STEP 2980/50000, loss = 1.370117 (0.031 sec/batch), lr: 0.003000
584
- 2026-01-16 07:01:15 INFO: Finished STEP 3000/50000, loss = 3.574947 (0.031 sec/batch), lr: 0.003000
585
- 2026-01-16 07:01:15 INFO: Evaluating on dev set...
586
- 2026-01-16 07:01:16 INFO: LAS MLAS BLEX
587
- 2026-01-16 07:01:16 INFO: 17.82 6.94 8.48
588
- 2026-01-16 07:01:16 INFO: step 3000: train_loss = 2.539027, dev_score = 0.1782
589
- 2026-01-16 07:01:16 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
590
- 2026-01-16 07:01:16 INFO: new model checkpoint saved.
591
- 2026-01-16 07:01:17 INFO: Finished STEP 3020/50000, loss = 1.453705 (0.031 sec/batch), lr: 0.003000
592
- 2026-01-16 07:01:18 INFO: Finished STEP 3040/50000, loss = 6.184045 (0.032 sec/batch), lr: 0.003000
593
- 2026-01-16 07:01:18 INFO: Finished STEP 3060/50000, loss = 4.618065 (0.030 sec/batch), lr: 0.003000
594
- 2026-01-16 07:01:19 INFO: Finished STEP 3080/50000, loss = 2.472624 (0.032 sec/batch), lr: 0.003000
595
- 2026-01-16 07:01:20 INFO: Finished STEP 3100/50000, loss = 2.051110 (0.032 sec/batch), lr: 0.003000
596
- 2026-01-16 07:01:20 INFO: Evaluating on dev set...
597
- 2026-01-16 07:01:20 INFO: LAS MLAS BLEX
598
- 2026-01-16 07:01:20 INFO: 23.51 9.39 13.88
599
- 2026-01-16 07:01:20 INFO: step 3100: train_loss = 2.557801, dev_score = 0.2351
600
- 2026-01-16 07:01:20 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
601
- 2026-01-16 07:01:20 INFO: new best model saved.
602
- 2026-01-16 07:01:21 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
603
- 2026-01-16 07:01:21 INFO: new model checkpoint saved.
604
- 2026-01-16 07:01:21 INFO: Finished STEP 3120/50000, loss = 1.929233 (0.031 sec/batch), lr: 0.003000
605
- 2026-01-16 07:01:22 INFO: Finished STEP 3140/50000, loss = 1.142895 (0.032 sec/batch), lr: 0.003000
606
- 2026-01-16 07:01:23 INFO: Finished STEP 3160/50000, loss = 2.250463 (0.031 sec/batch), lr: 0.003000
607
- 2026-01-16 07:01:23 INFO: Finished STEP 3180/50000, loss = 1.848980 (0.031 sec/batch), lr: 0.003000
608
- 2026-01-16 07:01:24 INFO: Finished STEP 3200/50000, loss = 1.592728 (0.031 sec/batch), lr: 0.003000
609
- 2026-01-16 07:01:24 INFO: Evaluating on dev set...
610
- 2026-01-16 07:01:24 INFO: LAS MLAS BLEX
611
- 2026-01-16 07:01:24 INFO: 14.11 7.38 10.10
612
- 2026-01-16 07:01:24 INFO: step 3200: train_loss = 2.549193, dev_score = 0.1411
613
- 2026-01-16 07:01:25 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
614
- 2026-01-16 07:01:25 INFO: new model checkpoint saved.
615
- 2026-01-16 07:01:26 INFO: Finished STEP 3220/50000, loss = 1.161050 (0.031 sec/batch), lr: 0.003000
616
- 2026-01-16 07:01:26 INFO: Finished STEP 3240/50000, loss = 2.996624 (0.031 sec/batch), lr: 0.003000
617
- 2026-01-16 07:01:27 INFO: Finished STEP 3260/50000, loss = 1.401396 (0.031 sec/batch), lr: 0.003000
618
- 2026-01-16 07:01:28 INFO: Finished STEP 3280/50000, loss = 3.537535 (0.032 sec/batch), lr: 0.003000
619
- 2026-01-16 07:01:28 INFO: Finished STEP 3300/50000, loss = 2.966022 (0.031 sec/batch), lr: 0.003000
620
- 2026-01-16 07:01:28 INFO: Evaluating on dev set...
621
- 2026-01-16 07:01:29 INFO: LAS MLAS BLEX
622
- 2026-01-16 07:01:29 INFO: 21.78 6.75 7.54
623
- 2026-01-16 07:01:29 INFO: step 3300: train_loss = 2.125264, dev_score = 0.2178
624
- 2026-01-16 07:01:29 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
625
- 2026-01-16 07:01:29 INFO: new model checkpoint saved.
626
- 2026-01-16 07:01:30 INFO: Finished STEP 3320/50000, loss = 1.093041 (0.032 sec/batch), lr: 0.003000
627
- 2026-01-16 07:01:30 INFO: Finished STEP 3340/50000, loss = 4.219791 (0.031 sec/batch), lr: 0.003000
628
- 2026-01-16 07:01:31 INFO: Finished STEP 3360/50000, loss = 1.543054 (0.032 sec/batch), lr: 0.003000
629
- 2026-01-16 07:01:32 INFO: Finished STEP 3380/50000, loss = 1.405899 (0.031 sec/batch), lr: 0.003000
630
- 2026-01-16 07:01:32 INFO: Finished STEP 3400/50000, loss = 1.838370 (0.032 sec/batch), lr: 0.003000
631
- 2026-01-16 07:01:32 INFO: Evaluating on dev set...
632
- 2026-01-16 07:01:33 INFO: LAS MLAS BLEX
633
- 2026-01-16 07:01:33 INFO: 11.39 5.43 9.69
634
- 2026-01-16 07:01:33 INFO: step 3400: train_loss = 2.268760, dev_score = 0.1139
635
- 2026-01-16 07:01:33 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
636
- 2026-01-16 07:01:33 INFO: new model checkpoint saved.
637
- 2026-01-16 07:01:34 INFO: Finished STEP 3420/50000, loss = 2.768859 (0.032 sec/batch), lr: 0.003000
638
- 2026-01-16 07:01:35 INFO: Finished STEP 3440/50000, loss = 1.197934 (0.031 sec/batch), lr: 0.003000
639
- 2026-01-16 07:01:35 INFO: Finished STEP 3460/50000, loss = 1.991076 (0.034 sec/batch), lr: 0.003000
640
- 2026-01-16 07:01:36 INFO: Finished STEP 3480/50000, loss = 1.895437 (0.032 sec/batch), lr: 0.003000
641
- 2026-01-16 07:01:37 INFO: Finished STEP 3500/50000, loss = 1.187534 (0.030 sec/batch), lr: 0.003000
642
- 2026-01-16 07:01:37 INFO: Evaluating on dev set...
643
- 2026-01-16 07:01:37 INFO: LAS MLAS BLEX
644
- 2026-01-16 07:01:37 INFO: 19.06 10.51 15.76
645
- 2026-01-16 07:01:37 INFO: step 3500: train_loss = 2.106533, dev_score = 0.1906
646
- 2026-01-16 07:01:38 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
647
- 2026-01-16 07:01:38 INFO: new model checkpoint saved.
648
- 2026-01-16 07:01:38 INFO: Finished STEP 3520/50000, loss = 1.585278 (0.032 sec/batch), lr: 0.003000
649
- 2026-01-16 07:01:39 INFO: Finished STEP 3540/50000, loss = 1.423405 (0.031 sec/batch), lr: 0.003000
650
- 2026-01-16 07:01:39 INFO: Finished STEP 3560/50000, loss = 0.421177 (0.030 sec/batch), lr: 0.003000
651
- 2026-01-16 07:01:40 INFO: Finished STEP 3580/50000, loss = 1.731428 (0.031 sec/batch), lr: 0.003000
652
- 2026-01-16 07:01:41 INFO: Finished STEP 3600/50000, loss = 0.526926 (0.031 sec/batch), lr: 0.003000
653
- 2026-01-16 07:01:41 INFO: Evaluating on dev set...
654
- 2026-01-16 07:01:41 INFO: LAS MLAS BLEX
655
- 2026-01-16 07:01:41 INFO: 19.55 8.20 10.55
656
- 2026-01-16 07:01:41 INFO: step 3600: train_loss = 2.619668, dev_score = 0.1955
657
- 2026-01-16 07:01:42 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
658
- 2026-01-16 07:01:42 INFO: new model checkpoint saved.
659
- 2026-01-16 07:01:42 INFO: Finished STEP 3620/50000, loss = 0.616525 (0.030 sec/batch), lr: 0.003000
660
- 2026-01-16 07:01:43 INFO: Finished STEP 3640/50000, loss = 1.884155 (0.030 sec/batch), lr: 0.003000
661
- 2026-01-16 07:01:44 INFO: Finished STEP 3660/50000, loss = 2.431090 (0.030 sec/batch), lr: 0.003000
662
- 2026-01-16 07:01:44 INFO: Finished STEP 3680/50000, loss = 4.406909 (0.029 sec/batch), lr: 0.003000
663
- 2026-01-16 07:01:45 INFO: Finished STEP 3700/50000, loss = 1.068943 (0.030 sec/batch), lr: 0.003000
664
- 2026-01-16 07:01:45 INFO: Evaluating on dev set...
665
- 2026-01-16 07:01:45 INFO: LAS MLAS BLEX
666
- 2026-01-16 07:01:45 INFO: 10.15 6.90 9.20
667
- 2026-01-16 07:01:45 INFO: step 3700: train_loss = 2.132062, dev_score = 0.1015
668
- 2026-01-16 07:01:46 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
669
- 2026-01-16 07:01:46 INFO: new model checkpoint saved.
670
- 2026-01-16 07:01:46 INFO: Finished STEP 3720/50000, loss = 1.305085 (0.030 sec/batch), lr: 0.003000
671
- 2026-01-16 07:01:47 INFO: Finished STEP 3740/50000, loss = 0.134945 (0.031 sec/batch), lr: 0.003000
672
- 2026-01-16 07:01:48 INFO: Finished STEP 3760/50000, loss = 0.634447 (0.030 sec/batch), lr: 0.003000
673
- 2026-01-16 07:01:48 INFO: Finished STEP 3780/50000, loss = 1.520501 (0.030 sec/batch), lr: 0.003000
674
- 2026-01-16 07:01:49 INFO: Finished STEP 3800/50000, loss = 2.499804 (0.030 sec/batch), lr: 0.003000
675
- 2026-01-16 07:01:49 INFO: Evaluating on dev set...
676
- 2026-01-16 07:01:49 INFO: LAS MLAS BLEX
677
- 2026-01-16 07:01:49 INFO: 12.87 8.68 11.44
678
- 2026-01-16 07:01:49 INFO: step 3800: train_loss = 2.251492, dev_score = 0.1287
679
- 2026-01-16 07:01:50 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
680
- 2026-01-16 07:01:50 INFO: new model checkpoint saved.
681
- 2026-01-16 07:01:50 INFO: Finished STEP 3820/50000, loss = 4.027611 (0.030 sec/batch), lr: 0.003000
682
- 2026-01-16 07:01:51 INFO: Finished STEP 3840/50000, loss = 0.331204 (0.029 sec/batch), lr: 0.003000
683
- 2026-01-16 07:01:52 INFO: Finished STEP 3860/50000, loss = 1.255922 (0.029 sec/batch), lr: 0.003000
684
- 2026-01-16 07:01:52 INFO: Finished STEP 3880/50000, loss = 1.898022 (0.030 sec/batch), lr: 0.003000
685
- 2026-01-16 07:01:53 INFO: Finished STEP 3900/50000, loss = 1.984990 (0.029 sec/batch), lr: 0.003000
686
- 2026-01-16 07:01:53 INFO: Evaluating on dev set...
687
- 2026-01-16 07:01:53 INFO: LAS MLAS BLEX
688
- 2026-01-16 07:01:53 INFO: 14.36 8.71 9.90
689
- 2026-01-16 07:01:53 INFO: step 3900: train_loss = 2.995180, dev_score = 0.1436
690
- 2026-01-16 07:01:54 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
691
- 2026-01-16 07:01:54 INFO: new model checkpoint saved.
692
- 2026-01-16 07:01:54 INFO: Finished STEP 3920/50000, loss = 2.342984 (0.031 sec/batch), lr: 0.003000
693
- 2026-01-16 07:01:55 INFO: Finished STEP 3940/50000, loss = 0.587340 (0.030 sec/batch), lr: 0.003000
694
- 2026-01-16 07:01:56 INFO: Finished STEP 3960/50000, loss = 1.429672 (0.029 sec/batch), lr: 0.003000
695
- 2026-01-16 07:01:56 INFO: Finished STEP 3980/50000, loss = 1.932409 (0.030 sec/batch), lr: 0.003000
696
- 2026-01-16 07:01:57 INFO: Finished STEP 4000/50000, loss = 2.074286 (0.031 sec/batch), lr: 0.003000
697
- 2026-01-16 07:01:57 INFO: Evaluating on dev set...
698
- 2026-01-16 07:01:57 INFO: LAS MLAS BLEX
699
- 2026-01-16 07:01:57 INFO: 13.61 7.11 9.49
700
- 2026-01-16 07:01:57 INFO: step 4000: train_loss = 2.342431, dev_score = 0.1361
701
- 2026-01-16 07:01:58 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
702
- 2026-01-16 07:01:58 INFO: new model checkpoint saved.
703
- 2026-01-16 07:01:58 INFO: Finished STEP 4020/50000, loss = 0.798264 (0.029 sec/batch), lr: 0.003000
704
- 2026-01-16 07:01:59 INFO: Finished STEP 4040/50000, loss = 0.128710 (0.029 sec/batch), lr: 0.003000
705
- 2026-01-16 07:02:00 INFO: Finished STEP 4060/50000, loss = 0.836676 (0.029 sec/batch), lr: 0.003000
706
- 2026-01-16 07:02:00 INFO: Finished STEP 4080/50000, loss = 1.385599 (0.030 sec/batch), lr: 0.003000
707
- 2026-01-16 07:02:01 INFO: Finished STEP 4100/50000, loss = 1.616785 (0.032 sec/batch), lr: 0.003000
708
- 2026-01-16 07:02:01 INFO: Evaluating on dev set...
709
- 2026-01-16 07:02:01 INFO: LAS MLAS BLEX
710
- 2026-01-16 07:02:01 INFO: 17.33 6.88 8.91
711
- 2026-01-16 07:02:01 INFO: step 4100: train_loss = 2.344978, dev_score = 0.1733
712
- 2026-01-16 07:02:01 INFO: Training ended with 4100 steps.
713
- 2026-01-16 07:02:01 INFO: Best dev F1 = 23.51, at iteration = 3100
714
- 2026-01-16 07:02:02 INFO: Running dev depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '8', '--dropout', '0.33']
715
- 2026-01-16 07:02:02 INFO: Running parser in predict mode
716
- 2026-01-16 07:02:02 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
717
- 2026-01-16 07:02:04 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
718
- 2026-01-16 07:02:04 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
719
- 2026-01-16 07:02:04 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
720
- 2026-01-16 07:02:04 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
721
- 2026-01-16 07:02:04 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
722
- 2026-01-16 07:02:04 INFO: Loading data with batch size 8...
723
- 2026-01-16 07:02:04 DEBUG: 9 batches created.
724
- 2026-01-16 07:02:05 INFO: F1 scores for each dependency:
725
- Note that unlabeled attachment errors hurt the labeled attachment scores
726
- acl: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
727
- acl:relcl: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
728
- advcl: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
729
- advmod: p 0.1000 r 0.1200 f1 0.1091 (25 actual)
730
- amod: p 0.0000 r 0.0000 f1 0.0000 (31 actual)
731
- appos: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
732
- aux: p 0.0000 r 0.0000 f1 0.0000 (11 actual)
733
- case: p 0.4557 r 0.6429 f1 0.5333 (56 actual)
734
- cc: p 0.0000 r 0.0000 f1 0.0000 (13 actual)
735
- ccomp: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
736
- conj: p 0.0000 r 0.0000 f1 0.0000 (12 actual)
737
- cop: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
738
- csubj: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
739
- det: p 0.5000 r 0.4091 f1 0.4500 (22 actual)
740
- expl: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
741
- flat:name: p 0.0000 r 0.0000 f1 0.0000 (0 actual)
742
- iobj: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
743
- mark: p 0.0000 r 0.0000 f1 0.0000 (12 actual)
744
- nmod: p 0.0000 r 0.0000 f1 0.0000 (15 actual)
745
- nmod:poss: p 0.0000 r 0.0000 f1 0.0000 (19 actual)
746
- nsubj: p 0.0641 r 0.5882 f1 0.1156 (17 actual)
747
- nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
748
- obj: p 0.0000 r 0.0000 f1 0.0000 (22 actual)
749
- obl: p 0.2373 r 0.3415 f1 0.2800 (41 actual)
750
- obl:agent: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
751
- orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
752
- parataxis: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
753
- punct: p 0.3077 r 0.3077 f1 0.3077 (52 actual)
754
- root: p 0.5556 r 0.5556 f1 0.5556 (9 actual)
755
- xcomp: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
756
- 2026-01-16 07:02:05 INFO: LAS MLAS BLEX
757
- 2026-01-16 07:02:05 INFO: 23.51 9.39 13.88
758
- 2026-01-16 07:02:05 INFO: Parser score:
759
- 2026-01-16 07:02:05 INFO: sv_diachronic 23.51
760
- 2026-01-16 07:02:05 INFO: Finished running dev set on
761
- UD_Swedish-diachronic
762
- UAS LAS CLAS MLAS BLEX
763
- 48.27 23.51 13.88 9.39 13.88
764
- 2026-01-16 07:02:05 INFO: Running test depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '8', '--dropout', '0.33']
765
- 2026-01-16 07:02:05 INFO: Running parser in predict mode
766
- 2026-01-16 07:02:05 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
767
- 2026-01-16 07:02:07 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
768
- 2026-01-16 07:02:07 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
769
- 2026-01-16 07:02:07 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
770
- 2026-01-16 07:02:07 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
771
- 2026-01-16 07:02:07 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
772
- 2026-01-16 07:02:07 INFO: Loading data with batch size 8...
773
- 2026-01-16 07:02:07 DEBUG: 99 batches created.
774
- 2026-01-16 07:02:11 INFO: F1 scores for each dependency:
775
- Note that unlabeled attachment errors hurt the labeled attachment scores
776
- acl: p 0.0000 r 0.0000 f1 0.0000 (32 actual)
777
- acl:cleft: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
778
- acl:relcl: p 0.0000 r 0.0000 f1 0.0000 (75 actual)
779
- advcl: p 0.0000 r 0.0000 f1 0.0000 (60 actual)
780
- advcl:relcl: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
781
- advmod: p 0.0662 r 0.1007 f1 0.0799 (268 actual)
782
- amod: p 0.0000 r 0.0000 f1 0.0000 (230 actual)
783
- appos: p 0.0000 r 0.0000 f1 0.0000 (13 actual)
784
- aux: p 0.0000 r 0.0000 f1 0.0000 (84 actual)
785
- aux:pass: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
786
- case: p 0.3323 r 0.5818 f1 0.4230 (373 actual)
787
- cc: p 0.0000 r 0.0000 f1 0.0000 (155 actual)
788
- ccomp: p 0.0000 r 0.0000 f1 0.0000 (35 actual)
789
- compound:prt: p 0.0000 r 0.0000 f1 0.0000 (21 actual)
790
- conj: p 0.0000 r 0.0000 f1 0.0000 (158 actual)
791
- cop: p 0.0000 r 0.0000 f1 0.0000 (46 actual)
792
- csubj: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
793
- dep: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
794
- det: p 0.2518 r 0.1683 f1 0.2017 (208 actual)
795
- discourse: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
796
- dislocated: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
797
- expl: p 0.0000 r 0.0000 f1 0.0000 (11 actual)
798
- expl:pv: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
799
- fixed: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
800
- flat: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
801
- flat:name: p 0.4000 r 0.1667 f1 0.2353 (12 actual)
802
- goeswith: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
803
- iobj: p 0.0000 r 0.0000 f1 0.0000 (14 actual)
804
- mark: p 0.0000 r 0.0000 f1 0.0000 (153 actual)
805
- nmod: p 0.0000 r 0.0000 f1 0.0000 (102 actual)
806
- nmod:poss: p 0.0000 r 0.0000 f1 0.0000 (142 actual)
807
- nsubj: p 0.1027 r 0.5464 f1 0.1729 (280 actual)
808
- nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (25 actual)
809
- nummod: p 0.0000 r 0.0000 f1 0.0000 (10 actual)
810
- obj: p 0.0000 r 0.0000 f1 0.0000 (183 actual)
811
- obl: p 0.2272 r 0.3309 f1 0.2694 (278 actual)
812
- obl:agent: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
813
- orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
814
- parataxis: p 0.0000 r 0.0000 f1 0.0000 (18 actual)
815
- punct: p 0.3411 r 0.3459 f1 0.3435 (425 actual)
816
- reparandum: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
817
- root: p 0.4545 r 0.4545 f1 0.4545 (99 actual)
818
- vocative: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
819
- xcomp: p 0.0000 r 0.0000 f1 0.0000 (75 actual)
820
- 2026-01-16 07:02:11 INFO: LAS MLAS BLEX
821
- 2026-01-16 07:02:11 INFO: 20.06 11.15 14.33
822
- 2026-01-16 07:02:11 INFO: Parser score:
823
- 2026-01-16 07:02:11 INFO: sv_diachronic 20.06
824
- 2026-01-16 07:02:11 INFO: Finished running test set on
825
- UD_Swedish-diachronic
826
- UAS LAS CLAS MLAS BLEX
827
- 48.02 20.06 14.33 11.15 14.33
828
- DONE.
829
- Full log saved to: logs/log_conll17.pt_sv_diachron_20260116_065352.txt
830
- Symlink updated: logs/latest.txt → log_conll17.pt_sv_diachron_20260116_065352.txt
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_conll17.pt_sv_diachron_20260116_073035.txt DELETED
@@ -1,868 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_diachron_20260116_073035.txt ===
2
- Language codes: sv diachron
3
- Using pretrained model: conll17.pt
4
-
5
- Running: python prepare-train-val-test.py sv diachron
6
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec991-spf148.conllu
7
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec988-spf145.conllu
8
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
9
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
10
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
11
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec252-BremerF_Teckningar1.conllu
12
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
13
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
14
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
15
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
16
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec452-NyblomH_FantasierFyra.conllu
17
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1102-spf259.conllu
18
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-letter141673-Stalhammar.conllu
19
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1033-spf190.conllu
20
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
21
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
22
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
23
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
24
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
25
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
26
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
27
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
28
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
29
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
30
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
31
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec25-Runius.conllu
32
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
33
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
34
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec987-spf144.conllu
35
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1063-spf220.conllu
36
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
37
- Including DigPhil MACHINE in TRAIN (minus gold)…
38
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
39
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
40
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
41
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
42
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
43
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
44
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
45
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
46
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
47
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
48
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
49
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
50
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
51
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
52
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
53
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
54
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
55
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
56
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
57
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
58
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
59
- Cleaning TRAIN...
60
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
61
- [REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
62
- [REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
63
- [REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
64
- [REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
65
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
66
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
67
- [REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
68
- [REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
69
- [REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
70
- [REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
71
- [REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
72
- [REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
73
- [REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
74
- [REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
75
- [REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
76
- [REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
77
- [REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
78
- [REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
79
- [REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
80
- [REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
81
- [REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
82
- [REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
83
- [REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
84
- [REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
85
- [REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
86
- [REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
87
- [REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
88
- [REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
89
- [REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
90
- [REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
91
- [REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
92
- [REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
93
- [REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
94
- [REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
95
- [REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
96
- [REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
97
- [REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
98
- [REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
99
- [REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
100
- [REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
101
- [REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
102
- [REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
103
- [REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
104
- [REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
105
- [REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
106
- [REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
107
- [REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
108
- [REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
109
- [REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
110
- [REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
111
- [REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
112
- [REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
113
- [REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
114
- [REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
115
- [REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
116
- [REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
117
- [REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
118
- [REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
119
- [REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
120
- [REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
121
- [REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
122
- [REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
123
- [REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
124
- Cleaning DEV...
125
- Cleaning TEST...
126
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
127
- Writing TRAIN → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (60334 valid sentences)
128
- Writing DEV → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (34 valid sentences)
129
- Writing TEST → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (74 valid sentences)
130
- Done.
131
- Sourcing scripts/config_alvis.sh
132
- Running stanza dataset preparation…
133
- 2026-01-16 07:30:43 INFO: Datasets program called with:
134
- /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
135
- 2026-01-16 07:30:43 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
136
-
137
- 2026-01-16 07:30:44 INFO: Downloaded file to /cephyr/users/cleland/Alvis/stanza_resources/resources.json
138
- 2026-01-16 07:30:44 DEBUG: Processing parameter "processors"...
139
- 2026-01-16 07:30:44 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
140
- 2026-01-16 07:30:44 INFO: Downloading these customized packages for language: sv (Swedish)...
141
- =======================
142
- | Processor | Package |
143
- -----------------------
144
- =======================
145
-
146
- 2026-01-16 07:30:44 INFO: Finished downloading models and saved to /cephyr/users/cleland/Alvis/stanza_resources
147
- 2026-01-16 07:30:44 INFO: Using tagger model in /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt for sv_diachronic
148
- 2026-01-16 07:30:44 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
149
- 2026-01-16 07:30:44 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
150
- Augmented 189 quotes: Counter({'""': 26, '「」': 23, '„”': 23, '»«': 22, '″″': 18, '””': 18, '““': 17, '《》': 17, '„“': 13, '«»': 12})
151
- 2026-01-16 07:30:46 INFO: Running tagger to retag /local/tmp.5658622/tmp0y7bjycl/sv_diachronic.train.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
152
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5658622/tmp0y7bjycl/sv_diachronic.train.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu']
153
- 2026-01-16 07:30:47 INFO: Running tagger in predict mode
154
- 2026-01-16 07:30:47 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
155
- 2026-01-16 07:30:48 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
156
- 2026-01-16 07:30:48 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
157
- 2026-01-16 07:30:48 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
158
- 2026-01-16 07:30:48 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
159
- 2026-01-16 07:30:49 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
160
- 2026-01-16 07:30:51 INFO: Loading data with batch size 250...
161
- 2026-01-16 07:31:23 INFO: Start evaluation...
162
- 2026-01-16 07:34:24 INFO: UPOS XPOS UFeats AllTags
163
- 2026-01-16 07:34:24 INFO: 99.37 87.38 98.56 87.09
164
- 2026-01-16 07:34:24 INFO: POS Tagger score: sv_diachronic 87.09
165
- 2026-01-16 07:34:25 INFO: Running tagger to retag /local/tmp.5658622/tmp0y7bjycl/sv_diachronic.dev.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
166
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5658622/tmp0y7bjycl/sv_diachronic.dev.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu']
167
- 2026-01-16 07:34:25 INFO: Running tagger in predict mode
168
- 2026-01-16 07:34:25 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
169
- 2026-01-16 07:34:26 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
170
- 2026-01-16 07:34:26 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
171
- 2026-01-16 07:34:26 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
172
- 2026-01-16 07:34:26 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
173
- 2026-01-16 07:34:26 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
174
- 2026-01-16 07:34:26 INFO: Loading data with batch size 250...
175
- 2026-01-16 07:34:26 INFO: Start evaluation...
176
- 2026-01-16 07:34:27 INFO: UPOS XPOS UFeats AllTags
177
- 2026-01-16 07:34:27 INFO: 90.62 97.55 94.98 89.40
178
- 2026-01-16 07:34:27 INFO: POS Tagger score: sv_diachronic 89.40
179
- 2026-01-16 07:34:27 INFO: Running tagger to retag /local/tmp.5658622/tmp0y7bjycl/sv_diachronic.test.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
180
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5658622/tmp0y7bjycl/sv_diachronic.test.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu']
181
- 2026-01-16 07:34:27 INFO: Running tagger in predict mode
182
- 2026-01-16 07:34:27 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
183
- 2026-01-16 07:34:29 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
184
- 2026-01-16 07:34:29 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
185
- 2026-01-16 07:34:29 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
186
- 2026-01-16 07:34:29 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
187
- 2026-01-16 07:34:29 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
188
- 2026-01-16 07:34:29 INFO: Loading data with batch size 250...
189
- 2026-01-16 07:34:29 INFO: Start evaluation...
190
- 2026-01-16 07:34:29 INFO: UPOS XPOS UFeats AllTags
191
- 2026-01-16 07:34:29 INFO: 94.88 95.25 95.21 90.09
192
- 2026-01-16 07:34:29 INFO: POS Tagger score: sv_diachronic 90.09
193
- Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
194
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /local/tmp.5658622/tmp0y7bjycl/sv_diachronic.train.gold.conllu
195
- Swapped 'w1, w2' for 'w1 ,w2' 106 times
196
- Added 520 new sentences with asdf, zzzz -> asdf,zzzz
197
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /local/tmp.5658622/tmp0y7bjycl/sv_diachronic.dev.gold.conllu
198
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /local/tmp.5658622/tmp0y7bjycl/sv_diachronic.test.gold.conllu
199
- Running stanza dependency parser training…
200
- 2026-01-16 07:34:42 INFO: Training program called with:
201
- /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt --batch_size 8 --dropout 0.15
202
- 2026-01-16 07:34:42 DEBUG: UD_Swedish-diachronic: sv_diachronic
203
- 2026-01-16 07:34:42 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
204
- 2026-01-16 07:34:42 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
205
- 2026-01-16 07:34:42 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_charlm_parser.pt does not exist, training new model
206
- 2026-01-16 07:34:42 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
207
- 2026-01-16 07:34:42 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
208
- 2026-01-16 07:34:42 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--train_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '8', '--dropout', '0.15']
209
- 2026-01-16 07:34:42 INFO: Running parser in train mode
210
- 2026-01-16 07:34:42 INFO: Using pretrained contextualized char embedding
211
- 2026-01-16 07:34:42 INFO: Loading data with batch size 8...
212
- 2026-01-16 07:34:48 INFO: Train File /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 60854
213
- 2026-01-16 07:34:48 INFO: Original data size: 60854
214
- 2026-01-16 07:34:48 INFO: Augmented data size: 60854
215
- 2026-01-16 07:35:03 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
216
- 2026-01-16 07:35:03 INFO: Original length = 60854
217
- 2026-01-16 07:35:03 INFO: Filtered length = 60854
218
- 2026-01-16 07:35:17 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='|') for the xpos factory for sv_diachronic
219
- 2026-01-16 07:35:22 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
220
- 2026-01-16 07:35:33 DEBUG: 58067 batches created.
221
- 2026-01-16 07:35:33 DEBUG: 34 batches created.
222
- 2026-01-16 07:35:33 INFO: Training parser...
223
- 2026-01-16 07:35:33 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
224
- 2026-01-16 07:35:33 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
225
- 2026-01-16 07:35:34 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
226
- 2026-01-16 07:35:34 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
227
- 2026-01-16 07:35:37 INFO: Finished STEP 20/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
228
- 2026-01-16 07:35:38 INFO: Finished STEP 40/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
229
- 2026-01-16 07:35:38 INFO: Finished STEP 60/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
230
- 2026-01-16 07:35:39 INFO: Finished STEP 80/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
231
- 2026-01-16 07:35:39 INFO: Finished STEP 100/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
232
- 2026-01-16 07:35:39 INFO: Evaluating on dev set...
233
- 2026-01-16 07:35:41 INFO: LAS MLAS BLEX
234
- 2026-01-16 07:35:41 INFO: 0.06 0.08 0.08
235
- 2026-01-16 07:35:41 INFO: step 100: train_loss = 379.523051, dev_score = 0.0006
236
- 2026-01-16 07:35:41 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
237
- 2026-01-16 07:35:41 INFO: new best model saved.
238
- 2026-01-16 07:35:42 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
239
- 2026-01-16 07:35:42 INFO: new model checkpoint saved.
240
- 2026-01-16 07:35:42 INFO: Finished STEP 120/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
241
- 2026-01-16 07:35:43 INFO: Finished STEP 140/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
242
- 2026-01-16 07:35:44 INFO: Finished STEP 160/50000, loss = 0.000000 (0.028 sec/batch), lr: 0.003000
243
- 2026-01-16 07:35:44 INFO: Finished STEP 180/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
244
- 2026-01-16 07:35:45 INFO: Finished STEP 200/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
245
- 2026-01-16 07:35:45 INFO: Evaluating on dev set...
246
- 2026-01-16 07:35:46 INFO: LAS MLAS BLEX
247
- 2026-01-16 07:35:46 INFO: 0.06 0.08 0.08
248
- 2026-01-16 07:35:46 INFO: step 200: train_loss = 0.000000, dev_score = 0.0006
249
- 2026-01-16 07:35:47 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
250
- 2026-01-16 07:35:47 INFO: new best model saved.
251
- 2026-01-16 07:35:47 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
252
- 2026-01-16 07:35:47 INFO: new model checkpoint saved.
253
- 2026-01-16 07:35:48 INFO: Finished STEP 220/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
254
- 2026-01-16 07:35:48 INFO: Finished STEP 240/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
255
- 2026-01-16 07:35:49 INFO: Finished STEP 260/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
256
- 2026-01-16 07:35:50 INFO: Finished STEP 280/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
257
- 2026-01-16 07:35:50 INFO: Finished STEP 300/50000, loss = 2.424209 (0.030 sec/batch), lr: 0.003000
258
- 2026-01-16 07:35:50 INFO: Evaluating on dev set...
259
- 2026-01-16 07:35:52 INFO: LAS MLAS BLEX
260
- 2026-01-16 07:35:52 INFO: 1.35 0.00 0.00
261
- 2026-01-16 07:35:52 INFO: step 300: train_loss = 158.649473, dev_score = 0.0135
262
- 2026-01-16 07:35:52 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
263
- 2026-01-16 07:35:52 INFO: new best model saved.
264
- 2026-01-16 07:35:53 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
265
- 2026-01-16 07:35:53 INFO: new model checkpoint saved.
266
- 2026-01-16 07:35:53 INFO: Finished STEP 320/50000, loss = 1.969488 (0.031 sec/batch), lr: 0.003000
267
- 2026-01-16 07:35:54 INFO: Finished STEP 340/50000, loss = 1.649248 (0.029 sec/batch), lr: 0.003000
268
- 2026-01-16 07:35:54 INFO: Finished STEP 360/50000, loss = 1.934138 (0.030 sec/batch), lr: 0.003000
269
- 2026-01-16 07:35:55 INFO: Finished STEP 380/50000, loss = 2.360754 (0.031 sec/batch), lr: 0.003000
270
- 2026-01-16 07:35:56 INFO: Finished STEP 400/50000, loss = 0.562345 (0.030 sec/batch), lr: 0.003000
271
- 2026-01-16 07:35:56 INFO: Evaluating on dev set...
272
- 2026-01-16 07:35:58 INFO: LAS MLAS BLEX
273
- 2026-01-16 07:35:58 INFO: 4.72 4.16 4.25
274
- 2026-01-16 07:35:58 INFO: step 400: train_loss = 2.121128, dev_score = 0.0472
275
- 2026-01-16 07:35:58 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
276
- 2026-01-16 07:35:58 INFO: new best model saved.
277
- 2026-01-16 07:35:59 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
278
- 2026-01-16 07:35:59 INFO: new model checkpoint saved.
279
- 2026-01-16 07:35:59 INFO: Finished STEP 420/50000, loss = 0.002566 (0.029 sec/batch), lr: 0.003000
280
- 2026-01-16 07:36:00 INFO: Finished STEP 440/50000, loss = 0.734853 (0.030 sec/batch), lr: 0.003000
281
- 2026-01-16 07:36:01 INFO: Finished STEP 460/50000, loss = 0.494820 (0.029 sec/batch), lr: 0.003000
282
- 2026-01-16 07:36:01 INFO: Finished STEP 480/50000, loss = 3.201444 (0.030 sec/batch), lr: 0.003000
283
- 2026-01-16 07:36:02 INFO: Finished STEP 500/50000, loss = 0.795342 (0.030 sec/batch), lr: 0.003000
284
- 2026-01-16 07:36:02 INFO: Evaluating on dev set...
285
- 2026-01-16 07:36:04 INFO: LAS MLAS BLEX
286
- 2026-01-16 07:36:04 INFO: 5.88 5.61 6.11
287
- 2026-01-16 07:36:04 INFO: step 500: train_loss = 1.037986, dev_score = 0.0588
288
- 2026-01-16 07:36:04 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
289
- 2026-01-16 07:36:04 INFO: new best model saved.
290
- 2026-01-16 07:36:05 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
291
- 2026-01-16 07:36:05 INFO: new model checkpoint saved.
292
- 2026-01-16 07:36:05 INFO: Finished STEP 520/50000, loss = 0.778292 (0.029 sec/batch), lr: 0.003000
293
- 2026-01-16 07:36:06 INFO: Finished STEP 540/50000, loss = 0.409450 (0.030 sec/batch), lr: 0.003000
294
- 2026-01-16 07:36:06 INFO: Finished STEP 560/50000, loss = 0.671611 (0.029 sec/batch), lr: 0.003000
295
- 2026-01-16 07:36:07 INFO: Finished STEP 580/50000, loss = 0.081582 (0.029 sec/batch), lr: 0.003000
296
- 2026-01-16 07:36:08 INFO: Finished STEP 600/50000, loss = 0.436447 (0.030 sec/batch), lr: 0.003000
297
- 2026-01-16 07:36:08 INFO: Evaluating on dev set...
298
- 2026-01-16 07:36:10 INFO: LAS MLAS BLEX
299
- 2026-01-16 07:36:10 INFO: 5.76 6.09 6.18
300
- 2026-01-16 07:36:10 INFO: step 600: train_loss = 1.044479, dev_score = 0.0576
301
- 2026-01-16 07:36:11 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
302
- 2026-01-16 07:36:11 INFO: new model checkpoint saved.
303
- 2026-01-16 07:36:11 INFO: Finished STEP 620/50000, loss = 0.887764 (0.031 sec/batch), lr: 0.003000
304
- 2026-01-16 07:36:12 INFO: Finished STEP 640/50000, loss = 0.527841 (0.030 sec/batch), lr: 0.003000
305
- 2026-01-16 07:36:12 INFO: Finished STEP 660/50000, loss = 0.269212 (0.030 sec/batch), lr: 0.003000
306
- 2026-01-16 07:36:13 INFO: Finished STEP 680/50000, loss = 2.426815 (0.029 sec/batch), lr: 0.003000
307
- 2026-01-16 07:36:14 INFO: Finished STEP 700/50000, loss = 0.000517 (0.029 sec/batch), lr: 0.003000
308
- 2026-01-16 07:36:14 INFO: Evaluating on dev set...
309
- 2026-01-16 07:36:15 INFO: LAS MLAS BLEX
310
- 2026-01-16 07:36:15 INFO: 5.21 4.95 5.37
311
- 2026-01-16 07:36:15 INFO: step 700: train_loss = 0.978077, dev_score = 0.0521
312
- 2026-01-16 07:36:16 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
313
- 2026-01-16 07:36:16 INFO: new model checkpoint saved.
314
- 2026-01-16 07:36:16 INFO: Finished STEP 720/50000, loss = 2.358844 (0.029 sec/batch), lr: 0.003000
315
- 2026-01-16 07:36:17 INFO: Finished STEP 740/50000, loss = 0.791893 (0.030 sec/batch), lr: 0.003000
316
- 2026-01-16 07:36:18 INFO: Finished STEP 760/50000, loss = 0.906499 (0.029 sec/batch), lr: 0.003000
317
- 2026-01-16 07:36:18 INFO: Finished STEP 780/50000, loss = 1.075008 (0.030 sec/batch), lr: 0.003000
318
- 2026-01-16 07:36:19 INFO: Finished STEP 800/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
319
- 2026-01-16 07:36:19 INFO: Evaluating on dev set...
320
- 2026-01-16 07:36:21 INFO: LAS MLAS BLEX
321
- 2026-01-16 07:36:21 INFO: 2.14 1.13 1.13
322
- 2026-01-16 07:36:21 INFO: step 800: train_loss = 0.975283, dev_score = 0.0214
323
- 2026-01-16 07:36:21 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
324
- 2026-01-16 07:36:21 INFO: new model checkpoint saved.
325
- 2026-01-16 07:36:22 INFO: Finished STEP 820/50000, loss = 0.678568 (0.030 sec/batch), lr: 0.003000
326
- 2026-01-16 07:36:22 INFO: Finished STEP 840/50000, loss = 0.000088 (0.029 sec/batch), lr: 0.003000
327
- 2026-01-16 07:36:23 INFO: Finished STEP 860/50000, loss = 1.472390 (0.029 sec/batch), lr: 0.003000
328
- 2026-01-16 07:36:24 INFO: Finished STEP 880/50000, loss = 0.818992 (0.029 sec/batch), lr: 0.003000
329
- 2026-01-16 07:36:24 INFO: Finished STEP 900/50000, loss = 1.486508 (0.031 sec/batch), lr: 0.003000
330
- 2026-01-16 07:36:24 INFO: Evaluating on dev set...
331
- 2026-01-16 07:36:26 INFO: LAS MLAS BLEX
332
- 2026-01-16 07:36:26 INFO: 3.98 1.89 3.04
333
- 2026-01-16 07:36:26 INFO: step 900: train_loss = 0.911072, dev_score = 0.0398
334
- 2026-01-16 07:36:27 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
335
- 2026-01-16 07:36:27 INFO: new model checkpoint saved.
336
- 2026-01-16 07:36:27 INFO: Finished STEP 920/50000, loss = 0.001453 (0.030 sec/batch), lr: 0.003000
337
- 2026-01-16 07:36:28 INFO: Finished STEP 940/50000, loss = 0.934357 (0.029 sec/batch), lr: 0.003000
338
- 2026-01-16 07:36:28 INFO: Finished STEP 960/50000, loss = 0.010020 (0.029 sec/batch), lr: 0.003000
339
- 2026-01-16 07:36:29 INFO: Finished STEP 980/50000, loss = 0.627853 (0.029 sec/batch), lr: 0.003000
340
- 2026-01-16 07:36:30 INFO: Finished STEP 1000/50000, loss = 0.240197 (0.029 sec/batch), lr: 0.003000
341
- 2026-01-16 07:36:30 INFO: Evaluating on dev set...
342
- 2026-01-16 07:36:31 INFO: LAS MLAS BLEX
343
- 2026-01-16 07:36:31 INFO: 2.76 1.78 1.78
344
- 2026-01-16 07:36:31 INFO: step 1000: train_loss = 0.752214, dev_score = 0.0276
345
- 2026-01-16 07:36:32 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
346
- 2026-01-16 07:36:32 INFO: new model checkpoint saved.
347
- 2026-01-16 07:36:32 INFO: Finished STEP 1020/50000, loss = 1.534254 (0.029 sec/batch), lr: 0.003000
348
- 2026-01-16 07:36:33 INFO: Finished STEP 1040/50000, loss = 0.000038 (0.029 sec/batch), lr: 0.003000
349
- 2026-01-16 07:36:34 INFO: Finished STEP 1060/50000, loss = 1.659585 (0.029 sec/batch), lr: 0.003000
350
- 2026-01-16 07:36:34 INFO: Finished STEP 1080/50000, loss = 0.403725 (0.029 sec/batch), lr: 0.003000
351
- 2026-01-16 07:36:35 INFO: Finished STEP 1100/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
352
- 2026-01-16 07:36:35 INFO: Evaluating on dev set...
353
- 2026-01-16 07:36:37 INFO: LAS MLAS BLEX
354
- 2026-01-16 07:36:37 INFO: 3.12 2.80 2.89
355
- 2026-01-16 07:36:37 INFO: step 1100: train_loss = 0.824820, dev_score = 0.0312
356
- 2026-01-16 07:36:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
357
- 2026-01-16 07:36:37 INFO: new model checkpoint saved.
358
- 2026-01-16 07:36:38 INFO: Finished STEP 1120/50000, loss = 6.452912 (0.031 sec/batch), lr: 0.003000
359
- 2026-01-16 07:36:38 INFO: Finished STEP 1140/50000, loss = 0.749853 (0.031 sec/batch), lr: 0.003000
360
- 2026-01-16 07:36:39 INFO: Finished STEP 1160/50000, loss = 4.511549 (0.030 sec/batch), lr: 0.003000
361
- 2026-01-16 07:36:40 INFO: Finished STEP 1180/50000, loss = 2.717284 (0.030 sec/batch), lr: 0.003000
362
- 2026-01-16 07:36:40 INFO: Finished STEP 1200/50000, loss = 0.523814 (0.031 sec/batch), lr: 0.003000
363
- 2026-01-16 07:36:40 INFO: Evaluating on dev set...
364
- 2026-01-16 07:36:42 INFO: LAS MLAS BLEX
365
- 2026-01-16 07:36:42 INFO: 4.53 1.41 1.50
366
- 2026-01-16 07:36:42 INFO: step 1200: train_loss = 2.303929, dev_score = 0.0453
367
- 2026-01-16 07:36:42 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
368
- 2026-01-16 07:36:42 INFO: new model checkpoint saved.
369
- 2026-01-16 07:36:43 INFO: Finished STEP 1220/50000, loss = 0.628675 (0.030 sec/batch), lr: 0.003000
370
- 2026-01-16 07:36:44 INFO: Finished STEP 1240/50000, loss = 1.167853 (0.031 sec/batch), lr: 0.003000
371
- 2026-01-16 07:36:44 INFO: Finished STEP 1260/50000, loss = 2.458635 (0.031 sec/batch), lr: 0.003000
372
- 2026-01-16 07:36:45 INFO: Finished STEP 1280/50000, loss = 1.727715 (0.030 sec/batch), lr: 0.003000
373
- 2026-01-16 07:36:46 INFO: Finished STEP 1300/50000, loss = 3.715376 (0.031 sec/batch), lr: 0.003000
374
- 2026-01-16 07:36:46 INFO: Evaluating on dev set...
375
- 2026-01-16 07:36:47 INFO: LAS MLAS BLEX
376
- 2026-01-16 07:36:47 INFO: 3.06 1.24 1.32
377
- 2026-01-16 07:36:47 INFO: step 1300: train_loss = 1.477564, dev_score = 0.0306
378
- 2026-01-16 07:36:48 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
379
- 2026-01-16 07:36:48 INFO: new model checkpoint saved.
380
- 2026-01-16 07:36:48 INFO: Finished STEP 1320/50000, loss = 1.768359 (0.030 sec/batch), lr: 0.003000
381
- 2026-01-16 07:36:49 INFO: Finished STEP 1340/50000, loss = 1.245610 (0.030 sec/batch), lr: 0.003000
382
- 2026-01-16 07:36:50 INFO: Finished STEP 1360/50000, loss = 0.451835 (0.031 sec/batch), lr: 0.003000
383
- 2026-01-16 07:36:50 INFO: Finished STEP 1380/50000, loss = 1.443755 (0.031 sec/batch), lr: 0.003000
384
- 2026-01-16 07:36:51 INFO: Finished STEP 1400/50000, loss = 1.238335 (0.030 sec/batch), lr: 0.003000
385
- 2026-01-16 07:36:51 INFO: Evaluating on dev set...
386
- 2026-01-16 07:36:53 INFO: LAS MLAS BLEX
387
- 2026-01-16 07:36:53 INFO: 7.11 6.09 6.96
388
- 2026-01-16 07:36:53 INFO: step 1400: train_loss = 1.361486, dev_score = 0.0711
389
- 2026-01-16 07:36:53 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
390
- 2026-01-16 07:36:53 INFO: new best model saved.
391
- 2026-01-16 07:36:53 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
392
- 2026-01-16 07:36:53 INFO: new model checkpoint saved.
393
- 2026-01-16 07:36:54 INFO: Finished STEP 1420/50000, loss = 1.591924 (0.031 sec/batch), lr: 0.003000
394
- 2026-01-16 07:36:55 INFO: Finished STEP 1440/50000, loss = 0.735520 (0.030 sec/batch), lr: 0.003000
395
- 2026-01-16 07:36:55 INFO: Finished STEP 1460/50000, loss = 1.212776 (0.032 sec/batch), lr: 0.003000
396
- 2026-01-16 07:36:56 INFO: Finished STEP 1480/50000, loss = 1.210722 (0.031 sec/batch), lr: 0.003000
397
- 2026-01-16 07:36:57 INFO: Finished STEP 1500/50000, loss = 3.332783 (0.031 sec/batch), lr: 0.003000
398
- 2026-01-16 07:36:57 INFO: Evaluating on dev set...
399
- 2026-01-16 07:36:58 INFO: LAS MLAS BLEX
400
- 2026-01-16 07:36:58 INFO: 6.74 3.03 3.88
401
- 2026-01-16 07:36:58 INFO: step 1500: train_loss = 1.381426, dev_score = 0.0674
402
- 2026-01-16 07:36:59 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
403
- 2026-01-16 07:36:59 INFO: new model checkpoint saved.
404
- 2026-01-16 07:37:00 INFO: Finished STEP 1520/50000, loss = 0.796412 (0.031 sec/batch), lr: 0.003000
405
- 2026-01-16 07:37:00 INFO: Finished STEP 1540/50000, loss = 0.730972 (0.031 sec/batch), lr: 0.003000
406
- 2026-01-16 07:37:01 INFO: Finished STEP 1560/50000, loss = 1.262556 (0.031 sec/batch), lr: 0.003000
407
- 2026-01-16 07:37:01 INFO: Finished STEP 1580/50000, loss = 0.734725 (0.030 sec/batch), lr: 0.003000
408
- 2026-01-16 07:37:02 INFO: Finished STEP 1600/50000, loss = 0.073490 (0.030 sec/batch), lr: 0.003000
409
- 2026-01-16 07:37:02 INFO: Evaluating on dev set...
410
- 2026-01-16 07:37:05 INFO: LAS MLAS BLEX
411
- 2026-01-16 07:37:05 INFO: 12.87 10.38 11.31
412
- 2026-01-16 07:37:05 INFO: step 1600: train_loss = 1.284919, dev_score = 0.1287
413
- 2026-01-16 07:37:06 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
414
- 2026-01-16 07:37:06 INFO: new best model saved.
415
- 2026-01-16 07:37:06 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
416
- 2026-01-16 07:37:06 INFO: new model checkpoint saved.
417
- 2026-01-16 07:37:07 INFO: Finished STEP 1620/50000, loss = 0.640151 (0.030 sec/batch), lr: 0.003000
418
- 2026-01-16 07:37:07 INFO: Finished STEP 1640/50000, loss = 1.584237 (0.030 sec/batch), lr: 0.003000
419
- 2026-01-16 07:37:08 INFO: Finished STEP 1660/50000, loss = 2.824079 (0.031 sec/batch), lr: 0.003000
420
- 2026-01-16 07:37:09 INFO: Finished STEP 1680/50000, loss = 1.456016 (0.030 sec/batch), lr: 0.003000
421
- 2026-01-16 07:37:09 INFO: Finished STEP 1700/50000, loss = 1.605577 (0.030 sec/batch), lr: 0.003000
422
- 2026-01-16 07:37:09 INFO: Evaluating on dev set...
423
- 2026-01-16 07:37:11 INFO: LAS MLAS BLEX
424
- 2026-01-16 07:37:11 INFO: 13.36 9.21 10.85
425
- 2026-01-16 07:37:11 INFO: step 1700: train_loss = 1.177195, dev_score = 0.1336
426
- 2026-01-16 07:37:11 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
427
- 2026-01-16 07:37:11 INFO: new best model saved.
428
- 2026-01-16 07:37:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
429
- 2026-01-16 07:37:12 INFO: new model checkpoint saved.
430
- 2026-01-16 07:37:12 INFO: Finished STEP 1720/50000, loss = 1.005753 (0.030 sec/batch), lr: 0.003000
431
- 2026-01-16 07:37:13 INFO: Finished STEP 1740/50000, loss = 1.587673 (0.029 sec/batch), lr: 0.003000
432
- 2026-01-16 07:37:14 INFO: Finished STEP 1760/50000, loss = 1.467617 (0.030 sec/batch), lr: 0.003000
433
- 2026-01-16 07:37:14 INFO: Finished STEP 1780/50000, loss = 2.158596 (0.034 sec/batch), lr: 0.003000
434
- 2026-01-16 07:37:15 INFO: Finished STEP 1800/50000, loss = 0.786507 (0.031 sec/batch), lr: 0.003000
435
- 2026-01-16 07:37:15 INFO: Evaluating on dev set...
436
- 2026-01-16 07:37:17 INFO: LAS MLAS BLEX
437
- 2026-01-16 07:37:17 INFO: 10.11 6.65 7.70
438
- 2026-01-16 07:37:17 INFO: step 1800: train_loss = 1.175222, dev_score = 0.1011
439
- 2026-01-16 07:37:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
440
- 2026-01-16 07:37:17 INFO: new model checkpoint saved.
441
- 2026-01-16 07:37:18 INFO: Finished STEP 1820/50000, loss = 0.861782 (0.031 sec/batch), lr: 0.003000
442
- 2026-01-16 07:37:19 INFO: Finished STEP 1840/50000, loss = 1.011274 (0.030 sec/batch), lr: 0.003000
443
- 2026-01-16 07:37:19 INFO: Finished STEP 1860/50000, loss = 0.641151 (0.030 sec/batch), lr: 0.003000
444
- 2026-01-16 07:37:20 INFO: Finished STEP 1880/50000, loss = 0.764767 (0.036 sec/batch), lr: 0.003000
445
- 2026-01-16 07:37:21 INFO: Finished STEP 1900/50000, loss = 1.876217 (0.030 sec/batch), lr: 0.003000
446
- 2026-01-16 07:37:21 INFO: Evaluating on dev set...
447
- 2026-01-16 07:37:25 INFO: LAS MLAS BLEX
448
- 2026-01-16 07:37:25 INFO: 7.66 4.98 5.64
449
- 2026-01-16 07:37:25 INFO: step 1900: train_loss = 1.227179, dev_score = 0.0766
450
- 2026-01-16 07:37:26 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
451
- 2026-01-16 07:37:26 INFO: new model checkpoint saved.
452
- 2026-01-16 07:37:27 INFO: Finished STEP 1920/50000, loss = 1.518084 (0.031 sec/batch), lr: 0.003000
453
- 2026-01-16 07:37:27 INFO: Finished STEP 1940/50000, loss = 0.689516 (0.031 sec/batch), lr: 0.003000
454
- 2026-01-16 07:37:28 INFO: Finished STEP 1960/50000, loss = 0.541874 (0.031 sec/batch), lr: 0.003000
455
- 2026-01-16 07:37:29 INFO: Finished STEP 1980/50000, loss = 1.434824 (0.030 sec/batch), lr: 0.003000
456
- 2026-01-16 07:37:29 INFO: Finished STEP 2000/50000, loss = 0.343702 (0.031 sec/batch), lr: 0.003000
457
- 2026-01-16 07:37:29 INFO: Evaluating on dev set...
458
- 2026-01-16 07:37:31 INFO: LAS MLAS BLEX
459
- 2026-01-16 07:37:31 INFO: 11.52 4.14 5.58
460
- 2026-01-16 07:37:31 INFO: step 2000: train_loss = 1.288903, dev_score = 0.1152
461
- 2026-01-16 07:37:32 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
462
- 2026-01-16 07:37:32 INFO: new model checkpoint saved.
463
- 2026-01-16 07:37:33 INFO: Finished STEP 2020/50000, loss = 1.551595 (0.030 sec/batch), lr: 0.003000
464
- 2026-01-16 07:37:33 INFO: Finished STEP 2040/50000, loss = 0.806195 (0.031 sec/batch), lr: 0.003000
465
- 2026-01-16 07:37:34 INFO: Finished STEP 2060/50000, loss = 0.586268 (0.031 sec/batch), lr: 0.003000
466
- 2026-01-16 07:37:34 INFO: Finished STEP 2080/50000, loss = 1.326350 (0.031 sec/batch), lr: 0.003000
467
- 2026-01-16 07:37:35 INFO: Finished STEP 2100/50000, loss = 1.068810 (0.031 sec/batch), lr: 0.003000
468
- 2026-01-16 07:37:35 INFO: Evaluating on dev set...
469
- 2026-01-16 07:37:37 INFO: LAS MLAS BLEX
470
- 2026-01-16 07:37:37 INFO: 3.49 1.97 2.55
471
- 2026-01-16 07:37:37 INFO: step 2100: train_loss = 1.064336, dev_score = 0.0349
472
- 2026-01-16 07:37:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
473
- 2026-01-16 07:37:37 INFO: new model checkpoint saved.
474
- 2026-01-16 07:37:38 INFO: Finished STEP 2120/50000, loss = 0.723737 (0.033 sec/batch), lr: 0.003000
475
- 2026-01-16 07:37:39 INFO: Finished STEP 2140/50000, loss = 1.762422 (0.031 sec/batch), lr: 0.003000
476
- 2026-01-16 07:37:39 INFO: Finished STEP 2160/50000, loss = 0.039095 (0.036 sec/batch), lr: 0.003000
477
- 2026-01-16 07:37:40 INFO: Finished STEP 2180/50000, loss = 0.746787 (0.030 sec/batch), lr: 0.003000
478
- 2026-01-16 07:37:41 INFO: Finished STEP 2200/50000, loss = 0.145349 (0.031 sec/batch), lr: 0.003000
479
- 2026-01-16 07:37:41 INFO: Evaluating on dev set...
480
- 2026-01-16 07:37:43 INFO: LAS MLAS BLEX
481
- 2026-01-16 07:37:43 INFO: 13.30 9.26 9.89
482
- 2026-01-16 07:37:43 INFO: step 2200: train_loss = 1.202942, dev_score = 0.1330
483
- 2026-01-16 07:37:43 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
484
- 2026-01-16 07:37:43 INFO: new model checkpoint saved.
485
- 2026-01-16 07:37:44 INFO: Finished STEP 2220/50000, loss = 2.421659 (0.030 sec/batch), lr: 0.003000
486
- 2026-01-16 07:37:44 INFO: Finished STEP 2240/50000, loss = 2.592595 (0.030 sec/batch), lr: 0.003000
487
- 2026-01-16 07:37:45 INFO: Finished STEP 2260/50000, loss = 3.920859 (0.030 sec/batch), lr: 0.003000
488
- 2026-01-16 07:37:46 INFO: Finished STEP 2280/50000, loss = 1.928826 (0.030 sec/batch), lr: 0.003000
489
- 2026-01-16 07:37:46 INFO: Finished STEP 2300/50000, loss = 5.176385 (0.030 sec/batch), lr: 0.003000
490
- 2026-01-16 07:37:46 INFO: Evaluating on dev set...
491
- 2026-01-16 07:37:48 INFO: LAS MLAS BLEX
492
- 2026-01-16 07:37:48 INFO: 17.83 9.07 10.24
493
- 2026-01-16 07:37:48 INFO: step 2300: train_loss = 2.391255, dev_score = 0.1783
494
- 2026-01-16 07:37:48 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
495
- 2026-01-16 07:37:48 INFO: new best model saved.
496
- 2026-01-16 07:37:49 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
497
- 2026-01-16 07:37:49 INFO: new model checkpoint saved.
498
- 2026-01-16 07:37:49 INFO: Finished STEP 2320/50000, loss = 2.267345 (0.030 sec/batch), lr: 0.003000
499
- 2026-01-16 07:37:50 INFO: Finished STEP 2340/50000, loss = 0.719973 (0.030 sec/batch), lr: 0.003000
500
- 2026-01-16 07:37:51 INFO: Finished STEP 2360/50000, loss = 1.036772 (0.031 sec/batch), lr: 0.003000
501
- 2026-01-16 07:37:51 INFO: Finished STEP 2380/50000, loss = 2.774978 (0.030 sec/batch), lr: 0.003000
502
- 2026-01-16 07:37:52 INFO: Finished STEP 2400/50000, loss = 1.541276 (0.030 sec/batch), lr: 0.003000
503
- 2026-01-16 07:37:52 INFO: Evaluating on dev set...
504
- 2026-01-16 07:37:54 INFO: LAS MLAS BLEX
505
- 2026-01-16 07:37:54 INFO: 5.76 3.46 4.14
506
- 2026-01-16 07:37:54 INFO: step 2400: train_loss = 2.107856, dev_score = 0.0576
507
- 2026-01-16 07:37:54 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
508
- 2026-01-16 07:37:54 INFO: new model checkpoint saved.
509
- 2026-01-16 07:37:55 INFO: Finished STEP 2420/50000, loss = 0.558537 (0.029 sec/batch), lr: 0.003000
510
- 2026-01-16 07:37:55 INFO: Finished STEP 2440/50000, loss = 0.417549 (0.029 sec/batch), lr: 0.003000
511
- 2026-01-16 07:37:56 INFO: Finished STEP 2460/50000, loss = 1.211123 (0.029 sec/batch), lr: 0.003000
512
- 2026-01-16 07:37:57 INFO: Finished STEP 2480/50000, loss = 1.066298 (0.030 sec/batch), lr: 0.003000
513
- 2026-01-16 07:37:57 INFO: Finished STEP 2500/50000, loss = 1.788547 (0.031 sec/batch), lr: 0.003000
514
- 2026-01-16 07:37:57 INFO: Evaluating on dev set...
515
- 2026-01-16 07:37:59 INFO: LAS MLAS BLEX
516
- 2026-01-16 07:37:59 INFO: 12.68 5.16 6.11
517
- 2026-01-16 07:37:59 INFO: step 2500: train_loss = 2.112291, dev_score = 0.1268
518
- 2026-01-16 07:38:00 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
519
- 2026-01-16 07:38:00 INFO: new model checkpoint saved.
520
- 2026-01-16 07:38:00 INFO: Finished STEP 2520/50000, loss = 1.717021 (0.029 sec/batch), lr: 0.003000
521
- 2026-01-16 07:38:01 INFO: Finished STEP 2540/50000, loss = 1.660125 (0.032 sec/batch), lr: 0.003000
522
- 2026-01-16 07:38:01 INFO: Finished STEP 2560/50000, loss = 1.768633 (0.030 sec/batch), lr: 0.003000
523
- 2026-01-16 07:38:02 INFO: Finished STEP 2580/50000, loss = 0.593540 (0.029 sec/batch), lr: 0.003000
524
- 2026-01-16 07:38:03 INFO: Finished STEP 2600/50000, loss = 0.938626 (0.030 sec/batch), lr: 0.003000
525
- 2026-01-16 07:38:03 INFO: Evaluating on dev set...
526
- 2026-01-16 07:38:05 INFO: LAS MLAS BLEX
527
- 2026-01-16 07:38:05 INFO: 19.12 9.02 10.39
528
- 2026-01-16 07:38:05 INFO: step 2600: train_loss = 1.933433, dev_score = 0.1912
529
- 2026-01-16 07:38:05 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
530
- 2026-01-16 07:38:05 INFO: new best model saved.
531
- 2026-01-16 07:38:06 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
532
- 2026-01-16 07:38:06 INFO: new model checkpoint saved.
533
- 2026-01-16 07:38:06 INFO: Finished STEP 2620/50000, loss = 2.193763 (0.029 sec/batch), lr: 0.003000
534
- 2026-01-16 07:38:07 INFO: Finished STEP 2640/50000, loss = 1.114236 (0.030 sec/batch), lr: 0.003000
535
- 2026-01-16 07:38:07 INFO: Finished STEP 2660/50000, loss = 1.077905 (0.030 sec/batch), lr: 0.003000
536
- 2026-01-16 07:38:08 INFO: Finished STEP 2680/50000, loss = 2.145391 (0.029 sec/batch), lr: 0.003000
537
- 2026-01-16 07:38:09 INFO: Finished STEP 2700/50000, loss = 4.426371 (0.030 sec/batch), lr: 0.003000
538
- 2026-01-16 07:38:09 INFO: Evaluating on dev set...
539
- 2026-01-16 07:38:11 INFO: LAS MLAS BLEX
540
- 2026-01-16 07:38:11 INFO: 11.76 7.72 9.78
541
- 2026-01-16 07:38:11 INFO: step 2700: train_loss = 1.947624, dev_score = 0.1176
542
- 2026-01-16 07:38:11 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
543
- 2026-01-16 07:38:11 INFO: new model checkpoint saved.
544
- 2026-01-16 07:38:12 INFO: Finished STEP 2720/50000, loss = 1.433867 (0.030 sec/batch), lr: 0.003000
545
- 2026-01-16 07:38:12 INFO: Finished STEP 2740/50000, loss = 0.081805 (0.030 sec/batch), lr: 0.003000
546
- 2026-01-16 07:38:13 INFO: Finished STEP 2760/50000, loss = 2.694524 (0.031 sec/batch), lr: 0.003000
547
- 2026-01-16 07:38:14 INFO: Finished STEP 2780/50000, loss = 2.323376 (0.030 sec/batch), lr: 0.003000
548
- 2026-01-16 07:38:14 INFO: Finished STEP 2800/50000, loss = 0.507773 (0.031 sec/batch), lr: 0.003000
549
- 2026-01-16 07:38:14 INFO: Evaluating on dev set...
550
- 2026-01-16 07:38:16 INFO: LAS MLAS BLEX
551
- 2026-01-16 07:38:16 INFO: 16.05 7.79 9.64
552
- 2026-01-16 07:38:16 INFO: step 2800: train_loss = 2.021911, dev_score = 0.1605
553
- 2026-01-16 07:38:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
554
- 2026-01-16 07:38:17 INFO: new model checkpoint saved.
555
- 2026-01-16 07:38:17 INFO: Finished STEP 2820/50000, loss = 2.562084 (0.030 sec/batch), lr: 0.003000
556
- 2026-01-16 07:38:18 INFO: Finished STEP 2840/50000, loss = 2.842648 (0.029 sec/batch), lr: 0.003000
557
- 2026-01-16 07:38:18 INFO: Finished STEP 2860/50000, loss = 0.741765 (0.031 sec/batch), lr: 0.003000
558
- 2026-01-16 07:38:19 INFO: Finished STEP 2880/50000, loss = 1.145487 (0.030 sec/batch), lr: 0.003000
559
- 2026-01-16 07:38:20 INFO: Finished STEP 2900/50000, loss = 1.986015 (0.029 sec/batch), lr: 0.003000
560
- 2026-01-16 07:38:20 INFO: Evaluating on dev set...
561
- 2026-01-16 07:38:22 INFO: LAS MLAS BLEX
562
- 2026-01-16 07:38:22 INFO: 19.42 11.02 14.05
563
- 2026-01-16 07:38:22 INFO: step 2900: train_loss = 1.980637, dev_score = 0.1942
564
- 2026-01-16 07:38:22 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
565
- 2026-01-16 07:38:22 INFO: new best model saved.
566
- 2026-01-16 07:38:23 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
567
- 2026-01-16 07:38:23 INFO: new model checkpoint saved.
568
- 2026-01-16 07:38:23 INFO: Finished STEP 2920/50000, loss = 2.065670 (0.029 sec/batch), lr: 0.003000
569
- 2026-01-16 07:38:24 INFO: Finished STEP 2940/50000, loss = 1.709635 (0.030 sec/batch), lr: 0.003000
570
- 2026-01-16 07:38:24 INFO: Finished STEP 2960/50000, loss = 3.801945 (0.030 sec/batch), lr: 0.003000
571
- 2026-01-16 07:38:25 INFO: Finished STEP 2980/50000, loss = 1.160876 (0.030 sec/batch), lr: 0.003000
572
- 2026-01-16 07:38:26 INFO: Finished STEP 3000/50000, loss = 3.921757 (0.030 sec/batch), lr: 0.003000
573
- 2026-01-16 07:38:26 INFO: Evaluating on dev set...
574
- 2026-01-16 07:38:28 INFO: LAS MLAS BLEX
575
- 2026-01-16 07:38:28 INFO: 12.19 4.47 6.57
576
- 2026-01-16 07:38:28 INFO: step 3000: train_loss = 1.918748, dev_score = 0.1219
577
- 2026-01-16 07:38:29 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
578
- 2026-01-16 07:38:29 INFO: new model checkpoint saved.
579
- 2026-01-16 07:38:29 INFO: Finished STEP 3020/50000, loss = 1.314914 (0.030 sec/batch), lr: 0.003000
580
- 2026-01-16 07:38:30 INFO: Finished STEP 3040/50000, loss = 3.862856 (0.031 sec/batch), lr: 0.003000
581
- 2026-01-16 07:38:31 INFO: Finished STEP 3060/50000, loss = 2.133690 (0.029 sec/batch), lr: 0.003000
582
- 2026-01-16 07:38:31 INFO: Finished STEP 3080/50000, loss = 2.644999 (0.031 sec/batch), lr: 0.003000
583
- 2026-01-16 07:38:32 INFO: Finished STEP 3100/50000, loss = 2.620658 (0.031 sec/batch), lr: 0.003000
584
- 2026-01-16 07:38:32 INFO: Evaluating on dev set...
585
- 2026-01-16 07:38:34 INFO: LAS MLAS BLEX
586
- 2026-01-16 07:38:34 INFO: 9.56 5.70 7.32
587
- 2026-01-16 07:38:34 INFO: step 3100: train_loss = 1.990910, dev_score = 0.0956
588
- 2026-01-16 07:38:34 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
589
- 2026-01-16 07:38:34 INFO: new model checkpoint saved.
590
- 2026-01-16 07:38:35 INFO: Finished STEP 3120/50000, loss = 1.526267 (0.029 sec/batch), lr: 0.003000
591
- 2026-01-16 07:38:35 INFO: Finished STEP 3140/50000, loss = 1.378218 (0.031 sec/batch), lr: 0.003000
592
- 2026-01-16 07:38:36 INFO: Finished STEP 3160/50000, loss = 1.390367 (0.030 sec/batch), lr: 0.003000
593
- 2026-01-16 07:38:37 INFO: Finished STEP 3180/50000, loss = 1.705234 (0.030 sec/batch), lr: 0.003000
594
- 2026-01-16 07:38:37 INFO: Finished STEP 3200/50000, loss = 1.402759 (0.030 sec/batch), lr: 0.003000
595
- 2026-01-16 07:38:37 INFO: Evaluating on dev set...
596
- 2026-01-16 07:38:40 INFO: LAS MLAS BLEX
597
- 2026-01-16 07:38:40 INFO: 17.65 10.33 12.02
598
- 2026-01-16 07:38:40 INFO: step 3200: train_loss = 2.082199, dev_score = 0.1765
599
- 2026-01-16 07:38:40 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
600
- 2026-01-16 07:38:40 INFO: new model checkpoint saved.
601
- 2026-01-16 07:38:41 INFO: Finished STEP 3220/50000, loss = 1.167270 (0.030 sec/batch), lr: 0.003000
602
- 2026-01-16 07:38:42 INFO: Finished STEP 3240/50000, loss = 3.461316 (0.031 sec/batch), lr: 0.003000
603
- 2026-01-16 07:38:42 INFO: Finished STEP 3260/50000, loss = 2.926430 (0.030 sec/batch), lr: 0.003000
604
- 2026-01-16 07:38:43 INFO: Finished STEP 3280/50000, loss = 3.029675 (0.030 sec/batch), lr: 0.003000
605
- 2026-01-16 07:38:44 INFO: Finished STEP 3300/50000, loss = 2.962304 (0.030 sec/batch), lr: 0.003000
606
- 2026-01-16 07:38:44 INFO: Evaluating on dev set...
607
- 2026-01-16 07:38:46 INFO: LAS MLAS BLEX
608
- 2026-01-16 07:38:46 INFO: 21.14 14.06 15.91
609
- 2026-01-16 07:38:46 INFO: step 3300: train_loss = 2.159422, dev_score = 0.2114
610
- 2026-01-16 07:38:46 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
611
- 2026-01-16 07:38:46 INFO: new best model saved.
612
- 2026-01-16 07:38:47 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
613
- 2026-01-16 07:38:47 INFO: new model checkpoint saved.
614
- 2026-01-16 07:38:47 INFO: Finished STEP 3320/50000, loss = 1.153689 (0.031 sec/batch), lr: 0.003000
615
- 2026-01-16 07:38:48 INFO: Finished STEP 3340/50000, loss = 6.680528 (0.030 sec/batch), lr: 0.003000
616
- 2026-01-16 07:38:49 INFO: Finished STEP 3360/50000, loss = 5.256881 (0.031 sec/batch), lr: 0.003000
617
- 2026-01-16 07:38:49 INFO: Finished STEP 3380/50000, loss = 2.249988 (0.030 sec/batch), lr: 0.003000
618
- 2026-01-16 07:38:50 INFO: Finished STEP 3400/50000, loss = 2.187397 (0.031 sec/batch), lr: 0.003000
619
- 2026-01-16 07:38:50 INFO: Evaluating on dev set...
620
- 2026-01-16 07:38:52 INFO: LAS MLAS BLEX
621
- 2026-01-16 07:38:52 INFO: 10.85 9.14 11.20
622
- 2026-01-16 07:38:52 INFO: step 3400: train_loss = 1.977379, dev_score = 0.1085
623
- 2026-01-16 07:38:52 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
624
- 2026-01-16 07:38:52 INFO: new model checkpoint saved.
625
- 2026-01-16 07:38:53 INFO: Finished STEP 3420/50000, loss = 1.144123 (0.031 sec/batch), lr: 0.003000
626
- 2026-01-16 07:38:53 INFO: Finished STEP 3440/50000, loss = 1.048997 (0.030 sec/batch), lr: 0.003000
627
- 2026-01-16 07:38:54 INFO: Finished STEP 3460/50000, loss = 2.239283 (0.033 sec/batch), lr: 0.003000
628
- 2026-01-16 07:38:55 INFO: Finished STEP 3480/50000, loss = 1.407322 (0.030 sec/batch), lr: 0.003000
629
- 2026-01-16 07:38:55 INFO: Finished STEP 3500/50000, loss = 1.793783 (0.029 sec/batch), lr: 0.003000
630
- 2026-01-16 07:38:55 INFO: Evaluating on dev set...
631
- 2026-01-16 07:38:57 INFO: LAS MLAS BLEX
632
- 2026-01-16 07:38:57 INFO: 25.55 14.33 15.18
633
- 2026-01-16 07:38:57 INFO: step 3500: train_loss = 1.776203, dev_score = 0.2555
634
- 2026-01-16 07:38:58 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
635
- 2026-01-16 07:38:58 INFO: new best model saved.
636
- 2026-01-16 07:38:58 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
637
- 2026-01-16 07:38:58 INFO: new model checkpoint saved.
638
- 2026-01-16 07:38:59 INFO: Finished STEP 3520/50000, loss = 1.173083 (0.031 sec/batch), lr: 0.003000
639
- 2026-01-16 07:38:59 INFO: Finished STEP 3540/50000, loss = 2.557373 (0.029 sec/batch), lr: 0.003000
640
- 2026-01-16 07:39:00 INFO: Finished STEP 3560/50000, loss = 0.362348 (0.029 sec/batch), lr: 0.003000
641
- 2026-01-16 07:39:01 INFO: Finished STEP 3580/50000, loss = 0.959906 (0.030 sec/batch), lr: 0.003000
642
- 2026-01-16 07:39:01 INFO: Finished STEP 3600/50000, loss = 0.376063 (0.029 sec/batch), lr: 0.003000
643
- 2026-01-16 07:39:01 INFO: Evaluating on dev set...
644
- 2026-01-16 07:39:03 INFO: LAS MLAS BLEX
645
- 2026-01-16 07:39:03 INFO: 10.60 5.75 9.91
646
- 2026-01-16 07:39:03 INFO: step 3600: train_loss = 1.918432, dev_score = 0.1060
647
- 2026-01-16 07:39:04 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
648
- 2026-01-16 07:39:04 INFO: new model checkpoint saved.
649
- 2026-01-16 07:39:04 INFO: Finished STEP 3620/50000, loss = 0.217418 (0.030 sec/batch), lr: 0.003000
650
- 2026-01-16 07:39:05 INFO: Finished STEP 3640/50000, loss = 1.392662 (0.030 sec/batch), lr: 0.003000
651
- 2026-01-16 07:39:05 INFO: Finished STEP 3660/50000, loss = 2.835163 (0.030 sec/batch), lr: 0.003000
652
- 2026-01-16 07:39:06 INFO: Finished STEP 3680/50000, loss = 2.799474 (0.030 sec/batch), lr: 0.003000
653
- 2026-01-16 07:39:07 INFO: Finished STEP 3700/50000, loss = 0.710460 (0.030 sec/batch), lr: 0.003000
654
- 2026-01-16 07:39:07 INFO: Evaluating on dev set...
655
- 2026-01-16 07:39:08 INFO: LAS MLAS BLEX
656
- 2026-01-16 07:39:08 INFO: 13.11 10.84 12.64
657
- 2026-01-16 07:39:08 INFO: step 3700: train_loss = 1.913150, dev_score = 0.1311
658
- 2026-01-16 07:39:09 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
659
- 2026-01-16 07:39:09 INFO: new model checkpoint saved.
660
- 2026-01-16 07:39:10 INFO: Finished STEP 3720/50000, loss = 2.065038 (0.030 sec/batch), lr: 0.003000
661
- 2026-01-16 07:39:10 INFO: Finished STEP 3740/50000, loss = 1.423430 (0.031 sec/batch), lr: 0.003000
662
- 2026-01-16 07:39:11 INFO: Finished STEP 3760/50000, loss = 0.174135 (0.031 sec/batch), lr: 0.003000
663
- 2026-01-16 07:39:11 INFO: Finished STEP 3780/50000, loss = 0.994005 (0.030 sec/batch), lr: 0.003000
664
- 2026-01-16 07:39:12 INFO: Finished STEP 3800/50000, loss = 2.113207 (0.030 sec/batch), lr: 0.003000
665
- 2026-01-16 07:39:12 INFO: Evaluating on dev set...
666
- 2026-01-16 07:39:14 INFO: LAS MLAS BLEX
667
- 2026-01-16 07:39:14 INFO: 11.70 9.55 12.36
668
- 2026-01-16 07:39:14 INFO: step 3800: train_loss = 1.940419, dev_score = 0.1170
669
- 2026-01-16 07:39:14 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
670
- 2026-01-16 07:39:14 INFO: new model checkpoint saved.
671
- 2026-01-16 07:39:15 INFO: Finished STEP 3820/50000, loss = 1.886823 (0.030 sec/batch), lr: 0.003000
672
- 2026-01-16 07:39:16 INFO: Finished STEP 3840/50000, loss = 0.537330 (0.029 sec/batch), lr: 0.003000
673
- 2026-01-16 07:39:16 INFO: Finished STEP 3860/50000, loss = 1.127199 (0.030 sec/batch), lr: 0.003000
674
- 2026-01-16 07:39:17 INFO: Finished STEP 3880/50000, loss = 1.348202 (0.030 sec/batch), lr: 0.003000
675
- 2026-01-16 07:39:17 INFO: Finished STEP 3900/50000, loss = 2.103697 (0.029 sec/batch), lr: 0.003000
676
- 2026-01-16 07:39:17 INFO: Evaluating on dev set...
677
- 2026-01-16 07:39:19 INFO: LAS MLAS BLEX
678
- 2026-01-16 07:39:19 INFO: 19.36 13.16 14.37
679
- 2026-01-16 07:39:19 INFO: step 3900: train_loss = 2.127213, dev_score = 0.1936
680
- 2026-01-16 07:39:20 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
681
- 2026-01-16 07:39:20 INFO: new model checkpoint saved.
682
- 2026-01-16 07:39:20 INFO: Finished STEP 3920/50000, loss = 2.552736 (0.031 sec/batch), lr: 0.003000
683
- 2026-01-16 07:39:21 INFO: Finished STEP 3940/50000, loss = 0.713954 (0.030 sec/batch), lr: 0.003000
684
- 2026-01-16 07:39:22 INFO: Finished STEP 3960/50000, loss = 0.829729 (0.030 sec/batch), lr: 0.003000
685
- 2026-01-16 07:39:22 INFO: Finished STEP 3980/50000, loss = 2.967856 (0.030 sec/batch), lr: 0.003000
686
- 2026-01-16 07:39:23 INFO: Finished STEP 4000/50000, loss = 2.044243 (0.031 sec/batch), lr: 0.003000
687
- 2026-01-16 07:39:23 INFO: Evaluating on dev set...
688
- 2026-01-16 07:39:25 INFO: LAS MLAS BLEX
689
- 2026-01-16 07:39:25 INFO: 13.91 7.27 8.25
690
- 2026-01-16 07:39:25 INFO: step 4000: train_loss = 1.918335, dev_score = 0.1391
691
- 2026-01-16 07:39:25 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
692
- 2026-01-16 07:39:25 INFO: new model checkpoint saved.
693
- 2026-01-16 07:39:26 INFO: Finished STEP 4020/50000, loss = 0.570849 (0.030 sec/batch), lr: 0.003000
694
- 2026-01-16 07:39:26 INFO: Finished STEP 4040/50000, loss = 0.004094 (0.030 sec/batch), lr: 0.003000
695
- 2026-01-16 07:39:27 INFO: Finished STEP 4060/50000, loss = 0.615223 (0.029 sec/batch), lr: 0.003000
696
- 2026-01-16 07:39:28 INFO: Finished STEP 4080/50000, loss = 1.598375 (0.031 sec/batch), lr: 0.003000
697
- 2026-01-16 07:39:28 INFO: Finished STEP 4100/50000, loss = 1.763149 (0.033 sec/batch), lr: 0.003000
698
- 2026-01-16 07:39:28 INFO: Evaluating on dev set...
699
- 2026-01-16 07:39:30 INFO: LAS MLAS BLEX
700
- 2026-01-16 07:39:30 INFO: 17.22 6.34 9.61
701
- 2026-01-16 07:39:30 INFO: step 4100: train_loss = 1.769413, dev_score = 0.1722
702
- 2026-01-16 07:39:31 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
703
- 2026-01-16 07:39:31 INFO: new model checkpoint saved.
704
- 2026-01-16 07:39:31 INFO: Finished STEP 4120/50000, loss = 4.137807 (0.031 sec/batch), lr: 0.003000
705
- 2026-01-16 07:39:32 INFO: Finished STEP 4140/50000, loss = 3.163526 (0.030 sec/batch), lr: 0.003000
706
- 2026-01-16 07:39:32 INFO: Finished STEP 4160/50000, loss = 7.217985 (0.030 sec/batch), lr: 0.003000
707
- 2026-01-16 07:39:33 INFO: Finished STEP 4180/50000, loss = 1.032177 (0.031 sec/batch), lr: 0.003000
708
- 2026-01-16 07:39:34 INFO: Finished STEP 4200/50000, loss = 0.475087 (0.030 sec/batch), lr: 0.003000
709
- 2026-01-16 07:39:34 INFO: Evaluating on dev set...
710
- 2026-01-16 07:39:36 INFO: LAS MLAS BLEX
711
- 2026-01-16 07:39:36 INFO: 10.11 6.28 11.64
712
- 2026-01-16 07:39:36 INFO: step 4200: train_loss = 1.876384, dev_score = 0.1011
713
- 2026-01-16 07:39:36 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
714
- 2026-01-16 07:39:36 INFO: new model checkpoint saved.
715
- 2026-01-16 07:39:37 INFO: Finished STEP 4220/50000, loss = 1.246820 (0.030 sec/batch), lr: 0.003000
716
- 2026-01-16 07:39:37 INFO: Finished STEP 4240/50000, loss = 1.793286 (0.030 sec/batch), lr: 0.003000
717
- 2026-01-16 07:39:38 INFO: Finished STEP 4260/50000, loss = 2.043184 (0.031 sec/batch), lr: 0.003000
718
- 2026-01-16 07:39:39 INFO: Finished STEP 4280/50000, loss = 2.374679 (0.030 sec/batch), lr: 0.003000
719
- 2026-01-16 07:39:39 INFO: Finished STEP 4300/50000, loss = 1.198715 (0.031 sec/batch), lr: 0.003000
720
- 2026-01-16 07:39:39 INFO: Evaluating on dev set...
721
- 2026-01-16 07:39:41 INFO: LAS MLAS BLEX
722
- 2026-01-16 07:39:41 INFO: 12.13 4.51 8.45
723
- 2026-01-16 07:39:41 INFO: step 4300: train_loss = 1.807878, dev_score = 0.1213
724
- 2026-01-16 07:39:42 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
725
- 2026-01-16 07:39:42 INFO: new model checkpoint saved.
726
- 2026-01-16 07:39:42 INFO: Finished STEP 4320/50000, loss = 1.510360 (0.030 sec/batch), lr: 0.003000
727
- 2026-01-16 07:39:43 INFO: Finished STEP 4340/50000, loss = 2.409623 (0.030 sec/batch), lr: 0.003000
728
- 2026-01-16 07:39:43 INFO: Finished STEP 4360/50000, loss = 1.254725 (0.031 sec/batch), lr: 0.003000
729
- 2026-01-16 07:39:44 INFO: Finished STEP 4380/50000, loss = 7.248045 (0.030 sec/batch), lr: 0.003000
730
- 2026-01-16 07:39:45 INFO: Finished STEP 4400/50000, loss = 2.985596 (0.030 sec/batch), lr: 0.003000
731
- 2026-01-16 07:39:45 INFO: Evaluating on dev set...
732
- 2026-01-16 07:39:46 INFO: LAS MLAS BLEX
733
- 2026-01-16 07:39:46 INFO: 11.95 7.40 9.70
734
- 2026-01-16 07:39:46 INFO: step 4400: train_loss = 1.787297, dev_score = 0.1195
735
- 2026-01-16 07:39:47 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
736
- 2026-01-16 07:39:47 INFO: new model checkpoint saved.
737
- 2026-01-16 07:39:47 INFO: Finished STEP 4420/50000, loss = 1.253725 (0.030 sec/batch), lr: 0.003000
738
- 2026-01-16 07:39:48 INFO: Finished STEP 4440/50000, loss = 2.283563 (0.029 sec/batch), lr: 0.003000
739
- 2026-01-16 07:39:49 INFO: Finished STEP 4460/50000, loss = 2.580783 (0.030 sec/batch), lr: 0.003000
740
- 2026-01-16 07:39:49 INFO: Finished STEP 4480/50000, loss = 1.721368 (0.032 sec/batch), lr: 0.003000
741
- 2026-01-16 07:39:50 INFO: Finished STEP 4500/50000, loss = 1.226920 (0.034 sec/batch), lr: 0.003000
742
- 2026-01-16 07:39:50 INFO: Evaluating on dev set...
743
- 2026-01-16 07:39:52 INFO: LAS MLAS BLEX
744
- 2026-01-16 07:39:52 INFO: 18.81 11.17 13.67
745
- 2026-01-16 07:39:52 INFO: step 4500: train_loss = 2.092089, dev_score = 0.1881
746
- 2026-01-16 07:39:52 INFO: Training ended with 4500 steps.
747
- 2026-01-16 07:39:52 INFO: Best dev F1 = 25.55, at iteration = 3500
748
- 2026-01-16 07:39:53 INFO: Running dev depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '8', '--dropout', '0.15']
749
- 2026-01-16 07:39:53 INFO: Running parser in predict mode
750
- 2026-01-16 07:39:53 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
751
- 2026-01-16 07:39:54 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
752
- 2026-01-16 07:39:54 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
753
- 2026-01-16 07:39:54 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
754
- 2026-01-16 07:39:54 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
755
- 2026-01-16 07:39:55 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
756
- 2026-01-16 07:39:55 INFO: Loading data with batch size 8...
757
- 2026-01-16 07:39:55 DEBUG: 34 batches created.
758
- 2026-01-16 07:39:57 INFO: F1 scores for each dependency:
759
- Note that unlabeled attachment errors hurt the labeled attachment scores
760
- acl: p 0.0000 r 0.0000 f1 0.0000 (14 actual)
761
- acl:relcl: p 0.0000 r 0.0000 f1 0.0000 (32 actual)
762
- advcl: p 0.0000 r 0.0000 f1 0.0000 (26 actual)
763
- advcl:relcl: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
764
- advmod: p 0.1391 r 0.5081 f1 0.2184 (124 actual)
765
- amod: p 0.1538 r 0.0194 f1 0.0345 (103 actual)
766
- appos: p 0.0000 r 0.0000 f1 0.0000 (12 actual)
767
- aux: p 0.0000 r 0.0000 f1 0.0000 (42 actual)
768
- aux:pass: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
769
- case: p 0.6946 r 0.6784 f1 0.6864 (171 actual)
770
- cc: p 0.0000 r 0.0000 f1 0.0000 (67 actual)
771
- ccomp: p 0.0000 r 0.0000 f1 0.0000 (17 actual)
772
- compound:prt: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
773
- conj: p 0.0000 r 0.0000 f1 0.0000 (64 actual)
774
- cop: p 0.0000 r 0.0000 f1 0.0000 (16 actual)
775
- csubj: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
776
- dep: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
777
- det: p 0.7222 r 0.7800 f1 0.7500 (100 actual)
778
- discourse: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
779
- expl: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
780
- fixed: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
781
- flat: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
782
- flat:name: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
783
- goeswith: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
784
- iobj: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
785
- mark: p 0.0000 r 0.0000 f1 0.0000 (68 actual)
786
- nmod: p 0.0000 r 0.0000 f1 0.0000 (55 actual)
787
- nmod:poss: p 0.3750 r 0.0484 f1 0.0857 (62 actual)
788
- nsubj: p 0.1684 r 0.5294 f1 0.2556 (119 actual)
789
- nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (10 actual)
790
- nummod: p 0.0000 r 0.0000 f1 0.0000 (10 actual)
791
- obj: p 0.0517 r 0.1481 f1 0.0767 (81 actual)
792
- obl: p 0.0270 r 0.0080 f1 0.0123 (125 actual)
793
- obl:agent: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
794
- parataxis: p 0.0000 r 0.0000 f1 0.0000 (13 actual)
795
- punct: p 0.3245 r 0.3333 f1 0.3288 (183 actual)
796
- reparandum: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
797
- root: p 0.4412 r 0.4412 f1 0.4412 (34 actual)
798
- vocative: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
799
- xcomp: p 0.0000 r 0.0000 f1 0.0000 (28 actual)
800
- 2026-01-16 07:39:57 INFO: LAS MLAS BLEX
801
- 2026-01-16 07:39:57 INFO: 25.55 14.33 15.18
802
- 2026-01-16 07:39:57 INFO: Parser score:
803
- 2026-01-16 07:39:57 INFO: sv_diachronic 25.55
804
- 2026-01-16 07:39:57 INFO: Finished running dev set on
805
- UD_Swedish-diachronic
806
- UAS LAS CLAS MLAS BLEX
807
- 49.08 25.55 15.18 14.33 15.18
808
- 2026-01-16 07:39:57 INFO: Running test depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '8', '--dropout', '0.15']
809
- 2026-01-16 07:39:57 INFO: Running parser in predict mode
810
- 2026-01-16 07:39:57 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
811
- 2026-01-16 07:39:59 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
812
- 2026-01-16 07:39:59 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
813
- 2026-01-16 07:39:59 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
814
- 2026-01-16 07:39:59 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
815
- 2026-01-16 07:39:59 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
816
- 2026-01-16 07:39:59 INFO: Loading data with batch size 8...
817
- 2026-01-16 07:39:59 DEBUG: 74 batches created.
818
- 2026-01-16 07:40:02 INFO: F1 scores for each dependency:
819
- Note that unlabeled attachment errors hurt the labeled attachment scores
820
- acl: p 0.0000 r 0.0000 f1 0.0000 (21 actual)
821
- acl:cleft: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
822
- acl:relcl: p 0.0000 r 0.0000 f1 0.0000 (50 actual)
823
- advcl: p 0.0000 r 0.0000 f1 0.0000 (39 actual)
824
- advmod: p 0.1618 r 0.5858 f1 0.2535 (169 actual)
825
- amod: p 0.1698 r 0.0570 f1 0.0853 (158 actual)
826
- appos: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
827
- aux: p 0.0000 r 0.0000 f1 0.0000 (53 actual)
828
- aux:pass: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
829
- case: p 0.7148 r 0.7868 f1 0.7491 (258 actual)
830
- cc: p 0.0000 r 0.0000 f1 0.0000 (101 actual)
831
- ccomp: p 0.0000 r 0.0000 f1 0.0000 (20 actual)
832
- compound:prt: p 0.0000 r 0.0000 f1 0.0000 (14 actual)
833
- conj: p 0.0000 r 0.0000 f1 0.0000 (106 actual)
834
- cop: p 0.0000 r 0.0000 f1 0.0000 (33 actual)
835
- csubj: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
836
- det: p 0.6382 r 0.7462 f1 0.6879 (130 actual)
837
- discourse: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
838
- dislocated: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
839
- expl: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
840
- expl:pv: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
841
- fixed: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
842
- flat:name: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
843
- iobj: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
844
- mark: p 0.0000 r 0.0000 f1 0.0000 (97 actual)
845
- nmod: p 0.0000 r 0.0000 f1 0.0000 (62 actual)
846
- nmod:poss: p 0.0588 r 0.0101 f1 0.0172 (99 actual)
847
- nsubj: p 0.1690 r 0.4719 f1 0.2489 (178 actual)
848
- nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (20 actual)
849
- obj: p 0.0508 r 0.1532 f1 0.0763 (124 actual)
850
- obl: p 0.0000 r 0.0000 f1 0.0000 (194 actual)
851
- obl:agent: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
852
- orphan: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
853
- parataxis: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
854
- punct: p 0.3435 r 0.3435 f1 0.3435 (294 actual)
855
- root: p 0.4189 r 0.4189 f1 0.4189 (74 actual)
856
- vocative: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
857
- xcomp: p 0.0000 r 0.0000 f1 0.0000 (55 actual)
858
- 2026-01-16 07:40:02 INFO: LAS MLAS BLEX
859
- 2026-01-16 07:40:02 INFO: 27.02 15.21 16.05
860
- 2026-01-16 07:40:02 INFO: Parser score:
861
- 2026-01-16 07:40:02 INFO: sv_diachronic 27.02
862
- 2026-01-16 07:40:03 INFO: Finished running test set on
863
- UD_Swedish-diachronic
864
- UAS LAS CLAS MLAS BLEX
865
- 49.63 27.02 16.05 15.21 16.05
866
- DONE.
867
- Full log saved to: logs/log_conll17.pt_sv_diachron_20260116_073035.txt
868
- Symlink updated: logs/latest.txt → log_conll17.pt_sv_diachron_20260116_073035.txt
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
logs/log_conll17.pt_sv_diachron_de_lit_20251203_154509.txt DELETED
@@ -1,740 +0,0 @@
1
- === LOGFILE: logs/log_conll17.pt_sv_diachron_de_lit_20251203_154509.txt ===
2
- Language codes: sv diachron de_lit
3
- Using pretrained model: conll17.pt
4
-
5
- Running: python prepare-train-val-test.py sv diachron de_lit
6
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
7
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
8
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
9
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
10
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
11
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
12
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/ucxn_ud_swedish-talbanken.conllu
13
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
14
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
15
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
16
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
17
- Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-de_lit/de_lit-ud-test.conllu
18
- Including DigPhil MACHINE in TRAIN (minus gold)…
19
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
20
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
21
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
22
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
23
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
24
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
25
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
26
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
27
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
28
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
29
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
30
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
31
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
32
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
33
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
34
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
35
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
36
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
37
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
38
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
39
- Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
40
- Cleaning TRAIN...
41
- [REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
42
- [REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
43
- [REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
44
- [REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
45
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
46
- [REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
47
- [REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
48
- [REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
49
- [REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
50
- [REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
51
- [REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
52
- [REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
53
- [REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
54
- [REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
55
- [REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
56
- [REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
57
- [REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
58
- [REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
59
- [REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
60
- [REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
61
- [REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
62
- [REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
63
- [REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
64
- [REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
65
- [REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
66
- [REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
67
- [REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
68
- [REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
69
- [REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
70
- [REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
71
- [REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
72
- [REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
73
- [REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
74
- [REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
75
- [REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
76
- [REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
77
- [REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
78
- [REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
79
- [REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
80
- [REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
81
- [REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
82
- [REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
83
- [REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
84
- [REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
85
- [REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
86
- [REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
87
- [REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
88
- [REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
89
- [REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
90
- [REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
91
- [REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
92
- [REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
93
- [REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
94
- [REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
95
- [REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
96
- [REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
97
- [REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
98
- [REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
99
- [REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
100
- [REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
101
- [REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
102
- [REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
103
- [REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
104
- Cleaning DEV...
105
- [REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
106
- Cleaning TEST...
107
- Writing TRAIN → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (68172 valid sentences)
108
- Writing DEV → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (9 valid sentences)
109
- Writing TEST → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (99 valid sentences)
110
- Done.
111
- Sourcing scripts/config_alvis.sh
112
- Running stanza dataset preparation…
113
- 2025-12-03 15:45:18 INFO: Datasets program called with:
114
- /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
115
- 2025-12-03 15:45:18 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
116
-
117
- 2025-12-03 15:45:18 INFO: Downloaded file to /cephyr/users/cleland/Alvis/stanza_resources/resources.json
118
- 2025-12-03 15:45:18 DEBUG: Processing parameter "processors"...
119
- 2025-12-03 15:45:18 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
120
- 2025-12-03 15:45:18 INFO: Downloading these customized packages for language: sv (Swedish)...
121
- =======================
122
- | Processor | Package |
123
- -----------------------
124
- =======================
125
-
126
- 2025-12-03 15:45:18 INFO: Finished downloading models and saved to /cephyr/users/cleland/Alvis/stanza_resources
127
- 2025-12-03 15:45:18 INFO: Using tagger model in /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt for sv_diachronic
128
- 2025-12-03 15:45:18 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
129
- 2025-12-03 15:45:18 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
130
- Augmented 188 quotes: Counter({'«»': 23, '»«': 22, '″″': 22, '""': 20, '““': 20, '„”': 19, '「」': 18, '《》': 16, '„“': 15, '””': 13})
131
- 2025-12-03 15:45:21 INFO: Running tagger to retag /local/tmp.5440223/tmplttwhhzq/sv_diachronic.train.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
132
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5440223/tmplttwhhzq/sv_diachronic.train.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu']
133
- 2025-12-03 15:45:22 INFO: Running tagger in predict mode
134
- 2025-12-03 15:45:22 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
135
- 2025-12-03 15:45:23 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
136
- 2025-12-03 15:45:23 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
137
- 2025-12-03 15:45:23 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
138
- 2025-12-03 15:45:23 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
139
- 2025-12-03 15:45:24 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
140
- 2025-12-03 15:45:26 INFO: Loading data with batch size 250...
141
- 2025-12-03 15:46:06 INFO: Start evaluation...
142
- 2025-12-03 15:49:44 INFO: UPOS XPOS UFeats AllTags
143
- 2025-12-03 15:49:44 INFO: 96.83 85.29 95.46 84.47
144
- 2025-12-03 15:49:44 INFO: POS Tagger score: sv_diachronic 84.47
145
- 2025-12-03 15:49:45 INFO: Running tagger to retag /local/tmp.5440223/tmplttwhhzq/sv_diachronic.dev.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
146
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5440223/tmplttwhhzq/sv_diachronic.dev.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu']
147
- 2025-12-03 15:49:45 INFO: Running tagger in predict mode
148
- 2025-12-03 15:49:45 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
149
- 2025-12-03 15:49:47 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
150
- 2025-12-03 15:49:47 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
151
- 2025-12-03 15:49:47 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
152
- 2025-12-03 15:49:47 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
153
- 2025-12-03 15:49:47 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
154
- 2025-12-03 15:49:47 INFO: Loading data with batch size 250...
155
- 2025-12-03 15:49:47 INFO: Start evaluation...
156
- 2025-12-03 15:49:47 INFO: UPOS XPOS UFeats AllTags
157
- 2025-12-03 15:49:47 INFO: 93.32 90.84 93.32 85.64
158
- 2025-12-03 15:49:47 INFO: POS Tagger score: sv_diachronic 85.64
159
- 2025-12-03 15:49:47 INFO: Running tagger to retag /local/tmp.5440223/tmplttwhhzq/sv_diachronic.test.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
160
- Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5440223/tmplttwhhzq/sv_diachronic.test.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu']
161
- 2025-12-03 15:49:47 INFO: Running tagger in predict mode
162
- 2025-12-03 15:49:47 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
163
- 2025-12-03 15:49:49 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
164
- 2025-12-03 15:49:49 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
165
- 2025-12-03 15:49:49 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
166
- 2025-12-03 15:49:49 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
167
- 2025-12-03 15:49:49 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
168
- 2025-12-03 15:49:49 INFO: Loading data with batch size 250...
169
- 2025-12-03 15:49:49 INFO: Start evaluation...
170
- 2025-12-03 15:49:50 INFO: UPOS XPOS UFeats AllTags
171
- 2025-12-03 15:49:50 INFO: 93.14 96.78 95.32 90.28
172
- 2025-12-03 15:49:50 INFO: POS Tagger score: sv_diachronic 90.28
173
- Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
174
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /local/tmp.5440223/tmplttwhhzq/sv_diachronic.train.gold.conllu
175
- Swapped 'w1, w2' for 'w1 ,w2' 141 times
176
- Added 573 new sentences with asdf, zzzz -> asdf,zzzz
177
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /local/tmp.5440223/tmplttwhhzq/sv_diachronic.dev.gold.conllu
178
- Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /local/tmp.5440223/tmplttwhhzq/sv_diachronic.test.gold.conllu
179
- Running stanza dependency parser training…
180
- 2025-12-03 15:50:04 INFO: Training program called with:
181
- /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt --batch_size 32 --dropout 0.33
182
- 2025-12-03 15:50:04 DEBUG: UD_Swedish-diachronic: sv_diachronic
183
- 2025-12-03 15:50:04 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
184
- 2025-12-03 15:50:04 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
185
- 2025-12-03 15:50:04 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_charlm_parser.pt does not exist, training new model
186
- 2025-12-03 15:50:04 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
187
- 2025-12-03 15:50:04 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
188
- 2025-12-03 15:50:04 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--train_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
189
- 2025-12-03 15:50:04 INFO: Running parser in train mode
190
- 2025-12-03 15:50:04 INFO: Using pretrained contextualized char embedding
191
- 2025-12-03 15:50:04 INFO: Loading data with batch size 32...
192
- 2025-12-03 15:50:11 INFO: Train File /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 68745
193
- 2025-12-03 15:50:11 INFO: Original data size: 68745
194
- 2025-12-03 15:50:12 INFO: Augmented data size: 69046
195
- 2025-12-03 15:50:31 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
196
- 2025-12-03 15:50:31 INFO: Original length = 69046
197
- 2025-12-03 15:50:31 INFO: Filtered length = 69046
198
- 2025-12-03 15:50:47 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='|') for the xpos factory for sv_diachronic
199
- 2025-12-03 15:50:55 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
200
- 2025-12-03 15:51:09 DEBUG: 40457 batches created.
201
- 2025-12-03 15:51:09 DEBUG: 9 batches created.
202
- 2025-12-03 15:51:09 INFO: Training parser...
203
- 2025-12-03 15:51:09 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
204
- 2025-12-03 15:51:09 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
205
- 2025-12-03 15:51:09 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
206
- 2025-12-03 15:51:10 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
207
- 2025-12-03 15:51:17 INFO: Finished STEP 20/50000, loss = 6.791992 (0.200 sec/batch), lr: 0.003000
208
- 2025-12-03 15:51:21 INFO: Finished STEP 40/50000, loss = 5.211706 (0.167 sec/batch), lr: 0.003000
209
- 2025-12-03 15:51:24 INFO: Finished STEP 60/50000, loss = 5.668643 (0.147 sec/batch), lr: 0.003000
210
- 2025-12-03 15:51:27 INFO: Finished STEP 80/50000, loss = 4.197357 (0.151 sec/batch), lr: 0.003000
211
- 2025-12-03 15:51:30 INFO: Finished STEP 100/50000, loss = 5.356755 (0.129 sec/batch), lr: 0.003000
212
- 2025-12-03 15:51:30 INFO: Evaluating on dev set...
213
- 2025-12-03 15:51:31 INFO: LAS MLAS BLEX
214
- 2025-12-03 15:51:31 INFO: 39.85 26.44 33.26
215
- 2025-12-03 15:51:31 INFO: step 100: train_loss = 8.587716, dev_score = 0.3985
216
- 2025-12-03 15:51:31 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
217
- 2025-12-03 15:51:31 INFO: new best model saved.
218
- 2025-12-03 15:51:32 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
219
- 2025-12-03 15:51:32 INFO: new model checkpoint saved.
220
- 2025-12-03 15:51:34 INFO: Finished STEP 120/50000, loss = 4.996027 (0.120 sec/batch), lr: 0.003000
221
- 2025-12-03 15:51:37 INFO: Finished STEP 140/50000, loss = 5.125143 (0.131 sec/batch), lr: 0.003000
222
- 2025-12-03 15:51:40 INFO: Finished STEP 160/50000, loss = 4.139769 (0.138 sec/batch), lr: 0.003000
223
- 2025-12-03 15:51:42 INFO: Finished STEP 180/50000, loss = 4.071211 (0.131 sec/batch), lr: 0.003000
224
- 2025-12-03 15:51:45 INFO: Finished STEP 200/50000, loss = 4.302421 (0.121 sec/batch), lr: 0.003000
225
- 2025-12-03 15:51:45 INFO: Evaluating on dev set...
226
- 2025-12-03 15:51:45 INFO: LAS MLAS BLEX
227
- 2025-12-03 15:51:45 INFO: 50.00 35.44 40.93
228
- 2025-12-03 15:51:45 INFO: step 200: train_loss = 4.776229, dev_score = 0.5000
229
- 2025-12-03 15:51:45 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
230
- 2025-12-03 15:51:45 INFO: new best model saved.
231
- 2025-12-03 15:51:46 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
232
- 2025-12-03 15:51:46 INFO: new model checkpoint saved.
233
- 2025-12-03 15:51:48 INFO: Finished STEP 220/50000, loss = 4.349181 (0.124 sec/batch), lr: 0.003000
234
- 2025-12-03 15:51:51 INFO: Finished STEP 240/50000, loss = 5.862440 (0.125 sec/batch), lr: 0.003000
235
- 2025-12-03 15:51:53 INFO: Finished STEP 260/50000, loss = 4.765276 (0.118 sec/batch), lr: 0.003000
236
- 2025-12-03 15:51:56 INFO: Finished STEP 280/50000, loss = 4.617913 (0.116 sec/batch), lr: 0.003000
237
- 2025-12-03 15:51:58 INFO: Finished STEP 300/50000, loss = 5.353960 (0.113 sec/batch), lr: 0.003000
238
- 2025-12-03 15:51:58 INFO: Evaluating on dev set...
239
- 2025-12-03 15:51:58 INFO: LAS MLAS BLEX
240
- 2025-12-03 15:51:58 INFO: 51.49 38.56 43.22
241
- 2025-12-03 15:51:58 INFO: step 300: train_loss = 4.421819, dev_score = 0.5149
242
- 2025-12-03 15:51:59 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
243
- 2025-12-03 15:51:59 INFO: new best model saved.
244
- 2025-12-03 15:51:59 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
245
- 2025-12-03 15:51:59 INFO: new model checkpoint saved.
246
- 2025-12-03 15:52:02 INFO: Finished STEP 320/50000, loss = 5.069269 (0.121 sec/batch), lr: 0.003000
247
- 2025-12-03 15:52:04 INFO: Finished STEP 340/50000, loss = 3.614627 (0.112 sec/batch), lr: 0.003000
248
- 2025-12-03 15:52:06 INFO: Finished STEP 360/50000, loss = 3.878788 (0.115 sec/batch), lr: 0.003000
249
- 2025-12-03 15:52:08 INFO: Finished STEP 380/50000, loss = 4.367438 (0.100 sec/batch), lr: 0.003000
250
- 2025-12-03 15:52:10 INFO: Finished STEP 400/50000, loss = 4.081972 (0.107 sec/batch), lr: 0.003000
251
- 2025-12-03 15:52:10 INFO: Evaluating on dev set...
252
- 2025-12-03 15:52:11 INFO: LAS MLAS BLEX
253
- 2025-12-03 15:52:11 INFO: 55.69 43.55 46.51
254
- 2025-12-03 15:52:11 INFO: step 400: train_loss = 4.280971, dev_score = 0.5569
255
- 2025-12-03 15:52:11 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
256
- 2025-12-03 15:52:11 INFO: new best model saved.
257
- 2025-12-03 15:52:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
258
- 2025-12-03 15:52:12 INFO: new model checkpoint saved.
259
- 2025-12-03 15:52:14 INFO: Finished STEP 420/50000, loss = 3.892664 (0.110 sec/batch), lr: 0.003000
260
- 2025-12-03 15:52:16 INFO: Finished STEP 440/50000, loss = 3.998952 (0.100 sec/batch), lr: 0.003000
261
- 2025-12-03 15:52:18 INFO: Finished STEP 460/50000, loss = 5.833274 (0.112 sec/batch), lr: 0.003000
262
- 2025-12-03 15:52:20 INFO: Finished STEP 480/50000, loss = 3.561453 (0.106 sec/batch), lr: 0.003000
263
- 2025-12-03 15:52:22 INFO: Finished STEP 500/50000, loss = 3.267099 (0.106 sec/batch), lr: 0.003000
264
- 2025-12-03 15:52:22 INFO: Evaluating on dev set...
265
- 2025-12-03 15:52:23 INFO: LAS MLAS BLEX
266
- 2025-12-03 15:52:23 INFO: 55.45 44.30 48.95
267
- 2025-12-03 15:52:23 INFO: step 500: train_loss = 4.074671, dev_score = 0.5545
268
- 2025-12-03 15:52:24 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
269
- 2025-12-03 15:52:24 INFO: new model checkpoint saved.
270
- 2025-12-03 15:52:26 INFO: Finished STEP 520/50000, loss = 3.987905 (0.101 sec/batch), lr: 0.003000
271
- 2025-12-03 15:52:28 INFO: Finished STEP 540/50000, loss = 4.040174 (0.105 sec/batch), lr: 0.003000
272
- 2025-12-03 15:52:30 INFO: Finished STEP 560/50000, loss = 5.425183 (0.096 sec/batch), lr: 0.003000
273
- 2025-12-03 15:52:32 INFO: Finished STEP 580/50000, loss = 3.171549 (0.098 sec/batch), lr: 0.003000
274
- 2025-12-03 15:52:34 INFO: Finished STEP 600/50000, loss = 5.054599 (0.087 sec/batch), lr: 0.003000
275
- 2025-12-03 15:52:34 INFO: Evaluating on dev set...
276
- 2025-12-03 15:52:34 INFO: LAS MLAS BLEX
277
- 2025-12-03 15:52:34 INFO: 59.41 50.85 53.39
278
- 2025-12-03 15:52:34 INFO: step 600: train_loss = 3.989031, dev_score = 0.5941
279
- 2025-12-03 15:52:34 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
280
- 2025-12-03 15:52:34 INFO: new best model saved.
281
- 2025-12-03 15:52:35 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
282
- 2025-12-03 15:52:35 INFO: new model checkpoint saved.
283
- 2025-12-03 15:52:37 INFO: Finished STEP 620/50000, loss = 5.185298 (0.107 sec/batch), lr: 0.003000
284
- 2025-12-03 15:52:39 INFO: Finished STEP 640/50000, loss = 3.282787 (0.098 sec/batch), lr: 0.003000
285
- 2025-12-03 15:52:41 INFO: Finished STEP 660/50000, loss = 7.106239 (0.095 sec/batch), lr: 0.003000
286
- 2025-12-03 15:52:43 INFO: Finished STEP 680/50000, loss = 3.663255 (0.099 sec/batch), lr: 0.003000
287
- 2025-12-03 15:52:45 INFO: Finished STEP 700/50000, loss = 4.011485 (0.098 sec/batch), lr: 0.003000
288
- 2025-12-03 15:52:45 INFO: Evaluating on dev set...
289
- 2025-12-03 15:52:46 INFO: LAS MLAS BLEX
290
- 2025-12-03 15:52:46 INFO: 59.16 46.74 52.21
291
- 2025-12-03 15:52:46 INFO: step 700: train_loss = 3.973323, dev_score = 0.5916
292
- 2025-12-03 15:52:46 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
293
- 2025-12-03 15:52:46 INFO: new model checkpoint saved.
294
- 2025-12-03 15:52:48 INFO: Finished STEP 720/50000, loss = 4.034673 (0.094 sec/batch), lr: 0.003000
295
- 2025-12-03 15:52:50 INFO: Finished STEP 740/50000, loss = 4.719307 (0.099 sec/batch), lr: 0.003000
296
- 2025-12-03 15:52:52 INFO: Finished STEP 760/50000, loss = 3.175733 (0.089 sec/batch), lr: 0.003000
297
- 2025-12-03 15:52:54 INFO: Finished STEP 780/50000, loss = 3.158538 (0.098 sec/batch), lr: 0.003000
298
- 2025-12-03 15:52:56 INFO: Finished STEP 800/50000, loss = 2.881503 (0.099 sec/batch), lr: 0.003000
299
- 2025-12-03 15:52:56 INFO: Evaluating on dev set...
300
- 2025-12-03 15:52:56 INFO: LAS MLAS BLEX
301
- 2025-12-03 15:52:56 INFO: 57.92 46.19 49.15
302
- 2025-12-03 15:52:56 INFO: step 800: train_loss = 3.763542, dev_score = 0.5792
303
- 2025-12-03 15:52:57 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
304
- 2025-12-03 15:52:57 INFO: new model checkpoint saved.
305
- 2025-12-03 15:52:59 INFO: Finished STEP 820/50000, loss = 4.045841 (0.087 sec/batch), lr: 0.003000
306
- 2025-12-03 15:53:01 INFO: Finished STEP 840/50000, loss = 3.735970 (0.088 sec/batch), lr: 0.003000
307
- 2025-12-03 15:53:02 INFO: Finished STEP 860/50000, loss = 4.020472 (0.080 sec/batch), lr: 0.003000
308
- 2025-12-03 15:53:04 INFO: Finished STEP 880/50000, loss = 5.290604 (0.091 sec/batch), lr: 0.003000
309
- 2025-12-03 15:53:06 INFO: Finished STEP 900/50000, loss = 3.524802 (0.089 sec/batch), lr: 0.003000
310
- 2025-12-03 15:53:06 INFO: Evaluating on dev set...
311
- 2025-12-03 15:53:07 INFO: LAS MLAS BLEX
312
- 2025-12-03 15:53:07 INFO: 56.93 45.67 49.47
313
- 2025-12-03 15:53:07 INFO: step 900: train_loss = 3.948068, dev_score = 0.5693
314
- 2025-12-03 15:53:07 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
315
- 2025-12-03 15:53:07 INFO: new model checkpoint saved.
316
- 2025-12-03 15:53:09 INFO: Finished STEP 920/50000, loss = 3.348033 (0.088 sec/batch), lr: 0.003000
317
- 2025-12-03 15:53:11 INFO: Finished STEP 940/50000, loss = 3.860394 (0.091 sec/batch), lr: 0.003000
318
- 2025-12-03 15:53:13 INFO: Finished STEP 960/50000, loss = 3.459932 (0.082 sec/batch), lr: 0.003000
319
- 2025-12-03 15:53:15 INFO: Finished STEP 980/50000, loss = 4.335596 (0.085 sec/batch), lr: 0.003000
320
- 2025-12-03 15:53:17 INFO: Finished STEP 1000/50000, loss = 3.108686 (0.089 sec/batch), lr: 0.003000
321
- 2025-12-03 15:53:17 INFO: Evaluating on dev set...
322
- 2025-12-03 15:53:17 INFO: LAS MLAS BLEX
323
- 2025-12-03 15:53:17 INFO: 60.40 51.27 53.81
324
- 2025-12-03 15:53:17 INFO: step 1000: train_loss = 3.976391, dev_score = 0.6040
325
- 2025-12-03 15:53:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
326
- 2025-12-03 15:53:17 INFO: new best model saved.
327
- 2025-12-03 15:53:18 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
328
- 2025-12-03 15:53:18 INFO: new model checkpoint saved.
329
- 2025-12-03 15:53:20 INFO: Finished STEP 1020/50000, loss = 6.081649 (0.089 sec/batch), lr: 0.003000
330
- 2025-12-03 15:53:21 INFO: Finished STEP 1040/50000, loss = 3.454382 (0.089 sec/batch), lr: 0.003000
331
- 2025-12-03 15:53:23 INFO: Finished STEP 1060/50000, loss = 3.093398 (0.085 sec/batch), lr: 0.003000
332
- 2025-12-03 15:53:25 INFO: Finished STEP 1080/50000, loss = 3.780082 (0.088 sec/batch), lr: 0.003000
333
- 2025-12-03 15:53:27 INFO: Finished STEP 1100/50000, loss = 3.758047 (0.075 sec/batch), lr: 0.003000
334
- 2025-12-03 15:53:27 INFO: Evaluating on dev set...
335
- 2025-12-03 15:53:27 INFO: LAS MLAS BLEX
336
- 2025-12-03 15:53:27 INFO: 59.41 49.05 52.85
337
- 2025-12-03 15:53:27 INFO: step 1100: train_loss = 3.779099, dev_score = 0.5941
338
- 2025-12-03 15:53:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
339
- 2025-12-03 15:53:28 INFO: new model checkpoint saved.
340
- 2025-12-03 15:53:30 INFO: Finished STEP 1120/50000, loss = 3.774802 (0.088 sec/batch), lr: 0.003000
341
- 2025-12-03 15:53:31 INFO: Finished STEP 1140/50000, loss = 2.977976 (0.089 sec/batch), lr: 0.003000
342
- 2025-12-03 15:53:33 INFO: Finished STEP 1160/50000, loss = 3.838738 (0.078 sec/batch), lr: 0.003000
343
- 2025-12-03 15:53:35 INFO: Finished STEP 1180/50000, loss = 4.660880 (0.080 sec/batch), lr: 0.003000
344
- 2025-12-03 15:53:36 INFO: Finished STEP 1200/50000, loss = 2.489745 (0.088 sec/batch), lr: 0.003000
345
- 2025-12-03 15:53:36 INFO: Evaluating on dev set...
346
- 2025-12-03 15:53:37 INFO: LAS MLAS BLEX
347
- 2025-12-03 15:53:37 INFO: 58.66 48.31 51.69
348
- 2025-12-03 15:53:37 INFO: step 1200: train_loss = 3.720397, dev_score = 0.5866
349
- 2025-12-03 15:53:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
350
- 2025-12-03 15:53:37 INFO: new model checkpoint saved.
351
- 2025-12-03 15:53:39 INFO: Finished STEP 1220/50000, loss = 3.247226 (0.090 sec/batch), lr: 0.003000
352
- 2025-12-03 15:53:41 INFO: Finished STEP 1240/50000, loss = 4.093720 (0.079 sec/batch), lr: 0.003000
353
- 2025-12-03 15:53:43 INFO: Finished STEP 1260/50000, loss = 2.970687 (0.080 sec/batch), lr: 0.003000
354
- 2025-12-03 15:53:44 INFO: Finished STEP 1280/50000, loss = 3.959898 (0.087 sec/batch), lr: 0.003000
355
- 2025-12-03 15:53:46 INFO: Finished STEP 1300/50000, loss = 3.890834 (0.081 sec/batch), lr: 0.003000
356
- 2025-12-03 15:53:46 INFO: Evaluating on dev set...
357
- 2025-12-03 15:53:47 INFO: LAS MLAS BLEX
358
- 2025-12-03 15:53:47 INFO: 57.67 50.00 52.97
359
- 2025-12-03 15:53:47 INFO: step 1300: train_loss = 3.945667, dev_score = 0.5767
360
- 2025-12-03 15:53:47 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
361
- 2025-12-03 15:53:47 INFO: new model checkpoint saved.
362
- 2025-12-03 15:53:49 INFO: Finished STEP 1320/50000, loss = 4.299366 (0.076 sec/batch), lr: 0.003000
363
- 2025-12-03 15:53:51 INFO: Finished STEP 1340/50000, loss = 3.080094 (0.082 sec/batch), lr: 0.003000
364
- 2025-12-03 15:53:52 INFO: Finished STEP 1360/50000, loss = 5.326971 (0.087 sec/batch), lr: 0.003000
365
- 2025-12-03 15:53:54 INFO: Finished STEP 1380/50000, loss = 4.148820 (0.083 sec/batch), lr: 0.003000
366
- 2025-12-03 15:53:56 INFO: Finished STEP 1400/50000, loss = 3.131835 (0.085 sec/batch), lr: 0.003000
367
- 2025-12-03 15:53:56 INFO: Evaluating on dev set...
368
- 2025-12-03 15:53:56 INFO: LAS MLAS BLEX
369
- 2025-12-03 15:53:56 INFO: 59.16 49.47 52.85
370
- 2025-12-03 15:53:56 INFO: step 1400: train_loss = 3.828261, dev_score = 0.5916
371
- 2025-12-03 15:53:57 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
372
- 2025-12-03 15:53:57 INFO: new model checkpoint saved.
373
- 2025-12-03 15:53:58 INFO: Finished STEP 1420/50000, loss = 3.877927 (0.079 sec/batch), lr: 0.003000
374
- 2025-12-03 15:54:00 INFO: Finished STEP 1440/50000, loss = 2.828221 (0.077 sec/batch), lr: 0.003000
375
- 2025-12-03 15:54:02 INFO: Finished STEP 1460/50000, loss = 2.771763 (0.088 sec/batch), lr: 0.003000
376
- 2025-12-03 15:54:03 INFO: Finished STEP 1480/50000, loss = 3.887072 (0.082 sec/batch), lr: 0.003000
377
- 2025-12-03 15:54:05 INFO: Finished STEP 1500/50000, loss = 3.089223 (0.080 sec/batch), lr: 0.003000
378
- 2025-12-03 15:54:05 INFO: Evaluating on dev set...
379
- 2025-12-03 15:54:05 INFO: LAS MLAS BLEX
380
- 2025-12-03 15:54:05 INFO: 60.89 49.47 54.12
381
- 2025-12-03 15:54:05 INFO: step 1500: train_loss = 3.662534, dev_score = 0.6089
382
- 2025-12-03 15:54:06 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
383
- 2025-12-03 15:54:06 INFO: new best model saved.
384
- 2025-12-03 15:54:06 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
385
- 2025-12-03 15:54:06 INFO: new model checkpoint saved.
386
- 2025-12-03 15:54:08 INFO: Finished STEP 1520/50000, loss = 3.424534 (0.079 sec/batch), lr: 0.003000
387
- 2025-12-03 15:54:10 INFO: Finished STEP 1540/50000, loss = 4.257630 (0.081 sec/batch), lr: 0.003000
388
- 2025-12-03 15:54:11 INFO: Finished STEP 1560/50000, loss = 4.004986 (0.078 sec/batch), lr: 0.003000
389
- 2025-12-03 15:54:13 INFO: Finished STEP 1580/50000, loss = 4.259016 (0.076 sec/batch), lr: 0.003000
390
- 2025-12-03 15:54:14 INFO: Finished STEP 1600/50000, loss = 4.886831 (0.071 sec/batch), lr: 0.003000
391
- 2025-12-03 15:54:14 INFO: Evaluating on dev set...
392
- 2025-12-03 15:54:15 INFO: LAS MLAS BLEX
393
- 2025-12-03 15:54:15 INFO: 58.66 49.47 52.01
394
- 2025-12-03 15:54:15 INFO: step 1600: train_loss = 3.940190, dev_score = 0.5866
395
- 2025-12-03 15:54:15 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
396
- 2025-12-03 15:54:15 INFO: new model checkpoint saved.
397
- 2025-12-03 15:54:17 INFO: Finished STEP 1620/50000, loss = 3.486598 (0.072 sec/batch), lr: 0.003000
398
- 2025-12-03 15:54:19 INFO: Finished STEP 1640/50000, loss = 3.426835 (0.076 sec/batch), lr: 0.003000
399
- 2025-12-03 15:54:20 INFO: Finished STEP 1660/50000, loss = 3.082500 (0.077 sec/batch), lr: 0.003000
400
- 2025-12-03 15:54:22 INFO: Finished STEP 1680/50000, loss = 2.587858 (0.080 sec/batch), lr: 0.003000
401
- 2025-12-03 15:54:23 INFO: Finished STEP 1700/50000, loss = 4.049259 (0.081 sec/batch), lr: 0.003000
402
- 2025-12-03 15:54:23 INFO: Evaluating on dev set...
403
- 2025-12-03 15:54:24 INFO: LAS MLAS BLEX
404
- 2025-12-03 15:54:24 INFO: 59.41 50.63 54.43
405
- 2025-12-03 15:54:24 INFO: step 1700: train_loss = 3.731144, dev_score = 0.5941
406
- 2025-12-03 15:54:24 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
407
- 2025-12-03 15:54:24 INFO: new model checkpoint saved.
408
- 2025-12-03 15:54:26 INFO: Finished STEP 1720/50000, loss = 2.259671 (0.076 sec/batch), lr: 0.003000
409
- 2025-12-03 15:54:28 INFO: Finished STEP 1740/50000, loss = 5.103413 (0.079 sec/batch), lr: 0.003000
410
- 2025-12-03 15:54:29 INFO: Finished STEP 1760/50000, loss = 3.651329 (0.072 sec/batch), lr: 0.003000
411
- 2025-12-03 15:54:31 INFO: Finished STEP 1780/50000, loss = 3.879523 (0.078 sec/batch), lr: 0.003000
412
- 2025-12-03 15:54:32 INFO: Finished STEP 1800/50000, loss = 3.246203 (0.078 sec/batch), lr: 0.003000
413
- 2025-12-03 15:54:32 INFO: Evaluating on dev set...
414
- 2025-12-03 15:54:33 INFO: LAS MLAS BLEX
415
- 2025-12-03 15:54:33 INFO: 60.89 53.39 55.93
416
- 2025-12-03 15:54:33 INFO: step 1800: train_loss = 3.826308, dev_score = 0.6089
417
- 2025-12-03 15:54:33 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
418
- 2025-12-03 15:54:33 INFO: new best model saved.
419
- 2025-12-03 15:54:34 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
420
- 2025-12-03 15:54:34 INFO: new model checkpoint saved.
421
- 2025-12-03 15:54:35 INFO: Finished STEP 1820/50000, loss = 4.498244 (0.072 sec/batch), lr: 0.003000
422
- 2025-12-03 15:54:37 INFO: Finished STEP 1840/50000, loss = 3.589321 (0.078 sec/batch), lr: 0.003000
423
- 2025-12-03 15:54:39 INFO: Finished STEP 1860/50000, loss = 2.884829 (0.077 sec/batch), lr: 0.003000
424
- 2025-12-03 15:54:40 INFO: Finished STEP 1880/50000, loss = 3.823880 (0.082 sec/batch), lr: 0.003000
425
- 2025-12-03 15:54:42 INFO: Finished STEP 1900/50000, loss = 2.696285 (0.082 sec/batch), lr: 0.003000
426
- 2025-12-03 15:54:42 INFO: Evaluating on dev set...
427
- 2025-12-03 15:54:42 INFO: LAS MLAS BLEX
428
- 2025-12-03 15:54:42 INFO: 60.64 51.16 55.39
429
- 2025-12-03 15:54:42 INFO: step 1900: train_loss = 4.029873, dev_score = 0.6064
430
- 2025-12-03 15:54:43 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
431
- 2025-12-03 15:54:43 INFO: new model checkpoint saved.
432
- 2025-12-03 15:54:44 INFO: Finished STEP 1920/50000, loss = 5.128090 (0.068 sec/batch), lr: 0.003000
433
- 2025-12-03 15:54:46 INFO: Finished STEP 1940/50000, loss = 3.145783 (0.078 sec/batch), lr: 0.003000
434
- 2025-12-03 15:54:47 INFO: Finished STEP 1960/50000, loss = 3.195472 (0.075 sec/batch), lr: 0.003000
435
- 2025-12-03 15:54:49 INFO: Finished STEP 1980/50000, loss = 3.154953 (0.072 sec/batch), lr: 0.003000
436
- 2025-12-03 15:54:51 INFO: Finished STEP 2000/50000, loss = 3.763106 (0.080 sec/batch), lr: 0.003000
437
- 2025-12-03 15:54:51 INFO: Evaluating on dev set...
438
- 2025-12-03 15:54:51 INFO: LAS MLAS BLEX
439
- 2025-12-03 15:54:51 INFO: 61.39 50.85 54.66
440
- 2025-12-03 15:54:51 INFO: step 2000: train_loss = 3.764299, dev_score = 0.6139
441
- 2025-12-03 15:54:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
442
- 2025-12-03 15:54:51 INFO: new best model saved.
443
- 2025-12-03 15:54:52 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
444
- 2025-12-03 15:54:52 INFO: new model checkpoint saved.
445
- 2025-12-03 15:54:53 INFO: Finished STEP 2020/50000, loss = 4.129125 (0.077 sec/batch), lr: 0.003000
446
- 2025-12-03 15:54:55 INFO: Finished STEP 2040/50000, loss = 4.000008 (0.076 sec/batch), lr: 0.003000
447
- 2025-12-03 15:54:56 INFO: Finished STEP 2060/50000, loss = 3.594474 (0.076 sec/batch), lr: 0.003000
448
- 2025-12-03 15:54:58 INFO: Finished STEP 2080/50000, loss = 5.156559 (0.070 sec/batch), lr: 0.003000
449
- 2025-12-03 15:54:59 INFO: Finished STEP 2100/50000, loss = 2.780583 (0.076 sec/batch), lr: 0.003000
450
- 2025-12-03 15:54:59 INFO: Evaluating on dev set...
451
- 2025-12-03 15:55:00 INFO: LAS MLAS BLEX
452
- 2025-12-03 15:55:00 INFO: 58.91 50.32 53.28
453
- 2025-12-03 15:55:00 INFO: step 2100: train_loss = 3.970306, dev_score = 0.5891
454
- 2025-12-03 15:55:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
455
- 2025-12-03 15:55:01 INFO: new model checkpoint saved.
456
- 2025-12-03 15:55:02 INFO: Finished STEP 2120/50000, loss = 3.461082 (0.078 sec/batch), lr: 0.003000
457
- 2025-12-03 15:55:04 INFO: Finished STEP 2140/50000, loss = 4.371090 (0.068 sec/batch), lr: 0.003000
458
- 2025-12-03 15:55:05 INFO: Finished STEP 2160/50000, loss = 3.734883 (0.074 sec/batch), lr: 0.003000
459
- 2025-12-03 15:55:07 INFO: Finished STEP 2180/50000, loss = 2.186298 (0.077 sec/batch), lr: 0.003000
460
- 2025-12-03 15:55:08 INFO: Finished STEP 2200/50000, loss = 3.871332 (0.078 sec/batch), lr: 0.003000
461
- 2025-12-03 15:55:08 INFO: Evaluating on dev set...
462
- 2025-12-03 15:55:09 INFO: LAS MLAS BLEX
463
- 2025-12-03 15:55:09 INFO: 58.66 48.73 52.97
464
- 2025-12-03 15:55:09 INFO: step 2200: train_loss = 3.926589, dev_score = 0.5866
465
- 2025-12-03 15:55:09 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
466
- 2025-12-03 15:55:09 INFO: new model checkpoint saved.
467
- 2025-12-03 15:55:11 INFO: Finished STEP 2220/50000, loss = 2.865967 (0.073 sec/batch), lr: 0.003000
468
- 2025-12-03 15:55:12 INFO: Finished STEP 2240/50000, loss = 2.578629 (0.069 sec/batch), lr: 0.003000
469
- 2025-12-03 15:55:14 INFO: Finished STEP 2260/50000, loss = 3.295215 (0.068 sec/batch), lr: 0.003000
470
- 2025-12-03 15:55:15 INFO: Finished STEP 2280/50000, loss = 3.903890 (0.074 sec/batch), lr: 0.003000
471
- 2025-12-03 15:55:17 INFO: Finished STEP 2300/50000, loss = 3.614969 (0.071 sec/batch), lr: 0.003000
472
- 2025-12-03 15:55:17 INFO: Evaluating on dev set...
473
- 2025-12-03 15:55:17 INFO: LAS MLAS BLEX
474
- 2025-12-03 15:55:17 INFO: 60.40 52.43 56.66
475
- 2025-12-03 15:55:17 INFO: step 2300: train_loss = 3.953848, dev_score = 0.6040
476
- 2025-12-03 15:55:18 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
477
- 2025-12-03 15:55:18 INFO: new model checkpoint saved.
478
- 2025-12-03 15:55:19 INFO: Finished STEP 2320/50000, loss = 3.836884 (0.076 sec/batch), lr: 0.003000
479
- 2025-12-03 15:55:21 INFO: Finished STEP 2340/50000, loss = 2.887274 (0.079 sec/batch), lr: 0.003000
480
- 2025-12-03 15:55:22 INFO: Finished STEP 2360/50000, loss = 4.795152 (0.079 sec/batch), lr: 0.003000
481
- 2025-12-03 15:55:24 INFO: Finished STEP 2380/50000, loss = 4.483939 (0.074 sec/batch), lr: 0.003000
482
- 2025-12-03 15:55:25 INFO: Finished STEP 2400/50000, loss = 4.851534 (0.075 sec/batch), lr: 0.003000
483
- 2025-12-03 15:55:25 INFO: Evaluating on dev set...
484
- 2025-12-03 15:55:26 INFO: LAS MLAS BLEX
485
- 2025-12-03 15:55:26 INFO: 58.91 50.21 54.85
486
- 2025-12-03 15:55:26 INFO: step 2400: train_loss = 3.932371, dev_score = 0.5891
487
- 2025-12-03 15:55:27 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
488
- 2025-12-03 15:55:27 INFO: new model checkpoint saved.
489
- 2025-12-03 15:55:28 INFO: Finished STEP 2420/50000, loss = 4.692697 (0.067 sec/batch), lr: 0.003000
490
- 2025-12-03 15:55:30 INFO: Finished STEP 2440/50000, loss = 2.805047 (0.068 sec/batch), lr: 0.003000
491
- 2025-12-03 15:55:31 INFO: Finished STEP 2460/50000, loss = 3.715670 (0.077 sec/batch), lr: 0.003000
492
- 2025-12-03 15:55:32 INFO: Finished STEP 2480/50000, loss = 5.003852 (0.072 sec/batch), lr: 0.003000
493
- 2025-12-03 15:55:34 INFO: Finished STEP 2500/50000, loss = 5.362862 (0.071 sec/batch), lr: 0.003000
494
- 2025-12-03 15:55:34 INFO: Evaluating on dev set...
495
- 2025-12-03 15:55:34 INFO: LAS MLAS BLEX
496
- 2025-12-03 15:55:34 INFO: 62.87 54.12 57.51
497
- 2025-12-03 15:55:34 INFO: step 2500: train_loss = 3.716296, dev_score = 0.6287
498
- 2025-12-03 15:55:35 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
499
- 2025-12-03 15:55:35 INFO: new best model saved.
500
- 2025-12-03 15:55:35 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
501
- 2025-12-03 15:55:35 INFO: new model checkpoint saved.
502
- 2025-12-03 15:55:37 INFO: Finished STEP 2520/50000, loss = 4.285573 (0.073 sec/batch), lr: 0.003000
503
- 2025-12-03 15:55:38 INFO: Finished STEP 2540/50000, loss = 3.220155 (0.071 sec/batch), lr: 0.003000
504
- 2025-12-03 15:55:40 INFO: Finished STEP 2560/50000, loss = 4.155645 (0.077 sec/batch), lr: 0.003000
505
- 2025-12-03 15:55:41 INFO: Finished STEP 2580/50000, loss = 3.434030 (0.073 sec/batch), lr: 0.003000
506
- 2025-12-03 15:55:43 INFO: Finished STEP 2600/50000, loss = 2.680579 (0.076 sec/batch), lr: 0.003000
507
- 2025-12-03 15:55:43 INFO: Evaluating on dev set...
508
- 2025-12-03 15:55:43 INFO: LAS MLAS BLEX
509
- 2025-12-03 15:55:43 INFO: 62.87 54.66 58.05
510
- 2025-12-03 15:55:43 INFO: step 2600: train_loss = 4.018973, dev_score = 0.6287
511
- 2025-12-03 15:55:43 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
512
- 2025-12-03 15:55:43 INFO: new best model saved.
513
- 2025-12-03 15:55:44 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
514
- 2025-12-03 15:55:44 INFO: new model checkpoint saved.
515
- 2025-12-03 15:55:45 INFO: Finished STEP 2620/50000, loss = 2.524696 (0.072 sec/batch), lr: 0.003000
516
- 2025-12-03 15:55:47 INFO: Finished STEP 2640/50000, loss = 3.829545 (0.065 sec/batch), lr: 0.003000
517
- 2025-12-03 15:55:48 INFO: Finished STEP 2660/50000, loss = 3.246467 (0.069 sec/batch), lr: 0.003000
518
- 2025-12-03 15:55:50 INFO: Finished STEP 2680/50000, loss = 2.396587 (0.068 sec/batch), lr: 0.003000
519
- 2025-12-03 15:55:51 INFO: Finished STEP 2700/50000, loss = 3.570580 (0.071 sec/batch), lr: 0.003000
520
- 2025-12-03 15:55:51 INFO: Evaluating on dev set...
521
- 2025-12-03 15:55:52 INFO: LAS MLAS BLEX
522
- 2025-12-03 15:55:52 INFO: 58.91 49.58 54.24
523
- 2025-12-03 15:55:52 INFO: step 2700: train_loss = 3.918322, dev_score = 0.5891
524
- 2025-12-03 15:55:52 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
525
- 2025-12-03 15:55:52 INFO: new model checkpoint saved.
526
- 2025-12-03 15:55:54 INFO: Finished STEP 2720/50000, loss = 2.712030 (0.072 sec/batch), lr: 0.003000
527
- 2025-12-03 15:55:55 INFO: Finished STEP 2740/50000, loss = 5.811014 (0.077 sec/batch), lr: 0.003000
528
- 2025-12-03 15:55:57 INFO: Finished STEP 2760/50000, loss = 3.773255 (0.076 sec/batch), lr: 0.003000
529
- 2025-12-03 15:55:58 INFO: Finished STEP 2780/50000, loss = 2.087427 (0.075 sec/batch), lr: 0.003000
530
- 2025-12-03 15:56:00 INFO: Finished STEP 2800/50000, loss = 3.811720 (0.063 sec/batch), lr: 0.003000
531
- 2025-12-03 15:56:00 INFO: Evaluating on dev set...
532
- 2025-12-03 15:56:00 INFO: LAS MLAS BLEX
533
- 2025-12-03 15:56:00 INFO: 61.63 51.59 55.39
534
- 2025-12-03 15:56:00 INFO: step 2800: train_loss = 3.902705, dev_score = 0.6163
535
- 2025-12-03 15:56:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
536
- 2025-12-03 15:56:01 INFO: new model checkpoint saved.
537
- 2025-12-03 15:56:02 INFO: Finished STEP 2820/50000, loss = 3.443505 (0.070 sec/batch), lr: 0.003000
538
- 2025-12-03 15:56:04 INFO: Finished STEP 2840/50000, loss = 3.684120 (0.064 sec/batch), lr: 0.003000
539
- 2025-12-03 15:56:05 INFO: Finished STEP 2860/50000, loss = 5.325919 (0.071 sec/batch), lr: 0.003000
540
- 2025-12-03 15:56:06 INFO: Finished STEP 2880/50000, loss = 3.516696 (0.068 sec/batch), lr: 0.003000
541
- 2025-12-03 15:56:08 INFO: Finished STEP 2900/50000, loss = 3.573216 (0.072 sec/batch), lr: 0.003000
542
- 2025-12-03 15:56:08 INFO: Evaluating on dev set...
543
- 2025-12-03 15:56:09 INFO: LAS MLAS BLEX
544
- 2025-12-03 15:56:09 INFO: 58.17 48.20 53.28
545
- 2025-12-03 15:56:09 INFO: step 2900: train_loss = 4.154457, dev_score = 0.5817
546
- 2025-12-03 15:56:09 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
547
- 2025-12-03 15:56:09 INFO: new model checkpoint saved.
548
- 2025-12-03 15:56:11 INFO: Finished STEP 2920/50000, loss = 2.522206 (0.077 sec/batch), lr: 0.003000
549
- 2025-12-03 15:56:12 INFO: Finished STEP 2940/50000, loss = 2.687997 (0.082 sec/batch), lr: 0.003000
550
- 2025-12-03 15:56:13 INFO: Finished STEP 2960/50000, loss = 3.291323 (0.078 sec/batch), lr: 0.003000
551
- 2025-12-03 15:56:15 INFO: Finished STEP 2980/50000, loss = 4.553081 (0.069 sec/batch), lr: 0.003000
552
- 2025-12-03 15:56:16 INFO: Finished STEP 3000/50000, loss = 3.136598 (0.071 sec/batch), lr: 0.003000
553
- 2025-12-03 15:56:16 INFO: Evaluating on dev set...
554
- 2025-12-03 15:56:17 INFO: LAS MLAS BLEX
555
- 2025-12-03 15:56:17 INFO: 61.14 52.43 55.39
556
- 2025-12-03 15:56:17 INFO: step 3000: train_loss = 3.949049, dev_score = 0.6114
557
- 2025-12-03 15:56:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
558
- 2025-12-03 15:56:17 INFO: new model checkpoint saved.
559
- 2025-12-03 15:56:19 INFO: Finished STEP 3020/50000, loss = 3.020242 (0.077 sec/batch), lr: 0.003000
560
- 2025-12-03 15:56:20 INFO: Finished STEP 3040/50000, loss = 4.876935 (0.069 sec/batch), lr: 0.003000
561
- 2025-12-03 15:56:22 INFO: Finished STEP 3060/50000, loss = 3.937078 (0.080 sec/batch), lr: 0.003000
562
- 2025-12-03 15:56:23 INFO: Finished STEP 3080/50000, loss = 4.811732 (0.064 sec/batch), lr: 0.003000
563
- 2025-12-03 15:56:25 INFO: Finished STEP 3100/50000, loss = 3.313049 (0.073 sec/batch), lr: 0.003000
564
- 2025-12-03 15:56:25 INFO: Evaluating on dev set...
565
- 2025-12-03 15:56:25 INFO: LAS MLAS BLEX
566
- 2025-12-03 15:56:25 INFO: 62.13 54.12 58.35
567
- 2025-12-03 15:56:25 INFO: step 3100: train_loss = 4.106587, dev_score = 0.6213
568
- 2025-12-03 15:56:26 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
569
- 2025-12-03 15:56:26 INFO: new model checkpoint saved.
570
- 2025-12-03 15:56:27 INFO: Finished STEP 3120/50000, loss = 3.817453 (0.067 sec/batch), lr: 0.003000
571
- 2025-12-03 15:56:29 INFO: Finished STEP 3140/50000, loss = 5.620703 (0.070 sec/batch), lr: 0.003000
572
- 2025-12-03 15:56:30 INFO: Finished STEP 3160/50000, loss = 4.408039 (0.078 sec/batch), lr: 0.003000
573
- 2025-12-03 15:56:32 INFO: Finished STEP 3180/50000, loss = 2.225734 (0.074 sec/batch), lr: 0.003000
574
- 2025-12-03 15:56:33 INFO: Finished STEP 3200/50000, loss = 6.085883 (0.082 sec/batch), lr: 0.003000
575
- 2025-12-03 15:56:33 INFO: Evaluating on dev set...
576
- 2025-12-03 15:56:34 INFO: LAS MLAS BLEX
577
- 2025-12-03 15:56:34 INFO: 58.91 51.59 53.70
578
- 2025-12-03 15:56:34 INFO: step 3200: train_loss = 4.045106, dev_score = 0.5891
579
- 2025-12-03 15:56:34 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
580
- 2025-12-03 15:56:34 INFO: new model checkpoint saved.
581
- 2025-12-03 15:56:36 INFO: Finished STEP 3220/50000, loss = 3.643058 (0.064 sec/batch), lr: 0.003000
582
- 2025-12-03 15:56:37 INFO: Finished STEP 3240/50000, loss = 3.506227 (0.068 sec/batch), lr: 0.003000
583
- 2025-12-03 15:56:38 INFO: Finished STEP 3260/50000, loss = 2.930928 (0.072 sec/batch), lr: 0.003000
584
- 2025-12-03 15:56:40 INFO: Finished STEP 3280/50000, loss = 4.896526 (0.071 sec/batch), lr: 0.003000
585
- 2025-12-03 15:56:41 INFO: Finished STEP 3300/50000, loss = 4.648397 (0.077 sec/batch), lr: 0.003000
586
- 2025-12-03 15:56:41 INFO: Evaluating on dev set...
587
- 2025-12-03 15:56:42 INFO: LAS MLAS BLEX
588
- 2025-12-03 15:56:42 INFO: 59.16 52.01 52.43
589
- 2025-12-03 15:56:42 INFO: step 3300: train_loss = 3.880126, dev_score = 0.5916
590
- 2025-12-03 15:56:42 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
591
- 2025-12-03 15:56:42 INFO: new model checkpoint saved.
592
- 2025-12-03 15:56:44 INFO: Finished STEP 3320/50000, loss = 4.656130 (0.065 sec/batch), lr: 0.003000
593
- 2025-12-03 15:56:45 INFO: Finished STEP 3340/50000, loss = 4.644033 (0.070 sec/batch), lr: 0.003000
594
- 2025-12-03 15:56:47 INFO: Finished STEP 3360/50000, loss = 4.342103 (0.070 sec/batch), lr: 0.003000
595
- 2025-12-03 15:56:48 INFO: Finished STEP 3380/50000, loss = 1.757741 (0.069 sec/batch), lr: 0.003000
596
- 2025-12-03 15:56:49 INFO: Finished STEP 3400/50000, loss = 3.296278 (0.077 sec/batch), lr: 0.003000
597
- 2025-12-03 15:56:49 INFO: Evaluating on dev set...
598
- 2025-12-03 15:56:50 INFO: LAS MLAS BLEX
599
- 2025-12-03 15:56:50 INFO: 57.92 48.63 52.43
600
- 2025-12-03 15:56:50 INFO: step 3400: train_loss = 3.761222, dev_score = 0.5792
601
- 2025-12-03 15:56:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
602
- 2025-12-03 15:56:51 INFO: new model checkpoint saved.
603
- 2025-12-03 15:56:52 INFO: Finished STEP 3420/50000, loss = 4.175681 (0.063 sec/batch), lr: 0.003000
604
- 2025-12-03 15:56:53 INFO: Finished STEP 3440/50000, loss = 3.109319 (0.072 sec/batch), lr: 0.003000
605
- 2025-12-03 15:56:55 INFO: Finished STEP 3460/50000, loss = 5.698625 (0.070 sec/batch), lr: 0.003000
606
- 2025-12-03 15:56:56 INFO: Finished STEP 3480/50000, loss = 4.906111 (0.069 sec/batch), lr: 0.003000
607
- 2025-12-03 15:56:58 INFO: Finished STEP 3500/50000, loss = 2.886077 (0.068 sec/batch), lr: 0.003000
608
- 2025-12-03 15:56:58 INFO: Evaluating on dev set...
609
- 2025-12-03 15:56:58 INFO: LAS MLAS BLEX
610
- 2025-12-03 15:56:58 INFO: 60.64 51.48 55.70
611
- 2025-12-03 15:56:58 INFO: step 3500: train_loss = 3.804499, dev_score = 0.6064
612
- 2025-12-03 15:56:59 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
613
- 2025-12-03 15:56:59 INFO: new model checkpoint saved.
614
- 2025-12-03 15:57:00 INFO: Finished STEP 3520/50000, loss = 2.301594 (0.065 sec/batch), lr: 0.003000
615
- 2025-12-03 15:57:01 INFO: Finished STEP 3540/50000, loss = 4.423633 (0.075 sec/batch), lr: 0.003000
616
- 2025-12-03 15:57:03 INFO: Finished STEP 3560/50000, loss = 3.613044 (0.063 sec/batch), lr: 0.003000
617
- 2025-12-03 15:57:04 INFO: Finished STEP 3580/50000, loss = 3.496284 (0.067 sec/batch), lr: 0.003000
618
- 2025-12-03 15:57:06 INFO: Finished STEP 3600/50000, loss = 6.527389 (0.066 sec/batch), lr: 0.003000
619
- 2025-12-03 15:57:06 INFO: Evaluating on dev set...
620
- 2025-12-03 15:57:06 INFO: LAS MLAS BLEX
621
- 2025-12-03 15:57:06 INFO: 60.64 51.27 55.08
622
- 2025-12-03 15:57:06 INFO: step 3600: train_loss = 4.020496, dev_score = 0.6064
623
- 2025-12-03 15:57:06 INFO: Training ended with 3600 steps.
624
- 2025-12-03 15:57:06 INFO: Best dev F1 = 62.87, at iteration = 2500
625
- 2025-12-03 15:57:07 INFO: Running dev depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
626
- 2025-12-03 15:57:07 INFO: Running parser in predict mode
627
- 2025-12-03 15:57:07 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
628
- 2025-12-03 15:57:09 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
629
- 2025-12-03 15:57:09 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
630
- 2025-12-03 15:57:09 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
631
- 2025-12-03 15:57:09 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
632
- 2025-12-03 15:57:10 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
633
- 2025-12-03 15:57:10 INFO: Loading data with batch size 32...
634
- 2025-12-03 15:57:10 DEBUG: 9 batches created.
635
- 2025-12-03 15:57:10 INFO: F1 scores for each dependency:
636
- Note that unlabeled attachment errors hurt the labeled attachment scores
637
- acl: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
638
- acl:relcl: p 0.2222 r 0.2857 f1 0.2500 (7 actual)
639
- advcl: p 0.1250 r 0.2000 f1 0.1538 (5 actual)
640
- advmod: p 0.5357 r 0.6000 f1 0.5660 (25 actual)
641
- amod: p 0.9231 r 0.7742 f1 0.8421 (31 actual)
642
- appos: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
643
- aux: p 0.8182 r 0.8182 f1 0.8182 (11 actual)
644
- case: p 0.9623 r 0.9107 f1 0.9358 (56 actual)
645
- cc: p 0.6429 r 0.6923 f1 0.6667 (13 actual)
646
- ccomp: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
647
- conj: p 0.2667 r 0.3333 f1 0.2963 (12 actual)
648
- cop: p 0.5000 r 0.3333 f1 0.4000 (3 actual)
649
- csubj: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
650
- det: p 0.8696 r 0.9091 f1 0.8889 (22 actual)
651
- expl: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
652
- iobj: p 0.5000 r 0.5000 f1 0.5000 (2 actual)
653
- mark: p 0.4167 r 0.4167 f1 0.4167 (12 actual)
654
- nmod: p 0.4000 r 0.5333 f1 0.4571 (15 actual)
655
- nmod:poss: p 1.0000 r 0.8947 f1 0.9444 (19 actual)
656
- nsubj: p 0.5417 r 0.7647 f1 0.6341 (17 actual)
657
- nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
658
- obj: p 0.6071 r 0.7727 f1 0.6800 (22 actual)
659
- obl: p 0.6154 r 0.5854 f1 0.6000 (41 actual)
660
- obl:agent: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
661
- orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
662
- parataxis: p 0.1111 r 0.3333 f1 0.1667 (3 actual)
663
- punct: p 0.4231 r 0.4231 f1 0.4231 (52 actual)
664
- root: p 0.4444 r 0.4444 f1 0.4444 (9 actual)
665
- xcomp: p 0.5000 r 0.1250 f1 0.2000 (8 actual)
666
- 2025-12-03 15:57:10 INFO: LAS MLAS BLEX
667
- 2025-12-03 15:57:10 INFO: 62.87 54.66 58.05
668
- 2025-12-03 15:57:10 INFO: Parser score:
669
- 2025-12-03 15:57:10 INFO: sv_diachronic 62.87
670
- 2025-12-03 15:57:10 INFO: Finished running dev set on
671
- UD_Swedish-diachronic
672
- UAS LAS CLAS MLAS BLEX
673
- 68.56 62.87 58.05 54.66 58.05
674
- 2025-12-03 15:57:10 INFO: Running test depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
675
- 2025-12-03 15:57:10 INFO: Running parser in predict mode
676
- 2025-12-03 15:57:10 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
677
- 2025-12-03 15:57:12 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
678
- 2025-12-03 15:57:12 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
679
- 2025-12-03 15:57:12 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
680
- 2025-12-03 15:57:12 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
681
- 2025-12-03 15:57:12 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
682
- 2025-12-03 15:57:12 INFO: Loading data with batch size 32...
683
- 2025-12-03 15:57:12 DEBUG: 93 batches created.
684
- 2025-12-03 15:57:17 INFO: F1 scores for each dependency:
685
- Note that unlabeled attachment errors hurt the labeled attachment scores
686
- acl: p 0.2778 r 0.1562 f1 0.2000 (32 actual)
687
- acl:cleft: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
688
- acl:relcl: p 0.2075 r 0.1467 f1 0.1719 (75 actual)
689
- advcl: p 0.0842 r 0.1333 f1 0.1032 (60 actual)
690
- advcl:relcl: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
691
- advmod: p 0.5520 r 0.5746 f1 0.5631 (268 actual)
692
- amod: p 0.8596 r 0.8522 f1 0.8559 (230 actual)
693
- appos: p 0.0000 r 0.0000 f1 0.0000 (13 actual)
694
- aux: p 0.8734 r 0.8214 f1 0.8466 (84 actual)
695
- aux:pass: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
696
- case: p 0.8481 r 0.8231 f1 0.8354 (373 actual)
697
- cc: p 0.5833 r 0.5871 f1 0.5852 (155 actual)
698
- ccomp: p 0.0000 r 0.0000 f1 0.0000 (35 actual)
699
- compound:prt: p 0.7727 r 0.8095 f1 0.7907 (21 actual)
700
- conj: p 0.2397 r 0.2215 f1 0.2303 (158 actual)
701
- cop: p 0.7714 r 0.5870 f1 0.6667 (46 actual)
702
- csubj: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
703
- dep: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
704
- det: p 0.8960 r 0.8702 f1 0.8829 (208 actual)
705
- discourse: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
706
- dislocated: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
707
- expl: p 0.0000 r 0.0000 f1 0.0000 (11 actual)
708
- expl:pv: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
709
- fixed: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
710
- flat: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
711
- flat:name: p 0.0000 r 0.0000 f1 0.0000 (12 actual)
712
- goeswith: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
713
- iobj: p 0.3333 r 0.0714 f1 0.1176 (14 actual)
714
- mark: p 0.6757 r 0.6536 f1 0.6645 (153 actual)
715
- nmod: p 0.3798 r 0.4804 f1 0.4242 (102 actual)
716
- nmod:poss: p 0.8707 r 0.9014 f1 0.8858 (142 actual)
717
- nsubj: p 0.5673 r 0.6321 f1 0.5980 (280 actual)
718
- nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (25 actual)
719
- nummod: p 0.7500 r 0.9000 f1 0.8182 (10 actual)
720
- obj: p 0.5628 r 0.7104 f1 0.6280 (183 actual)
721
- obl: p 0.4857 r 0.5504 f1 0.5160 (278 actual)
722
- obl:agent: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
723
- orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
724
- parataxis: p 0.0333 r 0.1667 f1 0.0556 (18 actual)
725
- punct: p 0.4262 r 0.4282 f1 0.4272 (425 actual)
726
- reparandum: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
727
- root: p 0.5354 r 0.5354 f1 0.5354 (99 actual)
728
- vocative: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
729
- xcomp: p 0.5750 r 0.3067 f1 0.4000 (75 actual)
730
- 2025-12-03 15:57:17 INFO: LAS MLAS BLEX
731
- 2025-12-03 15:57:17 INFO: 58.54 50.12 53.03
732
- 2025-12-03 15:57:17 INFO: Parser score:
733
- 2025-12-03 15:57:17 INFO: sv_diachronic 58.54
734
- 2025-12-03 15:57:17 INFO: Finished running test set on
735
- UD_Swedish-diachronic
736
- UAS LAS CLAS MLAS BLEX
737
- 67.16 58.54 53.03 50.12 53.03
738
- DONE.
739
- Full log saved to: logs/log_conll17.pt_sv_diachron_de_lit_20251203_154509.txt
740
- Symlink updated: logs/latest.txt → log_conll17.pt_sv_diachron_de_lit_20251203_154509.txt