refactor: only models here, training pipeline archive on gh
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- README.md +231 -17
- parsed/old-conll17-1/svediakorp-letter141673-Stalhammar.conllu → conll17_baseline_sv_only.pt +2 -2
- parsed/old-conll17-1/svediakorp-letter141673-Stalhammar_preprocessed.conllu → conll17_bm.pt +2 -2
- eval/gold_as_val.conllu → conll17_bm_dk.pt +2 -2
- parsed/old-conll17-1/svediakorp-sec1018-spf175.conllu → conll17_de_lit.pt +2 -2
- conll17_icepahc.pt +3 -0
- conll17_is-modern.pt +3 -0
- conll17_isPUD-pahc-gc.pt +3 -0
- conll17_isPUD.pt +3 -0
- conll17_nn.pt +3 -0
- conll17_sv_diachron.pt +3 -0
- digphil_inference_pipeline.py +0 -114
- eval/execute_eval.py +0 -273
- eval/mod_conll18_ud_eval.py +0 -672
- eval/scores/__init__.py +0 -0
- eval/scores/scores_talbanken.txt +0 -162
- eval/scores/scores_transformer_no_silver.txt +0 -162
- eval/scores/scores_transformer_silver.txt +0 -162
- logs/latest.txt +0 -1
- logs/log_bert-base-swedish-cased_sv_20260219_135555.txt +0 -114
- logs/log_bert-base-swedish-cased_sv_20260219_135610.txt +0 -138
- logs/log_bert-base-swedish-cased_sv_20260219_135635.txt +0 -0
- logs/log_bert-base-swedish-cased_sv_20260318_131623.txt +0 -271
- logs/log_bert-base-swedish-cased_sv_20260318_132248.txt +0 -109
- logs/log_bert-base-swedish-cased_sv_20260318_132404.txt +0 -109
- logs/log_bert-base-swedish-cased_sv_20260318_132733.txt +0 -325
- logs/log_bert-base-swedish-cased_sv_20260318_133119.txt +0 -0
- logs/log_bert-base-swedish-cased_sv_20260318_141654.txt +0 -0
- logs/log_bert-base-swedish-cased_sv_20260318_215701.txt +0 -0
- logs/log_bert-base-swedish-cased_sv_20260319_200948.txt +0 -0
- logs/log_bert-base-swedish-cased_sv_20260319_201349.txt +0 -0
- logs/log_bert-base-swedish-cased_sv_20260319_214621.txt +0 -0
- logs/log_conll17.pt_sv_20251203_232257.txt +0 -719
- logs/log_conll17.pt_sv_20260218_190034.txt +0 -120
- logs/log_conll17.pt_sv_20260218_190242.txt +0 -143
- logs/log_conll17.pt_sv_20260218_190400.txt +0 -38
- logs/log_conll17.pt_sv_20260218_190808.txt +0 -38
- logs/log_conll17.pt_sv_20260218_190956.txt +0 -38
- logs/log_conll17.pt_sv_20260218_191305.txt +0 -38
- logs/log_conll17.pt_sv_20260218_191719.txt +0 -38
- logs/log_conll17.pt_sv_20260219_104841.txt +0 -44
- logs/log_conll17.pt_sv_20260219_105228.txt +0 -315
- logs/log_conll17.pt_sv_20260219_105756.txt +0 -0
- logs/log_conll17.pt_sv_20260219_114115.txt +0 -0
- logs/log_conll17.pt_sv_diachron_20251203_223822.txt +0 -731
- logs/log_conll17.pt_sv_diachron_20251212_145854.txt +0 -161
- logs/log_conll17.pt_sv_diachron_20251212_150001.txt +0 -0
- logs/log_conll17.pt_sv_diachron_20260116_065352.txt +0 -830
- logs/log_conll17.pt_sv_diachron_20260116_073035.txt +0 -868
- logs/log_conll17.pt_sv_diachron_de_lit_20251203_154509.txt +0 -740
README.md
CHANGED
|
@@ -15,35 +15,249 @@ library_name: stanza
|
|
| 15 |
|
| 16 |
# Retraining Stanza to optimize dependency parsing on a diachronic Swedish corpus
|
| 17 |
|
| 18 |
-
This repository contains
|
| 19 |
|
| 20 |
-
##
|
| 21 |
|
| 22 |
-
|
|
|
|
| 23 |
|
| 24 |
-
|
| 25 |
|
| 26 |
-
|
| 27 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
|
| 29 |
-
|
| 30 |
|
| 31 |
-
|
| 32 |
|
| 33 |
-
|
| 34 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 35 |
|
| 36 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 37 |
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
|
| 41 |
|
| 42 |
-
|
| 43 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
```
|
| 45 |
-
./make_new_model.sh diachronic.pt sv diachron bm
|
| 46 |
-
```
|
| 47 |
|
| 48 |
-
##
|
|
|
|
|
|
|
| 49 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 15 |
|
| 16 |
# Retraining Stanza to optimize dependency parsing on a diachronic Swedish corpus
|
| 17 |
|
| 18 |
+
This repository contains Stanza BiLSTM models retrained on different combinations of UD treebanks relevant to historical Swedish. The models prefixed `conll17_` are trained with static embeddings, and the models prefixed `transformer_` are trained with dynamic embeddings from the transformer encoder "KBLab/bert-base-swedish-cased".
|
| 19 |
|
| 20 |
+
## LAS Scores
|
| 21 |
|
| 22 |
+
LAS scores for the models are computed against a gold set of 109 manually annoted sentences divided into five different periods.
|
| 23 |
+
For the models trained on static vectors, only the overall test scores is given:
|
| 24 |
|
| 25 |
+
### Models with static vector embeddings (`conll17.pt`)
|
| 26 |
|
| 27 |
+
| Languages | LAS |
|
| 28 |
+
|--------------------------------|-------|
|
| 29 |
+
| **Swedish (with diachronic)** | **61.95** |
|
| 30 |
+
| Icelandic (PUD) | 61.49 |
|
| 31 |
+
| German (LIT) | 61.43 |
|
| 32 |
+
| Icelandic (GC) | 61.43 |
|
| 33 |
+
| Bokmaal, Danish | 60.13 |
|
| 34 |
+
| Nynorsk | 50.46 |
|
| 35 |
+
| *Swedish (without diachronic)* | *50.34* |
|
| 36 |
+
| Icelandic (Modern) | 46.47 |
|
| 37 |
+
| Bokmaal | 45.96 |
|
| 38 |
+
| Icelandic (IcePaHC) | 44.60 |
|
| 39 |
+
|
| 40 |
+
For the transformer-fed models, more fine-grained scores on each period are given as a histogram. The model `transformer_seen_gold_no_silver.pt` was given the gold set during training and hence has no score, but is intuitively the best model. As a benchmark, an "out-of-the-box" Stanza trained only on Talbanken is given.
|
| 41 |
|
| 42 |
+
<img src="plot/las_f1_scores_comparison.png" width="40%">
|
| 43 |
|
| 44 |
+
## Inference
|
| 45 |
|
| 46 |
+
Example for how the models can be run:
|
| 47 |
```
|
| 48 |
+
import os
|
| 49 |
+
from pathlib import Path
|
| 50 |
+
import stanza
|
| 51 |
+
from stanza.utils.conll import CoNLL
|
| 52 |
+
import time
|
| 53 |
+
import gc
|
| 54 |
+
import torch
|
| 55 |
+
from tqdm import tqdm
|
| 56 |
|
| 57 |
+
if torch.xpu.is_available():
|
| 58 |
+
device = torch.device("xpu")
|
| 59 |
+
elif torch.cuda.is_available():
|
| 60 |
+
device = torch.device("cuda")
|
| 61 |
+
else:
|
| 62 |
+
device = torch.device("cpu")
|
| 63 |
|
| 64 |
+
###################################################################################################
|
| 65 |
+
##### SETTINGS ####################################################################################
|
| 66 |
+
###################################################################################################
|
| 67 |
|
| 68 |
+
vanilla = False
|
| 69 |
|
| 70 |
+
conllu_in_dir = Path("YOUR/CONLLUS/")
|
| 71 |
+
conllu_out_dir = Path("OUT/DIR")
|
| 72 |
+
|
| 73 |
+
depparse_model_path = "transformer_seen_gold_no_silver.pt"
|
| 74 |
+
|
| 75 |
+
os.makedirs(conllu_out_dir, exist_ok=True)
|
| 76 |
+
|
| 77 |
+
##### PREPARE #####################################################################################
|
| 78 |
+
|
| 79 |
+
if vanilla:
|
| 80 |
+
print("Using vanilla Swedish Stanza pipeline with default models.")
|
| 81 |
+
nlp = stanza.Pipeline(
|
| 82 |
+
"sv",
|
| 83 |
+
processors="tokenize,pos,lemma,depparse",
|
| 84 |
+
tokenize_pretokenized=True, # to keep original tokens
|
| 85 |
+
use_gpu=True,
|
| 86 |
+
pos_batch_size=3000,
|
| 87 |
+
package=None,
|
| 88 |
+
device=device,
|
| 89 |
+
)
|
| 90 |
+
|
| 91 |
+
else:
|
| 92 |
+
nlp = stanza.Pipeline(
|
| 93 |
+
"sv",
|
| 94 |
+
processors="tokenize,pos,lemma,depparse",
|
| 95 |
+
tokenize_pretokenized=True, # to keep original tokens
|
| 96 |
+
use_gpu=True,
|
| 97 |
+
pos_batch_size=3000,
|
| 98 |
+
package=None,
|
| 99 |
+
download_method=None,
|
| 100 |
+
depparse_model_path=depparse_model_path, # depparse_model_path becomes model_path inside the depparse processor config. explicit model_path is to control where normal Stanza models are stored
|
| 101 |
+
device=device,
|
| 102 |
+
)
|
| 103 |
+
|
| 104 |
+
##### INFERENCE ###################################################################################
|
| 105 |
+
|
| 106 |
+
t0 = time.time()
|
| 107 |
+
total_sentences = 0
|
| 108 |
+
|
| 109 |
+
for fname in tqdm(sorted(os.listdir(conllu_in_dir)), desc="Files"):
|
| 110 |
+
if not fname.endswith(".conllu"):
|
| 111 |
+
continue
|
| 112 |
+
|
| 113 |
+
in_path = conllu_in_dir / fname
|
| 114 |
+
out_path = conllu_out_dir / fname
|
| 115 |
+
|
| 116 |
+
if out_path.exists():
|
| 117 |
+
print(f"File already parsed, skipping: {out_path}")
|
| 118 |
+
continue
|
| 119 |
+
|
| 120 |
+
doc = CoNLL.conll2doc(in_path)
|
| 121 |
+
|
| 122 |
+
total_sentences += len(doc.sentences)
|
| 123 |
+
|
| 124 |
+
with torch.inference_mode():
|
| 125 |
+
parsed_doc = nlp(doc)
|
| 126 |
+
|
| 127 |
+
with open(out_path, "w", encoding="utf-8") as f:
|
| 128 |
+
CoNLL.write_doc2conll(parsed_doc, f)
|
| 129 |
+
|
| 130 |
+
# Release per-file objects and ask both Python and the torch allocator to
|
| 131 |
+
# reclaim memory that is no longer needed.
|
| 132 |
+
del doc
|
| 133 |
+
del parsed_doc
|
| 134 |
+
gc.collect()
|
| 135 |
+
if device.type == "cuda":
|
| 136 |
+
torch.cuda.empty_cache()
|
| 137 |
+
memory_max = torch.cuda.max_memory_allocated()
|
| 138 |
+
print(f"Max memory allocated so far: {memory_max / (1024 ** 3):.2f} GB")
|
| 139 |
+
elif device.type == "xpu":
|
| 140 |
+
torch.xpu.empty_cache()
|
| 141 |
+
memory_max = torch.xpu.max_memory_allocated()
|
| 142 |
+
print(f"Max memory allocated so far: {memory_max / (1024 ** 3):.2f} GB")
|
| 143 |
+
|
| 144 |
+
t = time.time() - t0
|
| 145 |
+
|
| 146 |
+
###################################################################################################
|
| 147 |
+
|
| 148 |
+
if device.type == "cuda":
|
| 149 |
+
memory_max = torch.cuda.max_memory_allocated()
|
| 150 |
+
torch.cuda.reset_peak_memory_stats()
|
| 151 |
+
elif device.type == "xpu":
|
| 152 |
+
memory_max = torch.xpu.max_memory_allocated()
|
| 153 |
+
torch.xpu.reset_peak_memory_stats()
|
| 154 |
+
else:
|
| 155 |
+
memory_max = 0
|
| 156 |
+
|
| 157 |
+
print(
|
| 158 |
+
"\nFinished parsing.\n"
|
| 159 |
+
f"Total sentences parsed: {total_sentences}\n"
|
| 160 |
+
f"Total time: {t / 60:.2f} minutes\n"
|
| 161 |
+
f"Max memory allocated: {memory_max / (1024 ** 3):.2f} GB\n"
|
| 162 |
+
)
|
| 163 |
```
|
|
|
|
|
|
|
| 164 |
|
| 165 |
+
## Training args
|
| 166 |
+
|
| 167 |
+
Full list of training args:
|
| 168 |
|
| 169 |
+
```
|
| 170 |
+
batch_size: 32
|
| 171 |
+
bert_finetune: False
|
| 172 |
+
bert_finetune_layers: None
|
| 173 |
+
bert_hidden_layers: 4
|
| 174 |
+
bert_learning_rate: 1.0
|
| 175 |
+
bert_model: KBLab/bert-base-swedish-cased
|
| 176 |
+
bert_start_finetuning: 200
|
| 177 |
+
bert_warmup_steps: 200
|
| 178 |
+
bert_weight_decay: 0.0
|
| 179 |
+
beta2: 0.999
|
| 180 |
+
char: True
|
| 181 |
+
char_emb_dim: 100
|
| 182 |
+
char_hidden_dim: 400
|
| 183 |
+
char_num_layers: 1
|
| 184 |
+
char_rec_dropout: 0
|
| 185 |
+
charlm: True
|
| 186 |
+
charlm_backward_file: /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
|
| 187 |
+
charlm_forward_file: /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt
|
| 188 |
+
charlm_save_dir: saved_models/charlm
|
| 189 |
+
charlm_shorthand: sv_conll17
|
| 190 |
+
checkpoint: True
|
| 191 |
+
checkpoint_interval: 500
|
| 192 |
+
checkpoint_save_name: None
|
| 193 |
+
continue_from: None
|
| 194 |
+
data_dir: data/depparse
|
| 195 |
+
deep_biaff_hidden_dim: 400
|
| 196 |
+
deep_biaff_output_dim: 160
|
| 197 |
+
device: xpu:0
|
| 198 |
+
distance: True
|
| 199 |
+
dropout: 0.33
|
| 200 |
+
eval_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 201 |
+
eval_interval: 100
|
| 202 |
+
gold_labels: True
|
| 203 |
+
hidden_dim: 400
|
| 204 |
+
lang: sv
|
| 205 |
+
linearization: True
|
| 206 |
+
log_norms: False
|
| 207 |
+
log_step: 20
|
| 208 |
+
lora_alpha: 128
|
| 209 |
+
lora_dropout: 0.1
|
| 210 |
+
lora_modules_to_save: []
|
| 211 |
+
lora_rank: 64
|
| 212 |
+
lora_target_modules: ['query', 'value', 'output.dense', 'intermediate.dense']
|
| 213 |
+
lr: 2.0
|
| 214 |
+
max_grad_norm: 1.0
|
| 215 |
+
max_steps: 50000
|
| 216 |
+
max_steps_before_stop: 2000
|
| 217 |
+
mode: train
|
| 218 |
+
model_type: graph
|
| 219 |
+
num_layers: 3
|
| 220 |
+
optim: adadelta
|
| 221 |
+
output_file: None
|
| 222 |
+
output_latex: False
|
| 223 |
+
pretrain: True
|
| 224 |
+
pretrain_max_vocab: 250000
|
| 225 |
+
rec_dropout: 0
|
| 226 |
+
reversed: False
|
| 227 |
+
sample_train: 1.0
|
| 228 |
+
save_dir: saved_models/depparse
|
| 229 |
+
save_name: {shorthand}_{embedding}_parser.pt
|
| 230 |
+
second_batch_size: None
|
| 231 |
+
second_bert_learning_rate: 0.001
|
| 232 |
+
second_lr: 0.0002
|
| 233 |
+
second_optim: adam
|
| 234 |
+
second_optim_start_step: 10000
|
| 235 |
+
second_warmup_steps: 200
|
| 236 |
+
seed: 1234
|
| 237 |
+
shorthand: sv_diachronic
|
| 238 |
+
silver_file: None
|
| 239 |
+
silver_weight: 0.5
|
| 240 |
+
tag_emb_dim: 50
|
| 241 |
+
train_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 242 |
+
train_size: None
|
| 243 |
+
transformed_dim: 125
|
| 244 |
+
transition_embedding_dim: 20
|
| 245 |
+
transition_hidden_dim: 20
|
| 246 |
+
transition_merge_hidden_dim: 200
|
| 247 |
+
transition_subtree_combination: SubtreeCombination.NONE
|
| 248 |
+
transition_subtree_nonlinearity: none
|
| 249 |
+
use_arc_embedding: False
|
| 250 |
+
use_peft: False
|
| 251 |
+
use_ufeats: True
|
| 252 |
+
use_upos: True
|
| 253 |
+
use_xpos: True
|
| 254 |
+
wandb: False
|
| 255 |
+
wandb_name: None
|
| 256 |
+
weight_decay: 1e-05
|
| 257 |
+
word_cutoff: 7
|
| 258 |
+
word_dropout: 0.33
|
| 259 |
+
word_emb_dim: 75
|
| 260 |
+
wordvec_dir: /home/urdatorn/stanza_resources/sv/pretrain
|
| 261 |
+
wordvec_file: None
|
| 262 |
+
wordvec_pretrain_file: /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
|
| 263 |
+
```
|
parsed/old-conll17-1/svediakorp-letter141673-Stalhammar.conllu → conll17_baseline_sv_only.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:50aee8f0eab8b4b0e28958c0e428f1aad29cf118e4320332862b8701fc6dd6e3
|
| 3 |
+
size 141291582
|
parsed/old-conll17-1/svediakorp-letter141673-Stalhammar_preprocessed.conllu → conll17_bm.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4b0848c6832ca155fcfbf78040530aeb27598562cda430821afddbead91ea0b1
|
| 3 |
+
size 148183635
|
eval/gold_as_val.conllu → conll17_bm_dk.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ab11092dd5795e51add850b4aa325a474b72f374b93df53a78fd247ad456497c
|
| 3 |
+
size 150762047
|
parsed/old-conll17-1/svediakorp-sec1018-spf175.conllu → conll17_de_lit.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b84862d0f728ba53e2076208a94d5e90bd1dc7e0f2884a12db3f21ffcaf073dc
|
| 3 |
+
size 146969172
|
conll17_icepahc.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f59e82a5e94514bdf086d19931a26f5cd6274c6ad29c667f3366e93be0bec585
|
| 3 |
+
size 151645170
|
conll17_is-modern.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e02dffc9ae0d419373d1ba597f4d7e0459d70d6909e18211d40c33472e5a6d46
|
| 3 |
+
size 148251198
|
conll17_isPUD-pahc-gc.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5fe5f0e97a6a4e257b124f8b229c59cc61d58124512421d56795ac6ffb67359e
|
| 3 |
+
size 146969174
|
conll17_isPUD.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dcdc23474d269b8f4dba7d246f184b42a446cae7294f1134121cf5da8511efa6
|
| 3 |
+
size 146117067
|
conll17_nn.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0d74ed335831d36ee6d4569d1496a62860779c378ed391a693a8aa09159f388e
|
| 3 |
+
size 147986243
|
conll17_sv_diachron.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a67914c27dcc9cdc33034cdbebed9eec4c55b10ab2aeffc5f5d6053a96c5107a
|
| 3 |
+
size 145283635
|
digphil_inference_pipeline.py
DELETED
|
@@ -1,114 +0,0 @@
|
|
| 1 |
-
import os
|
| 2 |
-
from pathlib import Path
|
| 3 |
-
import stanza
|
| 4 |
-
from stanza.utils.conll import CoNLL
|
| 5 |
-
import time
|
| 6 |
-
import gc
|
| 7 |
-
import torch
|
| 8 |
-
from tqdm import tqdm
|
| 9 |
-
|
| 10 |
-
if torch.xpu.is_available():
|
| 11 |
-
device = torch.device("xpu")
|
| 12 |
-
elif torch.cuda.is_available():
|
| 13 |
-
device = torch.device("cuda")
|
| 14 |
-
|
| 15 |
-
###################################################################################################
|
| 16 |
-
##### SETTINGS ####################################################################################
|
| 17 |
-
###################################################################################################
|
| 18 |
-
|
| 19 |
-
vanilla = False
|
| 20 |
-
|
| 21 |
-
conllu_in_dir = Path("/home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated")
|
| 22 |
-
#conllu_out_dir = Path("parsed/testset_talbanken_not_seen_gold")
|
| 23 |
-
conllu_out_dir = Path("parsed/testset_transformer_not_seen_gold_no_silver")
|
| 24 |
-
|
| 25 |
-
depparse_model_path = "saved_models/depparse/transformer_not_seen_gold_no_silver.pt"
|
| 26 |
-
|
| 27 |
-
os.makedirs(conllu_out_dir, exist_ok=True)
|
| 28 |
-
|
| 29 |
-
##### PREPARE #####################################################################################
|
| 30 |
-
|
| 31 |
-
if vanilla:
|
| 32 |
-
print("Using vanilla Swedish Stanza pipeline with default models.")
|
| 33 |
-
nlp = stanza.Pipeline(
|
| 34 |
-
"sv",
|
| 35 |
-
processors="tokenize,pos,lemma,depparse",
|
| 36 |
-
tokenize_pretokenized=True, # to keep original tokens
|
| 37 |
-
use_gpu=True,
|
| 38 |
-
pos_batch_size=3000,
|
| 39 |
-
package=None,
|
| 40 |
-
device=device,
|
| 41 |
-
)
|
| 42 |
-
|
| 43 |
-
else:
|
| 44 |
-
nlp = stanza.Pipeline(
|
| 45 |
-
"sv",
|
| 46 |
-
processors="tokenize,pos,lemma,depparse",
|
| 47 |
-
tokenize_pretokenized=True, # to keep original tokens
|
| 48 |
-
use_gpu=True,
|
| 49 |
-
pos_batch_size=3000,
|
| 50 |
-
package=None,
|
| 51 |
-
download_method=None,
|
| 52 |
-
depparse_model_path=depparse_model_path, # depparse_model_path becomes model_path inside the depparse processor config. explicit model_path is to control where normal Stanza models are stored
|
| 53 |
-
device=device,
|
| 54 |
-
)
|
| 55 |
-
|
| 56 |
-
##### INFERENCE ###################################################################################
|
| 57 |
-
|
| 58 |
-
t0 = time.time()
|
| 59 |
-
total_sentences = 0
|
| 60 |
-
|
| 61 |
-
for fname in tqdm(sorted(os.listdir(conllu_in_dir)), desc="Files"):
|
| 62 |
-
if not fname.endswith(".conllu"):
|
| 63 |
-
continue
|
| 64 |
-
|
| 65 |
-
in_path = conllu_in_dir / fname
|
| 66 |
-
out_path = conllu_out_dir / fname
|
| 67 |
-
|
| 68 |
-
if out_path.exists():
|
| 69 |
-
print(f"File already parsed, skipping: {out_path}")
|
| 70 |
-
continue
|
| 71 |
-
|
| 72 |
-
doc = CoNLL.conll2doc(in_path)
|
| 73 |
-
|
| 74 |
-
total_sentences += len(doc.sentences)
|
| 75 |
-
|
| 76 |
-
with torch.inference_mode():
|
| 77 |
-
parsed_doc = nlp(doc)
|
| 78 |
-
|
| 79 |
-
with open(out_path, "w", encoding="utf-8") as f:
|
| 80 |
-
CoNLL.write_doc2conll(parsed_doc, f)
|
| 81 |
-
|
| 82 |
-
# Release per-file objects and ask both Python and the torch allocator to
|
| 83 |
-
# reclaim memory that is no longer needed.
|
| 84 |
-
del doc
|
| 85 |
-
del parsed_doc
|
| 86 |
-
gc.collect()
|
| 87 |
-
if device.type == "cuda":
|
| 88 |
-
torch.cuda.empty_cache()
|
| 89 |
-
memory_max = torch.cuda.max_memory_allocated()
|
| 90 |
-
print(f"Max memory allocated so far: {memory_max / (1024 ** 3):.2f} GB")
|
| 91 |
-
elif device.type == "xpu":
|
| 92 |
-
torch.xpu.empty_cache()
|
| 93 |
-
memory_max = torch.xpu.max_memory_allocated()
|
| 94 |
-
print(f"Max memory allocated so far: {memory_max / (1024 ** 3):.2f} GB")
|
| 95 |
-
|
| 96 |
-
t = time.time() - t0
|
| 97 |
-
|
| 98 |
-
###################################################################################################
|
| 99 |
-
|
| 100 |
-
if device.type == "cuda":
|
| 101 |
-
memory_max = torch.cuda.max_memory_allocated()
|
| 102 |
-
torch.cuda.reset_peak_memory_stats()
|
| 103 |
-
elif device.type == "xpu":
|
| 104 |
-
memory_max = torch.xpu.max_memory_allocated()
|
| 105 |
-
torch.xpu.reset_peak_memory_stats()
|
| 106 |
-
else:
|
| 107 |
-
memory_max = 0
|
| 108 |
-
|
| 109 |
-
print(
|
| 110 |
-
"\nFinished parsing.\n"
|
| 111 |
-
f"Total sentences parsed: {total_sentences}\n"
|
| 112 |
-
f"Total time: {t / 60:.2f} minutes\n"
|
| 113 |
-
f"Max memory allocated: {memory_max / (1024 ** 3):.2f} GB\n"
|
| 114 |
-
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
eval/execute_eval.py
DELETED
|
@@ -1,273 +0,0 @@
|
|
| 1 |
-
from mod_conll18_ud_eval import load_conllu, evaluate
|
| 2 |
-
from collections import defaultdict
|
| 3 |
-
from pathlib import Path
|
| 4 |
-
import pyconll
|
| 5 |
-
import sys
|
| 6 |
-
import os
|
| 7 |
-
import numpy as np
|
| 8 |
-
|
| 9 |
-
ROOT = Path.cwd()
|
| 10 |
-
print(ROOT)
|
| 11 |
-
gold_dir = Path("/home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated") # mapp med vårt manuellt annoterade guldstandardset
|
| 12 |
-
val_file = ROOT / "eval/gold_as_val.conllu" # sentences used during validation - must be excluded
|
| 13 |
-
|
| 14 |
-
def load_val_sentence_ids(val_file_path):
|
| 15 |
-
"""Load sentence IDs from the validation file to exclude from evaluation."""
|
| 16 |
-
val_ids = set()
|
| 17 |
-
if not os.path.exists(val_file_path):
|
| 18 |
-
print(f"Warning: validation file {val_file_path} not found. No sentences will be excluded.")
|
| 19 |
-
return val_ids
|
| 20 |
-
val_data = pyconll.load_from_file(str(val_file_path))
|
| 21 |
-
for sentence in val_data:
|
| 22 |
-
val_ids.add(sentence.id)
|
| 23 |
-
print(f"Loaded {len(val_ids)} sentence IDs to exclude from evaluation (used in validation).")
|
| 24 |
-
return val_ids
|
| 25 |
-
|
| 26 |
-
def filter_gold_file(gold_file_path, excluded_ids):
|
| 27 |
-
"""Create a filtered gold file excluding sentences used during validation.
|
| 28 |
-
Returns path to filtered file and the number of excluded sentences."""
|
| 29 |
-
gold_data = pyconll.load_from_file(gold_file_path)
|
| 30 |
-
filtered_sentences = [s for s in gold_data if s.id not in excluded_ids]
|
| 31 |
-
n_excluded = len(gold_data) - len(filtered_sentences)
|
| 32 |
-
|
| 33 |
-
if n_excluded > 0:
|
| 34 |
-
print(f" Excluded {n_excluded} validation sentences from {os.path.basename(gold_file_path)}")
|
| 35 |
-
|
| 36 |
-
# Write filtered data to a temp file
|
| 37 |
-
filtered_path = gold_file_path.replace(".conllu", "_eval_filtered.conllu")
|
| 38 |
-
with open(filtered_path, "w") as f:
|
| 39 |
-
for sentence in filtered_sentences:
|
| 40 |
-
f.write(sentence.conll())
|
| 41 |
-
f.write("\n\n")
|
| 42 |
-
return filtered_path, len(filtered_sentences)
|
| 43 |
-
|
| 44 |
-
# Load validation sentence IDs at startup
|
| 45 |
-
val_sentence_ids = load_val_sentence_ids(val_file)
|
| 46 |
-
|
| 47 |
-
# Model configurations: (predicted_dir, output_filename)
|
| 48 |
-
MODEL_CONFIGS = [
|
| 49 |
-
(ROOT / "parsed/testset_talbanken_not_seen_gold", "eval/scores/scores_talbanken.txt"),
|
| 50 |
-
(ROOT / "parsed/testset_transformer_silver_not_seen_gold", "eval/scores/scores_transformer_silver.txt"),
|
| 51 |
-
(ROOT / "parsed/testset_transformer_not_seen_gold_no_silver", "eval/scores/scores_transformer_no_silver.txt"),
|
| 52 |
-
]
|
| 53 |
-
|
| 54 |
-
def check_valid_gold(gold_file_path):
|
| 55 |
-
""" Check that the gold trees are valid """
|
| 56 |
-
gold_data = pyconll.load_from_file(gold_file_path)
|
| 57 |
-
errors = []
|
| 58 |
-
for sentence in gold_data:
|
| 59 |
-
for token in sentence:
|
| 60 |
-
if token.head is None:
|
| 61 |
-
e = ValueError(f"Invalid gold tree in file {gold_file_path}, sentence ID {sentence.id}: token {token.id} '{token.form}' has no head.")
|
| 62 |
-
errors.append(e)
|
| 63 |
-
elif token.head == token.id:
|
| 64 |
-
e = ValueError(f"Invalid gold tree in file {gold_file_path}, sentence ID {sentence.id}: token {token.id} '{token.form}' has itself as head.")
|
| 65 |
-
errors.append(e)
|
| 66 |
-
roots = [token for token in sentence if token.head == '0']
|
| 67 |
-
if len(roots) != 1:
|
| 68 |
-
e = ValueError(f"Invalid gold tree in file {gold_file_path}, sentence ID {sentence.id}: expected 1 root, found {len(roots)}.")
|
| 69 |
-
errors.append(e)
|
| 70 |
-
return errors if len(errors) > 0 else None
|
| 71 |
-
|
| 72 |
-
def preprocess_system_file(system_file_path, gold_file_path):
|
| 73 |
-
""" Make a new base file with only the trees present in the gold file """
|
| 74 |
-
gold_conllu = pyconll.load_from_file(gold_file_path)
|
| 75 |
-
system_conllu = pyconll.load_from_file(system_file_path)
|
| 76 |
-
gold_ids = {sentence.id for sentence in gold_conllu}
|
| 77 |
-
filtered_sentences = [sentence for sentence in system_conllu if sentence.id in gold_ids]
|
| 78 |
-
preprocessed_file_path = system_file_path.replace(".conllu", "_preprocessed.conllu")
|
| 79 |
-
with open(preprocessed_file_path, "w") as outfile:
|
| 80 |
-
for sentence in filtered_sentences:
|
| 81 |
-
outfile.write(sentence.conll())
|
| 82 |
-
outfile.write("\n\n")
|
| 83 |
-
return preprocessed_file_path
|
| 84 |
-
|
| 85 |
-
def execute_evaluation(gold_file_path, predicted_file_path):
|
| 86 |
-
# Read goldfile
|
| 87 |
-
try:
|
| 88 |
-
gold_data = load_conllu(open(gold_file_path, "r", encoding="utf-8"))
|
| 89 |
-
except Exception as e:
|
| 90 |
-
print(f"Skipping text because of error reading gold file {gold_file_path}: {e}")
|
| 91 |
-
#raise e
|
| 92 |
-
return None
|
| 93 |
-
# Read predicted file and ignore some format errors
|
| 94 |
-
try:
|
| 95 |
-
predicted_data = load_conllu(open(predicted_file_path, "r", encoding="utf-8"), ignore_invalid_format=True)
|
| 96 |
-
except Exception as e:
|
| 97 |
-
print(f"Skipping text because of error reading predicted file {predicted_file_path}: {e}")
|
| 98 |
-
#raise e
|
| 99 |
-
return None
|
| 100 |
-
# Evaluate the predictions against the gold standard
|
| 101 |
-
try:
|
| 102 |
-
results = evaluate(gold_data, predicted_data)
|
| 103 |
-
except Exception as e:
|
| 104 |
-
print(f"Error during evaluation between {gold_file_path} and {predicted_file_path}: {e}")
|
| 105 |
-
raise e
|
| 106 |
-
return results
|
| 107 |
-
|
| 108 |
-
time_periods = ["1700-1750", "1750-1800", "1800-1850", "1850-1900", "1900-1950"]
|
| 109 |
-
metrics = ["UPOS", "UAS", "LAS"] # the relevant metrics
|
| 110 |
-
|
| 111 |
-
# To look up time period index for a sample
|
| 112 |
-
sample_period = {"svediakorp-letter141673-Stalhammar": 0, "svediakorp-sec25-Runius": 0, "svediakorp-sec330-GyllenborgC_SwenskaSpratthoken": 0,
|
| 113 |
-
"svediakorp-sec277-EnbomPU_MedborgeligtSkalde": 1, "svediakorp-sec268-DulciU_VitterhetsNojen3": 1,
|
| 114 |
-
"svediakorp-sec991-spf148": 2, "svediakorp-sec252-BremerF_Teckningar1": 2, "svediakorp-sec324-GranbergPA_Enslighetsalskaren": 2,
|
| 115 |
-
"svediakorp-sec254-CederborghF_BerattelseOmJohnHall": 2, "svediakorp-sec987-spf144": 2, "svediakorp-sec988-spf145": 2,
|
| 116 |
-
"svediakorp-sec452-NyblomH_FantasierFyra": 3, "svediakorp-sec486-SchwartzMS_BellmansSkor": 3, "svediakorp-sec1102-spf259": 3,
|
| 117 |
-
"svediakorp-sec208-Anonym_DetGrasligaMordet": 3, "svediakorp-sec1063-spf220": 3,
|
| 118 |
-
"svediakorp-sec631-HasselskogN_HallaHallaGronkoping": 4, "svediakorp-sec1033-spf190": 4, "svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol": 4,
|
| 119 |
-
"svediakorp-sec613-EngstromA_StrindbergOchJag": 4, "svediakorp-sec639-HeidenstamV_Proletarfilosofiens": 4}
|
| 120 |
-
|
| 121 |
-
|
| 122 |
-
def run_evaluation(predicted_dir, output_file):
|
| 123 |
-
"""Run evaluation for a single model configuration."""
|
| 124 |
-
print(f"\n{'='*70}")
|
| 125 |
-
print(f"Evaluating: {predicted_dir}")
|
| 126 |
-
print(f"Output: {output_file}")
|
| 127 |
-
print(f"{'='*70}\n")
|
| 128 |
-
|
| 129 |
-
# Check if predicted directory exists
|
| 130 |
-
if not os.path.exists(predicted_dir):
|
| 131 |
-
print(f"ERROR: Predicted directory does not exist: {predicted_dir}")
|
| 132 |
-
return
|
| 133 |
-
|
| 134 |
-
# Reset per-model state
|
| 135 |
-
score_arrays = [np.zeros((4, len(metrics))), np.zeros((4, len(metrics))), np.zeros((4, len(metrics))), np.zeros((4, len(metrics))), np.zeros((4, len(metrics)))]
|
| 136 |
-
skipped_samples = []
|
| 137 |
-
sample_nsents = defaultdict()
|
| 138 |
-
processed_files = [] # Track (gold_path, pred_path) pairs for overall evaluation
|
| 139 |
-
|
| 140 |
-
with open(output_file, "w") as outfile:
|
| 141 |
-
# Loop through each of the validated samples
|
| 142 |
-
for sample in os.listdir(gold_dir):
|
| 143 |
-
sample_name = sample.split("/")[-1].replace(".conllu", "")
|
| 144 |
-
|
| 145 |
-
print("DEBUG sample_name:", sample_name)
|
| 146 |
-
|
| 147 |
-
period = sample_period.get(sample_name) # int 0-4
|
| 148 |
-
print("DEBUG period value:", period)
|
| 149 |
-
|
| 150 |
-
# Filter gold file to exclude sentences used during validation
|
| 151 |
-
gold_file_path_orig = os.path.join(gold_dir, sample)
|
| 152 |
-
gold_file_path, n_filtered_sents = filter_gold_file(gold_file_path_orig, val_sentence_ids)
|
| 153 |
-
|
| 154 |
-
# Skip if all sentences were excluded
|
| 155 |
-
if n_filtered_sents == 0:
|
| 156 |
-
print(f"Skipping {sample_name}: all sentences were used in validation")
|
| 157 |
-
skipped_samples.append(sample_name)
|
| 158 |
-
continue
|
| 159 |
-
|
| 160 |
-
# Check that gold trees are valid (using filtered file)
|
| 161 |
-
e = check_valid_gold(gold_file_path)
|
| 162 |
-
if e: # skip invalid gold files
|
| 163 |
-
for error in e:
|
| 164 |
-
print(error)
|
| 165 |
-
skipped_samples.append(sample_name)
|
| 166 |
-
continue
|
| 167 |
-
# Always regenerate preprocessed predicted file to match filtered gold
|
| 168 |
-
unprocessed_predicted_file_path = os.path.join(predicted_dir, sample)
|
| 169 |
-
if os.path.exists(unprocessed_predicted_file_path):
|
| 170 |
-
# Use filtered gold file to preprocess predictions
|
| 171 |
-
predicted_file_path = preprocess_system_file(unprocessed_predicted_file_path, gold_file_path)
|
| 172 |
-
else:
|
| 173 |
-
print(f"Skipping {sample_name}: missing predicted file {unprocessed_predicted_file_path}")
|
| 174 |
-
skipped_samples.append(sample_name)
|
| 175 |
-
continue
|
| 176 |
-
# Score base tree (parser output) against validated tree (gold standard)
|
| 177 |
-
results = execute_evaluation(gold_file_path, predicted_file_path)
|
| 178 |
-
|
| 179 |
-
if not results: # skip invalid files
|
| 180 |
-
skipped_samples.append(sample_name)
|
| 181 |
-
# Write results to scores file and save to time period arrays
|
| 182 |
-
else:
|
| 183 |
-
# Save number of sentences in lookup dict (using filtered count)
|
| 184 |
-
sample_nsents[sample_name] = n_filtered_sents
|
| 185 |
-
# Track file paths for overall evaluation
|
| 186 |
-
processed_files.append((gold_file_path, predicted_file_path))
|
| 187 |
-
print("DEBUG period value:", period)
|
| 188 |
-
print("DEBUG time_periods:", time_periods)
|
| 189 |
-
|
| 190 |
-
outfile.write(f"Sample: {sample_name}, time period: {time_periods[period]}\n")
|
| 191 |
-
outfile.write("Metric\tPrecision\tRecall\tF1\tAligned Accuracy\n")
|
| 192 |
-
metric_n = 0 # counter for loop below
|
| 193 |
-
# Loop over each metric: UPOS, UAS, LAS etc.
|
| 194 |
-
for metric, score in results.items(): # get the metric and its 4 scores
|
| 195 |
-
if metric in metrics: # only save the relevant metrics
|
| 196 |
-
# Save scores weighted by the number of sentences to arrays
|
| 197 |
-
score_arrays[period][0][metric_n] += (score.precision * sample_nsents[sample_name])
|
| 198 |
-
score_arrays[period][1][metric_n] += (score.recall * sample_nsents[sample_name])
|
| 199 |
-
score_arrays[period][2][metric_n] += (score.f1 * sample_nsents[sample_name])
|
| 200 |
-
# Write the metric and its file-average scores
|
| 201 |
-
outfile.write(f"{metric}\t{score.precision:.4f}\t{score.recall:.4f}\t{score.f1:.4f}\t")
|
| 202 |
-
if score.aligned_accuracy is None:
|
| 203 |
-
outfile.write(f"{score.aligned_accuracy}\n")
|
| 204 |
-
else:
|
| 205 |
-
outfile.write(f"{score.aligned_accuracy:.4f}\n")
|
| 206 |
-
# Save aligned accuracy to score arrays
|
| 207 |
-
score_arrays[period][3][metric_n] += (score.aligned_accuracy * sample_nsents[sample_name])
|
| 208 |
-
metric_n += 1 # move to next metric in array
|
| 209 |
-
outfile.write("\n")
|
| 210 |
-
outfile.write(f"{'-'*70}\n")
|
| 211 |
-
|
| 212 |
-
# After all sample scores are written, compute average scores for each time period
|
| 213 |
-
for period, period_scores in enumerate(score_arrays):
|
| 214 |
-
# Get number of samples and sentences for the period
|
| 215 |
-
samples_per_period = [sample_name for sample_name in sample_period.keys() if sample_period[sample_name] == period and sample_name not in skipped_samples]
|
| 216 |
-
n_sents_per_period = sum([sample_nsents[sample_name] for sample_name in samples_per_period])
|
| 217 |
-
# Write
|
| 218 |
-
outfile.write(f"Average scores for time period {time_periods[period]} ({len(samples_per_period)} samples with a total of {n_sents_per_period} sentences):\n")
|
| 219 |
-
outfile.write("Metric\tPrecision\tRecall\tF1\tAligned Accuracy\n")
|
| 220 |
-
for metric_n, metric in enumerate(metrics):
|
| 221 |
-
# Normalize weighted averages by the total number of sentences for the time period
|
| 222 |
-
precision_avg = period_scores[0][metric_n] / n_sents_per_period
|
| 223 |
-
recall_avg = period_scores[1][metric_n] / n_sents_per_period
|
| 224 |
-
f1_avg = period_scores[2][metric_n] / n_sents_per_period
|
| 225 |
-
aligned_acc_avg = period_scores[3][metric_n] / n_sents_per_period
|
| 226 |
-
outfile.write(f"{metric}\t{precision_avg:.4f}\t{recall_avg:.4f}\t{f1_avg:.4f}\t{aligned_acc_avg:.4f}\n")
|
| 227 |
-
outfile.write("\n")
|
| 228 |
-
|
| 229 |
-
# Compute overall scores by concatenating all files and running single evaluation
|
| 230 |
-
valid_samples = [sample_name for sample_name in sample_period.keys() if sample_name in sample_nsents and sample_name not in skipped_samples]
|
| 231 |
-
total_sents = sum(sample_nsents[sample_name] for sample_name in valid_samples)
|
| 232 |
-
|
| 233 |
-
if len(processed_files) == 0:
|
| 234 |
-
outfile.write(f"Overall scores across all time periods ({len(valid_samples)} samples with a total of {total_sents} sentences):\n")
|
| 235 |
-
outfile.write("Metric\tPrecision\tRecall\tF1\tAligned Accuracy\n")
|
| 236 |
-
outfile.write("No valid samples available for overall evaluation.\n")
|
| 237 |
-
else:
|
| 238 |
-
# Concatenate all gold and predicted files
|
| 239 |
-
all_gold_path = os.path.join(predicted_dir, "_all_gold_combined.conllu")
|
| 240 |
-
all_pred_path = os.path.join(predicted_dir, "_all_pred_combined.conllu")
|
| 241 |
-
|
| 242 |
-
with open(all_gold_path, "w") as gold_out, open(all_pred_path, "w") as pred_out:
|
| 243 |
-
for gold_path, pred_path in processed_files:
|
| 244 |
-
with open(gold_path) as f:
|
| 245 |
-
gold_out.write(f.read())
|
| 246 |
-
with open(pred_path) as f:
|
| 247 |
-
pred_out.write(f.read())
|
| 248 |
-
|
| 249 |
-
# Run single evaluation on combined files
|
| 250 |
-
overall_results = execute_evaluation(all_gold_path, all_pred_path)
|
| 251 |
-
|
| 252 |
-
outfile.write(f"Overall scores across all time periods ({len(valid_samples)} samples with a total of {total_sents} sentences):\n")
|
| 253 |
-
outfile.write("Metric\tPrecision\tRecall\tF1\tAligned Accuracy\n")
|
| 254 |
-
|
| 255 |
-
if overall_results:
|
| 256 |
-
for metric, score in overall_results.items():
|
| 257 |
-
if metric in metrics:
|
| 258 |
-
outfile.write(f"{metric}\t{score.precision:.4f}\t{score.recall:.4f}\t{score.f1:.4f}\t")
|
| 259 |
-
if score.aligned_accuracy is None:
|
| 260 |
-
outfile.write(f"{score.aligned_accuracy}\n")
|
| 261 |
-
else:
|
| 262 |
-
outfile.write(f"{score.aligned_accuracy:.4f}\n")
|
| 263 |
-
else:
|
| 264 |
-
outfile.write("Error computing overall scores.\n")
|
| 265 |
-
|
| 266 |
-
print(f"Results saved to {output_file}")
|
| 267 |
-
|
| 268 |
-
|
| 269 |
-
# Run evaluation for all model configurations
|
| 270 |
-
if __name__ == "__main__":
|
| 271 |
-
for predicted_dir, output_file in MODEL_CONFIGS:
|
| 272 |
-
run_evaluation(predicted_dir, output_file)
|
| 273 |
-
print("\nAll evaluations complete!")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
eval/mod_conll18_ud_eval.py
DELETED
|
@@ -1,672 +0,0 @@
|
|
| 1 |
-
# This script is from the CoNLL 2018 Shared Task.
|
| 2 |
-
# It was downloaded from here: https://universaldependencies.org/conll18/evaluation.html
|
| 3 |
-
# in October 2025, and modified by Astrid B. Ingelstam for the group project about diachronic treebanks.
|
| 4 |
-
|
| 5 |
-
#!/usr/bin/env python3
|
| 6 |
-
|
| 7 |
-
# Compatible with Python 2.7 and 3.2+, can be used either as a module
|
| 8 |
-
# or a standalone executable.
|
| 9 |
-
#
|
| 10 |
-
# Copyright 2017, 2018 Institute of Formal and Applied Linguistics (UFAL),
|
| 11 |
-
# Faculty of Mathematics and Physics, Charles University, Czech Republic.
|
| 12 |
-
#
|
| 13 |
-
# This Source Code Form is subject to the terms of the Mozilla Public
|
| 14 |
-
# License, v. 2.0. If a copy of the MPL was not distributed with this
|
| 15 |
-
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
|
| 16 |
-
#
|
| 17 |
-
# Authors: Milan Straka, Martin Popel <surname@ufal.mff.cuni.cz>
|
| 18 |
-
#
|
| 19 |
-
# Changelog:
|
| 20 |
-
# - [12 Apr 2018] Version 0.9: Initial release.
|
| 21 |
-
# - [19 Apr 2018] Version 1.0: Fix bug in MLAS (duplicate entries in functional_children).
|
| 22 |
-
# Add --counts option.
|
| 23 |
-
# - [02 May 2018] Version 1.1: When removing spaces to match gold and system characters,
|
| 24 |
-
# consider all Unicode characters of category Zs instead of
|
| 25 |
-
# just ASCII space.
|
| 26 |
-
# - [25 Jun 2018] Version 1.2: Use python3 in the she-bang (instead of python).
|
| 27 |
-
# In Python2, make the whole computation use `unicode` strings.
|
| 28 |
-
|
| 29 |
-
# Command line usage
|
| 30 |
-
# ------------------
|
| 31 |
-
# conll18_ud_eval.py [-v] gold_conllu_file system_conllu_file
|
| 32 |
-
#
|
| 33 |
-
# - if no -v is given, only the official CoNLL18 UD Shared Task evaluation metrics
|
| 34 |
-
# are printed
|
| 35 |
-
# - if -v is given, more metrics are printed (as precision, recall, F1 score,
|
| 36 |
-
# and in case the metric is computed on aligned words also accuracy on these):
|
| 37 |
-
# - Tokens: how well do the gold tokens match system tokens
|
| 38 |
-
# - Sentences: how well do the gold sentences match system sentences
|
| 39 |
-
# - Words: how well can the gold words be aligned to system words
|
| 40 |
-
# - UPOS: using aligned words, how well does UPOS match
|
| 41 |
-
# - XPOS: using aligned words, how well does XPOS match
|
| 42 |
-
# - UFeats: using aligned words, how well does universal FEATS match
|
| 43 |
-
# - AllTags: using aligned words, how well does UPOS+XPOS+FEATS match
|
| 44 |
-
# - Lemmas: using aligned words, how well does LEMMA match
|
| 45 |
-
# - UAS: using aligned words, how well does HEAD match
|
| 46 |
-
# - LAS: using aligned words, how well does HEAD+DEPREL(ignoring subtypes) match
|
| 47 |
-
# - CLAS: using aligned words with content DEPREL, how well does
|
| 48 |
-
# HEAD+DEPREL(ignoring subtypes) match
|
| 49 |
-
# - MLAS: using aligned words with content DEPREL, how well does
|
| 50 |
-
# HEAD+DEPREL(ignoring subtypes)+UPOS+UFEATS+FunctionalChildren(DEPREL+UPOS+UFEATS) match
|
| 51 |
-
# - BLEX: using aligned words with content DEPREL, how well does
|
| 52 |
-
# HEAD+DEPREL(ignoring subtypes)+LEMMAS match
|
| 53 |
-
# - if -c is given, raw counts of correct/gold_total/system_total/aligned words are printed
|
| 54 |
-
# instead of precision/recall/F1/AlignedAccuracy for all metrics.
|
| 55 |
-
|
| 56 |
-
# API usage
|
| 57 |
-
# ---------
|
| 58 |
-
# - load_conllu(file)
|
| 59 |
-
# - loads CoNLL-U file from given file object to an internal representation
|
| 60 |
-
# - the file object should return str in both Python 2 and Python 3
|
| 61 |
-
# - raises UDError exception if the given file cannot be loaded
|
| 62 |
-
# - evaluate(gold_ud, system_ud)
|
| 63 |
-
# - evaluate the given gold and system CoNLL-U files (loaded with load_conllu)
|
| 64 |
-
# - raises UDError if the concatenated tokens of gold and system file do not match
|
| 65 |
-
# - returns a dictionary with the metrics described above, each metric having
|
| 66 |
-
# three fields: precision, recall and f1
|
| 67 |
-
|
| 68 |
-
# Description of token matching
|
| 69 |
-
# -----------------------------
|
| 70 |
-
# In order to match tokens of gold file and system file, we consider the text
|
| 71 |
-
# resulting from concatenation of gold tokens and text resulting from
|
| 72 |
-
# concatenation of system tokens. These texts should match -- if they do not,
|
| 73 |
-
# the evaluation fails.
|
| 74 |
-
#
|
| 75 |
-
# If the texts do match, every token is represented as a range in this original
|
| 76 |
-
# text, and tokens are equal only if their range is the same.
|
| 77 |
-
|
| 78 |
-
# Description of word matching
|
| 79 |
-
# ----------------------------
|
| 80 |
-
# When matching words of gold file and system file, we first match the tokens.
|
| 81 |
-
# The words which are also tokens are matched as tokens, but words in multi-word
|
| 82 |
-
# tokens have to be handled differently.
|
| 83 |
-
#
|
| 84 |
-
# To handle multi-word tokens, we start by finding "multi-word spans".
|
| 85 |
-
# Multi-word span is a span in the original text such that
|
| 86 |
-
# - it contains at least one multi-word token
|
| 87 |
-
# - all multi-word tokens in the span (considering both gold and system ones)
|
| 88 |
-
# are completely inside the span (i.e., they do not "stick out")
|
| 89 |
-
# - the multi-word span is as small as possible
|
| 90 |
-
#
|
| 91 |
-
# For every multi-word span, we align the gold and system words completely
|
| 92 |
-
# inside this span using LCS on their FORMs. The words not intersecting
|
| 93 |
-
# (even partially) any multi-word span are then aligned as tokens.
|
| 94 |
-
|
| 95 |
-
|
| 96 |
-
from __future__ import division
|
| 97 |
-
from __future__ import print_function
|
| 98 |
-
|
| 99 |
-
import argparse
|
| 100 |
-
import io
|
| 101 |
-
import sys
|
| 102 |
-
import unicodedata
|
| 103 |
-
import unittest
|
| 104 |
-
from collections import defaultdict
|
| 105 |
-
|
| 106 |
-
CONFUSIONS = defaultdict(lambda: defaultdict(int)) #ADDED
|
| 107 |
-
|
| 108 |
-
# CoNLL-U column names
|
| 109 |
-
ID, FORM, LEMMA, UPOS, XPOS, FEATS, HEAD, DEPREL, DEPS, MISC = range(10)
|
| 110 |
-
|
| 111 |
-
# Content and functional relations
|
| 112 |
-
CONTENT_DEPRELS = {
|
| 113 |
-
"acl", "advcl", "advmod", "amod", "appos", "aux", "case", "cc", "ccomp",
|
| 114 |
-
"clf", "compound", "conj", "cop", "csubj", "dep", "det", "discourse",
|
| 115 |
-
"dislocated", "expl", "fixed", "flat", "goeswith", "iobj", "list", "mark",
|
| 116 |
-
"nmod", "nsubj", "nummod", "obj", "obl", "orphan", "parataxis", "punct",
|
| 117 |
-
"reparandum", "root", "vocative", "xcomp"
|
| 118 |
-
}
|
| 119 |
-
# Add for detailed scores
|
| 120 |
-
UPOS_TAGS = {
|
| 121 |
-
"ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM",
|
| 122 |
-
"PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X"
|
| 123 |
-
}
|
| 124 |
-
##
|
| 125 |
-
FUNCTIONAL_DEPRELS = {
|
| 126 |
-
"aux", "cop", "mark", "det", "clf", "case", "cc"
|
| 127 |
-
}
|
| 128 |
-
|
| 129 |
-
UNIVERSAL_FEATURES = {
|
| 130 |
-
"PronType", "NumType", "Poss", "Reflex", "Foreign", "Abbr", "Gender",
|
| 131 |
-
"Animacy", "Number", "Case", "Definite", "Degree", "VerbForm", "Mood",
|
| 132 |
-
"Tense", "Aspect", "Voice", "Evident", "Polarity", "Person", "Polite"
|
| 133 |
-
}
|
| 134 |
-
|
| 135 |
-
# UD Error is used when raising exceptions in this module
|
| 136 |
-
class UDError(Exception):
|
| 137 |
-
pass
|
| 138 |
-
|
| 139 |
-
# Conversion methods handling `str` <-> `unicode` conversions in Python2
|
| 140 |
-
def _decode(text):
|
| 141 |
-
return text if sys.version_info[0] >= 3 or not isinstance(text, str) else text.decode("utf-8")
|
| 142 |
-
|
| 143 |
-
def _encode(text):
|
| 144 |
-
return text if sys.version_info[0] >= 3 or not isinstance(text, unicode) else text.encode("utf-8")
|
| 145 |
-
|
| 146 |
-
# Load given CoNLL-U file into internal representation
|
| 147 |
-
def load_conllu(file, ignore_invalid_format=False):
|
| 148 |
-
# Internal representation classes
|
| 149 |
-
class UDRepresentation:
|
| 150 |
-
def __init__(self):
|
| 151 |
-
# Characters of all the tokens in the whole file.
|
| 152 |
-
# Whitespace between tokens is not included.
|
| 153 |
-
self.characters = []
|
| 154 |
-
# List of UDSpan instances with start&end indices into `characters`.
|
| 155 |
-
self.tokens = []
|
| 156 |
-
# List of UDWord instances.
|
| 157 |
-
self.words = []
|
| 158 |
-
# List of UDSpan instances with start&end indices into `characters`.
|
| 159 |
-
self.sentences = []
|
| 160 |
-
class UDSpan:
|
| 161 |
-
def __init__(self, start, end):
|
| 162 |
-
self.start = start
|
| 163 |
-
# Note that self.end marks the first position **after the end** of span,
|
| 164 |
-
# so we can use characters[start:end] or range(start, end).
|
| 165 |
-
self.end = end
|
| 166 |
-
class UDWord:
|
| 167 |
-
def __init__(self, span, columns, is_multiword):
|
| 168 |
-
# Span of this word (or MWT, see below) within ud_representation.characters.
|
| 169 |
-
self.span = span
|
| 170 |
-
# 10 columns of the CoNLL-U file: ID, FORM, LEMMA,...
|
| 171 |
-
self.columns = columns
|
| 172 |
-
# is_multiword==True means that this word is part of a multi-word token.
|
| 173 |
-
# In that case, self.span marks the span of the whole multi-word token.
|
| 174 |
-
self.is_multiword = is_multiword
|
| 175 |
-
# Reference to the UDWord instance representing the HEAD (or None if root).
|
| 176 |
-
self.parent = None
|
| 177 |
-
# List of references to UDWord instances representing functional-deprel children.
|
| 178 |
-
self.functional_children = []
|
| 179 |
-
# Only consider universal FEATS.
|
| 180 |
-
self.columns[FEATS] = "|".join(sorted(feat for feat in columns[FEATS].split("|")
|
| 181 |
-
if feat.split("=", 1)[0] in UNIVERSAL_FEATURES))
|
| 182 |
-
# Let's ignore language-specific deprel subtypes.
|
| 183 |
-
self.columns[DEPREL] = columns[DEPREL].split(":")[0]
|
| 184 |
-
# Precompute which deprels are CONTENT_DEPRELS and which FUNCTIONAL_DEPRELS
|
| 185 |
-
self.is_content_deprel = self.columns[DEPREL] in CONTENT_DEPRELS
|
| 186 |
-
self.is_functional_deprel = self.columns[DEPREL] in FUNCTIONAL_DEPRELS
|
| 187 |
-
|
| 188 |
-
ud = UDRepresentation()
|
| 189 |
-
|
| 190 |
-
# Load the CoNLL-U file
|
| 191 |
-
index, sentence_start = 0, None
|
| 192 |
-
while True:
|
| 193 |
-
line = file.readline()
|
| 194 |
-
if not line:
|
| 195 |
-
break
|
| 196 |
-
line = _decode(line.rstrip("\r\n"))
|
| 197 |
-
|
| 198 |
-
# Handle sentence start boundaries
|
| 199 |
-
if sentence_start is None:
|
| 200 |
-
# Skip comments
|
| 201 |
-
if line.startswith("#"):
|
| 202 |
-
continue
|
| 203 |
-
# Start a new sentence
|
| 204 |
-
ud.sentences.append(UDSpan(index, 0))
|
| 205 |
-
sentence_start = len(ud.words)
|
| 206 |
-
if not line:
|
| 207 |
-
# Add parent and children UDWord links and check there are no cycles
|
| 208 |
-
def process_word(word):
|
| 209 |
-
if word.parent == "remapping":
|
| 210 |
-
if not ignore_invalid_format:
|
| 211 |
-
raise UDError("There is a cycle in a sentence")
|
| 212 |
-
if word.parent is None:
|
| 213 |
-
try:
|
| 214 |
-
head = int(word.columns[HEAD])
|
| 215 |
-
except ValueError:
|
| 216 |
-
head = 0 # Astrid: handle headless nodes as roots
|
| 217 |
-
if head < 0 or head > len(ud.words) - sentence_start:
|
| 218 |
-
if not ignore_invalid_format:
|
| 219 |
-
raise UDError("HEAD '{}' points outside of the sentence".format(_encode(word.columns[HEAD])))
|
| 220 |
-
if head:
|
| 221 |
-
parent = ud.words[sentence_start + head - 1]
|
| 222 |
-
word.parent = "remapping"
|
| 223 |
-
process_word(parent)
|
| 224 |
-
word.parent = parent
|
| 225 |
-
|
| 226 |
-
for word in ud.words[sentence_start:]:
|
| 227 |
-
process_word(word)
|
| 228 |
-
# func_children cannot be assigned within process_word
|
| 229 |
-
# because it is called recursively and may result in adding one child twice.
|
| 230 |
-
for word in ud.words[sentence_start:]:
|
| 231 |
-
if word.parent and word.is_functional_deprel:
|
| 232 |
-
word.parent.functional_children.append(word)
|
| 233 |
-
|
| 234 |
-
# Check there is a single root node
|
| 235 |
-
if len([word for word in ud.words[sentence_start:] if word.parent is None]) != 1:
|
| 236 |
-
#print(f"Warning: There are {len([word for word in ud.words[sentence_start:] if word.parent is None])} roots in a sentence.")
|
| 237 |
-
if not ignore_invalid_format:
|
| 238 |
-
raise UDError("There are multiple roots in a sentence")
|
| 239 |
-
|
| 240 |
-
# End the sentence
|
| 241 |
-
ud.sentences[-1].end = index
|
| 242 |
-
sentence_start = None
|
| 243 |
-
continue
|
| 244 |
-
|
| 245 |
-
# Read next token/word
|
| 246 |
-
columns = line.split("\t")
|
| 247 |
-
if len(columns) != 10:
|
| 248 |
-
raise UDError("The CoNLL-U line does not contain 10 tab-separated columns: '{}'".format(_encode(line)))
|
| 249 |
-
|
| 250 |
-
# Skip empty nodes
|
| 251 |
-
if "." in columns[ID]:
|
| 252 |
-
continue
|
| 253 |
-
|
| 254 |
-
# Delete spaces from FORM, so gold.characters == system.characters
|
| 255 |
-
# even if one of them tokenizes the space. Use any Unicode character
|
| 256 |
-
# with category Zs.
|
| 257 |
-
columns[FORM] = "".join(filter(lambda c: unicodedata.category(c) != "Zs", columns[FORM]))
|
| 258 |
-
if not columns[FORM]:
|
| 259 |
-
raise UDError("There is an empty FORM in the CoNLL-U file")
|
| 260 |
-
|
| 261 |
-
# Save token
|
| 262 |
-
ud.characters.extend(columns[FORM])
|
| 263 |
-
ud.tokens.append(UDSpan(index, index + len(columns[FORM])))
|
| 264 |
-
index += len(columns[FORM])
|
| 265 |
-
|
| 266 |
-
# Handle multi-word tokens to save word(s)
|
| 267 |
-
if "-" in columns[ID]:
|
| 268 |
-
try:
|
| 269 |
-
start, end = map(int, columns[ID].split("-"))
|
| 270 |
-
except:
|
| 271 |
-
raise UDError("Cannot parse multi-word token ID '{}'".format(_encode(columns[ID])))
|
| 272 |
-
|
| 273 |
-
for _ in range(start, end + 1):
|
| 274 |
-
word_line = _decode(file.readline().rstrip("\r\n"))
|
| 275 |
-
word_columns = word_line.split("\t")
|
| 276 |
-
if len(word_columns) != 10:
|
| 277 |
-
raise UDError("The CoNLL-U line does not contain 10 tab-separated columns: '{}'".format(_encode(word_line)))
|
| 278 |
-
ud.words.append(UDWord(ud.tokens[-1], word_columns, is_multiword=True))
|
| 279 |
-
# Basic tokens/words
|
| 280 |
-
else:
|
| 281 |
-
try:
|
| 282 |
-
word_id = int(columns[ID])
|
| 283 |
-
except:
|
| 284 |
-
raise UDError("Cannot parse word ID '{}'".format(_encode(columns[ID])))
|
| 285 |
-
if word_id != len(ud.words) - sentence_start + 1:
|
| 286 |
-
raise UDError("Incorrect word ID '{}' for word '{}', expected '{}'".format(
|
| 287 |
-
_encode(columns[ID]), _encode(columns[FORM]), len(ud.words) - sentence_start + 1))
|
| 288 |
-
|
| 289 |
-
try:
|
| 290 |
-
head_id = int(columns[HEAD])
|
| 291 |
-
except:
|
| 292 |
-
if ignore_invalid_format:
|
| 293 |
-
head_id = 0 # Astrid: handle headless nodes as roots
|
| 294 |
-
else:
|
| 295 |
-
raise UDError("Cannot parse HEAD '{}'".format(_encode(columns[HEAD])))
|
| 296 |
-
if head_id < 0:
|
| 297 |
-
raise UDError("HEAD cannot be negative")
|
| 298 |
-
|
| 299 |
-
ud.words.append(UDWord(ud.tokens[-1], columns, is_multiword=False))
|
| 300 |
-
|
| 301 |
-
if sentence_start is not None:
|
| 302 |
-
raise UDError("The CoNLL-U file does not end with empty line")
|
| 303 |
-
|
| 304 |
-
return ud
|
| 305 |
-
|
| 306 |
-
# Evaluate the gold and system treebanks (loaded using load_conllu).
|
| 307 |
-
def evaluate(gold_ud, system_ud, detailed=False, cm=False): #ADDED cm=False
|
| 308 |
-
class Score:
|
| 309 |
-
def __init__(self, gold_total, system_total, correct, aligned_total=None):
|
| 310 |
-
self.correct = correct
|
| 311 |
-
self.gold_total = gold_total
|
| 312 |
-
self.system_total = system_total
|
| 313 |
-
self.aligned_total = aligned_total
|
| 314 |
-
self.precision = correct / system_total if system_total else 0.0
|
| 315 |
-
self.recall = correct / gold_total if gold_total else 0.0
|
| 316 |
-
self.f1 = 2 * correct / (system_total + gold_total) if system_total + gold_total else 0.0
|
| 317 |
-
self.aligned_accuracy = correct / aligned_total if aligned_total else aligned_total
|
| 318 |
-
class AlignmentWord:
|
| 319 |
-
def __init__(self, gold_word, system_word):
|
| 320 |
-
self.gold_word = gold_word
|
| 321 |
-
self.system_word = system_word
|
| 322 |
-
class Alignment:
|
| 323 |
-
def __init__(self, gold_words, system_words):
|
| 324 |
-
self.gold_words = gold_words
|
| 325 |
-
self.system_words = system_words
|
| 326 |
-
self.matched_words = []
|
| 327 |
-
self.matched_words_map = {}
|
| 328 |
-
def append_aligned_words(self, gold_word, system_word):
|
| 329 |
-
self.matched_words.append(AlignmentWord(gold_word, system_word))
|
| 330 |
-
self.matched_words_map[system_word] = gold_word
|
| 331 |
-
|
| 332 |
-
def spans_score(gold_spans, system_spans):
|
| 333 |
-
correct, gi, si = 0, 0, 0
|
| 334 |
-
while gi < len(gold_spans) and si < len(system_spans):
|
| 335 |
-
if system_spans[si].start < gold_spans[gi].start:
|
| 336 |
-
si += 1
|
| 337 |
-
elif gold_spans[gi].start < system_spans[si].start:
|
| 338 |
-
gi += 1
|
| 339 |
-
else:
|
| 340 |
-
correct += gold_spans[gi].end == system_spans[si].end
|
| 341 |
-
si += 1
|
| 342 |
-
gi += 1
|
| 343 |
-
|
| 344 |
-
return Score(len(gold_spans), len(system_spans), correct)
|
| 345 |
-
|
| 346 |
-
def alignment_score(alignment, key_fn=None, filter_fn=None, name=None): # added "name" flag for confusion matrix
|
| 347 |
-
if filter_fn is not None:
|
| 348 |
-
gold = sum(1 for gold in alignment.gold_words if filter_fn(gold))
|
| 349 |
-
system = sum(1 for system in alignment.system_words if filter_fn(system))
|
| 350 |
-
aligned = sum(
|
| 351 |
-
1 for pair in alignment.matched_words
|
| 352 |
-
if filter_fn(pair.gold_word) and filter_fn(pair.system_word))
|
| 353 |
-
else:
|
| 354 |
-
gold = len(alignment.gold_words)
|
| 355 |
-
system = len(alignment.system_words)
|
| 356 |
-
aligned = len(alignment.matched_words)
|
| 357 |
-
|
| 358 |
-
if key_fn is None:
|
| 359 |
-
# Return score for whole aligned words
|
| 360 |
-
return Score(gold, system, aligned)
|
| 361 |
-
|
| 362 |
-
def gold_aligned_gold(word):
|
| 363 |
-
return word
|
| 364 |
-
def gold_aligned_system(word):
|
| 365 |
-
return alignment.matched_words_map.get(word, "NotAligned") if word is not None else None
|
| 366 |
-
correct = 0
|
| 367 |
-
for words in alignment.matched_words:
|
| 368 |
-
if filter_fn is None or filter_fn(words.gold_word):
|
| 369 |
-
gold_label = key_fn(words.gold_word, gold_aligned_gold) # ADDED
|
| 370 |
-
system_label = key_fn(words.system_word, gold_aligned_system) # ADDED
|
| 371 |
-
if key_fn(words.gold_word, gold_aligned_gold) == key_fn(words.system_word, gold_aligned_system):
|
| 372 |
-
correct += 1
|
| 373 |
-
|
| 374 |
-
if name is not None:
|
| 375 |
-
CONFUSIONS[name][(gold_label, system_label)] += 1
|
| 376 |
-
|
| 377 |
-
return Score(gold, system, correct, aligned)
|
| 378 |
-
|
| 379 |
-
def beyond_end(words, i, multiword_span_end):
|
| 380 |
-
if i >= len(words):
|
| 381 |
-
return True
|
| 382 |
-
if words[i].is_multiword:
|
| 383 |
-
return words[i].span.start >= multiword_span_end
|
| 384 |
-
return words[i].span.end > multiword_span_end
|
| 385 |
-
|
| 386 |
-
def extend_end(word, multiword_span_end):
|
| 387 |
-
if word.is_multiword and word.span.end > multiword_span_end:
|
| 388 |
-
return word.span.end
|
| 389 |
-
return multiword_span_end
|
| 390 |
-
|
| 391 |
-
def find_multiword_span(gold_words, system_words, gi, si):
|
| 392 |
-
# We know gold_words[gi].is_multiword or system_words[si].is_multiword.
|
| 393 |
-
# Find the start of the multiword span (gs, ss), so the multiword span is minimal.
|
| 394 |
-
# Initialize multiword_span_end characters index.
|
| 395 |
-
if gold_words[gi].is_multiword:
|
| 396 |
-
multiword_span_end = gold_words[gi].span.end
|
| 397 |
-
if not system_words[si].is_multiword and system_words[si].span.start < gold_words[gi].span.start:
|
| 398 |
-
si += 1
|
| 399 |
-
else: # if system_words[si].is_multiword
|
| 400 |
-
multiword_span_end = system_words[si].span.end
|
| 401 |
-
if not gold_words[gi].is_multiword and gold_words[gi].span.start < system_words[si].span.start:
|
| 402 |
-
gi += 1
|
| 403 |
-
gs, ss = gi, si
|
| 404 |
-
|
| 405 |
-
# Find the end of the multiword span
|
| 406 |
-
# (so both gi and si are pointing to the word following the multiword span end).
|
| 407 |
-
while not beyond_end(gold_words, gi, multiword_span_end) or \
|
| 408 |
-
not beyond_end(system_words, si, multiword_span_end):
|
| 409 |
-
if gi < len(gold_words) and (si >= len(system_words) or
|
| 410 |
-
gold_words[gi].span.start <= system_words[si].span.start):
|
| 411 |
-
multiword_span_end = extend_end(gold_words[gi], multiword_span_end)
|
| 412 |
-
gi += 1
|
| 413 |
-
else:
|
| 414 |
-
multiword_span_end = extend_end(system_words[si], multiword_span_end)
|
| 415 |
-
si += 1
|
| 416 |
-
return gs, ss, gi, si
|
| 417 |
-
|
| 418 |
-
def compute_lcs(gold_words, system_words, gi, si, gs, ss):
|
| 419 |
-
lcs = [[0] * (si - ss) for i in range(gi - gs)]
|
| 420 |
-
for g in reversed(range(gi - gs)):
|
| 421 |
-
for s in reversed(range(si - ss)):
|
| 422 |
-
if gold_words[gs + g].columns[FORM].lower() == system_words[ss + s].columns[FORM].lower():
|
| 423 |
-
lcs[g][s] = 1 + (lcs[g+1][s+1] if g+1 < gi-gs and s+1 < si-ss else 0)
|
| 424 |
-
lcs[g][s] = max(lcs[g][s], lcs[g+1][s] if g+1 < gi-gs else 0)
|
| 425 |
-
lcs[g][s] = max(lcs[g][s], lcs[g][s+1] if s+1 < si-ss else 0)
|
| 426 |
-
return lcs
|
| 427 |
-
|
| 428 |
-
def align_words(gold_words, system_words):
|
| 429 |
-
alignment = Alignment(gold_words, system_words)
|
| 430 |
-
|
| 431 |
-
gi, si = 0, 0
|
| 432 |
-
while gi < len(gold_words) and si < len(system_words):
|
| 433 |
-
if gold_words[gi].is_multiword or system_words[si].is_multiword:
|
| 434 |
-
# A: Multi-word tokens => align via LCS within the whole "multiword span".
|
| 435 |
-
gs, ss, gi, si = find_multiword_span(gold_words, system_words, gi, si)
|
| 436 |
-
|
| 437 |
-
if si > ss and gi > gs:
|
| 438 |
-
lcs = compute_lcs(gold_words, system_words, gi, si, gs, ss)
|
| 439 |
-
|
| 440 |
-
# Store aligned words
|
| 441 |
-
s, g = 0, 0
|
| 442 |
-
while g < gi - gs and s < si - ss:
|
| 443 |
-
if gold_words[gs + g].columns[FORM].lower() == system_words[ss + s].columns[FORM].lower():
|
| 444 |
-
alignment.append_aligned_words(gold_words[gs+g], system_words[ss+s])
|
| 445 |
-
g += 1
|
| 446 |
-
s += 1
|
| 447 |
-
elif lcs[g][s] == (lcs[g+1][s] if g+1 < gi-gs else 0):
|
| 448 |
-
g += 1
|
| 449 |
-
else:
|
| 450 |
-
s += 1
|
| 451 |
-
else:
|
| 452 |
-
# B: No multi-word token => align according to spans.
|
| 453 |
-
if (gold_words[gi].span.start, gold_words[gi].span.end) == (system_words[si].span.start, system_words[si].span.end):
|
| 454 |
-
alignment.append_aligned_words(gold_words[gi], system_words[si])
|
| 455 |
-
gi += 1
|
| 456 |
-
si += 1
|
| 457 |
-
elif gold_words[gi].span.start <= system_words[si].span.start:
|
| 458 |
-
gi += 1
|
| 459 |
-
else:
|
| 460 |
-
si += 1
|
| 461 |
-
|
| 462 |
-
return alignment
|
| 463 |
-
|
| 464 |
-
# Check that the underlying character sequences do match.
|
| 465 |
-
if gold_ud.characters != system_ud.characters:
|
| 466 |
-
"""index = 0
|
| 467 |
-
while index < len(gold_ud.characters) and index < len(system_ud.characters) and \
|
| 468 |
-
gold_ud.characters[index] == system_ud.characters[index]:
|
| 469 |
-
index += 1
|
| 470 |
-
|
| 471 |
-
raise UDError(
|
| 472 |
-
"The concatenation of tokens in gold file and in system file differ!\n" +
|
| 473 |
-
"First 20 differing characters in gold file: '{}' and system file: '{}'".format(
|
| 474 |
-
"".join(map(_encode, gold_ud.characters[index:index + 20])),
|
| 475 |
-
"".join(map(_encode, system_ud.characters[index:index + 20]))
|
| 476 |
-
)
|
| 477 |
-
)"""
|
| 478 |
-
gold_ud.characters = system_ud.characters # Astrid: token form is altered - go with base tree form
|
| 479 |
-
|
| 480 |
-
|
| 481 |
-
# Added to compute detailed UPOS scores:
|
| 482 |
-
def detailed_score(alignment, labels, column_ind):
|
| 483 |
-
r = {}
|
| 484 |
-
total_gold, total_system, total_correct = 0, 0, 0
|
| 485 |
-
for label in labels:
|
| 486 |
-
score = alignment_score(
|
| 487 |
-
alignment,
|
| 488 |
-
filter_fn=lambda w, label=label: w.columns[column_ind] == label
|
| 489 |
-
)
|
| 490 |
-
|
| 491 |
-
r[label] = {
|
| 492 |
-
"precision": score.precision,
|
| 493 |
-
"recall": score.recall,
|
| 494 |
-
"f1": score.f1
|
| 495 |
-
}
|
| 496 |
-
total_gold += score.gold_total
|
| 497 |
-
total_system += score.system_total
|
| 498 |
-
total_correct += score.correct
|
| 499 |
-
|
| 500 |
-
r[f'TOTAL'] = {
|
| 501 |
-
"precision": total_correct / total_system if total_system else 0.0,
|
| 502 |
-
"recall": total_correct / total_gold if total_gold else 0.0,
|
| 503 |
-
"f1": 2 * total_correct / (total_system + total_gold) if (total_system + total_gold) else 0.0
|
| 504 |
-
}
|
| 505 |
-
return r
|
| 506 |
-
|
| 507 |
-
# Added to compute detailed UPOS scores:
|
| 508 |
-
|
| 509 |
-
## ADDED
|
| 510 |
-
def dump_confusion(confusion, filename):
|
| 511 |
-
labels = sorted(set(g for g, s in confusion) | set(s for g, s in confusion))
|
| 512 |
-
with open(filename, "w", encoding="utf-8") as f:
|
| 513 |
-
f.write("gold\\system\t" + "\t".join(map(str, labels)) + "\n")
|
| 514 |
-
for g in labels:
|
| 515 |
-
row = [str(confusion.get((g, s), 0)) for s in labels]
|
| 516 |
-
f.write(str(g) + "\t" + "\t".join(row) + "\n")
|
| 517 |
-
|
| 518 |
-
#for name, matrix in CONFUSIONS.items():
|
| 519 |
-
# dump_confusion(matrix, f"{name}_confusion.tsv")
|
| 520 |
-
## ADDED
|
| 521 |
-
|
| 522 |
-
|
| 523 |
-
# Align words
|
| 524 |
-
alignment = align_words(gold_ud.words, system_ud.words)
|
| 525 |
-
|
| 526 |
-
# Compute the F1-scores
|
| 527 |
-
if detailed: #The initial F1-scores
|
| 528 |
-
upos_scores = detailed_score(alignment, UPOS_TAGS, UPOS)
|
| 529 |
-
deprel_scores = detailed_score(alignment, CONTENT_DEPRELS, DEPREL)
|
| 530 |
-
return upos_scores, deprel_scores
|
| 531 |
-
|
| 532 |
-
# For Confusion matrix
|
| 533 |
-
if cm:
|
| 534 |
-
sc = {
|
| 535 |
-
"UPOS": alignment_score(alignment,lambda w, _: w.columns[UPOS], name = "UPOS"),
|
| 536 |
-
"DEPREL": alignment_score(alignment, lambda w, _: w.columns[DEPREL], name ="DEPREL"),
|
| 537 |
-
}
|
| 538 |
-
return CONFUSIONS
|
| 539 |
-
|
| 540 |
-
|
| 541 |
-
return {
|
| 542 |
-
"Tokens": spans_score(gold_ud.tokens, system_ud.tokens),
|
| 543 |
-
"Sentences": spans_score(gold_ud.sentences, system_ud.sentences),
|
| 544 |
-
"Words": alignment_score(alignment),
|
| 545 |
-
"UPOS": alignment_score(alignment, lambda w, _: w.columns[UPOS]),
|
| 546 |
-
"XPOS": alignment_score(alignment, lambda w, _: w.columns[XPOS]),
|
| 547 |
-
"UFeats": alignment_score(alignment, lambda w, _: w.columns[FEATS]),
|
| 548 |
-
"AllTags": alignment_score(alignment, lambda w, _: (w.columns[UPOS], w.columns[XPOS], w.columns[FEATS])),
|
| 549 |
-
"Lemmas": alignment_score(alignment, lambda w, ga: w.columns[LEMMA] if ga(w).columns[LEMMA] != "_" else "_"),
|
| 550 |
-
"UAS": alignment_score(alignment, lambda w, ga: ga(w.parent)),
|
| 551 |
-
"LAS": alignment_score(alignment, lambda w, ga: (ga(w.parent), w.columns[DEPREL])),
|
| 552 |
-
"CLAS": alignment_score(alignment, lambda w, ga: (ga(w.parent), w.columns[DEPREL]),
|
| 553 |
-
filter_fn=lambda w: w.is_content_deprel),
|
| 554 |
-
"MLAS": alignment_score(alignment, lambda w, ga: (ga(w.parent), w.columns[DEPREL], w.columns[UPOS], w.columns[FEATS],
|
| 555 |
-
[(ga(c), c.columns[DEPREL], c.columns[UPOS], c.columns[FEATS])
|
| 556 |
-
for c in w.functional_children]),
|
| 557 |
-
filter_fn=lambda w: w.is_content_deprel),
|
| 558 |
-
"BLEX": alignment_score(alignment, lambda w, ga: (ga(w.parent), w.columns[DEPREL],
|
| 559 |
-
w.columns[LEMMA] if ga(w).columns[LEMMA] != "_" else "_"),
|
| 560 |
-
filter_fn=lambda w: w.is_content_deprel),
|
| 561 |
-
}
|
| 562 |
-
|
| 563 |
-
|
| 564 |
-
def load_conllu_file(path, ignore_invalid_format=False):
|
| 565 |
-
_file = open(path, mode="r", **({"encoding": "utf-8"} if sys.version_info >= (3, 0) else {}))
|
| 566 |
-
return load_conllu(_file, ignore_invalid_format=ignore_invalid_format)
|
| 567 |
-
|
| 568 |
-
def evaluate_wrapper(args):
|
| 569 |
-
# Load CoNLL-U files
|
| 570 |
-
gold_ud = load_conllu_file(args.gold_file)
|
| 571 |
-
system_ud = load_conllu_file(args.system_file, ignore_invalid_format=args.ignore_invalid_format)
|
| 572 |
-
return evaluate(gold_ud, system_ud)
|
| 573 |
-
|
| 574 |
-
def main():
|
| 575 |
-
# Parse arguments
|
| 576 |
-
parser = argparse.ArgumentParser()
|
| 577 |
-
parser.add_argument("gold_file", type=str,
|
| 578 |
-
help="Name of the CoNLL-U file with the gold data.")
|
| 579 |
-
parser.add_argument("system_file", type=str,
|
| 580 |
-
help="Name of the CoNLL-U file with the predicted data.")
|
| 581 |
-
parser.add_argument("--verbose", "-v", default=False, action="store_true",
|
| 582 |
-
help="Print all metrics.")
|
| 583 |
-
parser.add_argument("--counts", "-c", default=False, action="store_true",
|
| 584 |
-
help="Print raw counts of correct/gold/system/aligned words instead of prec/rec/F1 for all metrics.")
|
| 585 |
-
parser.add_argument("--ignore-invalid-format", "-i", default=False, action="store_true",
|
| 586 |
-
help="Ignore invalid format errors, such as multiple roots, when loading CoNLL-U files.")
|
| 587 |
-
args = parser.parse_args()
|
| 588 |
-
|
| 589 |
-
# Evaluate
|
| 590 |
-
evaluation = evaluate_wrapper(args)
|
| 591 |
-
|
| 592 |
-
# Print the evaluation
|
| 593 |
-
if not args.verbose and not args.counts:
|
| 594 |
-
print("LAS F1 Score: {:.2f}".format(100 * evaluation["LAS"].f1))
|
| 595 |
-
print("MLAS Score: {:.2f}".format(100 * evaluation["MLAS"].f1))
|
| 596 |
-
print("BLEX Score: {:.2f}".format(100 * evaluation["BLEX"].f1))
|
| 597 |
-
else:
|
| 598 |
-
if args.counts:
|
| 599 |
-
print("Metric | Correct | Gold | Predicted | Aligned")
|
| 600 |
-
else:
|
| 601 |
-
print("Metric | Precision | Recall | F1 Score | AligndAcc")
|
| 602 |
-
print("-----------+-----------+-----------+-----------+-----------")
|
| 603 |
-
for metric in["Tokens", "Sentences", "Words", "UPOS", "XPOS", "UFeats", "AllTags", "Lemmas", "UAS", "LAS", "CLAS", "MLAS", "BLEX"]:
|
| 604 |
-
if args.counts:
|
| 605 |
-
print("{:11}|{:10} |{:10} |{:10} |{:10}".format(
|
| 606 |
-
metric,
|
| 607 |
-
evaluation[metric].correct,
|
| 608 |
-
evaluation[metric].gold_total,
|
| 609 |
-
evaluation[metric].system_total,
|
| 610 |
-
evaluation[metric].aligned_total or (evaluation[metric].correct if metric == "Words" else "")
|
| 611 |
-
))
|
| 612 |
-
else:
|
| 613 |
-
print("{:11}|{:10.2f} |{:10.2f} |{:10.2f} |{}".format(
|
| 614 |
-
metric,
|
| 615 |
-
100 * evaluation[metric].precision,
|
| 616 |
-
100 * evaluation[metric].recall,
|
| 617 |
-
100 * evaluation[metric].f1,
|
| 618 |
-
"{:10.2f}".format(100 * evaluation[metric].aligned_accuracy) if evaluation[metric].aligned_accuracy is not None else ""
|
| 619 |
-
))
|
| 620 |
-
|
| 621 |
-
if __name__ == "__main__":
|
| 622 |
-
main()
|
| 623 |
-
|
| 624 |
-
# Tests, which can be executed with `python -m unittest conll18_ud_eval`.
|
| 625 |
-
class TestAlignment(unittest.TestCase):
|
| 626 |
-
@staticmethod
|
| 627 |
-
def _load_words(words):
|
| 628 |
-
"""Prepare fake CoNLL-U files with fake HEAD to prevent multiple roots errors."""
|
| 629 |
-
lines, num_words = [], 0
|
| 630 |
-
for w in words:
|
| 631 |
-
parts = w.split(" ")
|
| 632 |
-
if len(parts) == 1:
|
| 633 |
-
num_words += 1
|
| 634 |
-
lines.append("{}\t{}\t_\t_\t_\t_\t{}\t_\t_\t_".format(num_words, parts[0], int(num_words>1)))
|
| 635 |
-
else:
|
| 636 |
-
lines.append("{}-{}\t{}\t_\t_\t_\t_\t_\t_\t_\t_".format(num_words + 1, num_words + len(parts) - 1, parts[0]))
|
| 637 |
-
for part in parts[1:]:
|
| 638 |
-
num_words += 1
|
| 639 |
-
lines.append("{}\t{}\t_\t_\t_\t_\t{}\t_\t_\t_".format(num_words, part, int(num_words>1)))
|
| 640 |
-
return load_conllu((io.StringIO if sys.version_info >= (3, 0) else io.BytesIO)("\n".join(lines+["\n"])))
|
| 641 |
-
|
| 642 |
-
def _test_exception(self, gold, system):
|
| 643 |
-
self.assertRaises(UDError, evaluate, self._load_words(gold), self._load_words(system))
|
| 644 |
-
|
| 645 |
-
def _test_ok(self, gold, system, correct):
|
| 646 |
-
metrics = evaluate(self._load_words(gold), self._load_words(system))
|
| 647 |
-
gold_words = sum((max(1, len(word.split(" ")) - 1) for word in gold))
|
| 648 |
-
system_words = sum((max(1, len(word.split(" ")) - 1) for word in system))
|
| 649 |
-
self.assertEqual((metrics["Words"].precision, metrics["Words"].recall, metrics["Words"].f1),
|
| 650 |
-
(correct / system_words, correct / gold_words, 2 * correct / (gold_words + system_words)))
|
| 651 |
-
|
| 652 |
-
def test_exception(self):
|
| 653 |
-
self._test_exception(["a"], ["b"])
|
| 654 |
-
|
| 655 |
-
def test_equal(self):
|
| 656 |
-
self._test_ok(["a"], ["a"], 1)
|
| 657 |
-
self._test_ok(["a", "b", "c"], ["a", "b", "c"], 3)
|
| 658 |
-
|
| 659 |
-
def test_equal_with_multiword(self):
|
| 660 |
-
self._test_ok(["abc a b c"], ["a", "b", "c"], 3)
|
| 661 |
-
self._test_ok(["a", "bc b c", "d"], ["a", "b", "c", "d"], 4)
|
| 662 |
-
self._test_ok(["abcd a b c d"], ["ab a b", "cd c d"], 4)
|
| 663 |
-
self._test_ok(["abc a b c", "de d e"], ["a", "bcd b c d", "e"], 5)
|
| 664 |
-
|
| 665 |
-
def test_alignment(self):
|
| 666 |
-
self._test_ok(["abcd"], ["a", "b", "c", "d"], 0)
|
| 667 |
-
self._test_ok(["abc", "d"], ["a", "b", "c", "d"], 1)
|
| 668 |
-
self._test_ok(["a", "bc", "d"], ["a", "b", "c", "d"], 2)
|
| 669 |
-
self._test_ok(["a", "bc b c", "d"], ["a", "b", "cd"], 2)
|
| 670 |
-
self._test_ok(["abc a BX c", "def d EX f"], ["ab a b", "cd c d", "ef e f"], 4)
|
| 671 |
-
self._test_ok(["ab a b", "cd bc d"], ["a", "bc", "d"], 2)
|
| 672 |
-
self._test_ok(["a", "bc b c", "d"], ["ab AX BX", "cd CX a"], 1)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
eval/scores/__init__.py
DELETED
|
File without changes
|
eval/scores/scores_talbanken.txt
DELETED
|
@@ -1,162 +0,0 @@
|
|
| 1 |
-
Sample: svediakorp-sec987-spf144, time period: 1800-1850
|
| 2 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 3 |
-
UPOS 1.0000 1.0000 1.0000 1.0000
|
| 4 |
-
UAS 1.0000 1.0000 1.0000 1.0000
|
| 5 |
-
LAS 0.9615 0.9615 0.9615 0.9615
|
| 6 |
-
|
| 7 |
-
Sample: svediakorp-sec613-EngstromA_StrindbergOchJag, time period: 1900-1950
|
| 8 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 9 |
-
UPOS 1.0000 1.0000 1.0000 1.0000
|
| 10 |
-
UAS 0.9146 0.9146 0.9146 0.9146
|
| 11 |
-
LAS 0.8659 0.8659 0.8659 0.8659
|
| 12 |
-
|
| 13 |
-
Sample: svediakorp-sec268-DulciU_VitterhetsNojen3, time period: 1750-1800
|
| 14 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 15 |
-
UPOS 0.9663 0.9663 0.9663 0.9663
|
| 16 |
-
UAS 0.8798 0.8798 0.8798 0.8798
|
| 17 |
-
LAS 0.8365 0.8365 0.8365 0.8365
|
| 18 |
-
|
| 19 |
-
Sample: svediakorp-sec25-Runius, time period: 1700-1750
|
| 20 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 21 |
-
UPOS 0.8750 0.8750 0.8750 0.8750
|
| 22 |
-
UAS 0.7188 0.7188 0.7188 0.7188
|
| 23 |
-
LAS 0.6562 0.6562 0.6562 0.6562
|
| 24 |
-
|
| 25 |
-
Sample: svediakorp-sec208-Anonym_DetGrasligaMordet, time period: 1850-1900
|
| 26 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 27 |
-
UPOS 0.9865 0.9865 0.9865 0.9865
|
| 28 |
-
UAS 0.8784 0.8784 0.8784 0.8784
|
| 29 |
-
LAS 0.8243 0.8243 0.8243 0.8243
|
| 30 |
-
|
| 31 |
-
Sample: svediakorp-sec324-GranbergPA_Enslighetsalskaren, time period: 1800-1850
|
| 32 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 33 |
-
UPOS 0.9400 0.9400 0.9400 0.9400
|
| 34 |
-
UAS 0.8400 0.8400 0.8400 0.8400
|
| 35 |
-
LAS 0.7400 0.7400 0.7400 0.7400
|
| 36 |
-
|
| 37 |
-
Sample: svediakorp-sec254-CederborghF_BerattelseOmJohnHall, time period: 1800-1850
|
| 38 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 39 |
-
UPOS 0.9706 0.9706 0.9706 0.9706
|
| 40 |
-
UAS 0.5412 0.5412 0.5412 0.5412
|
| 41 |
-
LAS 0.5000 0.5000 0.5000 0.5000
|
| 42 |
-
|
| 43 |
-
Sample: svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol, time period: 1900-1950
|
| 44 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 45 |
-
UPOS 0.9800 0.9800 0.9800 0.9800
|
| 46 |
-
UAS 0.7200 0.7200 0.7200 0.7200
|
| 47 |
-
LAS 0.7200 0.7200 0.7200 0.7200
|
| 48 |
-
|
| 49 |
-
Sample: svediakorp-sec452-NyblomH_FantasierFyra, time period: 1850-1900
|
| 50 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 51 |
-
UPOS 0.9464 0.9464 0.9464 0.9464
|
| 52 |
-
UAS 0.7500 0.7500 0.7500 0.7500
|
| 53 |
-
LAS 0.6786 0.6786 0.6786 0.6786
|
| 54 |
-
|
| 55 |
-
Sample: svediakorp-sec252-BremerF_Teckningar1, time period: 1800-1850
|
| 56 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 57 |
-
UPOS 1.0000 1.0000 1.0000 1.0000
|
| 58 |
-
UAS 0.8750 0.8750 0.8750 0.8750
|
| 59 |
-
LAS 0.8750 0.8750 0.8750 0.8750
|
| 60 |
-
|
| 61 |
-
Sample: svediakorp-sec486-SchwartzMS_BellmansSkor, time period: 1850-1900
|
| 62 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 63 |
-
UPOS 0.9873 0.9873 0.9873 0.9873
|
| 64 |
-
UAS 0.7722 0.7722 0.7722 0.7722
|
| 65 |
-
LAS 0.6962 0.6962 0.6962 0.6962
|
| 66 |
-
|
| 67 |
-
Sample: svediakorp-sec330-GyllenborgC_SwenskaSpratthoken, time period: 1700-1750
|
| 68 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 69 |
-
UPOS 0.8650 0.8650 0.8650 0.8650
|
| 70 |
-
UAS 0.6687 0.6687 0.6687 0.6687
|
| 71 |
-
LAS 0.5828 0.5828 0.5828 0.5828
|
| 72 |
-
|
| 73 |
-
Sample: svediakorp-sec631-HasselskogN_HallaHallaGronkoping, time period: 1900-1950
|
| 74 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 75 |
-
UPOS 0.9605 0.9605 0.9605 0.9605
|
| 76 |
-
UAS 0.7368 0.7368 0.7368 0.7368
|
| 77 |
-
LAS 0.6974 0.6974 0.6974 0.6974
|
| 78 |
-
|
| 79 |
-
Sample: svediakorp-sec277-EnbomPU_MedborgeligtSkalde, time period: 1750-1800
|
| 80 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 81 |
-
UPOS 0.9464 0.9464 0.9464 0.9464
|
| 82 |
-
UAS 0.7500 0.7500 0.7500 0.7500
|
| 83 |
-
LAS 0.6786 0.6786 0.6786 0.6786
|
| 84 |
-
|
| 85 |
-
Sample: svediakorp-sec1102-spf259, time period: 1850-1900
|
| 86 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 87 |
-
UPOS 0.9444 0.9444 0.9444 0.9444
|
| 88 |
-
UAS 0.8889 0.8889 0.8889 0.8889
|
| 89 |
-
LAS 0.8889 0.8889 0.8889 0.8889
|
| 90 |
-
|
| 91 |
-
Sample: svediakorp-sec991-spf148, time period: 1800-1850
|
| 92 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 93 |
-
UPOS 0.9561 0.9561 0.9561 0.9561
|
| 94 |
-
UAS 0.7982 0.7982 0.7982 0.7982
|
| 95 |
-
LAS 0.7368 0.7368 0.7368 0.7368
|
| 96 |
-
|
| 97 |
-
Sample: svediakorp-sec639-HeidenstamV_Proletarfilosofiens, time period: 1900-1950
|
| 98 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 99 |
-
UPOS 0.9861 0.9861 0.9861 0.9861
|
| 100 |
-
UAS 1.0000 1.0000 1.0000 1.0000
|
| 101 |
-
LAS 1.0000 1.0000 1.0000 1.0000
|
| 102 |
-
|
| 103 |
-
Sample: svediakorp-sec988-spf145, time period: 1800-1850
|
| 104 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 105 |
-
UPOS 0.9147 0.9147 0.9147 0.9147
|
| 106 |
-
UAS 0.8450 0.8450 0.8450 0.8450
|
| 107 |
-
LAS 0.7829 0.7829 0.7829 0.7829
|
| 108 |
-
|
| 109 |
-
Sample: svediakorp-sec1063-spf220, time period: 1850-1900
|
| 110 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 111 |
-
UPOS 0.9660 0.9660 0.9660 0.9660
|
| 112 |
-
UAS 0.9184 0.9184 0.9184 0.9184
|
| 113 |
-
LAS 0.8707 0.8707 0.8707 0.8707
|
| 114 |
-
|
| 115 |
-
Sample: svediakorp-sec1033-spf190, time period: 1900-1950
|
| 116 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 117 |
-
UPOS 0.9692 0.9692 0.9692 0.9692
|
| 118 |
-
UAS 0.8923 0.8923 0.8923 0.8923
|
| 119 |
-
LAS 0.8462 0.8462 0.8462 0.8462
|
| 120 |
-
|
| 121 |
-
Sample: svediakorp-letter141673-Stalhammar, time period: 1700-1750
|
| 122 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 123 |
-
UPOS 0.7059 0.7059 0.7059 0.7059
|
| 124 |
-
UAS 0.5294 0.5294 0.5294 0.5294
|
| 125 |
-
LAS 0.4118 0.4118 0.4118 0.4118
|
| 126 |
-
|
| 127 |
-
----------------------------------------------------------------------
|
| 128 |
-
Average scores for time period 1700-1750 (3 samples with a total of 8 sentences):
|
| 129 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 130 |
-
UPOS 0.8302 0.8302 0.8302 0.8302
|
| 131 |
-
UAS 0.6589 0.6589 0.6589 0.6589
|
| 132 |
-
LAS 0.5768 0.5768 0.5768 0.5768
|
| 133 |
-
|
| 134 |
-
Average scores for time period 1750-1800 (2 samples with a total of 5 sentences):
|
| 135 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 136 |
-
UPOS 0.9584 0.9584 0.9584 0.9584
|
| 137 |
-
UAS 0.8279 0.8279 0.8279 0.8279
|
| 138 |
-
LAS 0.7734 0.7734 0.7734 0.7734
|
| 139 |
-
|
| 140 |
-
Average scores for time period 1800-1850 (6 samples with a total of 19 sentences):
|
| 141 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 142 |
-
UPOS 0.9517 0.9517 0.9517 0.9517
|
| 143 |
-
UAS 0.7774 0.7774 0.7774 0.7774
|
| 144 |
-
LAS 0.7164 0.7164 0.7164 0.7164
|
| 145 |
-
|
| 146 |
-
Average scores for time period 1850-1900 (5 samples with a total of 14 sentences):
|
| 147 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 148 |
-
UPOS 0.9692 0.9692 0.9692 0.9692
|
| 149 |
-
UAS 0.8552 0.8552 0.8552 0.8552
|
| 150 |
-
LAS 0.8006 0.8006 0.8006 0.8006
|
| 151 |
-
|
| 152 |
-
Average scores for time period 1900-1950 (5 samples with a total of 13 sentences):
|
| 153 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 154 |
-
UPOS 0.9805 0.9805 0.9805 0.9805
|
| 155 |
-
UAS 0.8719 0.8719 0.8719 0.8719
|
| 156 |
-
LAS 0.8439 0.8439 0.8439 0.8439
|
| 157 |
-
|
| 158 |
-
Overall scores across all time periods (21 samples with a total of 59 sentences):
|
| 159 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 160 |
-
UPOS 0.9511 0.9511 0.9511 0.9511
|
| 161 |
-
UAS 0.7992 0.7992 0.7992 0.7992
|
| 162 |
-
LAS 0.7452 0.7452 0.7452 0.7452
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
eval/scores/scores_transformer_no_silver.txt
DELETED
|
@@ -1,162 +0,0 @@
|
|
| 1 |
-
Sample: svediakorp-sec987-spf144, time period: 1800-1850
|
| 2 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 3 |
-
UPOS 1.0000 1.0000 1.0000 1.0000
|
| 4 |
-
UAS 0.9615 0.9615 0.9615 0.9615
|
| 5 |
-
LAS 0.9231 0.9231 0.9231 0.9231
|
| 6 |
-
|
| 7 |
-
Sample: svediakorp-sec613-EngstromA_StrindbergOchJag, time period: 1900-1950
|
| 8 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 9 |
-
UPOS 1.0000 1.0000 1.0000 1.0000
|
| 10 |
-
UAS 0.9024 0.9024 0.9024 0.9024
|
| 11 |
-
LAS 0.8537 0.8537 0.8537 0.8537
|
| 12 |
-
|
| 13 |
-
Sample: svediakorp-sec268-DulciU_VitterhetsNojen3, time period: 1750-1800
|
| 14 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 15 |
-
UPOS 0.9663 0.9663 0.9663 0.9663
|
| 16 |
-
UAS 0.8317 0.8317 0.8317 0.8317
|
| 17 |
-
LAS 0.7837 0.7837 0.7837 0.7837
|
| 18 |
-
|
| 19 |
-
Sample: svediakorp-sec25-Runius, time period: 1700-1750
|
| 20 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 21 |
-
UPOS 0.8750 0.8750 0.8750 0.8750
|
| 22 |
-
UAS 0.6875 0.6875 0.6875 0.6875
|
| 23 |
-
LAS 0.5625 0.5625 0.5625 0.5625
|
| 24 |
-
|
| 25 |
-
Sample: svediakorp-sec208-Anonym_DetGrasligaMordet, time period: 1850-1900
|
| 26 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 27 |
-
UPOS 0.9865 0.9865 0.9865 0.9865
|
| 28 |
-
UAS 0.7973 0.7973 0.7973 0.7973
|
| 29 |
-
LAS 0.7432 0.7432 0.7432 0.7432
|
| 30 |
-
|
| 31 |
-
Sample: svediakorp-sec324-GranbergPA_Enslighetsalskaren, time period: 1800-1850
|
| 32 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 33 |
-
UPOS 0.9400 0.9400 0.9400 0.9400
|
| 34 |
-
UAS 0.8800 0.8800 0.8800 0.8800
|
| 35 |
-
LAS 0.8100 0.8100 0.8100 0.8100
|
| 36 |
-
|
| 37 |
-
Sample: svediakorp-sec254-CederborghF_BerattelseOmJohnHall, time period: 1800-1850
|
| 38 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 39 |
-
UPOS 0.9706 0.9706 0.9706 0.9706
|
| 40 |
-
UAS 0.5882 0.5882 0.5882 0.5882
|
| 41 |
-
LAS 0.5118 0.5118 0.5118 0.5118
|
| 42 |
-
|
| 43 |
-
Sample: svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol, time period: 1900-1950
|
| 44 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 45 |
-
UPOS 0.9800 0.9800 0.9800 0.9800
|
| 46 |
-
UAS 0.7800 0.7800 0.7800 0.7800
|
| 47 |
-
LAS 0.7800 0.7800 0.7800 0.7800
|
| 48 |
-
|
| 49 |
-
Sample: svediakorp-sec452-NyblomH_FantasierFyra, time period: 1850-1900
|
| 50 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 51 |
-
UPOS 0.9464 0.9464 0.9464 0.9464
|
| 52 |
-
UAS 0.8393 0.8393 0.8393 0.8393
|
| 53 |
-
LAS 0.8036 0.8036 0.8036 0.8036
|
| 54 |
-
|
| 55 |
-
Sample: svediakorp-sec252-BremerF_Teckningar1, time period: 1800-1850
|
| 56 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 57 |
-
UPOS 1.0000 1.0000 1.0000 1.0000
|
| 58 |
-
UAS 0.8750 0.8750 0.8750 0.8750
|
| 59 |
-
LAS 0.8750 0.8750 0.8750 0.8750
|
| 60 |
-
|
| 61 |
-
Sample: svediakorp-sec486-SchwartzMS_BellmansSkor, time period: 1850-1900
|
| 62 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 63 |
-
UPOS 0.9873 0.9873 0.9873 0.9873
|
| 64 |
-
UAS 0.7848 0.7848 0.7848 0.7848
|
| 65 |
-
LAS 0.6835 0.6835 0.6835 0.6835
|
| 66 |
-
|
| 67 |
-
Sample: svediakorp-sec330-GyllenborgC_SwenskaSpratthoken, time period: 1700-1750
|
| 68 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 69 |
-
UPOS 0.8650 0.8650 0.8650 0.8650
|
| 70 |
-
UAS 0.7301 0.7301 0.7301 0.7301
|
| 71 |
-
LAS 0.6135 0.6135 0.6135 0.6135
|
| 72 |
-
|
| 73 |
-
Sample: svediakorp-sec631-HasselskogN_HallaHallaGronkoping, time period: 1900-1950
|
| 74 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 75 |
-
UPOS 0.9605 0.9605 0.9605 0.9605
|
| 76 |
-
UAS 0.8684 0.8684 0.8684 0.8684
|
| 77 |
-
LAS 0.7632 0.7632 0.7632 0.7632
|
| 78 |
-
|
| 79 |
-
Sample: svediakorp-sec277-EnbomPU_MedborgeligtSkalde, time period: 1750-1800
|
| 80 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 81 |
-
UPOS 0.9464 0.9464 0.9464 0.9464
|
| 82 |
-
UAS 0.7857 0.7857 0.7857 0.7857
|
| 83 |
-
LAS 0.7500 0.7500 0.7500 0.7500
|
| 84 |
-
|
| 85 |
-
Sample: svediakorp-sec1102-spf259, time period: 1850-1900
|
| 86 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 87 |
-
UPOS 0.9444 0.9444 0.9444 0.9444
|
| 88 |
-
UAS 0.8889 0.8889 0.8889 0.8889
|
| 89 |
-
LAS 0.8889 0.8889 0.8889 0.8889
|
| 90 |
-
|
| 91 |
-
Sample: svediakorp-sec991-spf148, time period: 1800-1850
|
| 92 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 93 |
-
UPOS 0.9561 0.9561 0.9561 0.9561
|
| 94 |
-
UAS 0.7982 0.7982 0.7982 0.7982
|
| 95 |
-
LAS 0.7807 0.7807 0.7807 0.7807
|
| 96 |
-
|
| 97 |
-
Sample: svediakorp-sec639-HeidenstamV_Proletarfilosofiens, time period: 1900-1950
|
| 98 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 99 |
-
UPOS 0.9861 0.9861 0.9861 0.9861
|
| 100 |
-
UAS 0.9583 0.9583 0.9583 0.9583
|
| 101 |
-
LAS 0.9444 0.9444 0.9444 0.9444
|
| 102 |
-
|
| 103 |
-
Sample: svediakorp-sec988-spf145, time period: 1800-1850
|
| 104 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 105 |
-
UPOS 0.9147 0.9147 0.9147 0.9147
|
| 106 |
-
UAS 0.8682 0.8682 0.8682 0.8682
|
| 107 |
-
LAS 0.7984 0.7984 0.7984 0.7984
|
| 108 |
-
|
| 109 |
-
Sample: svediakorp-sec1063-spf220, time period: 1850-1900
|
| 110 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 111 |
-
UPOS 0.9660 0.9660 0.9660 0.9660
|
| 112 |
-
UAS 0.9116 0.9116 0.9116 0.9116
|
| 113 |
-
LAS 0.8503 0.8503 0.8503 0.8503
|
| 114 |
-
|
| 115 |
-
Sample: svediakorp-sec1033-spf190, time period: 1900-1950
|
| 116 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 117 |
-
UPOS 0.9692 0.9692 0.9692 0.9692
|
| 118 |
-
UAS 0.9077 0.9077 0.9077 0.9077
|
| 119 |
-
LAS 0.8154 0.8154 0.8154 0.8154
|
| 120 |
-
|
| 121 |
-
Sample: svediakorp-letter141673-Stalhammar, time period: 1700-1750
|
| 122 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 123 |
-
UPOS 0.7059 0.7059 0.7059 0.7059
|
| 124 |
-
UAS 0.6471 0.6471 0.6471 0.6471
|
| 125 |
-
LAS 0.4706 0.4706 0.4706 0.4706
|
| 126 |
-
|
| 127 |
-
----------------------------------------------------------------------
|
| 128 |
-
Average scores for time period 1700-1750 (3 samples with a total of 8 sentences):
|
| 129 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 130 |
-
UPOS 0.8302 0.8302 0.8302 0.8302
|
| 131 |
-
UAS 0.6880 0.6880 0.6880 0.6880
|
| 132 |
-
LAS 0.5523 0.5523 0.5523 0.5523
|
| 133 |
-
|
| 134 |
-
Average scores for time period 1750-1800 (2 samples with a total of 5 sentences):
|
| 135 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 136 |
-
UPOS 0.9584 0.9584 0.9584 0.9584
|
| 137 |
-
UAS 0.8133 0.8133 0.8133 0.8133
|
| 138 |
-
LAS 0.7702 0.7702 0.7702 0.7702
|
| 139 |
-
|
| 140 |
-
Average scores for time period 1800-1850 (6 samples with a total of 19 sentences):
|
| 141 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 142 |
-
UPOS 0.9517 0.9517 0.9517 0.9517
|
| 143 |
-
UAS 0.7986 0.7986 0.7986 0.7986
|
| 144 |
-
LAS 0.7464 0.7464 0.7464 0.7464
|
| 145 |
-
|
| 146 |
-
Average scores for time period 1850-1900 (5 samples with a total of 14 sentences):
|
| 147 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 148 |
-
UPOS 0.9692 0.9692 0.9692 0.9692
|
| 149 |
-
UAS 0.8561 0.8561 0.8561 0.8561
|
| 150 |
-
LAS 0.7954 0.7954 0.7954 0.7954
|
| 151 |
-
|
| 152 |
-
Average scores for time period 1900-1950 (5 samples with a total of 13 sentences):
|
| 153 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 154 |
-
UPOS 0.9805 0.9805 0.9805 0.9805
|
| 155 |
-
UAS 0.8925 0.8925 0.8925 0.8925
|
| 156 |
-
LAS 0.8405 0.8405 0.8405 0.8405
|
| 157 |
-
|
| 158 |
-
Overall scores across all time periods (21 samples with a total of 59 sentences):
|
| 159 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 160 |
-
UPOS 0.9511 0.9511 0.9511 0.9511
|
| 161 |
-
UAS 0.8140 0.8140 0.8140 0.8140
|
| 162 |
-
LAS 0.7503 0.7503 0.7503 0.7503
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
eval/scores/scores_transformer_silver.txt
DELETED
|
@@ -1,162 +0,0 @@
|
|
| 1 |
-
Sample: svediakorp-sec987-spf144, time period: 1800-1850
|
| 2 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 3 |
-
UPOS 1.0000 1.0000 1.0000 1.0000
|
| 4 |
-
UAS 0.9615 0.9615 0.9615 0.9615
|
| 5 |
-
LAS 0.8846 0.8846 0.8846 0.8846
|
| 6 |
-
|
| 7 |
-
Sample: svediakorp-sec613-EngstromA_StrindbergOchJag, time period: 1900-1950
|
| 8 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 9 |
-
UPOS 1.0000 1.0000 1.0000 1.0000
|
| 10 |
-
UAS 0.8780 0.8780 0.8780 0.8780
|
| 11 |
-
LAS 0.8171 0.8171 0.8171 0.8171
|
| 12 |
-
|
| 13 |
-
Sample: svediakorp-sec268-DulciU_VitterhetsNojen3, time period: 1750-1800
|
| 14 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 15 |
-
UPOS 0.9663 0.9663 0.9663 0.9663
|
| 16 |
-
UAS 0.8221 0.8221 0.8221 0.8221
|
| 17 |
-
LAS 0.7692 0.7692 0.7692 0.7692
|
| 18 |
-
|
| 19 |
-
Sample: svediakorp-sec25-Runius, time period: 1700-1750
|
| 20 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 21 |
-
UPOS 0.8750 0.8750 0.8750 0.8750
|
| 22 |
-
UAS 0.6562 0.6562 0.6562 0.6562
|
| 23 |
-
LAS 0.5625 0.5625 0.5625 0.5625
|
| 24 |
-
|
| 25 |
-
Sample: svediakorp-sec208-Anonym_DetGrasligaMordet, time period: 1850-1900
|
| 26 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 27 |
-
UPOS 0.9865 0.9865 0.9865 0.9865
|
| 28 |
-
UAS 0.7568 0.7568 0.7568 0.7568
|
| 29 |
-
LAS 0.7027 0.7027 0.7027 0.7027
|
| 30 |
-
|
| 31 |
-
Sample: svediakorp-sec324-GranbergPA_Enslighetsalskaren, time period: 1800-1850
|
| 32 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 33 |
-
UPOS 0.9400 0.9400 0.9400 0.9400
|
| 34 |
-
UAS 0.8500 0.8500 0.8500 0.8500
|
| 35 |
-
LAS 0.7500 0.7500 0.7500 0.7500
|
| 36 |
-
|
| 37 |
-
Sample: svediakorp-sec254-CederborghF_BerattelseOmJohnHall, time period: 1800-1850
|
| 38 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 39 |
-
UPOS 0.9706 0.9706 0.9706 0.9706
|
| 40 |
-
UAS 0.5765 0.5765 0.5765 0.5765
|
| 41 |
-
LAS 0.5000 0.5000 0.5000 0.5000
|
| 42 |
-
|
| 43 |
-
Sample: svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol, time period: 1900-1950
|
| 44 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 45 |
-
UPOS 0.9800 0.9800 0.9800 0.9800
|
| 46 |
-
UAS 0.8000 0.8000 0.8000 0.8000
|
| 47 |
-
LAS 0.7600 0.7600 0.7600 0.7600
|
| 48 |
-
|
| 49 |
-
Sample: svediakorp-sec452-NyblomH_FantasierFyra, time period: 1850-1900
|
| 50 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 51 |
-
UPOS 0.9464 0.9464 0.9464 0.9464
|
| 52 |
-
UAS 0.8571 0.8571 0.8571 0.8571
|
| 53 |
-
LAS 0.7679 0.7679 0.7679 0.7679
|
| 54 |
-
|
| 55 |
-
Sample: svediakorp-sec252-BremerF_Teckningar1, time period: 1800-1850
|
| 56 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 57 |
-
UPOS 1.0000 1.0000 1.0000 1.0000
|
| 58 |
-
UAS 0.8333 0.8333 0.8333 0.8333
|
| 59 |
-
LAS 0.7917 0.7917 0.7917 0.7917
|
| 60 |
-
|
| 61 |
-
Sample: svediakorp-sec486-SchwartzMS_BellmansSkor, time period: 1850-1900
|
| 62 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 63 |
-
UPOS 0.9873 0.9873 0.9873 0.9873
|
| 64 |
-
UAS 0.7722 0.7722 0.7722 0.7722
|
| 65 |
-
LAS 0.6582 0.6582 0.6582 0.6582
|
| 66 |
-
|
| 67 |
-
Sample: svediakorp-sec330-GyllenborgC_SwenskaSpratthoken, time period: 1700-1750
|
| 68 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 69 |
-
UPOS 0.8650 0.8650 0.8650 0.8650
|
| 70 |
-
UAS 0.7178 0.7178 0.7178 0.7178
|
| 71 |
-
LAS 0.6074 0.6074 0.6074 0.6074
|
| 72 |
-
|
| 73 |
-
Sample: svediakorp-sec631-HasselskogN_HallaHallaGronkoping, time period: 1900-1950
|
| 74 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 75 |
-
UPOS 0.9605 0.9605 0.9605 0.9605
|
| 76 |
-
UAS 0.8026 0.8026 0.8026 0.8026
|
| 77 |
-
LAS 0.7105 0.7105 0.7105 0.7105
|
| 78 |
-
|
| 79 |
-
Sample: svediakorp-sec277-EnbomPU_MedborgeligtSkalde, time period: 1750-1800
|
| 80 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 81 |
-
UPOS 0.9464 0.9464 0.9464 0.9464
|
| 82 |
-
UAS 0.8214 0.8214 0.8214 0.8214
|
| 83 |
-
LAS 0.7500 0.7500 0.7500 0.7500
|
| 84 |
-
|
| 85 |
-
Sample: svediakorp-sec1102-spf259, time period: 1850-1900
|
| 86 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 87 |
-
UPOS 0.9444 0.9444 0.9444 0.9444
|
| 88 |
-
UAS 0.8889 0.8889 0.8889 0.8889
|
| 89 |
-
LAS 0.8889 0.8889 0.8889 0.8889
|
| 90 |
-
|
| 91 |
-
Sample: svediakorp-sec991-spf148, time period: 1800-1850
|
| 92 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 93 |
-
UPOS 0.9561 0.9561 0.9561 0.9561
|
| 94 |
-
UAS 0.7982 0.7982 0.7982 0.7982
|
| 95 |
-
LAS 0.7544 0.7544 0.7544 0.7544
|
| 96 |
-
|
| 97 |
-
Sample: svediakorp-sec639-HeidenstamV_Proletarfilosofiens, time period: 1900-1950
|
| 98 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 99 |
-
UPOS 0.9861 0.9861 0.9861 0.9861
|
| 100 |
-
UAS 0.9306 0.9306 0.9306 0.9306
|
| 101 |
-
LAS 0.8889 0.8889 0.8889 0.8889
|
| 102 |
-
|
| 103 |
-
Sample: svediakorp-sec988-spf145, time period: 1800-1850
|
| 104 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 105 |
-
UPOS 0.9147 0.9147 0.9147 0.9147
|
| 106 |
-
UAS 0.7442 0.7442 0.7442 0.7442
|
| 107 |
-
LAS 0.6822 0.6822 0.6822 0.6822
|
| 108 |
-
|
| 109 |
-
Sample: svediakorp-sec1063-spf220, time period: 1850-1900
|
| 110 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 111 |
-
UPOS 0.9660 0.9660 0.9660 0.9660
|
| 112 |
-
UAS 0.8435 0.8435 0.8435 0.8435
|
| 113 |
-
LAS 0.7551 0.7551 0.7551 0.7551
|
| 114 |
-
|
| 115 |
-
Sample: svediakorp-sec1033-spf190, time period: 1900-1950
|
| 116 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 117 |
-
UPOS 0.9692 0.9692 0.9692 0.9692
|
| 118 |
-
UAS 0.9077 0.9077 0.9077 0.9077
|
| 119 |
-
LAS 0.8000 0.8000 0.8000 0.8000
|
| 120 |
-
|
| 121 |
-
Sample: svediakorp-letter141673-Stalhammar, time period: 1700-1750
|
| 122 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 123 |
-
UPOS 0.7059 0.7059 0.7059 0.7059
|
| 124 |
-
UAS 0.6471 0.6471 0.6471 0.6471
|
| 125 |
-
LAS 0.4706 0.4706 0.4706 0.4706
|
| 126 |
-
|
| 127 |
-
----------------------------------------------------------------------
|
| 128 |
-
Average scores for time period 1700-1750 (3 samples with a total of 8 sentences):
|
| 129 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 130 |
-
UPOS 0.8302 0.8302 0.8302 0.8302
|
| 131 |
-
UAS 0.6693 0.6693 0.6693 0.6693
|
| 132 |
-
LAS 0.5507 0.5507 0.5507 0.5507
|
| 133 |
-
|
| 134 |
-
Average scores for time period 1750-1800 (2 samples with a total of 5 sentences):
|
| 135 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 136 |
-
UPOS 0.9584 0.9584 0.9584 0.9584
|
| 137 |
-
UAS 0.8218 0.8218 0.8218 0.8218
|
| 138 |
-
LAS 0.7615 0.7615 0.7615 0.7615
|
| 139 |
-
|
| 140 |
-
Average scores for time period 1800-1850 (6 samples with a total of 19 sentences):
|
| 141 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 142 |
-
UPOS 0.9517 0.9517 0.9517 0.9517
|
| 143 |
-
UAS 0.7615 0.7615 0.7615 0.7615
|
| 144 |
-
LAS 0.6935 0.6935 0.6935 0.6935
|
| 145 |
-
|
| 146 |
-
Average scores for time period 1850-1900 (5 samples with a total of 14 sentences):
|
| 147 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 148 |
-
UPOS 0.9692 0.9692 0.9692 0.9692
|
| 149 |
-
UAS 0.8210 0.8210 0.8210 0.8210
|
| 150 |
-
LAS 0.7382 0.7382 0.7382 0.7382
|
| 151 |
-
|
| 152 |
-
Average scores for time period 1900-1950 (5 samples with a total of 13 sentences):
|
| 153 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 154 |
-
UPOS 0.9805 0.9805 0.9805 0.9805
|
| 155 |
-
UAS 0.8734 0.8734 0.8734 0.8734
|
| 156 |
-
LAS 0.8045 0.8045 0.8045 0.8045
|
| 157 |
-
|
| 158 |
-
Overall scores across all time periods (21 samples with a total of 59 sentences):
|
| 159 |
-
Metric Precision Recall F1 Aligned Accuracy
|
| 160 |
-
UPOS 0.9511 0.9511 0.9511 0.9511
|
| 161 |
-
UAS 0.7878 0.7878 0.7878 0.7878
|
| 162 |
-
LAS 0.7122 0.7122 0.7122 0.7122
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/latest.txt
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
log_bert-base-swedish-cased_sv_20260319_214621.txt
|
|
|
|
|
|
logs/log_bert-base-swedish-cased_sv_20260219_135555.txt
DELETED
|
@@ -1,114 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_bert-base-swedish-cased_sv_20260219_135555.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: KBLab/bert-base-swedish-cased
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-SILVER.py sv
|
| 6 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 7 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 8 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 9 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 10 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 11 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 12 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 13 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 14 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 15 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 16 |
-
Loading DigPhil MACHINE for SILVER (minus gold overlap)…
|
| 17 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
|
| 18 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 19 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 20 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
|
| 21 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 22 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 23 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 24 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 25 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 26 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 27 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 28 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 29 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 30 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 31 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
|
| 32 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
|
| 33 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 34 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
|
| 35 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
|
| 36 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
|
| 37 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
|
| 38 |
-
Adding 79 remaining gold sentences to TRAIN
|
| 39 |
-
Cleaning TRAIN...
|
| 40 |
-
Cleaning SILVER...
|
| 41 |
-
[REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
|
| 42 |
-
[REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
|
| 43 |
-
[REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
|
| 44 |
-
[REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
|
| 45 |
-
[REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
|
| 46 |
-
[REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
|
| 47 |
-
[REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
|
| 48 |
-
[REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
|
| 49 |
-
[REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
|
| 50 |
-
[REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
|
| 51 |
-
[REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
|
| 52 |
-
[REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
|
| 53 |
-
[REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
|
| 54 |
-
[REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
|
| 55 |
-
[REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
|
| 56 |
-
[REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
|
| 57 |
-
[REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
|
| 58 |
-
[REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
|
| 59 |
-
[REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
|
| 60 |
-
[REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
|
| 61 |
-
[REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
|
| 62 |
-
[REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
|
| 63 |
-
[REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
|
| 64 |
-
[REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
|
| 65 |
-
[REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
|
| 66 |
-
[REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
|
| 67 |
-
[REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
|
| 68 |
-
[REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
|
| 69 |
-
[REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
|
| 70 |
-
[REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
|
| 71 |
-
[REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
|
| 72 |
-
[REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
|
| 73 |
-
[REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
|
| 74 |
-
[REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
|
| 75 |
-
[REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
|
| 76 |
-
[REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
|
| 77 |
-
[REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
|
| 78 |
-
[REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
|
| 79 |
-
[REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
|
| 80 |
-
[REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
|
| 81 |
-
[REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
|
| 82 |
-
[REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
|
| 83 |
-
[REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
|
| 84 |
-
[REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
|
| 85 |
-
[REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
|
| 86 |
-
[REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
|
| 87 |
-
[REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
|
| 88 |
-
[REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
|
| 89 |
-
[REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
|
| 90 |
-
[REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
|
| 91 |
-
[REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
|
| 92 |
-
[REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
|
| 93 |
-
[REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
|
| 94 |
-
[REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
|
| 95 |
-
[REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
|
| 96 |
-
[REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
|
| 97 |
-
[REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
|
| 98 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
|
| 99 |
-
[REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
|
| 100 |
-
[REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
|
| 101 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
|
| 102 |
-
[REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
|
| 103 |
-
[REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
|
| 104 |
-
Cleaning DEV...
|
| 105 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 106 |
-
Cleaning TEST...
|
| 107 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 108 |
-
Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13873 valid sentences)
|
| 109 |
-
Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (46432 valid sentences)
|
| 110 |
-
Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (29 valid sentences)
|
| 111 |
-
Writing TEST → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (all DIGPHIL_GOLD, 108 valid sentences)
|
| 112 |
-
Done.
|
| 113 |
-
Sourcing scripts/config.sh
|
| 114 |
-
Running stanza dataset preparation…
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_bert-base-swedish-cased_sv_20260219_135610.txt
DELETED
|
@@ -1,138 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_bert-base-swedish-cased_sv_20260219_135610.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: KBLab/bert-base-swedish-cased
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-SILVER.py sv
|
| 6 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 7 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 8 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 9 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 10 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 11 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 12 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 13 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 14 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 15 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 16 |
-
Loading DigPhil MACHINE for SILVER (minus gold overlap)…
|
| 17 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
|
| 18 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 19 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 20 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
|
| 21 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 22 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 23 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 24 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 25 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 26 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 27 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 28 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 29 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 30 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 31 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
|
| 32 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
|
| 33 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 34 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
|
| 35 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
|
| 36 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
|
| 37 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
|
| 38 |
-
Adding 75 remaining gold sentences to TRAIN
|
| 39 |
-
Cleaning TRAIN...
|
| 40 |
-
Cleaning SILVER...
|
| 41 |
-
[REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
|
| 42 |
-
[REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
|
| 43 |
-
[REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
|
| 44 |
-
[REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
|
| 45 |
-
[REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
|
| 46 |
-
[REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
|
| 47 |
-
[REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
|
| 48 |
-
[REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
|
| 49 |
-
[REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
|
| 50 |
-
[REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
|
| 51 |
-
[REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
|
| 52 |
-
[REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
|
| 53 |
-
[REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
|
| 54 |
-
[REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
|
| 55 |
-
[REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
|
| 56 |
-
[REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
|
| 57 |
-
[REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
|
| 58 |
-
[REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
|
| 59 |
-
[REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
|
| 60 |
-
[REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
|
| 61 |
-
[REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
|
| 62 |
-
[REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
|
| 63 |
-
[REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
|
| 64 |
-
[REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
|
| 65 |
-
[REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
|
| 66 |
-
[REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
|
| 67 |
-
[REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
|
| 68 |
-
[REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
|
| 69 |
-
[REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
|
| 70 |
-
[REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
|
| 71 |
-
[REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
|
| 72 |
-
[REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
|
| 73 |
-
[REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
|
| 74 |
-
[REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
|
| 75 |
-
[REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
|
| 76 |
-
[REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
|
| 77 |
-
[REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
|
| 78 |
-
[REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
|
| 79 |
-
[REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
|
| 80 |
-
[REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
|
| 81 |
-
[REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
|
| 82 |
-
[REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
|
| 83 |
-
[REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
|
| 84 |
-
[REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
|
| 85 |
-
[REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
|
| 86 |
-
[REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
|
| 87 |
-
[REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
|
| 88 |
-
[REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
|
| 89 |
-
[REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
|
| 90 |
-
[REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
|
| 91 |
-
[REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
|
| 92 |
-
[REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
|
| 93 |
-
[REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
|
| 94 |
-
[REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
|
| 95 |
-
[REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
|
| 96 |
-
[REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
|
| 97 |
-
[REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
|
| 98 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
|
| 99 |
-
[REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
|
| 100 |
-
[REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
|
| 101 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
|
| 102 |
-
[REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
|
| 103 |
-
[REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
|
| 104 |
-
Cleaning DEV...
|
| 105 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 106 |
-
Cleaning TEST...
|
| 107 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 108 |
-
Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13869 valid sentences)
|
| 109 |
-
Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (46432 valid sentences)
|
| 110 |
-
Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (33 valid sentences)
|
| 111 |
-
Writing TEST → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (all DIGPHIL_GOLD, 108 valid sentences)
|
| 112 |
-
Done.
|
| 113 |
-
Sourcing scripts/config.sh
|
| 114 |
-
Running stanza dataset preparation…
|
| 115 |
-
2026-02-19 13:56:13 INFO: Datasets program called with:
|
| 116 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --gold
|
| 117 |
-
Augmented 60 quotes: Counter({'《》': 9, '»«': 8, '″″': 8, '““': 7, '„”': 6, '""': 6, '„“': 6, '«»': 4, '””': 4, '「」': 2})
|
| 118 |
-
Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
|
| 119 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /tmp/tmpzh8h6jmq/sv_diachronic.train.gold.conllu
|
| 120 |
-
Swapped 'w1, w2' for 'w1 ,w2' 94 times
|
| 121 |
-
Added 85 new sentences with asdf, zzzz -> asdf,zzzz
|
| 122 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /tmp/tmpzh8h6jmq/sv_diachronic.dev.gold.conllu
|
| 123 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /tmp/tmpzh8h6jmq/sv_diachronic.test.gold.conllu
|
| 124 |
-
Copying from /tmp/tmpzh8h6jmq/sv_diachronic.train.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 125 |
-
Copying from /tmp/tmpzh8h6jmq/sv_diachronic.dev.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 126 |
-
Copying from /tmp/tmpzh8h6jmq/sv_diachronic.test.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
|
| 127 |
-
Running stanza dependency parser training…
|
| 128 |
-
2026-02-19 13:56:16 INFO: Training program called with:
|
| 129 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --batch_size 32 --dropout 0.33 --use_bert --bert_model KBLab/bert-base-swedish-cased --silver_file ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu
|
| 130 |
-
2026-02-19 13:56:16 DEBUG: UD_Swedish-diachronic: sv_diachronic
|
| 131 |
-
2026-02-19 13:56:16 INFO: Using model /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 132 |
-
2026-02-19 13:56:16 INFO: Using model /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 133 |
-
2026-02-19 13:56:16 INFO: Using default pretrain for language sv, found in /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt To use a different pretrain, specify --wordvec_pretrain_file
|
| 134 |
-
2026-02-19 13:56:16 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_transformer_parser.pt exists, skipping!
|
| 135 |
-
DONE.
|
| 136 |
-
Full log saved to: logs/log_bert-base-swedish-cased_sv_20260219_135610.txt
|
| 137 |
-
Symlink updated: logs/latest.txt → log_bert-base-swedish-cased_sv_20260219_135610.txt
|
| 138 |
-
Latest log: log_bert-base-swedish-cased_sv_20260219_135610.txt
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_bert-base-swedish-cased_sv_20260219_135635.txt
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
logs/log_bert-base-swedish-cased_sv_20260318_131623.txt
DELETED
|
@@ -1,271 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_bert-base-swedish-cased_sv_20260318_131623.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: KBLab/bert-base-swedish-cased
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-SILVER.py sv
|
| 6 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 7 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 8 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 9 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 10 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 11 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 12 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 13 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 14 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 15 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 16 |
-
Loading DigPhil MACHINE for SILVER (minus gold overlap)…
|
| 17 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec987-spf144.conllu
|
| 18 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 19 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 20 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec25-Runius.conllu
|
| 21 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 22 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 23 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 24 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 25 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 26 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 27 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 28 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 29 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 30 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 31 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1102-spf259.conllu
|
| 32 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec991-spf148.conllu
|
| 33 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 34 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec988-spf145.conllu
|
| 35 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1063-spf220.conllu
|
| 36 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1033-spf190.conllu
|
| 37 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-letter141673-Stalhammar.conllu
|
| 38 |
-
Adding 76 remaining gold sentences to TRAIN
|
| 39 |
-
Cleaning TRAIN...
|
| 40 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 41 |
-
Cleaning SILVER...
|
| 42 |
-
[REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
|
| 43 |
-
[REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
|
| 44 |
-
[REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
|
| 45 |
-
[REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
|
| 46 |
-
[REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
|
| 47 |
-
[REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
|
| 48 |
-
[REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
|
| 49 |
-
[REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
|
| 50 |
-
[REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
|
| 51 |
-
[REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
|
| 52 |
-
[REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
|
| 53 |
-
[REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
|
| 54 |
-
[REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
|
| 55 |
-
[REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
|
| 56 |
-
[REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
|
| 57 |
-
[REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
|
| 58 |
-
[REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
|
| 59 |
-
[REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
|
| 60 |
-
[REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
|
| 61 |
-
[REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
|
| 62 |
-
[REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
|
| 63 |
-
[REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
|
| 64 |
-
[REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
|
| 65 |
-
[REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
|
| 66 |
-
[REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
|
| 67 |
-
[REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
|
| 68 |
-
[REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
|
| 69 |
-
[REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
|
| 70 |
-
[REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
|
| 71 |
-
[REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
|
| 72 |
-
[REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
|
| 73 |
-
[REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
|
| 74 |
-
[REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
|
| 75 |
-
[REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
|
| 76 |
-
[REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
|
| 77 |
-
[REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
|
| 78 |
-
[REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
|
| 79 |
-
[REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
|
| 80 |
-
[REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
|
| 81 |
-
[REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
|
| 82 |
-
[REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
|
| 83 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
|
| 84 |
-
[REMOVED] sent_id=1266 ERRORS=['Token 17: Missing form']
|
| 85 |
-
[REMOVED] sent_id=4897 ERRORS=['Token 46: Missing form', 'Token 47: Missing form']
|
| 86 |
-
Cleaning DEV...
|
| 87 |
-
Cleaning TEST...
|
| 88 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 89 |
-
Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13869 valid sentences)
|
| 90 |
-
Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (51431 valid sentences)
|
| 91 |
-
Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (33 valid sentences)
|
| 92 |
-
Writing TEST → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (all DIGPHIL_GOLD, 108 valid sentences)
|
| 93 |
-
Done.
|
| 94 |
-
Sourcing scripts/config.sh
|
| 95 |
-
Running stanza dataset preparation…
|
| 96 |
-
2026-03-18 13:16:26 INFO: Datasets program called with:
|
| 97 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --gold
|
| 98 |
-
Augmented 60 quotes: Counter({'《》': 9, '»«': 8, '″″': 8, '““': 7, '„”': 6, '""': 6, '„“': 6, '«»': 4, '””': 4, '「」': 2})
|
| 99 |
-
Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
|
| 100 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /tmp/tmpz3gm6g6g/sv_diachronic.train.gold.conllu
|
| 101 |
-
Swapped 'w1, w2' for 'w1 ,w2' 94 times
|
| 102 |
-
Added 85 new sentences with asdf, zzzz -> asdf,zzzz
|
| 103 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /tmp/tmpz3gm6g6g/sv_diachronic.dev.gold.conllu
|
| 104 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /tmp/tmpz3gm6g6g/sv_diachronic.test.gold.conllu
|
| 105 |
-
Copying from /tmp/tmpz3gm6g6g/sv_diachronic.train.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 106 |
-
Copying from /tmp/tmpz3gm6g6g/sv_diachronic.dev.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 107 |
-
Copying from /tmp/tmpz3gm6g6g/sv_diachronic.test.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
|
| 108 |
-
Running stanza dependency parser training…
|
| 109 |
-
2026-03-18 13:16:28 INFO: Training program called with:
|
| 110 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --batch_size 32 --dropout 0.33 --use_bert --bert_model KBLab/bert-base-swedish-cased --silver_file ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu
|
| 111 |
-
2026-03-18 13:16:28 DEBUG: UD_Swedish-diachronic: sv_diachronic
|
| 112 |
-
2026-03-18 13:16:28 INFO: Using model /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 113 |
-
2026-03-18 13:16:28 INFO: Using model /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 114 |
-
2026-03-18 13:16:28 INFO: Using default pretrain for language sv, found in /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt To use a different pretrain, specify --wordvec_pretrain_file
|
| 115 |
-
2026-03-18 13:16:28 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_transformer_parser.pt does not exist, training new model
|
| 116 |
-
2026-03-18 13:16:28 INFO: Using model /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 117 |
-
2026-03-18 13:16:28 INFO: Using model /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 118 |
-
2026-03-18 13:16:28 INFO: Using default pretrain for language sv, found in /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt To use a different pretrain, specify --wordvec_pretrain_file
|
| 119 |
-
2026-03-18 13:16:28 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/home/urdatorn/stanza_resources/sv/pretrain', '--train_file', '/home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--wordvec_pretrain_file', '/home/urdatorn/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt', '--batch_size', '32', '--dropout', '0.33', '--bert_model', 'KBLab/bert-base-swedish-cased', '--silver_file', 'ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu']
|
| 120 |
-
2026-03-18 13:16:28 INFO: Running parser in train mode
|
| 121 |
-
2026-03-18 13:16:29 DEBUG: Loaded pretrain from /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
|
| 122 |
-
2026-03-18 13:16:29 INFO: Using 7 as the word cutoff based on the size of the pretrain (250000)
|
| 123 |
-
2026-03-18 13:16:29 INFO: Using pretrained contextualized char embedding
|
| 124 |
-
2026-03-18 13:16:29 INFO: ARGS USED AT TRAINING TIME:
|
| 125 |
-
augment_nopunct: None
|
| 126 |
-
batch_size: 32
|
| 127 |
-
bert_finetune: False
|
| 128 |
-
bert_finetune_layers: None
|
| 129 |
-
bert_hidden_layers: 4
|
| 130 |
-
bert_learning_rate: 1.0
|
| 131 |
-
bert_model: KBLab/bert-base-swedish-cased
|
| 132 |
-
bert_start_finetuning: 200
|
| 133 |
-
bert_warmup_steps: 200
|
| 134 |
-
bert_weight_decay: 0.0
|
| 135 |
-
beta2: 0.999
|
| 136 |
-
char: True
|
| 137 |
-
char_emb_dim: 100
|
| 138 |
-
char_hidden_dim: 400
|
| 139 |
-
char_num_layers: 1
|
| 140 |
-
char_rec_dropout: 0
|
| 141 |
-
charlm: True
|
| 142 |
-
charlm_backward_file: /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
|
| 143 |
-
charlm_forward_file: /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt
|
| 144 |
-
charlm_save_dir: saved_models/charlm
|
| 145 |
-
charlm_shorthand: sv_conll17
|
| 146 |
-
checkpoint: True
|
| 147 |
-
checkpoint_interval: 500
|
| 148 |
-
checkpoint_save_name: None
|
| 149 |
-
continue_from: None
|
| 150 |
-
data_dir: data/depparse
|
| 151 |
-
deep_biaff_hidden_dim: 400
|
| 152 |
-
deep_biaff_output_dim: 160
|
| 153 |
-
device: cpu
|
| 154 |
-
distance: True
|
| 155 |
-
dropout: 0.33
|
| 156 |
-
eval_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 157 |
-
eval_interval: 100
|
| 158 |
-
gold_labels: True
|
| 159 |
-
hidden_dim: 400
|
| 160 |
-
lang: sv
|
| 161 |
-
linearization: True
|
| 162 |
-
log_norms: False
|
| 163 |
-
log_step: 20
|
| 164 |
-
lora_alpha: 128
|
| 165 |
-
lora_dropout: 0.1
|
| 166 |
-
lora_modules_to_save: []
|
| 167 |
-
lora_rank: 64
|
| 168 |
-
lora_target_modules: ['query', 'value', 'output.dense', 'intermediate.dense']
|
| 169 |
-
lr: 2.0
|
| 170 |
-
max_grad_norm: 1.0
|
| 171 |
-
max_steps: 50000
|
| 172 |
-
max_steps_before_stop: 2000
|
| 173 |
-
mode: train
|
| 174 |
-
model_type: graph
|
| 175 |
-
num_layers: 3
|
| 176 |
-
optim: adadelta
|
| 177 |
-
output_file: None
|
| 178 |
-
output_latex: False
|
| 179 |
-
pretrain: True
|
| 180 |
-
pretrain_max_vocab: 250000
|
| 181 |
-
rec_dropout: 0
|
| 182 |
-
reversed: False
|
| 183 |
-
sample_train: 1.0
|
| 184 |
-
save_dir: saved_models/depparse
|
| 185 |
-
save_name: {shorthand}_{embedding}_parser.pt
|
| 186 |
-
second_batch_size: None
|
| 187 |
-
second_bert_learning_rate: 0.001
|
| 188 |
-
second_lr: 0.0002
|
| 189 |
-
second_optim: adam
|
| 190 |
-
second_optim_start_step: 10000
|
| 191 |
-
second_warmup_steps: 200
|
| 192 |
-
seed: 1234
|
| 193 |
-
shorthand: sv_diachronic
|
| 194 |
-
silver_file: ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu
|
| 195 |
-
silver_weight: 0.5
|
| 196 |
-
tag_emb_dim: 50
|
| 197 |
-
train_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 198 |
-
train_size: None
|
| 199 |
-
transformed_dim: 125
|
| 200 |
-
transition_embedding_dim: 20
|
| 201 |
-
transition_hidden_dim: 20
|
| 202 |
-
transition_merge_hidden_dim: 200
|
| 203 |
-
transition_subtree_combination: SubtreeCombination.NONE
|
| 204 |
-
transition_subtree_nonlinearity: none
|
| 205 |
-
use_arc_embedding: False
|
| 206 |
-
use_peft: False
|
| 207 |
-
use_ufeats: True
|
| 208 |
-
use_upos: True
|
| 209 |
-
use_xpos: True
|
| 210 |
-
wandb: False
|
| 211 |
-
wandb_name: None
|
| 212 |
-
weight_decay: 1e-05
|
| 213 |
-
word_cutoff: 7
|
| 214 |
-
word_dropout: 0.33
|
| 215 |
-
word_emb_dim: 75
|
| 216 |
-
wordvec_dir: /home/urdatorn/stanza_resources/sv/pretrain
|
| 217 |
-
wordvec_file: None
|
| 218 |
-
wordvec_pretrain_file: /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
|
| 219 |
-
|
| 220 |
-
2026-03-18 13:16:29 INFO: Loading data with batch size 32...
|
| 221 |
-
2026-03-18 13:16:30 INFO: Train File /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 13954
|
| 222 |
-
2026-03-18 13:16:30 INFO: Original data size: 13954
|
| 223 |
-
2026-03-18 13:16:34 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
|
| 224 |
-
2026-03-18 13:16:34 INFO: Original length = 13954
|
| 225 |
-
2026-03-18 13:16:34 INFO: Filtered length = 13954
|
| 226 |
-
2026-03-18 13:16:37 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='-') for the xpos factory for sv_diachronic
|
| 227 |
-
2026-03-18 13:16:41 DEBUG: 9513 batches created.
|
| 228 |
-
2026-03-18 13:16:41 INFO: Augmented data size: 14712
|
| 229 |
-
2026-03-18 13:16:47 DEBUG: 10020 batches created.
|
| 230 |
-
2026-03-18 13:16:48 DEBUG: 32 batches created.
|
| 231 |
-
2026-03-18 13:17:10 DEBUG: 28588 batches created.
|
| 232 |
-
2026-03-18 13:17:10 INFO: Training parser...
|
| 233 |
-
/home/urdatorn/.pyenv/versions/xpu/lib/python3.13/site-packages/huggingface_hub/file_download.py:949: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
|
| 234 |
-
warnings.warn(
|
| 235 |
-
2026-03-18 13:17:12 DEBUG: Depparse model loading charmodels: /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt and /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
|
| 236 |
-
2026-03-18 13:17:12 DEBUG: Loading charlm from /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt
|
| 237 |
-
2026-03-18 13:17:12 DEBUG: Loading charlm from /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
|
| 238 |
-
2026-03-18 13:17:13 DEBUG: Not using arc embedding enhancement
|
| 239 |
-
2026-03-18 13:17:13 DEBUG: Building Adadelta with lr=2.000000, weight_decay=1e-05
|
| 240 |
-
2026-03-18 13:17:29 INFO: Finished STEP 20/50000, loss = 2.292368 (0.203 sec/batch), lr: 2.000000
|
| 241 |
-
2026-03-18 13:17:43 INFO: Finished STEP 40/50000, loss = 7.419842 (1.468 sec/batch), lr: 2.000000
|
| 242 |
-
2026-03-18 13:17:51 INFO: Finished STEP 60/50000, loss = 7.826311 (1.111 sec/batch), lr: 2.000000
|
| 243 |
-
2026-03-18 13:18:01 INFO: Finished STEP 80/50000, loss = 4.187514 (0.203 sec/batch), lr: 2.000000
|
| 244 |
-
2026-03-18 13:18:09 INFO: Finished STEP 100/50000, loss = 3.687605 (0.196 sec/batch), lr: 2.000000
|
| 245 |
-
2026-03-18 13:18:09 INFO: Evaluating on dev set...
|
| 246 |
-
2026-03-18 13:18:16 INFO: LAS MLAS BLEX
|
| 247 |
-
2026-03-18 13:18:16 INFO: 14.77 8.39 14.11
|
| 248 |
-
2026-03-18 13:18:16 INFO: step 100: train_loss = 5.880744, dev_score = 0.1477
|
| 249 |
-
2026-03-18 13:18:16 DEBUG: training batch usage: Counter({0: 68, 1: 32})
|
| 250 |
-
2026-03-18 13:18:16 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser.pt
|
| 251 |
-
2026-03-18 13:18:16 INFO: new best model saved.
|
| 252 |
-
2026-03-18 13:18:16 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
|
| 253 |
-
2026-03-18 13:18:16 INFO: new model checkpoint saved.
|
| 254 |
-
2026-03-18 13:18:27 INFO: Finished STEP 120/50000, loss = 5.895749 (0.951 sec/batch), lr: 2.000000
|
| 255 |
-
2026-03-18 13:18:38 INFO: Finished STEP 140/50000, loss = 1.609854 (0.200 sec/batch), lr: 2.000000
|
| 256 |
-
2026-03-18 13:18:45 INFO: Finished STEP 160/50000, loss = 1.833139 (0.220 sec/batch), lr: 2.000000
|
| 257 |
-
2026-03-18 13:18:56 INFO: Finished STEP 180/50000, loss = 1.341930 (0.204 sec/batch), lr: 2.000000
|
| 258 |
-
2026-03-18 13:19:04 INFO: Finished STEP 200/50000, loss = 2.078672 (0.224 sec/batch), lr: 2.000000
|
| 259 |
-
2026-03-18 13:19:04 INFO: Evaluating on dev set...
|
| 260 |
-
2026-03-18 13:19:11 INFO: LAS MLAS BLEX
|
| 261 |
-
2026-03-18 13:19:11 INFO: 42.27 32.23 37.34
|
| 262 |
-
2026-03-18 13:19:11 INFO: step 200: train_loss = 3.482911, dev_score = 0.4227
|
| 263 |
-
2026-03-18 13:19:11 DEBUG: training batch usage: Counter({0: 132, 1: 68})
|
| 264 |
-
2026-03-18 13:19:11 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser.pt
|
| 265 |
-
2026-03-18 13:19:11 INFO: new best model saved.
|
| 266 |
-
2026-03-18 13:19:11 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
|
| 267 |
-
2026-03-18 13:19:11 INFO: new model checkpoint saved.
|
| 268 |
-
2026-03-18 13:19:19 INFO: Finished STEP 220/50000, loss = 1.613766 (0.210 sec/batch), lr: 2.000000
|
| 269 |
-
2026-03-18 13:19:27 INFO: Finished STEP 240/50000, loss = 4.277971 (0.822 sec/batch), lr: 2.000000
|
| 270 |
-
2026-03-18 13:19:39 INFO: Finished STEP 260/50000, loss = 4.801499 (0.823 sec/batch), lr: 2.000000
|
| 271 |
-
2026-03-18 13:19:48 INFO: Finished STEP 280/50000, loss = 1.559014 (0.204 sec/batch), lr: 2.000000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_bert-base-swedish-cased_sv_20260318_132248.txt
DELETED
|
@@ -1,109 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_bert-base-swedish-cased_sv_20260318_132248.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: KBLab/bert-base-swedish-cased
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-SILVER.py sv
|
| 6 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 7 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 8 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 9 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 10 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 11 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 12 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 13 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 14 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 15 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 16 |
-
Loading DigPhil MACHINE for SILVER (minus gold overlap)…
|
| 17 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec987-spf144.conllu
|
| 18 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 19 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 20 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec25-Runius.conllu
|
| 21 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 22 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 23 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 24 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 25 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 26 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 27 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 28 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 29 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 30 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 31 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1102-spf259.conllu
|
| 32 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec991-spf148.conllu
|
| 33 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 34 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec988-spf145.conllu
|
| 35 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1063-spf220.conllu
|
| 36 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1033-spf190.conllu
|
| 37 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-letter141673-Stalhammar.conllu
|
| 38 |
-
Adding 77 remaining gold sentences to TRAIN
|
| 39 |
-
Cleaning TRAIN...
|
| 40 |
-
Cleaning SILVER...
|
| 41 |
-
[REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
|
| 42 |
-
[REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
|
| 43 |
-
[REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
|
| 44 |
-
[REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
|
| 45 |
-
[REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
|
| 46 |
-
[REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
|
| 47 |
-
[REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
|
| 48 |
-
[REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
|
| 49 |
-
[REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
|
| 50 |
-
[REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
|
| 51 |
-
[REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
|
| 52 |
-
[REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
|
| 53 |
-
[REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
|
| 54 |
-
[REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
|
| 55 |
-
[REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
|
| 56 |
-
[REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
|
| 57 |
-
[REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
|
| 58 |
-
[REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
|
| 59 |
-
[REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
|
| 60 |
-
[REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
|
| 61 |
-
[REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
|
| 62 |
-
[REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
|
| 63 |
-
[REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
|
| 64 |
-
[REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
|
| 65 |
-
[REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
|
| 66 |
-
[REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
|
| 67 |
-
[REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
|
| 68 |
-
[REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
|
| 69 |
-
[REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
|
| 70 |
-
[REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
|
| 71 |
-
[REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
|
| 72 |
-
[REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
|
| 73 |
-
[REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
|
| 74 |
-
[REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
|
| 75 |
-
[REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
|
| 76 |
-
[REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
|
| 77 |
-
[REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
|
| 78 |
-
[REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
|
| 79 |
-
[REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
|
| 80 |
-
[REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
|
| 81 |
-
[REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
|
| 82 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
|
| 83 |
-
[REMOVED] sent_id=1266 ERRORS=['Token 17: Missing form']
|
| 84 |
-
[REMOVED] sent_id=4897 ERRORS=['Token 46: Missing form', 'Token 47: Missing form']
|
| 85 |
-
Cleaning DEV...
|
| 86 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 87 |
-
Cleaning TEST...
|
| 88 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 89 |
-
Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13871 valid sentences)
|
| 90 |
-
Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (51431 valid sentences)
|
| 91 |
-
Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (31 valid sentences)
|
| 92 |
-
Writing TEST → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (all DIGPHIL_GOLD, 108 valid sentences)
|
| 93 |
-
Done.
|
| 94 |
-
Sourcing scripts/config.sh
|
| 95 |
-
Running stanza dataset preparation…
|
| 96 |
-
2026-03-18 13:22:51 INFO: Datasets program called with:
|
| 97 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --gold
|
| 98 |
-
Augmented 60 quotes: Counter({'《》': 9, '»«': 8, '″″': 8, '““': 7, '„”': 6, '""': 6, '„“': 6, '«»': 4, '””': 4, '「」': 2})
|
| 99 |
-
Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
|
| 100 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /tmp/tmpw00sc1dk/sv_diachronic.train.gold.conllu
|
| 101 |
-
Swapped 'w1, w2' for 'w1 ,w2' 94 times
|
| 102 |
-
Added 86 new sentences with asdf, zzzz -> asdf,zzzz
|
| 103 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /tmp/tmpw00sc1dk/sv_diachronic.dev.gold.conllu
|
| 104 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /tmp/tmpw00sc1dk/sv_diachronic.test.gold.conllu
|
| 105 |
-
Copying from /tmp/tmpw00sc1dk/sv_diachronic.train.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 106 |
-
Copying from /tmp/tmpw00sc1dk/sv_diachronic.dev.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 107 |
-
Copying from /tmp/tmpw00sc1dk/sv_diachronic.test.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
|
| 108 |
-
Running stanza dependency parser training…
|
| 109 |
-
./make_new_model_BERT_SILVER.sh: line 72: unexpected EOF while looking for matching `"'
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_bert-base-swedish-cased_sv_20260318_132404.txt
DELETED
|
@@ -1,109 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_bert-base-swedish-cased_sv_20260318_132404.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: KBLab/bert-base-swedish-cased
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-SILVER.py sv
|
| 6 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 7 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 8 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 9 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 10 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 11 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 12 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 13 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 14 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 15 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 16 |
-
Loading DigPhil MACHINE for SILVER (minus gold overlap)…
|
| 17 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec987-spf144.conllu
|
| 18 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 19 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 20 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec25-Runius.conllu
|
| 21 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 22 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 23 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 24 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 25 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 26 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 27 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 28 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 29 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 30 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 31 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1102-spf259.conllu
|
| 32 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec991-spf148.conllu
|
| 33 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 34 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec988-spf145.conllu
|
| 35 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1063-spf220.conllu
|
| 36 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1033-spf190.conllu
|
| 37 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-letter141673-Stalhammar.conllu
|
| 38 |
-
Adding 74 remaining gold sentences to TRAIN
|
| 39 |
-
Cleaning TRAIN...
|
| 40 |
-
Cleaning SILVER...
|
| 41 |
-
[REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
|
| 42 |
-
[REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
|
| 43 |
-
[REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
|
| 44 |
-
[REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
|
| 45 |
-
[REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
|
| 46 |
-
[REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
|
| 47 |
-
[REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
|
| 48 |
-
[REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
|
| 49 |
-
[REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
|
| 50 |
-
[REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
|
| 51 |
-
[REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
|
| 52 |
-
[REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
|
| 53 |
-
[REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
|
| 54 |
-
[REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
|
| 55 |
-
[REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
|
| 56 |
-
[REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
|
| 57 |
-
[REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
|
| 58 |
-
[REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
|
| 59 |
-
[REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
|
| 60 |
-
[REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
|
| 61 |
-
[REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
|
| 62 |
-
[REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
|
| 63 |
-
[REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
|
| 64 |
-
[REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
|
| 65 |
-
[REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
|
| 66 |
-
[REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
|
| 67 |
-
[REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
|
| 68 |
-
[REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
|
| 69 |
-
[REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
|
| 70 |
-
[REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
|
| 71 |
-
[REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
|
| 72 |
-
[REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
|
| 73 |
-
[REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
|
| 74 |
-
[REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
|
| 75 |
-
[REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
|
| 76 |
-
[REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
|
| 77 |
-
[REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
|
| 78 |
-
[REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
|
| 79 |
-
[REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
|
| 80 |
-
[REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
|
| 81 |
-
[REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
|
| 82 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
|
| 83 |
-
[REMOVED] sent_id=1266 ERRORS=['Token 17: Missing form']
|
| 84 |
-
[REMOVED] sent_id=4897 ERRORS=['Token 46: Missing form', 'Token 47: Missing form']
|
| 85 |
-
Cleaning DEV...
|
| 86 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 87 |
-
Cleaning TEST...
|
| 88 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 89 |
-
Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13868 valid sentences)
|
| 90 |
-
Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (51431 valid sentences)
|
| 91 |
-
Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (34 valid sentences)
|
| 92 |
-
Writing TEST → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (all DIGPHIL_GOLD, 108 valid sentences)
|
| 93 |
-
Done.
|
| 94 |
-
Sourcing scripts/config.sh
|
| 95 |
-
Running stanza dataset preparation…
|
| 96 |
-
2026-03-18 13:24:07 INFO: Datasets program called with:
|
| 97 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --gold
|
| 98 |
-
Augmented 60 quotes: Counter({'《》': 9, '»«': 8, '″″': 8, '““': 7, '„”': 6, '""': 6, '„“': 6, '«»': 4, '””': 4, '「」': 2})
|
| 99 |
-
Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
|
| 100 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /tmp/tmp0l99psu3/sv_diachronic.train.gold.conllu
|
| 101 |
-
Swapped 'w1, w2' for 'w1 ,w2' 93 times
|
| 102 |
-
Added 85 new sentences with asdf, zzzz -> asdf,zzzz
|
| 103 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /tmp/tmp0l99psu3/sv_diachronic.dev.gold.conllu
|
| 104 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /tmp/tmp0l99psu3/sv_diachronic.test.gold.conllu
|
| 105 |
-
Copying from /tmp/tmp0l99psu3/sv_diachronic.train.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 106 |
-
Copying from /tmp/tmp0l99psu3/sv_diachronic.dev.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 107 |
-
Copying from /tmp/tmp0l99psu3/sv_diachronic.test.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
|
| 108 |
-
Running stanza dependency parser training…
|
| 109 |
-
./make_new_model_BERT_SILVER.sh: line 72: unexpected EOF while looking for matching `"'
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_bert-base-swedish-cased_sv_20260318_132733.txt
DELETED
|
@@ -1,325 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_bert-base-swedish-cased_sv_20260318_132733.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: KBLab/bert-base-swedish-cased
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-SILVER.py sv
|
| 6 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 7 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 8 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 9 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 10 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 11 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 12 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 13 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 14 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 15 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 16 |
-
Loading DigPhil MACHINE for SILVER (minus gold overlap)…
|
| 17 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec987-spf144.conllu
|
| 18 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 19 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 20 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec25-Runius.conllu
|
| 21 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 22 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 23 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 24 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 25 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 26 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 27 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 28 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 29 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 30 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 31 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1102-spf259.conllu
|
| 32 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec991-spf148.conllu
|
| 33 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 34 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec988-spf145.conllu
|
| 35 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1063-spf220.conllu
|
| 36 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-sec1033-spf190.conllu
|
| 37 |
-
Reading GOLD: /home/urdatorn/git/Diachronic_Treebanks_DigPhil/parsed_data/validated/svediakorp-letter141673-Stalhammar.conllu
|
| 38 |
-
Adding 73 remaining gold sentences to TRAIN
|
| 39 |
-
Cleaning TRAIN...
|
| 40 |
-
Cleaning SILVER...
|
| 41 |
-
[REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
|
| 42 |
-
[REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
|
| 43 |
-
[REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
|
| 44 |
-
[REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
|
| 45 |
-
[REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
|
| 46 |
-
[REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
|
| 47 |
-
[REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
|
| 48 |
-
[REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
|
| 49 |
-
[REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
|
| 50 |
-
[REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
|
| 51 |
-
[REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
|
| 52 |
-
[REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
|
| 53 |
-
[REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
|
| 54 |
-
[REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
|
| 55 |
-
[REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
|
| 56 |
-
[REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
|
| 57 |
-
[REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
|
| 58 |
-
[REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
|
| 59 |
-
[REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
|
| 60 |
-
[REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
|
| 61 |
-
[REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
|
| 62 |
-
[REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
|
| 63 |
-
[REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
|
| 64 |
-
[REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
|
| 65 |
-
[REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
|
| 66 |
-
[REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
|
| 67 |
-
[REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
|
| 68 |
-
[REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
|
| 69 |
-
[REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
|
| 70 |
-
[REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
|
| 71 |
-
[REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
|
| 72 |
-
[REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
|
| 73 |
-
[REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
|
| 74 |
-
[REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
|
| 75 |
-
[REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
|
| 76 |
-
[REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
|
| 77 |
-
[REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
|
| 78 |
-
[REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
|
| 79 |
-
[REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
|
| 80 |
-
[REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
|
| 81 |
-
[REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
|
| 82 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
|
| 83 |
-
[REMOVED] sent_id=1266 ERRORS=['Token 17: Missing form']
|
| 84 |
-
[REMOVED] sent_id=4897 ERRORS=['Token 46: Missing form', 'Token 47: Missing form']
|
| 85 |
-
Cleaning DEV...
|
| 86 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 87 |
-
Cleaning TEST...
|
| 88 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 89 |
-
Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13867 valid sentences)
|
| 90 |
-
Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (51431 valid sentences)
|
| 91 |
-
Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (35 valid sentences)
|
| 92 |
-
Writing TEST → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (all DIGPHIL_GOLD, 108 valid sentences)
|
| 93 |
-
Done.
|
| 94 |
-
Sourcing scripts/config.sh
|
| 95 |
-
Running stanza dataset preparation…
|
| 96 |
-
2026-03-18 13:27:36 INFO: Datasets program called with:
|
| 97 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --gold
|
| 98 |
-
Augmented 60 quotes: Counter({'《》': 9, '»«': 8, '″″': 8, '““': 7, '„”': 6, '""': 6, '„“': 6, '«»': 4, '””': 4, '「」': 2})
|
| 99 |
-
Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
|
| 100 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /tmp/tmpl2s5qxw1/sv_diachronic.train.gold.conllu
|
| 101 |
-
Swapped 'w1, w2' for 'w1 ,w2' 98 times
|
| 102 |
-
Added 85 new sentences with asdf, zzzz -> asdf,zzzz
|
| 103 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /tmp/tmpl2s5qxw1/sv_diachronic.dev.gold.conllu
|
| 104 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /tmp/tmpl2s5qxw1/sv_diachronic.test.gold.conllu
|
| 105 |
-
Copying from /tmp/tmpl2s5qxw1/sv_diachronic.train.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 106 |
-
Copying from /tmp/tmpl2s5qxw1/sv_diachronic.dev.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 107 |
-
Copying from /tmp/tmpl2s5qxw1/sv_diachronic.test.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
|
| 108 |
-
Running stanza dependency parser training…
|
| 109 |
-
2026-03-18 13:27:38 INFO: Training program called with:
|
| 110 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --batch_size 32 --dropout 0.33 --use_bert --bert_model KBLab/bert-base-swedish-cased --silver_file ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu --device xpu:0
|
| 111 |
-
2026-03-18 13:27:38 DEBUG: UD_Swedish-diachronic: sv_diachronic
|
| 112 |
-
2026-03-18 13:27:38 INFO: Using model /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 113 |
-
2026-03-18 13:27:38 INFO: Using model /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 114 |
-
2026-03-18 13:27:38 INFO: Using default pretrain for language sv, found in /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt To use a different pretrain, specify --wordvec_pretrain_file
|
| 115 |
-
2026-03-18 13:27:38 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_transformer_parser.pt does not exist, training new model
|
| 116 |
-
2026-03-18 13:27:38 INFO: Using model /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 117 |
-
2026-03-18 13:27:38 INFO: Using model /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 118 |
-
2026-03-18 13:27:38 INFO: Using default pretrain for language sv, found in /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt To use a different pretrain, specify --wordvec_pretrain_file
|
| 119 |
-
2026-03-18 13:27:38 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/home/urdatorn/stanza_resources/sv/pretrain', '--train_file', '/home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--wordvec_pretrain_file', '/home/urdatorn/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt', '--batch_size', '32', '--dropout', '0.33', '--bert_model', 'KBLab/bert-base-swedish-cased', '--silver_file', 'ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu', '--device', 'xpu:0']
|
| 120 |
-
2026-03-18 13:27:38 INFO: Running parser in train mode
|
| 121 |
-
2026-03-18 13:27:39 DEBUG: Loaded pretrain from /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
|
| 122 |
-
2026-03-18 13:27:39 INFO: Using 7 as the word cutoff based on the size of the pretrain (250000)
|
| 123 |
-
2026-03-18 13:27:39 INFO: Using pretrained contextualized char embedding
|
| 124 |
-
2026-03-18 13:27:39 INFO: ARGS USED AT TRAINING TIME:
|
| 125 |
-
augment_nopunct: None
|
| 126 |
-
batch_size: 32
|
| 127 |
-
bert_finetune: False
|
| 128 |
-
bert_finetune_layers: None
|
| 129 |
-
bert_hidden_layers: 4
|
| 130 |
-
bert_learning_rate: 1.0
|
| 131 |
-
bert_model: KBLab/bert-base-swedish-cased
|
| 132 |
-
bert_start_finetuning: 200
|
| 133 |
-
bert_warmup_steps: 200
|
| 134 |
-
bert_weight_decay: 0.0
|
| 135 |
-
beta2: 0.999
|
| 136 |
-
char: True
|
| 137 |
-
char_emb_dim: 100
|
| 138 |
-
char_hidden_dim: 400
|
| 139 |
-
char_num_layers: 1
|
| 140 |
-
char_rec_dropout: 0
|
| 141 |
-
charlm: True
|
| 142 |
-
charlm_backward_file: /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
|
| 143 |
-
charlm_forward_file: /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt
|
| 144 |
-
charlm_save_dir: saved_models/charlm
|
| 145 |
-
charlm_shorthand: sv_conll17
|
| 146 |
-
checkpoint: True
|
| 147 |
-
checkpoint_interval: 500
|
| 148 |
-
checkpoint_save_name: None
|
| 149 |
-
continue_from: None
|
| 150 |
-
data_dir: data/depparse
|
| 151 |
-
deep_biaff_hidden_dim: 400
|
| 152 |
-
deep_biaff_output_dim: 160
|
| 153 |
-
device: xpu:0
|
| 154 |
-
distance: True
|
| 155 |
-
dropout: 0.33
|
| 156 |
-
eval_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 157 |
-
eval_interval: 100
|
| 158 |
-
gold_labels: True
|
| 159 |
-
hidden_dim: 400
|
| 160 |
-
lang: sv
|
| 161 |
-
linearization: True
|
| 162 |
-
log_norms: False
|
| 163 |
-
log_step: 20
|
| 164 |
-
lora_alpha: 128
|
| 165 |
-
lora_dropout: 0.1
|
| 166 |
-
lora_modules_to_save: []
|
| 167 |
-
lora_rank: 64
|
| 168 |
-
lora_target_modules: ['query', 'value', 'output.dense', 'intermediate.dense']
|
| 169 |
-
lr: 2.0
|
| 170 |
-
max_grad_norm: 1.0
|
| 171 |
-
max_steps: 50000
|
| 172 |
-
max_steps_before_stop: 2000
|
| 173 |
-
mode: train
|
| 174 |
-
model_type: graph
|
| 175 |
-
num_layers: 3
|
| 176 |
-
optim: adadelta
|
| 177 |
-
output_file: None
|
| 178 |
-
output_latex: False
|
| 179 |
-
pretrain: True
|
| 180 |
-
pretrain_max_vocab: 250000
|
| 181 |
-
rec_dropout: 0
|
| 182 |
-
reversed: False
|
| 183 |
-
sample_train: 1.0
|
| 184 |
-
save_dir: saved_models/depparse
|
| 185 |
-
save_name: {shorthand}_{embedding}_parser.pt
|
| 186 |
-
second_batch_size: None
|
| 187 |
-
second_bert_learning_rate: 0.001
|
| 188 |
-
second_lr: 0.0002
|
| 189 |
-
second_optim: adam
|
| 190 |
-
second_optim_start_step: 10000
|
| 191 |
-
second_warmup_steps: 200
|
| 192 |
-
seed: 1234
|
| 193 |
-
shorthand: sv_diachronic
|
| 194 |
-
silver_file: ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu
|
| 195 |
-
silver_weight: 0.5
|
| 196 |
-
tag_emb_dim: 50
|
| 197 |
-
train_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 198 |
-
train_size: None
|
| 199 |
-
transformed_dim: 125
|
| 200 |
-
transition_embedding_dim: 20
|
| 201 |
-
transition_hidden_dim: 20
|
| 202 |
-
transition_merge_hidden_dim: 200
|
| 203 |
-
transition_subtree_combination: SubtreeCombination.NONE
|
| 204 |
-
transition_subtree_nonlinearity: none
|
| 205 |
-
use_arc_embedding: False
|
| 206 |
-
use_peft: False
|
| 207 |
-
use_ufeats: True
|
| 208 |
-
use_upos: True
|
| 209 |
-
use_xpos: True
|
| 210 |
-
wandb: False
|
| 211 |
-
wandb_name: None
|
| 212 |
-
weight_decay: 1e-05
|
| 213 |
-
word_cutoff: 7
|
| 214 |
-
word_dropout: 0.33
|
| 215 |
-
word_emb_dim: 75
|
| 216 |
-
wordvec_dir: /home/urdatorn/stanza_resources/sv/pretrain
|
| 217 |
-
wordvec_file: None
|
| 218 |
-
wordvec_pretrain_file: /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
|
| 219 |
-
|
| 220 |
-
2026-03-18 13:27:39 INFO: Loading data with batch size 32...
|
| 221 |
-
2026-03-18 13:27:40 INFO: Train File /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 13952
|
| 222 |
-
2026-03-18 13:27:40 INFO: Original data size: 13952
|
| 223 |
-
2026-03-18 13:27:43 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
|
| 224 |
-
2026-03-18 13:27:43 INFO: Original length = 13952
|
| 225 |
-
2026-03-18 13:27:43 INFO: Filtered length = 13952
|
| 226 |
-
2026-03-18 13:27:47 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='-') for the xpos factory for sv_diachronic
|
| 227 |
-
2026-03-18 13:27:50 DEBUG: 9512 batches created.
|
| 228 |
-
2026-03-18 13:27:50 INFO: Augmented data size: 14707
|
| 229 |
-
2026-03-18 13:27:57 DEBUG: 10016 batches created.
|
| 230 |
-
2026-03-18 13:27:57 DEBUG: 34 batches created.
|
| 231 |
-
2026-03-18 13:28:19 DEBUG: 28588 batches created.
|
| 232 |
-
2026-03-18 13:28:19 INFO: Training parser...
|
| 233 |
-
/home/urdatorn/.pyenv/versions/xpu/lib/python3.13/site-packages/huggingface_hub/file_download.py:949: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
|
| 234 |
-
warnings.warn(
|
| 235 |
-
2026-03-18 13:28:20 DEBUG: Depparse model loading charmodels: /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt and /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
|
| 236 |
-
2026-03-18 13:28:20 DEBUG: Loading charlm from /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt
|
| 237 |
-
2026-03-18 13:28:20 DEBUG: Loading charlm from /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
|
| 238 |
-
2026-03-18 13:28:20 DEBUG: Not using arc embedding enhancement
|
| 239 |
-
2026-03-18 13:28:21 DEBUG: Building Adadelta with lr=2.000000, weight_decay=1e-05
|
| 240 |
-
2026-03-18 13:28:31 INFO: Finished STEP 20/50000, loss = 5.042704 (0.060 sec/batch), lr: 2.000000
|
| 241 |
-
2026-03-18 13:28:38 INFO: Finished STEP 40/50000, loss = 7.420871 (0.629 sec/batch), lr: 2.000000
|
| 242 |
-
2026-03-18 13:28:42 INFO: Finished STEP 60/50000, loss = 7.130751 (0.585 sec/batch), lr: 2.000000
|
| 243 |
-
2026-03-18 13:28:46 INFO: Finished STEP 80/50000, loss = 2.640507 (0.061 sec/batch), lr: 2.000000
|
| 244 |
-
2026-03-18 13:28:49 INFO: Finished STEP 100/50000, loss = 2.198331 (0.062 sec/batch), lr: 2.000000
|
| 245 |
-
2026-03-18 13:28:49 INFO: Evaluating on dev set...
|
| 246 |
-
2026-03-18 13:28:52 INFO: LAS MLAS BLEX
|
| 247 |
-
2026-03-18 13:28:52 INFO: 14.92 7.39 9.17
|
| 248 |
-
2026-03-18 13:28:52 INFO: step 100: train_loss = 6.003559, dev_score = 0.1492
|
| 249 |
-
2026-03-18 13:28:52 DEBUG: training batch usage: Counter({0: 68, 1: 32})
|
| 250 |
-
2026-03-18 13:28:52 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser.pt
|
| 251 |
-
2026-03-18 13:28:52 INFO: new best model saved.
|
| 252 |
-
2026-03-18 13:28:52 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
|
| 253 |
-
2026-03-18 13:28:52 INFO: new model checkpoint saved.
|
| 254 |
-
2026-03-18 13:28:58 INFO: Finished STEP 120/50000, loss = 5.791277 (0.479 sec/batch), lr: 2.000000
|
| 255 |
-
2026-03-18 13:29:02 INFO: Finished STEP 140/50000, loss = 1.561704 (0.065 sec/batch), lr: 2.000000
|
| 256 |
-
2026-03-18 13:29:05 INFO: Finished STEP 160/50000, loss = 2.749855 (0.064 sec/batch), lr: 2.000000
|
| 257 |
-
2026-03-18 13:29:10 INFO: Finished STEP 180/50000, loss = 2.644507 (0.062 sec/batch), lr: 2.000000
|
| 258 |
-
2026-03-18 13:29:14 INFO: Finished STEP 200/50000, loss = 2.221299 (0.062 sec/batch), lr: 2.000000
|
| 259 |
-
2026-03-18 13:29:14 INFO: Evaluating on dev set...
|
| 260 |
-
2026-03-18 13:29:17 INFO: LAS MLAS BLEX
|
| 261 |
-
2026-03-18 13:29:17 INFO: 36.67 23.23 32.82
|
| 262 |
-
2026-03-18 13:29:17 INFO: step 200: train_loss = 3.408124, dev_score = 0.3667
|
| 263 |
-
2026-03-18 13:29:17 DEBUG: training batch usage: Counter({0: 132, 1: 68})
|
| 264 |
-
2026-03-18 13:29:17 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser.pt
|
| 265 |
-
2026-03-18 13:29:17 INFO: new best model saved.
|
| 266 |
-
2026-03-18 13:29:17 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
|
| 267 |
-
2026-03-18 13:29:17 INFO: new model checkpoint saved.
|
| 268 |
-
2026-03-18 13:29:20 INFO: Finished STEP 220/50000, loss = 2.403125 (0.064 sec/batch), lr: 2.000000
|
| 269 |
-
2026-03-18 13:29:24 INFO: Finished STEP 240/50000, loss = 3.965761 (0.412 sec/batch), lr: 2.000000
|
| 270 |
-
2026-03-18 13:29:30 INFO: Finished STEP 260/50000, loss = 4.813926 (0.403 sec/batch), lr: 2.000000
|
| 271 |
-
2026-03-18 13:29:33 INFO: Finished STEP 280/50000, loss = 2.433779 (0.067 sec/batch), lr: 2.000000
|
| 272 |
-
2026-03-18 13:29:37 INFO: Finished STEP 300/50000, loss = 5.255473 (0.374 sec/batch), lr: 2.000000
|
| 273 |
-
2026-03-18 13:29:37 INFO: Evaluating on dev set...
|
| 274 |
-
2026-03-18 13:29:40 INFO: LAS MLAS BLEX
|
| 275 |
-
2026-03-18 13:29:40 INFO: 51.20 39.90 43.72
|
| 276 |
-
2026-03-18 13:29:40 INFO: step 300: train_loss = 3.172424, dev_score = 0.5120
|
| 277 |
-
2026-03-18 13:29:40 DEBUG: training batch usage: Counter({0: 191, 1: 109})
|
| 278 |
-
2026-03-18 13:29:40 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser.pt
|
| 279 |
-
2026-03-18 13:29:40 INFO: new best model saved.
|
| 280 |
-
2026-03-18 13:29:41 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
|
| 281 |
-
2026-03-18 13:29:41 INFO: new model checkpoint saved.
|
| 282 |
-
2026-03-18 13:29:42 INFO: Finished STEP 320/50000, loss = 4.718216 (0.382 sec/batch), lr: 2.000000
|
| 283 |
-
2026-03-18 13:29:46 INFO: Finished STEP 340/50000, loss = 2.242086 (0.065 sec/batch), lr: 2.000000
|
| 284 |
-
2026-03-18 13:29:48 INFO: Finished STEP 360/50000, loss = 1.605185 (0.065 sec/batch), lr: 2.000000
|
| 285 |
-
2026-03-18 13:29:50 INFO: Finished STEP 380/50000, loss = 1.821524 (0.063 sec/batch), lr: 2.000000
|
| 286 |
-
2026-03-18 13:29:54 INFO: Finished STEP 400/50000, loss = 1.597555 (0.065 sec/batch), lr: 2.000000
|
| 287 |
-
2026-03-18 13:29:54 INFO: Evaluating on dev set...
|
| 288 |
-
2026-03-18 13:29:57 INFO: LAS MLAS BLEX
|
| 289 |
-
2026-03-18 13:29:57 INFO: 50.46 38.59 43.93
|
| 290 |
-
2026-03-18 13:29:57 INFO: step 400: train_loss = 2.549639, dev_score = 0.5046
|
| 291 |
-
2026-03-18 13:29:57 DEBUG: training batch usage: Counter({0: 269, 1: 131})
|
| 292 |
-
2026-03-18 13:29:57 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
|
| 293 |
-
2026-03-18 13:29:57 INFO: new model checkpoint saved.
|
| 294 |
-
2026-03-18 13:30:00 INFO: Finished STEP 420/50000, loss = 2.220649 (0.065 sec/batch), lr: 2.000000
|
| 295 |
-
2026-03-18 13:30:04 INFO: Finished STEP 440/50000, loss = 1.550643 (0.064 sec/batch), lr: 2.000000
|
| 296 |
-
2026-03-18 13:30:08 INFO: Finished STEP 460/50000, loss = 3.319951 (0.344 sec/batch), lr: 2.000000
|
| 297 |
-
2026-03-18 13:30:12 INFO: Finished STEP 480/50000, loss = 2.615420 (0.068 sec/batch), lr: 2.000000
|
| 298 |
-
2026-03-18 13:30:14 INFO: Finished STEP 500/50000, loss = 1.976453 (0.068 sec/batch), lr: 2.000000
|
| 299 |
-
2026-03-18 13:30:14 INFO: Evaluating on dev set...
|
| 300 |
-
2026-03-18 13:30:17 INFO: LAS MLAS BLEX
|
| 301 |
-
2026-03-18 13:30:17 INFO: 52.11 42.78 46.03
|
| 302 |
-
2026-03-18 13:30:17 INFO: step 500: train_loss = 2.721934, dev_score = 0.5211
|
| 303 |
-
2026-03-18 13:30:17 DEBUG: training batch usage: Counter({0: 332, 1: 168})
|
| 304 |
-
2026-03-18 13:30:17 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser.pt
|
| 305 |
-
2026-03-18 13:30:17 INFO: new best model saved.
|
| 306 |
-
2026-03-18 13:30:18 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
|
| 307 |
-
2026-03-18 13:30:18 INFO: new model checkpoint saved.
|
| 308 |
-
2026-03-18 13:30:21 INFO: Finished STEP 520/50000, loss = 4.217991 (0.338 sec/batch), lr: 2.000000
|
| 309 |
-
2026-03-18 13:30:24 INFO: Finished STEP 540/50000, loss = 3.852498 (0.337 sec/batch), lr: 2.000000
|
| 310 |
-
2026-03-18 13:30:28 INFO: Finished STEP 560/50000, loss = 1.689470 (0.069 sec/batch), lr: 2.000000
|
| 311 |
-
2026-03-18 13:30:32 INFO: Finished STEP 580/50000, loss = 3.136248 (0.339 sec/batch), lr: 2.000000
|
| 312 |
-
2026-03-18 13:30:35 INFO: Finished STEP 600/50000, loss = 3.013068 (0.068 sec/batch), lr: 2.000000
|
| 313 |
-
2026-03-18 13:30:35 INFO: Evaluating on dev set...
|
| 314 |
-
2026-03-18 13:30:38 INFO: LAS MLAS BLEX
|
| 315 |
-
2026-03-18 13:30:38 INFO: 56.89 46.97 50.02
|
| 316 |
-
2026-03-18 13:30:38 INFO: step 600: train_loss = 2.802572, dev_score = 0.5689
|
| 317 |
-
2026-03-18 13:30:38 DEBUG: training batch usage: Counter({0: 392, 1: 208})
|
| 318 |
-
2026-03-18 13:30:38 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser.pt
|
| 319 |
-
2026-03-18 13:30:38 INFO: new best model saved.
|
| 320 |
-
2026-03-18 13:30:38 INFO: Model saved to saved_models/depparse/sv_diachronic_transformer_parser_checkpoint.pt
|
| 321 |
-
2026-03-18 13:30:38 INFO: new model checkpoint saved.
|
| 322 |
-
2026-03-18 13:30:41 INFO: Finished STEP 620/50000, loss = 3.173692 (0.327 sec/batch), lr: 2.000000
|
| 323 |
-
2026-03-18 13:30:44 INFO: Finished STEP 640/50000, loss = 1.843263 (0.070 sec/batch), lr: 2.000000
|
| 324 |
-
2026-03-18 13:30:47 INFO: Finished STEP 660/50000, loss = 1.417565 (0.069 sec/batch), lr: 2.000000
|
| 325 |
-
2026-03-18 13:30:51 INFO: Finished STEP 680/50000, loss = 4.149919 (0.325 sec/batch), lr: 2.000000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_bert-base-swedish-cased_sv_20260318_133119.txt
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
logs/log_bert-base-swedish-cased_sv_20260318_141654.txt
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
logs/log_bert-base-swedish-cased_sv_20260318_215701.txt
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
logs/log_bert-base-swedish-cased_sv_20260319_200948.txt
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
logs/log_bert-base-swedish-cased_sv_20260319_201349.txt
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
logs/log_bert-base-swedish-cased_sv_20260319_214621.txt
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
logs/log_conll17.pt_sv_20251203_232257.txt
DELETED
|
@@ -1,719 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_20251203_232257.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-test.py sv
|
| 6 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 7 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 8 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 9 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 10 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 11 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 12 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/ucxn_ud_swedish-talbanken.conllu
|
| 13 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 14 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 15 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 16 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 17 |
-
Skipping DigPhil MACHINE (diachron not requested).
|
| 18 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 19 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 20 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 21 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 22 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
|
| 23 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 24 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 25 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 26 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
|
| 27 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
|
| 28 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 29 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
|
| 30 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
|
| 31 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
|
| 32 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 33 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 34 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 35 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 36 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 37 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
|
| 38 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
|
| 39 |
-
Cleaning TRAIN...
|
| 40 |
-
Cleaning DEV...
|
| 41 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 42 |
-
Cleaning TEST...
|
| 43 |
-
Writing TRAIN → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (19820 valid sentences)
|
| 44 |
-
Writing DEV → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (9 valid sentences)
|
| 45 |
-
Writing TEST → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (99 valid sentences)
|
| 46 |
-
Done.
|
| 47 |
-
Sourcing scripts/config_alvis.sh
|
| 48 |
-
Running stanza dataset preparation…
|
| 49 |
-
2025-12-03 23:23:04 INFO: Datasets program called with:
|
| 50 |
-
/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 51 |
-
2025-12-03 23:23:04 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
|
| 52 |
-
|
| 53 |
-
2025-12-03 23:23:04 INFO: Downloaded file to /cephyr/users/cleland/Alvis/stanza_resources/resources.json
|
| 54 |
-
2025-12-03 23:23:04 DEBUG: Processing parameter "processors"...
|
| 55 |
-
2025-12-03 23:23:04 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
|
| 56 |
-
2025-12-03 23:23:04 INFO: Downloading these customized packages for language: sv (Swedish)...
|
| 57 |
-
=======================
|
| 58 |
-
| Processor | Package |
|
| 59 |
-
-----------------------
|
| 60 |
-
=======================
|
| 61 |
-
|
| 62 |
-
2025-12-03 23:23:04 INFO: Finished downloading models and saved to /cephyr/users/cleland/Alvis/stanza_resources
|
| 63 |
-
2025-12-03 23:23:04 INFO: Using tagger model in /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt for sv_diachronic
|
| 64 |
-
2025-12-03 23:23:04 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 65 |
-
2025-12-03 23:23:04 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 66 |
-
Augmented 56 quotes: Counter({'„”': 9, '″″': 9, '""': 8, '「」': 8, '””': 5, '““': 4, '《》': 4, '»«': 3, '„“': 3, '«»': 3})
|
| 67 |
-
2025-12-03 23:23:05 INFO: Running tagger to retag /local/tmp.5441282/tmp4sg9id1k/sv_diachronic.train.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 68 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5441282/tmp4sg9id1k/sv_diachronic.train.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu']
|
| 69 |
-
2025-12-03 23:23:05 INFO: Running tagger in predict mode
|
| 70 |
-
2025-12-03 23:23:05 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 71 |
-
2025-12-03 23:23:07 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 72 |
-
2025-12-03 23:23:07 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 73 |
-
2025-12-03 23:23:07 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 74 |
-
2025-12-03 23:23:07 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 75 |
-
2025-12-03 23:23:07 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 76 |
-
2025-12-03 23:23:10 INFO: Loading data with batch size 250...
|
| 77 |
-
2025-12-03 23:23:24 INFO: Start evaluation...
|
| 78 |
-
2025-12-03 23:24:31 INFO: UPOS XPOS UFeats AllTags
|
| 79 |
-
2025-12-03 23:24:31 INFO: 98.01 62.93 94.04 60.49
|
| 80 |
-
2025-12-03 23:24:31 INFO: POS Tagger score: sv_diachronic 60.49
|
| 81 |
-
2025-12-03 23:24:32 INFO: Running tagger to retag /local/tmp.5441282/tmp4sg9id1k/sv_diachronic.dev.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 82 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5441282/tmp4sg9id1k/sv_diachronic.dev.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu']
|
| 83 |
-
2025-12-03 23:24:32 INFO: Running tagger in predict mode
|
| 84 |
-
2025-12-03 23:24:32 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 85 |
-
2025-12-03 23:24:33 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 86 |
-
2025-12-03 23:24:33 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 87 |
-
2025-12-03 23:24:33 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 88 |
-
2025-12-03 23:24:33 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 89 |
-
2025-12-03 23:24:34 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 90 |
-
2025-12-03 23:24:34 INFO: Loading data with batch size 250...
|
| 91 |
-
2025-12-03 23:24:34 INFO: Start evaluation...
|
| 92 |
-
2025-12-03 23:24:34 INFO: UPOS XPOS UFeats AllTags
|
| 93 |
-
2025-12-03 23:24:34 INFO: 93.32 90.84 93.32 85.64
|
| 94 |
-
2025-12-03 23:24:34 INFO: POS Tagger score: sv_diachronic 85.64
|
| 95 |
-
2025-12-03 23:24:34 INFO: Running tagger to retag /local/tmp.5441282/tmp4sg9id1k/sv_diachronic.test.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
|
| 96 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5441282/tmp4sg9id1k/sv_diachronic.test.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu']
|
| 97 |
-
2025-12-03 23:24:34 INFO: Running tagger in predict mode
|
| 98 |
-
2025-12-03 23:24:34 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 99 |
-
2025-12-03 23:24:36 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 100 |
-
2025-12-03 23:24:36 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 101 |
-
2025-12-03 23:24:36 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 102 |
-
2025-12-03 23:24:36 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 103 |
-
2025-12-03 23:24:36 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 104 |
-
2025-12-03 23:24:36 INFO: Loading data with batch size 250...
|
| 105 |
-
2025-12-03 23:24:36 INFO: Start evaluation...
|
| 106 |
-
2025-12-03 23:24:37 INFO: UPOS XPOS UFeats AllTags
|
| 107 |
-
2025-12-03 23:24:37 INFO: 93.14 96.78 95.32 90.28
|
| 108 |
-
2025-12-03 23:24:37 INFO: POS Tagger score: sv_diachronic 90.28
|
| 109 |
-
Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
|
| 110 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /local/tmp.5441282/tmp4sg9id1k/sv_diachronic.train.gold.conllu
|
| 111 |
-
Swapped 'w1, w2' for 'w1 ,w2' 122 times
|
| 112 |
-
Added 100 new sentences with asdf, zzzz -> asdf,zzzz
|
| 113 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /local/tmp.5441282/tmp4sg9id1k/sv_diachronic.dev.gold.conllu
|
| 114 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /local/tmp.5441282/tmp4sg9id1k/sv_diachronic.test.gold.conllu
|
| 115 |
-
Running stanza dependency parser training…
|
| 116 |
-
2025-12-03 23:24:46 INFO: Training program called with:
|
| 117 |
-
/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt --batch_size 32 --dropout 0.33
|
| 118 |
-
2025-12-03 23:24:46 DEBUG: UD_Swedish-diachronic: sv_diachronic
|
| 119 |
-
2025-12-03 23:24:46 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 120 |
-
2025-12-03 23:24:46 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 121 |
-
2025-12-03 23:24:46 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_charlm_parser.pt does not exist, training new model
|
| 122 |
-
2025-12-03 23:24:46 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 123 |
-
2025-12-03 23:24:46 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 124 |
-
2025-12-03 23:24:46 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--train_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
|
| 125 |
-
2025-12-03 23:24:46 INFO: Running parser in train mode
|
| 126 |
-
2025-12-03 23:24:46 INFO: Using pretrained contextualized char embedding
|
| 127 |
-
2025-12-03 23:24:46 INFO: Loading data with batch size 32...
|
| 128 |
-
2025-12-03 23:24:49 INFO: Train File /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 19920
|
| 129 |
-
2025-12-03 23:24:49 INFO: Original data size: 19920
|
| 130 |
-
2025-12-03 23:24:49 INFO: Augmented data size: 20844
|
| 131 |
-
2025-12-03 23:24:57 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
|
| 132 |
-
2025-12-03 23:24:57 INFO: Original length = 20844
|
| 133 |
-
2025-12-03 23:24:57 INFO: Filtered length = 20844
|
| 134 |
-
2025-12-03 23:25:02 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='|') for the xpos factory for sv_diachronic
|
| 135 |
-
2025-12-03 23:25:06 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 136 |
-
2025-12-03 23:25:11 DEBUG: 13986 batches created.
|
| 137 |
-
2025-12-03 23:25:11 DEBUG: 9 batches created.
|
| 138 |
-
2025-12-03 23:25:11 INFO: Training parser...
|
| 139 |
-
2025-12-03 23:25:11 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 140 |
-
2025-12-03 23:25:11 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 141 |
-
2025-12-03 23:25:11 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 142 |
-
2025-12-03 23:25:11 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 143 |
-
2025-12-03 23:25:14 INFO: Finished STEP 20/50000, loss = 3.699069 (0.033 sec/batch), lr: 0.003000
|
| 144 |
-
2025-12-03 23:25:15 INFO: Finished STEP 40/50000, loss = 3.841336 (0.035 sec/batch), lr: 0.003000
|
| 145 |
-
2025-12-03 23:25:16 INFO: Finished STEP 60/50000, loss = 2.719965 (0.034 sec/batch), lr: 0.003000
|
| 146 |
-
2025-12-03 23:25:16 INFO: Finished STEP 80/50000, loss = 3.832992 (0.035 sec/batch), lr: 0.003000
|
| 147 |
-
2025-12-03 23:25:17 INFO: Finished STEP 100/50000, loss = 3.050168 (0.035 sec/batch), lr: 0.003000
|
| 148 |
-
2025-12-03 23:25:17 INFO: Evaluating on dev set...
|
| 149 |
-
2025-12-03 23:25:18 INFO: LAS MLAS BLEX
|
| 150 |
-
2025-12-03 23:25:18 INFO: 3.22 0.82 1.64
|
| 151 |
-
2025-12-03 23:25:18 INFO: step 100: train_loss = 1815.927894, dev_score = 0.0322
|
| 152 |
-
2025-12-03 23:25:18 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 153 |
-
2025-12-03 23:25:18 INFO: new best model saved.
|
| 154 |
-
2025-12-03 23:25:19 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 155 |
-
2025-12-03 23:25:19 INFO: new model checkpoint saved.
|
| 156 |
-
2025-12-03 23:25:19 INFO: Finished STEP 120/50000, loss = 3.082294 (0.035 sec/batch), lr: 0.003000
|
| 157 |
-
2025-12-03 23:25:20 INFO: Finished STEP 140/50000, loss = 2.829382 (0.035 sec/batch), lr: 0.003000
|
| 158 |
-
2025-12-03 23:25:21 INFO: Finished STEP 160/50000, loss = 2.907810 (0.036 sec/batch), lr: 0.003000
|
| 159 |
-
2025-12-03 23:25:22 INFO: Finished STEP 180/50000, loss = 3.411438 (0.035 sec/batch), lr: 0.003000
|
| 160 |
-
2025-12-03 23:25:22 INFO: Finished STEP 200/50000, loss = 2.634063 (0.037 sec/batch), lr: 0.003000
|
| 161 |
-
2025-12-03 23:25:22 INFO: Evaluating on dev set...
|
| 162 |
-
2025-12-03 23:25:23 INFO: LAS MLAS BLEX
|
| 163 |
-
2025-12-03 23:25:23 INFO: 7.92 6.14 7.51
|
| 164 |
-
2025-12-03 23:25:23 INFO: step 200: train_loss = 2.962158, dev_score = 0.0792
|
| 165 |
-
2025-12-03 23:25:23 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 166 |
-
2025-12-03 23:25:23 INFO: new best model saved.
|
| 167 |
-
2025-12-03 23:25:24 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 168 |
-
2025-12-03 23:25:24 INFO: new model checkpoint saved.
|
| 169 |
-
2025-12-03 23:25:25 INFO: Finished STEP 220/50000, loss = 2.243211 (0.036 sec/batch), lr: 0.003000
|
| 170 |
-
2025-12-03 23:25:25 INFO: Finished STEP 240/50000, loss = 3.037997 (0.034 sec/batch), lr: 0.003000
|
| 171 |
-
2025-12-03 23:25:26 INFO: Finished STEP 260/50000, loss = 2.535078 (0.034 sec/batch), lr: 0.003000
|
| 172 |
-
2025-12-03 23:25:27 INFO: Finished STEP 280/50000, loss = 3.026233 (0.035 sec/batch), lr: 0.003000
|
| 173 |
-
2025-12-03 23:25:27 INFO: Finished STEP 300/50000, loss = 3.630493 (0.034 sec/batch), lr: 0.003000
|
| 174 |
-
2025-12-03 23:25:27 INFO: Evaluating on dev set...
|
| 175 |
-
2025-12-03 23:25:28 INFO: LAS MLAS BLEX
|
| 176 |
-
2025-12-03 23:25:28 INFO: 13.61 10.54 12.24
|
| 177 |
-
2025-12-03 23:25:28 INFO: step 300: train_loss = 3.006871, dev_score = 0.1361
|
| 178 |
-
2025-12-03 23:25:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 179 |
-
2025-12-03 23:25:28 INFO: new best model saved.
|
| 180 |
-
2025-12-03 23:25:29 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 181 |
-
2025-12-03 23:25:29 INFO: new model checkpoint saved.
|
| 182 |
-
2025-12-03 23:25:30 INFO: Finished STEP 320/50000, loss = 2.784619 (0.034 sec/batch), lr: 0.003000
|
| 183 |
-
2025-12-03 23:25:30 INFO: Finished STEP 340/50000, loss = 3.004477 (0.038 sec/batch), lr: 0.003000
|
| 184 |
-
2025-12-03 23:25:31 INFO: Finished STEP 360/50000, loss = 2.542431 (0.035 sec/batch), lr: 0.003000
|
| 185 |
-
2025-12-03 23:25:32 INFO: Finished STEP 380/50000, loss = 3.084781 (0.036 sec/batch), lr: 0.003000
|
| 186 |
-
2025-12-03 23:25:32 INFO: Finished STEP 400/50000, loss = 2.454229 (0.035 sec/batch), lr: 0.003000
|
| 187 |
-
2025-12-03 23:25:32 INFO: Evaluating on dev set...
|
| 188 |
-
2025-12-03 23:25:33 INFO: LAS MLAS BLEX
|
| 189 |
-
2025-12-03 23:25:33 INFO: 18.81 6.64 10.55
|
| 190 |
-
2025-12-03 23:25:33 INFO: step 400: train_loss = 3.048551, dev_score = 0.1881
|
| 191 |
-
2025-12-03 23:25:33 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 192 |
-
2025-12-03 23:25:33 INFO: new best model saved.
|
| 193 |
-
2025-12-03 23:25:34 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 194 |
-
2025-12-03 23:25:34 INFO: new model checkpoint saved.
|
| 195 |
-
2025-12-03 23:25:35 INFO: Finished STEP 420/50000, loss = 2.214305 (0.034 sec/batch), lr: 0.003000
|
| 196 |
-
2025-12-03 23:25:35 INFO: Finished STEP 440/50000, loss = 2.953549 (0.034 sec/batch), lr: 0.003000
|
| 197 |
-
2025-12-03 23:25:36 INFO: Finished STEP 460/50000, loss = 2.711811 (0.037 sec/batch), lr: 0.003000
|
| 198 |
-
2025-12-03 23:25:37 INFO: Finished STEP 480/50000, loss = 2.823795 (0.036 sec/batch), lr: 0.003000
|
| 199 |
-
2025-12-03 23:25:38 INFO: Finished STEP 500/50000, loss = 4.227708 (0.035 sec/batch), lr: 0.003000
|
| 200 |
-
2025-12-03 23:25:38 INFO: Evaluating on dev set...
|
| 201 |
-
2025-12-03 23:25:38 INFO: LAS MLAS BLEX
|
| 202 |
-
2025-12-03 23:25:38 INFO: 18.07 7.92 9.43
|
| 203 |
-
2025-12-03 23:25:38 INFO: step 500: train_loss = 3.085661, dev_score = 0.1807
|
| 204 |
-
2025-12-03 23:25:39 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 205 |
-
2025-12-03 23:25:39 INFO: new model checkpoint saved.
|
| 206 |
-
2025-12-03 23:25:39 INFO: Finished STEP 520/50000, loss = 3.152939 (0.039 sec/batch), lr: 0.003000
|
| 207 |
-
2025-12-03 23:25:40 INFO: Finished STEP 540/50000, loss = 2.722816 (0.037 sec/batch), lr: 0.003000
|
| 208 |
-
2025-12-03 23:25:41 INFO: Finished STEP 560/50000, loss = 1.797972 (0.034 sec/batch), lr: 0.003000
|
| 209 |
-
2025-12-03 23:25:42 INFO: Finished STEP 580/50000, loss = 1.902476 (0.037 sec/batch), lr: 0.003000
|
| 210 |
-
2025-12-03 23:25:42 INFO: Finished STEP 600/50000, loss = 2.532953 (0.037 sec/batch), lr: 0.003000
|
| 211 |
-
2025-12-03 23:25:42 INFO: Evaluating on dev set...
|
| 212 |
-
2025-12-03 23:25:43 INFO: LAS MLAS BLEX
|
| 213 |
-
2025-12-03 23:25:43 INFO: 24.75 12.64 15.33
|
| 214 |
-
2025-12-03 23:25:43 INFO: step 600: train_loss = 3.006814, dev_score = 0.2475
|
| 215 |
-
2025-12-03 23:25:43 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 216 |
-
2025-12-03 23:25:43 INFO: new best model saved.
|
| 217 |
-
2025-12-03 23:25:44 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 218 |
-
2025-12-03 23:25:44 INFO: new model checkpoint saved.
|
| 219 |
-
2025-12-03 23:25:44 INFO: Finished STEP 620/50000, loss = 3.911072 (0.037 sec/batch), lr: 0.003000
|
| 220 |
-
2025-12-03 23:25:45 INFO: Finished STEP 640/50000, loss = 3.704555 (0.039 sec/batch), lr: 0.003000
|
| 221 |
-
2025-12-03 23:25:46 INFO: Finished STEP 660/50000, loss = 2.692690 (0.037 sec/batch), lr: 0.003000
|
| 222 |
-
2025-12-03 23:25:47 INFO: Finished STEP 680/50000, loss = 2.771069 (0.038 sec/batch), lr: 0.003000
|
| 223 |
-
2025-12-03 23:25:47 INFO: Finished STEP 700/50000, loss = 4.281591 (0.036 sec/batch), lr: 0.003000
|
| 224 |
-
2025-12-03 23:25:47 INFO: Evaluating on dev set...
|
| 225 |
-
2025-12-03 23:25:48 INFO: LAS MLAS BLEX
|
| 226 |
-
2025-12-03 23:25:48 INFO: 34.65 21.99 25.31
|
| 227 |
-
2025-12-03 23:25:48 INFO: step 700: train_loss = 3.216450, dev_score = 0.3465
|
| 228 |
-
2025-12-03 23:25:48 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 229 |
-
2025-12-03 23:25:48 INFO: new best model saved.
|
| 230 |
-
2025-12-03 23:25:49 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 231 |
-
2025-12-03 23:25:49 INFO: new model checkpoint saved.
|
| 232 |
-
2025-12-03 23:25:50 INFO: Finished STEP 720/50000, loss = 3.350647 (0.040 sec/batch), lr: 0.003000
|
| 233 |
-
2025-12-03 23:25:50 INFO: Finished STEP 740/50000, loss = 2.873540 (0.038 sec/batch), lr: 0.003000
|
| 234 |
-
2025-12-03 23:25:51 INFO: Finished STEP 760/50000, loss = 3.564713 (0.038 sec/batch), lr: 0.003000
|
| 235 |
-
2025-12-03 23:25:52 INFO: Finished STEP 780/50000, loss = 3.640228 (0.041 sec/batch), lr: 0.003000
|
| 236 |
-
2025-12-03 23:25:53 INFO: Finished STEP 800/50000, loss = 2.930106 (0.036 sec/batch), lr: 0.003000
|
| 237 |
-
2025-12-03 23:25:53 INFO: Evaluating on dev set...
|
| 238 |
-
2025-12-03 23:25:53 INFO: LAS MLAS BLEX
|
| 239 |
-
2025-12-03 23:25:53 INFO: 24.75 8.60 12.90
|
| 240 |
-
2025-12-03 23:25:53 INFO: step 800: train_loss = 3.300487, dev_score = 0.2475
|
| 241 |
-
2025-12-03 23:25:54 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 242 |
-
2025-12-03 23:25:54 INFO: new model checkpoint saved.
|
| 243 |
-
2025-12-03 23:25:54 INFO: Finished STEP 820/50000, loss = 2.968561 (0.035 sec/batch), lr: 0.003000
|
| 244 |
-
2025-12-03 23:25:55 INFO: Finished STEP 840/50000, loss = 2.491823 (0.037 sec/batch), lr: 0.003000
|
| 245 |
-
2025-12-03 23:25:56 INFO: Finished STEP 860/50000, loss = 3.991972 (0.035 sec/batch), lr: 0.003000
|
| 246 |
-
2025-12-03 23:25:57 INFO: Finished STEP 880/50000, loss = 2.541115 (0.037 sec/batch), lr: 0.003000
|
| 247 |
-
2025-12-03 23:25:58 INFO: Finished STEP 900/50000, loss = 3.015432 (0.038 sec/batch), lr: 0.003000
|
| 248 |
-
2025-12-03 23:25:58 INFO: Evaluating on dev set...
|
| 249 |
-
2025-12-03 23:25:58 INFO: LAS MLAS BLEX
|
| 250 |
-
2025-12-03 23:25:58 INFO: 25.50 11.48 13.52
|
| 251 |
-
2025-12-03 23:25:58 INFO: step 900: train_loss = 3.187365, dev_score = 0.2550
|
| 252 |
-
2025-12-03 23:25:59 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 253 |
-
2025-12-03 23:25:59 INFO: new model checkpoint saved.
|
| 254 |
-
2025-12-03 23:25:59 INFO: Finished STEP 920/50000, loss = 2.031316 (0.038 sec/batch), lr: 0.003000
|
| 255 |
-
2025-12-03 23:26:00 INFO: Finished STEP 940/50000, loss = 2.938839 (0.036 sec/batch), lr: 0.003000
|
| 256 |
-
2025-12-03 23:26:01 INFO: Finished STEP 960/50000, loss = 3.606135 (0.036 sec/batch), lr: 0.003000
|
| 257 |
-
2025-12-03 23:26:02 INFO: Finished STEP 980/50000, loss = 5.427132 (0.038 sec/batch), lr: 0.003000
|
| 258 |
-
2025-12-03 23:26:02 INFO: Finished STEP 1000/50000, loss = 2.710342 (0.039 sec/batch), lr: 0.003000
|
| 259 |
-
2025-12-03 23:26:02 INFO: Evaluating on dev set...
|
| 260 |
-
2025-12-03 23:26:03 INFO: LAS MLAS BLEX
|
| 261 |
-
2025-12-03 23:26:03 INFO: 37.87 21.71 24.22
|
| 262 |
-
2025-12-03 23:26:03 INFO: step 1000: train_loss = 3.299336, dev_score = 0.3787
|
| 263 |
-
2025-12-03 23:26:03 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 264 |
-
2025-12-03 23:26:03 INFO: new best model saved.
|
| 265 |
-
2025-12-03 23:26:04 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 266 |
-
2025-12-03 23:26:04 INFO: new model checkpoint saved.
|
| 267 |
-
2025-12-03 23:26:05 INFO: Finished STEP 1020/50000, loss = 3.274420 (0.040 sec/batch), lr: 0.003000
|
| 268 |
-
2025-12-03 23:26:05 INFO: Finished STEP 1040/50000, loss = 3.172289 (0.037 sec/batch), lr: 0.003000
|
| 269 |
-
2025-12-03 23:26:06 INFO: Finished STEP 1060/50000, loss = 2.884028 (0.040 sec/batch), lr: 0.003000
|
| 270 |
-
2025-12-03 23:26:07 INFO: Finished STEP 1080/50000, loss = 4.205043 (0.038 sec/batch), lr: 0.003000
|
| 271 |
-
2025-12-03 23:26:08 INFO: Finished STEP 1100/50000, loss = 3.608851 (0.039 sec/batch), lr: 0.003000
|
| 272 |
-
2025-12-03 23:26:08 INFO: Evaluating on dev set...
|
| 273 |
-
2025-12-03 23:26:08 INFO: LAS MLAS BLEX
|
| 274 |
-
2025-12-03 23:26:08 INFO: 38.61 24.19 26.78
|
| 275 |
-
2025-12-03 23:26:08 INFO: step 1100: train_loss = 3.198060, dev_score = 0.3861
|
| 276 |
-
2025-12-03 23:26:09 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 277 |
-
2025-12-03 23:26:09 INFO: new best model saved.
|
| 278 |
-
2025-12-03 23:26:09 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 279 |
-
2025-12-03 23:26:09 INFO: new model checkpoint saved.
|
| 280 |
-
2025-12-03 23:26:10 INFO: Finished STEP 1120/50000, loss = 2.438182 (0.039 sec/batch), lr: 0.003000
|
| 281 |
-
2025-12-03 23:26:11 INFO: Finished STEP 1140/50000, loss = 4.315670 (0.038 sec/batch), lr: 0.003000
|
| 282 |
-
2025-12-03 23:26:11 INFO: Finished STEP 1160/50000, loss = 3.651882 (0.038 sec/batch), lr: 0.003000
|
| 283 |
-
2025-12-03 23:26:12 INFO: Finished STEP 1180/50000, loss = 3.139420 (0.037 sec/batch), lr: 0.003000
|
| 284 |
-
2025-12-03 23:26:13 INFO: Finished STEP 1200/50000, loss = 2.040530 (0.038 sec/batch), lr: 0.003000
|
| 285 |
-
2025-12-03 23:26:13 INFO: Evaluating on dev set...
|
| 286 |
-
2025-12-03 23:26:14 INFO: LAS MLAS BLEX
|
| 287 |
-
2025-12-03 23:26:14 INFO: 39.85 23.98 26.98
|
| 288 |
-
2025-12-03 23:26:14 INFO: step 1200: train_loss = 3.227944, dev_score = 0.3985
|
| 289 |
-
2025-12-03 23:26:14 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 290 |
-
2025-12-03 23:26:14 INFO: new best model saved.
|
| 291 |
-
2025-12-03 23:26:14 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 292 |
-
2025-12-03 23:26:14 INFO: new model checkpoint saved.
|
| 293 |
-
2025-12-03 23:26:15 INFO: Finished STEP 1220/50000, loss = 2.697804 (0.041 sec/batch), lr: 0.003000
|
| 294 |
-
2025-12-03 23:26:16 INFO: Finished STEP 1240/50000, loss = 2.911831 (0.038 sec/batch), lr: 0.003000
|
| 295 |
-
2025-12-03 23:26:17 INFO: Finished STEP 1260/50000, loss = 1.945185 (0.040 sec/batch), lr: 0.003000
|
| 296 |
-
2025-12-03 23:26:18 INFO: Finished STEP 1280/50000, loss = 3.025609 (0.038 sec/batch), lr: 0.003000
|
| 297 |
-
2025-12-03 23:26:18 INFO: Finished STEP 1300/50000, loss = 4.230402 (0.037 sec/batch), lr: 0.003000
|
| 298 |
-
2025-12-03 23:26:18 INFO: Evaluating on dev set...
|
| 299 |
-
2025-12-03 23:26:19 INFO: LAS MLAS BLEX
|
| 300 |
-
2025-12-03 23:26:19 INFO: 45.30 30.97 33.55
|
| 301 |
-
2025-12-03 23:26:19 INFO: step 1300: train_loss = 3.111315, dev_score = 0.4530
|
| 302 |
-
2025-12-03 23:26:19 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 303 |
-
2025-12-03 23:26:19 INFO: new best model saved.
|
| 304 |
-
2025-12-03 23:26:20 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 305 |
-
2025-12-03 23:26:20 INFO: new model checkpoint saved.
|
| 306 |
-
2025-12-03 23:26:21 INFO: Finished STEP 1320/50000, loss = 1.966291 (0.038 sec/batch), lr: 0.003000
|
| 307 |
-
2025-12-03 23:26:21 INFO: Finished STEP 1340/50000, loss = 3.216881 (0.040 sec/batch), lr: 0.003000
|
| 308 |
-
2025-12-03 23:26:22 INFO: Finished STEP 1360/50000, loss = 2.379959 (0.038 sec/batch), lr: 0.003000
|
| 309 |
-
2025-12-03 23:26:23 INFO: Finished STEP 1380/50000, loss = 4.992296 (0.037 sec/batch), lr: 0.003000
|
| 310 |
-
2025-12-03 23:26:24 INFO: Finished STEP 1400/50000, loss = 3.349003 (0.042 sec/batch), lr: 0.003000
|
| 311 |
-
2025-12-03 23:26:24 INFO: Evaluating on dev set...
|
| 312 |
-
2025-12-03 23:26:24 INFO: LAS MLAS BLEX
|
| 313 |
-
2025-12-03 23:26:24 INFO: 36.88 23.21 24.89
|
| 314 |
-
2025-12-03 23:26:24 INFO: step 1400: train_loss = 3.364135, dev_score = 0.3688
|
| 315 |
-
2025-12-03 23:26:25 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 316 |
-
2025-12-03 23:26:25 INFO: new model checkpoint saved.
|
| 317 |
-
2025-12-03 23:26:26 INFO: Finished STEP 1420/50000, loss = 4.139513 (0.036 sec/batch), lr: 0.003000
|
| 318 |
-
2025-12-03 23:26:26 INFO: Finished STEP 1440/50000, loss = 2.905265 (0.037 sec/batch), lr: 0.003000
|
| 319 |
-
2025-12-03 23:26:27 INFO: Finished STEP 1460/50000, loss = 3.610150 (0.038 sec/batch), lr: 0.003000
|
| 320 |
-
2025-12-03 23:26:28 INFO: Finished STEP 1480/50000, loss = 4.759534 (0.038 sec/batch), lr: 0.003000
|
| 321 |
-
2025-12-03 23:26:29 INFO: Finished STEP 1500/50000, loss = 2.334270 (0.040 sec/batch), lr: 0.003000
|
| 322 |
-
2025-12-03 23:26:29 INFO: Evaluating on dev set...
|
| 323 |
-
2025-12-03 23:26:29 INFO: LAS MLAS BLEX
|
| 324 |
-
2025-12-03 23:26:29 INFO: 33.66 20.92 32.22
|
| 325 |
-
2025-12-03 23:26:29 INFO: step 1500: train_loss = 3.377781, dev_score = 0.3366
|
| 326 |
-
2025-12-03 23:26:30 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 327 |
-
2025-12-03 23:26:30 INFO: new model checkpoint saved.
|
| 328 |
-
2025-12-03 23:26:31 INFO: Finished STEP 1520/50000, loss = 2.824460 (0.037 sec/batch), lr: 0.003000
|
| 329 |
-
2025-12-03 23:26:32 INFO: Finished STEP 1540/50000, loss = 4.636815 (0.039 sec/batch), lr: 0.003000
|
| 330 |
-
2025-12-03 23:26:32 INFO: Finished STEP 1560/50000, loss = 3.666232 (0.039 sec/batch), lr: 0.003000
|
| 331 |
-
2025-12-03 23:26:33 INFO: Finished STEP 1580/50000, loss = 2.960546 (0.039 sec/batch), lr: 0.003000
|
| 332 |
-
2025-12-03 23:26:34 INFO: Finished STEP 1600/50000, loss = 2.503325 (0.039 sec/batch), lr: 0.003000
|
| 333 |
-
2025-12-03 23:26:34 INFO: Evaluating on dev set...
|
| 334 |
-
2025-12-03 23:26:34 INFO: LAS MLAS BLEX
|
| 335 |
-
2025-12-03 23:26:34 INFO: 42.57 27.91 31.29
|
| 336 |
-
2025-12-03 23:26:34 INFO: step 1600: train_loss = 3.303970, dev_score = 0.4257
|
| 337 |
-
2025-12-03 23:26:35 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 338 |
-
2025-12-03 23:26:35 INFO: new model checkpoint saved.
|
| 339 |
-
2025-12-03 23:26:36 INFO: Finished STEP 1620/50000, loss = 3.455626 (0.042 sec/batch), lr: 0.003000
|
| 340 |
-
2025-12-03 23:26:37 INFO: Finished STEP 1640/50000, loss = 3.333741 (0.038 sec/batch), lr: 0.003000
|
| 341 |
-
2025-12-03 23:26:37 INFO: Finished STEP 1660/50000, loss = 3.761724 (0.041 sec/batch), lr: 0.003000
|
| 342 |
-
2025-12-03 23:26:38 INFO: Finished STEP 1680/50000, loss = 3.171466 (0.039 sec/batch), lr: 0.003000
|
| 343 |
-
2025-12-03 23:26:39 INFO: Finished STEP 1700/50000, loss = 3.169466 (0.040 sec/batch), lr: 0.003000
|
| 344 |
-
2025-12-03 23:26:39 INFO: Evaluating on dev set...
|
| 345 |
-
2025-12-03 23:26:39 INFO: LAS MLAS BLEX
|
| 346 |
-
2025-12-03 23:26:39 INFO: 45.05 33.96 39.41
|
| 347 |
-
2025-12-03 23:26:39 INFO: step 1700: train_loss = 3.350673, dev_score = 0.4505
|
| 348 |
-
2025-12-03 23:26:40 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 349 |
-
2025-12-03 23:26:40 INFO: new model checkpoint saved.
|
| 350 |
-
2025-12-03 23:26:41 INFO: Finished STEP 1720/50000, loss = 2.898433 (0.037 sec/batch), lr: 0.003000
|
| 351 |
-
2025-12-03 23:26:42 INFO: Finished STEP 1740/50000, loss = 4.455222 (0.039 sec/batch), lr: 0.003000
|
| 352 |
-
2025-12-03 23:26:42 INFO: Finished STEP 1760/50000, loss = 2.290793 (0.041 sec/batch), lr: 0.003000
|
| 353 |
-
2025-12-03 23:26:43 INFO: Finished STEP 1780/50000, loss = 3.614108 (0.041 sec/batch), lr: 0.003000
|
| 354 |
-
2025-12-03 23:26:44 INFO: Finished STEP 1800/50000, loss = 2.709010 (0.039 sec/batch), lr: 0.003000
|
| 355 |
-
2025-12-03 23:26:44 INFO: Evaluating on dev set...
|
| 356 |
-
2025-12-03 23:26:45 INFO: LAS MLAS BLEX
|
| 357 |
-
2025-12-03 23:26:45 INFO: 46.04 29.96 33.76
|
| 358 |
-
2025-12-03 23:26:45 INFO: step 1800: train_loss = 3.388469, dev_score = 0.4604
|
| 359 |
-
2025-12-03 23:26:45 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 360 |
-
2025-12-03 23:26:45 INFO: new best model saved.
|
| 361 |
-
2025-12-03 23:26:45 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 362 |
-
2025-12-03 23:26:45 INFO: new model checkpoint saved.
|
| 363 |
-
2025-12-03 23:26:46 INFO: Finished STEP 1820/50000, loss = 3.027663 (0.039 sec/batch), lr: 0.003000
|
| 364 |
-
2025-12-03 23:26:47 INFO: Finished STEP 1840/50000, loss = 4.950170 (0.041 sec/batch), lr: 0.003000
|
| 365 |
-
2025-12-03 23:26:48 INFO: Finished STEP 1860/50000, loss = 3.627461 (0.037 sec/batch), lr: 0.003000
|
| 366 |
-
2025-12-03 23:26:49 INFO: Finished STEP 1880/50000, loss = 3.900440 (0.041 sec/batch), lr: 0.003000
|
| 367 |
-
2025-12-03 23:26:49 INFO: Finished STEP 1900/50000, loss = 3.206836 (0.037 sec/batch), lr: 0.003000
|
| 368 |
-
2025-12-03 23:26:49 INFO: Evaluating on dev set...
|
| 369 |
-
2025-12-03 23:26:50 INFO: LAS MLAS BLEX
|
| 370 |
-
2025-12-03 23:26:50 INFO: 41.09 28.63 34.11
|
| 371 |
-
2025-12-03 23:26:50 INFO: step 1900: train_loss = 3.372791, dev_score = 0.4109
|
| 372 |
-
2025-12-03 23:26:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 373 |
-
2025-12-03 23:26:51 INFO: new model checkpoint saved.
|
| 374 |
-
2025-12-03 23:26:51 INFO: Finished STEP 1920/50000, loss = 4.990382 (0.038 sec/batch), lr: 0.003000
|
| 375 |
-
2025-12-03 23:26:52 INFO: Finished STEP 1940/50000, loss = 3.309713 (0.041 sec/batch), lr: 0.003000
|
| 376 |
-
2025-12-03 23:26:53 INFO: Finished STEP 1960/50000, loss = 3.610244 (0.039 sec/batch), lr: 0.003000
|
| 377 |
-
2025-12-03 23:26:54 INFO: Finished STEP 1980/50000, loss = 3.427943 (0.039 sec/batch), lr: 0.003000
|
| 378 |
-
2025-12-03 23:26:55 INFO: Finished STEP 2000/50000, loss = 3.286536 (0.039 sec/batch), lr: 0.003000
|
| 379 |
-
2025-12-03 23:26:55 INFO: Evaluating on dev set...
|
| 380 |
-
2025-12-03 23:26:55 INFO: LAS MLAS BLEX
|
| 381 |
-
2025-12-03 23:26:55 INFO: 46.04 34.82 37.79
|
| 382 |
-
2025-12-03 23:26:55 INFO: step 2000: train_loss = 3.315330, dev_score = 0.4604
|
| 383 |
-
2025-12-03 23:26:55 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 384 |
-
2025-12-03 23:26:55 INFO: new best model saved.
|
| 385 |
-
2025-12-03 23:26:56 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 386 |
-
2025-12-03 23:26:56 INFO: new model checkpoint saved.
|
| 387 |
-
2025-12-03 23:26:57 INFO: Finished STEP 2020/50000, loss = 2.214467 (0.041 sec/batch), lr: 0.003000
|
| 388 |
-
2025-12-03 23:26:58 INFO: Finished STEP 2040/50000, loss = 3.128998 (0.041 sec/batch), lr: 0.003000
|
| 389 |
-
2025-12-03 23:26:58 INFO: Finished STEP 2060/50000, loss = 3.400111 (0.041 sec/batch), lr: 0.003000
|
| 390 |
-
2025-12-03 23:26:59 INFO: Finished STEP 2080/50000, loss = 5.836899 (0.042 sec/batch), lr: 0.003000
|
| 391 |
-
2025-12-03 23:27:00 INFO: Finished STEP 2100/50000, loss = 2.544196 (0.042 sec/batch), lr: 0.003000
|
| 392 |
-
2025-12-03 23:27:00 INFO: Evaluating on dev set...
|
| 393 |
-
2025-12-03 23:27:01 INFO: LAS MLAS BLEX
|
| 394 |
-
2025-12-03 23:27:01 INFO: 41.34 27.60 36.09
|
| 395 |
-
2025-12-03 23:27:01 INFO: step 2100: train_loss = 3.418321, dev_score = 0.4134
|
| 396 |
-
2025-12-03 23:27:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 397 |
-
2025-12-03 23:27:01 INFO: new model checkpoint saved.
|
| 398 |
-
2025-12-03 23:27:02 INFO: Finished STEP 2120/50000, loss = 4.017645 (0.041 sec/batch), lr: 0.003000
|
| 399 |
-
2025-12-03 23:27:03 INFO: Finished STEP 2140/50000, loss = 4.332951 (0.040 sec/batch), lr: 0.003000
|
| 400 |
-
2025-12-03 23:27:04 INFO: Finished STEP 2160/50000, loss = 2.522452 (0.038 sec/batch), lr: 0.003000
|
| 401 |
-
2025-12-03 23:27:04 INFO: Finished STEP 2180/50000, loss = 3.793148 (0.040 sec/batch), lr: 0.003000
|
| 402 |
-
2025-12-03 23:27:05 INFO: Finished STEP 2200/50000, loss = 2.889349 (0.041 sec/batch), lr: 0.003000
|
| 403 |
-
2025-12-03 23:27:05 INFO: Evaluating on dev set...
|
| 404 |
-
2025-12-03 23:27:06 INFO: LAS MLAS BLEX
|
| 405 |
-
2025-12-03 23:27:06 INFO: 39.60 26.64 35.10
|
| 406 |
-
2025-12-03 23:27:06 INFO: step 2200: train_loss = 3.486956, dev_score = 0.3960
|
| 407 |
-
2025-12-03 23:27:06 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 408 |
-
2025-12-03 23:27:06 INFO: new model checkpoint saved.
|
| 409 |
-
2025-12-03 23:27:07 INFO: Finished STEP 2220/50000, loss = 3.812532 (0.039 sec/batch), lr: 0.003000
|
| 410 |
-
2025-12-03 23:27:08 INFO: Finished STEP 2240/50000, loss = 4.048740 (0.039 sec/batch), lr: 0.003000
|
| 411 |
-
2025-12-03 23:27:09 INFO: Finished STEP 2260/50000, loss = 3.259020 (0.038 sec/batch), lr: 0.003000
|
| 412 |
-
2025-12-03 23:27:10 INFO: Finished STEP 2280/50000, loss = 3.957003 (0.038 sec/batch), lr: 0.003000
|
| 413 |
-
2025-12-03 23:27:10 INFO: Finished STEP 2300/50000, loss = 3.192188 (0.044 sec/batch), lr: 0.003000
|
| 414 |
-
2025-12-03 23:27:10 INFO: Evaluating on dev set...
|
| 415 |
-
2025-12-03 23:27:11 INFO: LAS MLAS BLEX
|
| 416 |
-
2025-12-03 23:27:11 INFO: 43.81 27.85 33.33
|
| 417 |
-
2025-12-03 23:27:11 INFO: step 2300: train_loss = 3.390584, dev_score = 0.4381
|
| 418 |
-
2025-12-03 23:27:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 419 |
-
2025-12-03 23:27:12 INFO: new model checkpoint saved.
|
| 420 |
-
2025-12-03 23:27:12 INFO: Finished STEP 2320/50000, loss = 4.818910 (0.040 sec/batch), lr: 0.003000
|
| 421 |
-
2025-12-03 23:27:13 INFO: Finished STEP 2340/50000, loss = 2.406329 (0.042 sec/batch), lr: 0.003000
|
| 422 |
-
2025-12-03 23:27:14 INFO: Finished STEP 2360/50000, loss = 2.298516 (0.040 sec/batch), lr: 0.003000
|
| 423 |
-
2025-12-03 23:27:15 INFO: Finished STEP 2380/50000, loss = 2.806631 (0.041 sec/batch), lr: 0.003000
|
| 424 |
-
2025-12-03 23:27:16 INFO: Finished STEP 2400/50000, loss = 3.427664 (0.040 sec/batch), lr: 0.003000
|
| 425 |
-
2025-12-03 23:27:16 INFO: Evaluating on dev set...
|
| 426 |
-
2025-12-03 23:27:16 INFO: LAS MLAS BLEX
|
| 427 |
-
2025-12-03 23:27:16 INFO: 43.07 31.09 37.39
|
| 428 |
-
2025-12-03 23:27:16 INFO: step 2400: train_loss = 3.488748, dev_score = 0.4307
|
| 429 |
-
2025-12-03 23:27:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 430 |
-
2025-12-03 23:27:17 INFO: new model checkpoint saved.
|
| 431 |
-
2025-12-03 23:27:18 INFO: Finished STEP 2420/50000, loss = 2.539723 (0.041 sec/batch), lr: 0.003000
|
| 432 |
-
2025-12-03 23:27:19 INFO: Finished STEP 2440/50000, loss = 4.533563 (0.043 sec/batch), lr: 0.003000
|
| 433 |
-
2025-12-03 23:27:19 INFO: Finished STEP 2460/50000, loss = 4.236969 (0.043 sec/batch), lr: 0.003000
|
| 434 |
-
2025-12-03 23:27:20 INFO: Finished STEP 2480/50000, loss = 2.358161 (0.044 sec/batch), lr: 0.003000
|
| 435 |
-
2025-12-03 23:27:21 INFO: Finished STEP 2500/50000, loss = 3.165836 (0.040 sec/batch), lr: 0.003000
|
| 436 |
-
2025-12-03 23:27:21 INFO: Evaluating on dev set...
|
| 437 |
-
2025-12-03 23:27:22 INFO: LAS MLAS BLEX
|
| 438 |
-
2025-12-03 23:27:22 INFO: 46.53 30.44 35.52
|
| 439 |
-
2025-12-03 23:27:22 INFO: step 2500: train_loss = 3.440427, dev_score = 0.4653
|
| 440 |
-
2025-12-03 23:27:22 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 441 |
-
2025-12-03 23:27:22 INFO: new best model saved.
|
| 442 |
-
2025-12-03 23:27:22 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 443 |
-
2025-12-03 23:27:22 INFO: new model checkpoint saved.
|
| 444 |
-
2025-12-03 23:27:23 INFO: Finished STEP 2520/50000, loss = 3.732689 (0.039 sec/batch), lr: 0.003000
|
| 445 |
-
2025-12-03 23:27:24 INFO: Finished STEP 2540/50000, loss = 5.190745 (0.040 sec/batch), lr: 0.003000
|
| 446 |
-
2025-12-03 23:27:25 INFO: Finished STEP 2560/50000, loss = 4.239511 (0.041 sec/batch), lr: 0.003000
|
| 447 |
-
2025-12-03 23:27:26 INFO: Finished STEP 2580/50000, loss = 4.053186 (0.040 sec/batch), lr: 0.003000
|
| 448 |
-
2025-12-03 23:27:27 INFO: Finished STEP 2600/50000, loss = 2.809570 (0.042 sec/batch), lr: 0.003000
|
| 449 |
-
2025-12-03 23:27:27 INFO: Evaluating on dev set...
|
| 450 |
-
2025-12-03 23:27:27 INFO: LAS MLAS BLEX
|
| 451 |
-
2025-12-03 23:27:27 INFO: 49.50 37.55 41.77
|
| 452 |
-
2025-12-03 23:27:27 INFO: step 2600: train_loss = 3.507095, dev_score = 0.4950
|
| 453 |
-
2025-12-03 23:27:27 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 454 |
-
2025-12-03 23:27:27 INFO: new best model saved.
|
| 455 |
-
2025-12-03 23:27:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 456 |
-
2025-12-03 23:27:28 INFO: new model checkpoint saved.
|
| 457 |
-
2025-12-03 23:27:29 INFO: Finished STEP 2620/50000, loss = 4.161043 (0.040 sec/batch), lr: 0.003000
|
| 458 |
-
2025-12-03 23:27:30 INFO: Finished STEP 2640/50000, loss = 3.440454 (0.041 sec/batch), lr: 0.003000
|
| 459 |
-
2025-12-03 23:27:31 INFO: Finished STEP 2660/50000, loss = 3.371374 (0.039 sec/batch), lr: 0.003000
|
| 460 |
-
2025-12-03 23:27:31 INFO: Finished STEP 2680/50000, loss = 2.921980 (0.039 sec/batch), lr: 0.003000
|
| 461 |
-
2025-12-03 23:27:32 INFO: Finished STEP 2700/50000, loss = 2.089964 (0.040 sec/batch), lr: 0.003000
|
| 462 |
-
2025-12-03 23:27:32 INFO: Evaluating on dev set...
|
| 463 |
-
2025-12-03 23:27:33 INFO: LAS MLAS BLEX
|
| 464 |
-
2025-12-03 23:27:33 INFO: 45.54 33.12 36.94
|
| 465 |
-
2025-12-03 23:27:33 INFO: step 2700: train_loss = 3.381599, dev_score = 0.4554
|
| 466 |
-
2025-12-03 23:27:33 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 467 |
-
2025-12-03 23:27:33 INFO: new model checkpoint saved.
|
| 468 |
-
2025-12-03 23:27:34 INFO: Finished STEP 2720/50000, loss = 5.878531 (0.041 sec/batch), lr: 0.003000
|
| 469 |
-
2025-12-03 23:27:35 INFO: Finished STEP 2740/50000, loss = 3.237665 (0.040 sec/batch), lr: 0.003000
|
| 470 |
-
2025-12-03 23:27:36 INFO: Finished STEP 2760/50000, loss = 2.492691 (0.042 sec/batch), lr: 0.003000
|
| 471 |
-
2025-12-03 23:27:37 INFO: Finished STEP 2780/50000, loss = 4.720194 (0.039 sec/batch), lr: 0.003000
|
| 472 |
-
2025-12-03 23:27:38 INFO: Finished STEP 2800/50000, loss = 3.760880 (0.042 sec/batch), lr: 0.003000
|
| 473 |
-
2025-12-03 23:27:38 INFO: Evaluating on dev set...
|
| 474 |
-
2025-12-03 23:27:38 INFO: LAS MLAS BLEX
|
| 475 |
-
2025-12-03 23:27:38 INFO: 51.24 38.24 43.28
|
| 476 |
-
2025-12-03 23:27:38 INFO: step 2800: train_loss = 3.545646, dev_score = 0.5124
|
| 477 |
-
2025-12-03 23:27:38 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 478 |
-
2025-12-03 23:27:38 INFO: new best model saved.
|
| 479 |
-
2025-12-03 23:27:39 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 480 |
-
2025-12-03 23:27:39 INFO: new model checkpoint saved.
|
| 481 |
-
2025-12-03 23:27:40 INFO: Finished STEP 2820/50000, loss = 3.466887 (0.040 sec/batch), lr: 0.003000
|
| 482 |
-
2025-12-03 23:27:41 INFO: Finished STEP 2840/50000, loss = 2.830301 (0.040 sec/batch), lr: 0.003000
|
| 483 |
-
2025-12-03 23:27:42 INFO: Finished STEP 2860/50000, loss = 3.183891 (0.043 sec/batch), lr: 0.003000
|
| 484 |
-
2025-12-03 23:27:43 INFO: Finished STEP 2880/50000, loss = 3.444857 (0.043 sec/batch), lr: 0.003000
|
| 485 |
-
2025-12-03 23:27:43 INFO: Finished STEP 2900/50000, loss = 3.642260 (0.044 sec/batch), lr: 0.003000
|
| 486 |
-
2025-12-03 23:27:43 INFO: Evaluating on dev set...
|
| 487 |
-
2025-12-03 23:27:44 INFO: LAS MLAS BLEX
|
| 488 |
-
2025-12-03 23:27:44 INFO: 53.71 40.59 44.82
|
| 489 |
-
2025-12-03 23:27:44 INFO: step 2900: train_loss = 3.663115, dev_score = 0.5371
|
| 490 |
-
2025-12-03 23:27:44 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 491 |
-
2025-12-03 23:27:44 INFO: new best model saved.
|
| 492 |
-
2025-12-03 23:27:45 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 493 |
-
2025-12-03 23:27:45 INFO: new model checkpoint saved.
|
| 494 |
-
2025-12-03 23:27:46 INFO: Finished STEP 2920/50000, loss = 2.488231 (0.043 sec/batch), lr: 0.003000
|
| 495 |
-
2025-12-03 23:27:46 INFO: Finished STEP 2940/50000, loss = 3.432896 (0.044 sec/batch), lr: 0.003000
|
| 496 |
-
2025-12-03 23:27:47 INFO: Finished STEP 2960/50000, loss = 2.781652 (0.041 sec/batch), lr: 0.003000
|
| 497 |
-
2025-12-03 23:27:48 INFO: Finished STEP 2980/50000, loss = 2.840132 (0.042 sec/batch), lr: 0.003000
|
| 498 |
-
2025-12-03 23:27:49 INFO: Finished STEP 3000/50000, loss = 4.171174 (0.039 sec/batch), lr: 0.003000
|
| 499 |
-
2025-12-03 23:27:49 INFO: Evaluating on dev set...
|
| 500 |
-
2025-12-03 23:27:49 INFO: LAS MLAS BLEX
|
| 501 |
-
2025-12-03 23:27:49 INFO: 51.98 39.50 45.80
|
| 502 |
-
2025-12-03 23:27:49 INFO: step 3000: train_loss = 3.476293, dev_score = 0.5198
|
| 503 |
-
2025-12-03 23:27:50 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 504 |
-
2025-12-03 23:27:50 INFO: new model checkpoint saved.
|
| 505 |
-
2025-12-03 23:27:51 INFO: Finished STEP 3020/50000, loss = 3.225040 (0.043 sec/batch), lr: 0.003000
|
| 506 |
-
2025-12-03 23:27:52 INFO: Finished STEP 3040/50000, loss = 3.064417 (0.042 sec/batch), lr: 0.003000
|
| 507 |
-
2025-12-03 23:27:53 INFO: Finished STEP 3060/50000, loss = 3.886524 (0.041 sec/batch), lr: 0.003000
|
| 508 |
-
2025-12-03 23:27:53 INFO: Finished STEP 3080/50000, loss = 3.704923 (0.041 sec/batch), lr: 0.003000
|
| 509 |
-
2025-12-03 23:27:54 INFO: Finished STEP 3100/50000, loss = 4.191244 (0.042 sec/batch), lr: 0.003000
|
| 510 |
-
2025-12-03 23:27:54 INFO: Evaluating on dev set...
|
| 511 |
-
2025-12-03 23:27:55 INFO: LAS MLAS BLEX
|
| 512 |
-
2025-12-03 23:27:55 INFO: 49.75 37.32 40.67
|
| 513 |
-
2025-12-03 23:27:55 INFO: step 3100: train_loss = 3.466918, dev_score = 0.4975
|
| 514 |
-
2025-12-03 23:27:55 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 515 |
-
2025-12-03 23:27:55 INFO: new model checkpoint saved.
|
| 516 |
-
2025-12-03 23:27:56 INFO: Finished STEP 3120/50000, loss = 3.950969 (0.042 sec/batch), lr: 0.003000
|
| 517 |
-
2025-12-03 23:27:57 INFO: Finished STEP 3140/50000, loss = 2.399348 (0.039 sec/batch), lr: 0.003000
|
| 518 |
-
2025-12-03 23:27:58 INFO: Finished STEP 3160/50000, loss = 3.456824 (0.042 sec/batch), lr: 0.003000
|
| 519 |
-
2025-12-03 23:27:59 INFO: Finished STEP 3180/50000, loss = 3.288764 (0.041 sec/batch), lr: 0.003000
|
| 520 |
-
2025-12-03 23:28:00 INFO: Finished STEP 3200/50000, loss = 3.057923 (0.041 sec/batch), lr: 0.003000
|
| 521 |
-
2025-12-03 23:28:00 INFO: Evaluating on dev set...
|
| 522 |
-
2025-12-03 23:28:00 INFO: LAS MLAS BLEX
|
| 523 |
-
2025-12-03 23:28:00 INFO: 47.77 36.55 40.76
|
| 524 |
-
2025-12-03 23:28:00 INFO: step 3200: train_loss = 3.563295, dev_score = 0.4777
|
| 525 |
-
2025-12-03 23:28:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 526 |
-
2025-12-03 23:28:01 INFO: new model checkpoint saved.
|
| 527 |
-
2025-12-03 23:28:02 INFO: Finished STEP 3220/50000, loss = 4.334808 (0.042 sec/batch), lr: 0.003000
|
| 528 |
-
2025-12-03 23:28:03 INFO: Finished STEP 3240/50000, loss = 2.773743 (0.040 sec/batch), lr: 0.003000
|
| 529 |
-
2025-12-03 23:28:03 INFO: Finished STEP 3260/50000, loss = 3.285001 (0.042 sec/batch), lr: 0.003000
|
| 530 |
-
2025-12-03 23:28:04 INFO: Finished STEP 3280/50000, loss = 3.142590 (0.046 sec/batch), lr: 0.003000
|
| 531 |
-
2025-12-03 23:28:05 INFO: Finished STEP 3300/50000, loss = 2.988616 (0.044 sec/batch), lr: 0.003000
|
| 532 |
-
2025-12-03 23:28:05 INFO: Evaluating on dev set...
|
| 533 |
-
2025-12-03 23:28:06 INFO: LAS MLAS BLEX
|
| 534 |
-
2025-12-03 23:28:06 INFO: 51.98 39.08 41.60
|
| 535 |
-
2025-12-03 23:28:06 INFO: step 3300: train_loss = 3.589203, dev_score = 0.5198
|
| 536 |
-
2025-12-03 23:28:06 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 537 |
-
2025-12-03 23:28:06 INFO: new model checkpoint saved.
|
| 538 |
-
2025-12-03 23:28:07 INFO: Finished STEP 3320/50000, loss = 5.763182 (0.045 sec/batch), lr: 0.003000
|
| 539 |
-
2025-12-03 23:28:08 INFO: Finished STEP 3340/50000, loss = 3.421094 (0.042 sec/batch), lr: 0.003000
|
| 540 |
-
2025-12-03 23:28:09 INFO: Finished STEP 3360/50000, loss = 5.406409 (0.043 sec/batch), lr: 0.003000
|
| 541 |
-
2025-12-03 23:28:10 INFO: Finished STEP 3380/50000, loss = 3.032209 (0.045 sec/batch), lr: 0.003000
|
| 542 |
-
2025-12-03 23:28:11 INFO: Finished STEP 3400/50000, loss = 3.139112 (0.042 sec/batch), lr: 0.003000
|
| 543 |
-
2025-12-03 23:28:11 INFO: Evaluating on dev set...
|
| 544 |
-
2025-12-03 23:28:11 INFO: LAS MLAS BLEX
|
| 545 |
-
2025-12-03 23:28:11 INFO: 49.75 36.29 40.08
|
| 546 |
-
2025-12-03 23:28:11 INFO: step 3400: train_loss = 3.722110, dev_score = 0.4975
|
| 547 |
-
2025-12-03 23:28:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 548 |
-
2025-12-03 23:28:12 INFO: new model checkpoint saved.
|
| 549 |
-
2025-12-03 23:28:13 INFO: Finished STEP 3420/50000, loss = 3.515473 (0.044 sec/batch), lr: 0.003000
|
| 550 |
-
2025-12-03 23:28:14 INFO: Finished STEP 3440/50000, loss = 4.545405 (0.042 sec/batch), lr: 0.003000
|
| 551 |
-
2025-12-03 23:28:15 INFO: Finished STEP 3460/50000, loss = 3.489767 (0.043 sec/batch), lr: 0.003000
|
| 552 |
-
2025-12-03 23:28:15 INFO: Finished STEP 3480/50000, loss = 4.931797 (0.042 sec/batch), lr: 0.003000
|
| 553 |
-
2025-12-03 23:28:16 INFO: Finished STEP 3500/50000, loss = 3.305768 (0.042 sec/batch), lr: 0.003000
|
| 554 |
-
2025-12-03 23:28:16 INFO: Evaluating on dev set...
|
| 555 |
-
2025-12-03 23:28:17 INFO: LAS MLAS BLEX
|
| 556 |
-
2025-12-03 23:28:17 INFO: 48.76 37.05 41.68
|
| 557 |
-
2025-12-03 23:28:17 INFO: step 3500: train_loss = 3.662503, dev_score = 0.4876
|
| 558 |
-
2025-12-03 23:28:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 559 |
-
2025-12-03 23:28:17 INFO: new model checkpoint saved.
|
| 560 |
-
2025-12-03 23:28:18 INFO: Finished STEP 3520/50000, loss = 4.402299 (0.043 sec/batch), lr: 0.003000
|
| 561 |
-
2025-12-03 23:28:19 INFO: Finished STEP 3540/50000, loss = 2.635880 (0.042 sec/batch), lr: 0.003000
|
| 562 |
-
2025-12-03 23:28:20 INFO: Finished STEP 3560/50000, loss = 2.857255 (0.043 sec/batch), lr: 0.003000
|
| 563 |
-
2025-12-03 23:28:21 INFO: Finished STEP 3580/50000, loss = 3.507267 (0.043 sec/batch), lr: 0.003000
|
| 564 |
-
2025-12-03 23:28:22 INFO: Finished STEP 3600/50000, loss = 5.196735 (0.042 sec/batch), lr: 0.003000
|
| 565 |
-
2025-12-03 23:28:22 INFO: Evaluating on dev set...
|
| 566 |
-
2025-12-03 23:28:22 INFO: LAS MLAS BLEX
|
| 567 |
-
2025-12-03 23:28:22 INFO: 51.24 40.84 44.21
|
| 568 |
-
2025-12-03 23:28:22 INFO: step 3600: train_loss = 3.681615, dev_score = 0.5124
|
| 569 |
-
2025-12-03 23:28:23 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 570 |
-
2025-12-03 23:28:23 INFO: new model checkpoint saved.
|
| 571 |
-
2025-12-03 23:28:24 INFO: Finished STEP 3620/50000, loss = 4.539114 (0.042 sec/batch), lr: 0.003000
|
| 572 |
-
2025-12-03 23:28:25 INFO: Finished STEP 3640/50000, loss = 6.413163 (0.042 sec/batch), lr: 0.003000
|
| 573 |
-
2025-12-03 23:28:25 INFO: Finished STEP 3660/50000, loss = 4.375307 (0.041 sec/batch), lr: 0.003000
|
| 574 |
-
2025-12-03 23:28:26 INFO: Finished STEP 3680/50000, loss = 5.120213 (0.041 sec/batch), lr: 0.003000
|
| 575 |
-
2025-12-03 23:28:27 INFO: Finished STEP 3700/50000, loss = 3.109243 (0.044 sec/batch), lr: 0.003000
|
| 576 |
-
2025-12-03 23:28:27 INFO: Evaluating on dev set...
|
| 577 |
-
2025-12-03 23:28:28 INFO: LAS MLAS BLEX
|
| 578 |
-
2025-12-03 23:28:28 INFO: 49.26 38.99 44.03
|
| 579 |
-
2025-12-03 23:28:28 INFO: step 3700: train_loss = 3.778901, dev_score = 0.4926
|
| 580 |
-
2025-12-03 23:28:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 581 |
-
2025-12-03 23:28:28 INFO: new model checkpoint saved.
|
| 582 |
-
2025-12-03 23:28:29 INFO: Finished STEP 3720/50000, loss = 3.468835 (0.045 sec/batch), lr: 0.003000
|
| 583 |
-
2025-12-03 23:28:30 INFO: Finished STEP 3740/50000, loss = 3.420460 (0.044 sec/batch), lr: 0.003000
|
| 584 |
-
2025-12-03 23:28:31 INFO: Finished STEP 3760/50000, loss = 2.856275 (0.042 sec/batch), lr: 0.003000
|
| 585 |
-
2025-12-03 23:28:32 INFO: Finished STEP 3780/50000, loss = 2.668692 (0.041 sec/batch), lr: 0.003000
|
| 586 |
-
2025-12-03 23:28:33 INFO: Finished STEP 3800/50000, loss = 4.511623 (0.044 sec/batch), lr: 0.003000
|
| 587 |
-
2025-12-03 23:28:33 INFO: Evaluating on dev set...
|
| 588 |
-
2025-12-03 23:28:33 INFO: LAS MLAS BLEX
|
| 589 |
-
2025-12-03 23:28:33 INFO: 45.30 35.79 39.16
|
| 590 |
-
2025-12-03 23:28:33 INFO: step 3800: train_loss = 3.920723, dev_score = 0.4530
|
| 591 |
-
2025-12-03 23:28:34 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 592 |
-
2025-12-03 23:28:34 INFO: new model checkpoint saved.
|
| 593 |
-
2025-12-03 23:28:35 INFO: Finished STEP 3820/50000, loss = 4.449274 (0.042 sec/batch), lr: 0.003000
|
| 594 |
-
2025-12-03 23:28:36 INFO: Finished STEP 3840/50000, loss = 2.533618 (0.042 sec/batch), lr: 0.003000
|
| 595 |
-
2025-12-03 23:28:36 INFO: Finished STEP 3860/50000, loss = 3.834488 (0.043 sec/batch), lr: 0.003000
|
| 596 |
-
2025-12-03 23:28:37 INFO: Finished STEP 3880/50000, loss = 3.449773 (0.043 sec/batch), lr: 0.003000
|
| 597 |
-
2025-12-03 23:28:38 INFO: Finished STEP 3900/50000, loss = 3.354862 (0.047 sec/batch), lr: 0.003000
|
| 598 |
-
2025-12-03 23:28:38 INFO: Evaluating on dev set...
|
| 599 |
-
2025-12-03 23:28:39 INFO: LAS MLAS BLEX
|
| 600 |
-
2025-12-03 23:28:39 INFO: 52.48 39.58 44.63
|
| 601 |
-
2025-12-03 23:28:39 INFO: step 3900: train_loss = 3.766196, dev_score = 0.5248
|
| 602 |
-
2025-12-03 23:28:39 INFO: Training ended with 3900 steps.
|
| 603 |
-
2025-12-03 23:28:39 INFO: Best dev F1 = 53.71, at iteration = 2900
|
| 604 |
-
2025-12-03 23:28:39 INFO: Running dev depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
|
| 605 |
-
2025-12-03 23:28:39 INFO: Running parser in predict mode
|
| 606 |
-
2025-12-03 23:28:39 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 607 |
-
2025-12-03 23:28:41 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 608 |
-
2025-12-03 23:28:41 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 609 |
-
2025-12-03 23:28:41 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 610 |
-
2025-12-03 23:28:41 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 611 |
-
2025-12-03 23:28:41 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 612 |
-
2025-12-03 23:28:41 INFO: Loading data with batch size 32...
|
| 613 |
-
2025-12-03 23:28:41 DEBUG: 9 batches created.
|
| 614 |
-
2025-12-03 23:28:42 INFO: F1 scores for each dependency:
|
| 615 |
-
Note that unlabeled attachment errors hurt the labeled attachment scores
|
| 616 |
-
acl: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
|
| 617 |
-
acl:relcl: p 0.1667 r 0.2857 f1 0.2105 (7 actual)
|
| 618 |
-
advcl: p 0.0526 r 0.2000 f1 0.0833 (5 actual)
|
| 619 |
-
advmod: p 0.4839 r 0.6000 f1 0.5357 (25 actual)
|
| 620 |
-
amod: p 0.8148 r 0.7097 f1 0.7586 (31 actual)
|
| 621 |
-
appos: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 622 |
-
aux: p 0.8889 r 0.7273 f1 0.8000 (11 actual)
|
| 623 |
-
case: p 0.7544 r 0.7679 f1 0.7611 (56 actual)
|
| 624 |
-
cc: p 0.7692 r 0.7692 f1 0.7692 (13 actual)
|
| 625 |
-
ccomp: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 626 |
-
conj: p 0.5000 r 0.0833 f1 0.1429 (12 actual)
|
| 627 |
-
cop: p 0.2500 r 0.3333 f1 0.2857 (3 actual)
|
| 628 |
-
csubj: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 629 |
-
det: p 0.8696 r 0.9091 f1 0.8889 (22 actual)
|
| 630 |
-
expl: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 631 |
-
iobj: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 632 |
-
mark: p 0.5000 r 0.3333 f1 0.4000 (12 actual)
|
| 633 |
-
nmod: p 0.2609 r 0.4000 f1 0.3158 (15 actual)
|
| 634 |
-
nmod:poss: p 1.0000 r 0.8947 f1 0.9444 (19 actual)
|
| 635 |
-
nsubj: p 0.3421 r 0.7647 f1 0.4727 (17 actual)
|
| 636 |
-
nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
|
| 637 |
-
obj: p 0.6667 r 0.2727 f1 0.3871 (22 actual)
|
| 638 |
-
obl: p 0.3333 r 0.4146 f1 0.3696 (41 actual)
|
| 639 |
-
obl:agent: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 640 |
-
orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 641 |
-
parataxis: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
|
| 642 |
-
punct: p 0.4808 r 0.4808 f1 0.4808 (52 actual)
|
| 643 |
-
root: p 0.4444 r 0.4444 f1 0.4444 (9 actual)
|
| 644 |
-
xcomp: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
|
| 645 |
-
2025-12-03 23:28:42 INFO: LAS MLAS BLEX
|
| 646 |
-
2025-12-03 23:28:42 INFO: 53.71 40.59 44.82
|
| 647 |
-
2025-12-03 23:28:42 INFO: Parser score:
|
| 648 |
-
2025-12-03 23:28:42 INFO: sv_diachronic 53.71
|
| 649 |
-
2025-12-03 23:28:42 INFO: Finished running dev set on
|
| 650 |
-
UD_Swedish-diachronic
|
| 651 |
-
UAS LAS CLAS MLAS BLEX
|
| 652 |
-
66.58 53.71 44.82 40.59 44.82
|
| 653 |
-
2025-12-03 23:28:42 INFO: Running test depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
|
| 654 |
-
2025-12-03 23:28:42 INFO: Running parser in predict mode
|
| 655 |
-
2025-12-03 23:28:42 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 656 |
-
2025-12-03 23:28:44 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 657 |
-
2025-12-03 23:28:44 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 658 |
-
2025-12-03 23:28:44 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 659 |
-
2025-12-03 23:28:44 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 660 |
-
2025-12-03 23:28:44 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 661 |
-
2025-12-03 23:28:44 INFO: Loading data with batch size 32...
|
| 662 |
-
2025-12-03 23:28:44 DEBUG: 93 batches created.
|
| 663 |
-
2025-12-03 23:28:49 INFO: F1 scores for each dependency:
|
| 664 |
-
Note that unlabeled attachment errors hurt the labeled attachment scores
|
| 665 |
-
acl: p 0.0000 r 0.0000 f1 0.0000 (32 actual)
|
| 666 |
-
acl:cleft: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 667 |
-
acl:relcl: p 0.1167 r 0.0933 f1 0.1037 (75 actual)
|
| 668 |
-
advcl: p 0.0544 r 0.2667 f1 0.0904 (60 actual)
|
| 669 |
-
advcl:relcl: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 670 |
-
advmod: p 0.5083 r 0.5746 f1 0.5394 (268 actual)
|
| 671 |
-
amod: p 0.6795 r 0.6913 f1 0.6853 (230 actual)
|
| 672 |
-
appos: p 0.0000 r 0.0000 f1 0.0000 (13 actual)
|
| 673 |
-
aux: p 0.7556 r 0.8095 f1 0.7816 (84 actual)
|
| 674 |
-
aux:pass: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 675 |
-
case: p 0.7245 r 0.7051 f1 0.7147 (373 actual)
|
| 676 |
-
cc: p 0.5759 r 0.5871 f1 0.5815 (155 actual)
|
| 677 |
-
ccomp: p 0.0000 r 0.0000 f1 0.0000 (35 actual)
|
| 678 |
-
compound:prt: p 0.0000 r 0.0000 f1 0.0000 (21 actual)
|
| 679 |
-
conj: p 0.1053 r 0.0253 f1 0.0408 (158 actual)
|
| 680 |
-
cop: p 0.7619 r 0.3478 f1 0.4776 (46 actual)
|
| 681 |
-
csubj: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 682 |
-
dep: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 683 |
-
det: p 0.7833 r 0.7644 f1 0.7737 (208 actual)
|
| 684 |
-
discourse: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
|
| 685 |
-
dislocated: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 686 |
-
expl: p 0.0000 r 0.0000 f1 0.0000 (11 actual)
|
| 687 |
-
expl:pv: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 688 |
-
fixed: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
|
| 689 |
-
flat: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 690 |
-
flat:name: p 0.0000 r 0.0000 f1 0.0000 (12 actual)
|
| 691 |
-
goeswith: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 692 |
-
iobj: p 0.0000 r 0.0000 f1 0.0000 (14 actual)
|
| 693 |
-
mark: p 0.6569 r 0.5882 f1 0.6207 (153 actual)
|
| 694 |
-
nmod: p 0.2320 r 0.2843 f1 0.2555 (102 actual)
|
| 695 |
-
nmod:poss: p 0.8102 r 0.7817 f1 0.7957 (142 actual)
|
| 696 |
-
nsubj: p 0.3947 r 0.6429 f1 0.4891 (280 actual)
|
| 697 |
-
nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (25 actual)
|
| 698 |
-
nummod: p 0.0000 r 0.0000 f1 0.0000 (10 actual)
|
| 699 |
-
obj: p 0.6111 r 0.1803 f1 0.2785 (183 actual)
|
| 700 |
-
obl: p 0.2623 r 0.4029 f1 0.3177 (278 actual)
|
| 701 |
-
obl:agent: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 702 |
-
orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 703 |
-
parataxis: p 0.0000 r 0.0000 f1 0.0000 (18 actual)
|
| 704 |
-
punct: p 0.3852 r 0.3906 f1 0.3879 (425 actual)
|
| 705 |
-
reparandum: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 706 |
-
root: p 0.5253 r 0.5253 f1 0.5253 (99 actual)
|
| 707 |
-
vocative: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
|
| 708 |
-
xcomp: p 0.0000 r 0.0000 f1 0.0000 (75 actual)
|
| 709 |
-
2025-12-03 23:28:49 INFO: LAS MLAS BLEX
|
| 710 |
-
2025-12-03 23:28:49 INFO: 47.60 36.14 39.67
|
| 711 |
-
2025-12-03 23:28:49 INFO: Parser score:
|
| 712 |
-
2025-12-03 23:28:49 INFO: sv_diachronic 47.60
|
| 713 |
-
2025-12-03 23:28:49 INFO: Finished running test set on
|
| 714 |
-
UD_Swedish-diachronic
|
| 715 |
-
UAS LAS CLAS MLAS BLEX
|
| 716 |
-
62.40 47.60 39.67 36.14 39.67
|
| 717 |
-
DONE.
|
| 718 |
-
Full log saved to: logs/log_conll17.pt_sv_20251203_232257.txt
|
| 719 |
-
Symlink updated: logs/latest.txt → log_conll17.pt_sv_20251203_232257.txt
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_conll17.pt_sv_20260218_190034.txt
DELETED
|
@@ -1,120 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_20260218_190034.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-SILVER.py sv
|
| 6 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 7 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 8 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 9 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 10 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 11 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 12 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 13 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 14 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 15 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 16 |
-
Loading DigPhil MACHINE for SILVER (minus gold overlap)…
|
| 17 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
|
| 18 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 19 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 20 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
|
| 21 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 22 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 23 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 24 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 25 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 26 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 27 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 28 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 29 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 30 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 31 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
|
| 32 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
|
| 33 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 34 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
|
| 35 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
|
| 36 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
|
| 37 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
|
| 38 |
-
Adding 75 remaining gold sentences to TRAIN
|
| 39 |
-
Cleaning TRAIN...
|
| 40 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 41 |
-
Cleaning SILVER...
|
| 42 |
-
[REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
|
| 43 |
-
[REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
|
| 44 |
-
[REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
|
| 45 |
-
[REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
|
| 46 |
-
[REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
|
| 47 |
-
[REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
|
| 48 |
-
[REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
|
| 49 |
-
[REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
|
| 50 |
-
[REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
|
| 51 |
-
[REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
|
| 52 |
-
[REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
|
| 53 |
-
[REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
|
| 54 |
-
[REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
|
| 55 |
-
[REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
|
| 56 |
-
[REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
|
| 57 |
-
[REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
|
| 58 |
-
[REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
|
| 59 |
-
[REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
|
| 60 |
-
[REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
|
| 61 |
-
[REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
|
| 62 |
-
[REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
|
| 63 |
-
[REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
|
| 64 |
-
[REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
|
| 65 |
-
[REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
|
| 66 |
-
[REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
|
| 67 |
-
[REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
|
| 68 |
-
[REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
|
| 69 |
-
[REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
|
| 70 |
-
[REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
|
| 71 |
-
[REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
|
| 72 |
-
[REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
|
| 73 |
-
[REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
|
| 74 |
-
[REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
|
| 75 |
-
[REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
|
| 76 |
-
[REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
|
| 77 |
-
[REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
|
| 78 |
-
[REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
|
| 79 |
-
[REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
|
| 80 |
-
[REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
|
| 81 |
-
[REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
|
| 82 |
-
[REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
|
| 83 |
-
[REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
|
| 84 |
-
[REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
|
| 85 |
-
[REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
|
| 86 |
-
[REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
|
| 87 |
-
[REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
|
| 88 |
-
[REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
|
| 89 |
-
[REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
|
| 90 |
-
[REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
|
| 91 |
-
[REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
|
| 92 |
-
[REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
|
| 93 |
-
[REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
|
| 94 |
-
[REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
|
| 95 |
-
[REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
|
| 96 |
-
[REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
|
| 97 |
-
[REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
|
| 98 |
-
[REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
|
| 99 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
|
| 100 |
-
[REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
|
| 101 |
-
[REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
|
| 102 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
|
| 103 |
-
[REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
|
| 104 |
-
[REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
|
| 105 |
-
Cleaning DEV...
|
| 106 |
-
Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13868 valid sentences)
|
| 107 |
-
Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (46432 valid sentences)
|
| 108 |
-
Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (34 valid sentences)
|
| 109 |
-
Done.
|
| 110 |
-
Sourcing scripts/config.sh
|
| 111 |
-
Running stanza dataset preparation…
|
| 112 |
-
2026-02-18 19:00:38 INFO: Datasets program called with:
|
| 113 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --use_bert --bert_model vesteinn/ScandiBERT --silver_file ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu
|
| 114 |
-
usage: python -m stanza.utils.datasets.prepare_depparse_treebank
|
| 115 |
-
[-h] [--gold] [--predicted]
|
| 116 |
-
[--wordvec_pretrain_file WORDVEC_PRETRAIN_FILE]
|
| 117 |
-
[--tagger_model TAGGER_MODEL] [--save_dir SAVE_DIR]
|
| 118 |
-
[--no_download_tagger] [--charlm CHARLM] [--no_charlm]
|
| 119 |
-
treebanks [treebanks ...]
|
| 120 |
-
python -m stanza.utils.datasets.prepare_depparse_treebank: error: unrecognized arguments: --use_bert --bert_model vesteinn/ScandiBERT --silver_file ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_conll17.pt_sv_20260218_190242.txt
DELETED
|
@@ -1,143 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_20260218_190242.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-SILVER.py sv
|
| 6 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 7 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 8 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 9 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 10 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 11 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 12 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 13 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 14 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 15 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 16 |
-
Loading DigPhil MACHINE for SILVER (minus gold overlap)…
|
| 17 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
|
| 18 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 19 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 20 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
|
| 21 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 22 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 23 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 24 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 25 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 26 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 27 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 28 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 29 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 30 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 31 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
|
| 32 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
|
| 33 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 34 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
|
| 35 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
|
| 36 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
|
| 37 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
|
| 38 |
-
Adding 75 remaining gold sentences to TRAIN
|
| 39 |
-
Cleaning TRAIN...
|
| 40 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 41 |
-
Cleaning SILVER...
|
| 42 |
-
[REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
|
| 43 |
-
[REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
|
| 44 |
-
[REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
|
| 45 |
-
[REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
|
| 46 |
-
[REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
|
| 47 |
-
[REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
|
| 48 |
-
[REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
|
| 49 |
-
[REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
|
| 50 |
-
[REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
|
| 51 |
-
[REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
|
| 52 |
-
[REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
|
| 53 |
-
[REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
|
| 54 |
-
[REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
|
| 55 |
-
[REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
|
| 56 |
-
[REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
|
| 57 |
-
[REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
|
| 58 |
-
[REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
|
| 59 |
-
[REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
|
| 60 |
-
[REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
|
| 61 |
-
[REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
|
| 62 |
-
[REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
|
| 63 |
-
[REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
|
| 64 |
-
[REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
|
| 65 |
-
[REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
|
| 66 |
-
[REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
|
| 67 |
-
[REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
|
| 68 |
-
[REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
|
| 69 |
-
[REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
|
| 70 |
-
[REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
|
| 71 |
-
[REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
|
| 72 |
-
[REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
|
| 73 |
-
[REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
|
| 74 |
-
[REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
|
| 75 |
-
[REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
|
| 76 |
-
[REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
|
| 77 |
-
[REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
|
| 78 |
-
[REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
|
| 79 |
-
[REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
|
| 80 |
-
[REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
|
| 81 |
-
[REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
|
| 82 |
-
[REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
|
| 83 |
-
[REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
|
| 84 |
-
[REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
|
| 85 |
-
[REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
|
| 86 |
-
[REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
|
| 87 |
-
[REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
|
| 88 |
-
[REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
|
| 89 |
-
[REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
|
| 90 |
-
[REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
|
| 91 |
-
[REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
|
| 92 |
-
[REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
|
| 93 |
-
[REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
|
| 94 |
-
[REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
|
| 95 |
-
[REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
|
| 96 |
-
[REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
|
| 97 |
-
[REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
|
| 98 |
-
[REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
|
| 99 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
|
| 100 |
-
[REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
|
| 101 |
-
[REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
|
| 102 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
|
| 103 |
-
[REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
|
| 104 |
-
[REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
|
| 105 |
-
Cleaning DEV...
|
| 106 |
-
Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13868 valid sentences)
|
| 107 |
-
Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (46432 valid sentences)
|
| 108 |
-
Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (34 valid sentences)
|
| 109 |
-
Done.
|
| 110 |
-
Sourcing scripts/config.sh
|
| 111 |
-
Running stanza dataset preparation…
|
| 112 |
-
2026-02-18 19:02:45 INFO: Datasets program called with:
|
| 113 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic
|
| 114 |
-
2026-02-18 19:02:45 DEBUG: Looking for tagger for lang |sv| dataset |diachronic|. Suggested model |None|. Looking first in |saved_models/pos|.
|
| 115 |
-
2026-02-18 19:02:45 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
|
| 116 |
-
|
| 117 |
-
2026-02-18 19:02:46 INFO: Downloaded file to /home/urdatorn/stanza_resources/resources.json
|
| 118 |
-
2026-02-18 19:02:46 DEBUG: Processing parameter "processors"...
|
| 119 |
-
2026-02-18 19:02:46 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
|
| 120 |
-
2026-02-18 19:02:46 INFO: Downloading these customized packages for language: sv (Swedish)...
|
| 121 |
-
=======================
|
| 122 |
-
| Processor | Package |
|
| 123 |
-
-----------------------
|
| 124 |
-
=======================
|
| 125 |
-
|
| 126 |
-
2026-02-18 19:02:46 INFO: Finished downloading models and saved to /home/urdatorn/stanza_resources
|
| 127 |
-
Traceback (most recent call last):
|
| 128 |
-
File "<frozen runpy>", line 198, in _run_module_as_main
|
| 129 |
-
File "<frozen runpy>", line 88, in _run_code
|
| 130 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 151, in <module>
|
| 131 |
-
main()
|
| 132 |
-
~~~~^^
|
| 133 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 147, in main
|
| 134 |
-
common.main(process_treebank, common.ModelType.DEPPARSE, add_specific_args)
|
| 135 |
-
~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 136 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 335, in main
|
| 137 |
-
process_treebank(treebank, model_type, paths, args)
|
| 138 |
-
~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 139 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 109, in process_treebank
|
| 140 |
-
tagger_model = choose_tagger_model(short_language, dataset, args.tagger_model, args)
|
| 141 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 90, in choose_tagger_model
|
| 142 |
-
raise FileNotFoundError("Could not figure out which model file to use for %s. Just tried to download to %s the models %s" % (short_language, args.save_dir, download_list))
|
| 143 |
-
FileNotFoundError: Could not figure out which model file to use for sv. Just tried to download to saved_models/pos the models []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_conll17.pt_sv_20260218_190400.txt
DELETED
|
@@ -1,38 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_20260218_190400.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Sourcing scripts/config.sh
|
| 6 |
-
Running stanza dataset preparation…
|
| 7 |
-
2026-02-18 19:04:02 INFO: Datasets program called with:
|
| 8 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic
|
| 9 |
-
2026-02-18 19:04:02 DEBUG: Looking for tagger for lang |sv| dataset |diachronic|. Suggested model |None|. Looking first in |saved_models/pos|.
|
| 10 |
-
2026-02-18 19:04:02 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
|
| 11 |
-
|
| 12 |
-
2026-02-18 19:04:02 INFO: Downloaded file to /home/urdatorn/stanza_resources/resources.json
|
| 13 |
-
2026-02-18 19:04:02 DEBUG: Processing parameter "processors"...
|
| 14 |
-
2026-02-18 19:04:02 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
|
| 15 |
-
2026-02-18 19:04:02 INFO: Downloading these customized packages for language: sv (Swedish)...
|
| 16 |
-
=======================
|
| 17 |
-
| Processor | Package |
|
| 18 |
-
-----------------------
|
| 19 |
-
=======================
|
| 20 |
-
|
| 21 |
-
2026-02-18 19:04:02 INFO: Finished downloading models and saved to /home/urdatorn/stanza_resources
|
| 22 |
-
Traceback (most recent call last):
|
| 23 |
-
File "<frozen runpy>", line 198, in _run_module_as_main
|
| 24 |
-
File "<frozen runpy>", line 88, in _run_code
|
| 25 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 151, in <module>
|
| 26 |
-
main()
|
| 27 |
-
~~~~^^
|
| 28 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 147, in main
|
| 29 |
-
common.main(process_treebank, common.ModelType.DEPPARSE, add_specific_args)
|
| 30 |
-
~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 31 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 335, in main
|
| 32 |
-
process_treebank(treebank, model_type, paths, args)
|
| 33 |
-
~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 34 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 109, in process_treebank
|
| 35 |
-
tagger_model = choose_tagger_model(short_language, dataset, args.tagger_model, args)
|
| 36 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 90, in choose_tagger_model
|
| 37 |
-
raise FileNotFoundError("Could not figure out which model file to use for %s. Just tried to download to %s the models %s" % (short_language, args.save_dir, download_list))
|
| 38 |
-
FileNotFoundError: Could not figure out which model file to use for sv. Just tried to download to saved_models/pos the models []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_conll17.pt_sv_20260218_190808.txt
DELETED
|
@@ -1,38 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_20260218_190808.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Sourcing scripts/config.sh
|
| 6 |
-
Running stanza dataset preparation…
|
| 7 |
-
2026-02-18 19:08:10 INFO: Datasets program called with:
|
| 8 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic
|
| 9 |
-
2026-02-18 19:08:10 DEBUG: Looking for tagger for lang |sv| dataset |diachronic|. Suggested model |None|. Looking first in |saved_models/pos|.
|
| 10 |
-
2026-02-18 19:08:10 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
|
| 11 |
-
|
| 12 |
-
2026-02-18 19:08:11 INFO: Downloaded file to /home/urdatorn/stanza_resources/resources.json
|
| 13 |
-
2026-02-18 19:08:11 DEBUG: Processing parameter "processors"...
|
| 14 |
-
2026-02-18 19:08:11 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
|
| 15 |
-
2026-02-18 19:08:11 INFO: Downloading these customized packages for language: sv (Swedish)...
|
| 16 |
-
=======================
|
| 17 |
-
| Processor | Package |
|
| 18 |
-
-----------------------
|
| 19 |
-
=======================
|
| 20 |
-
|
| 21 |
-
2026-02-18 19:08:11 INFO: Finished downloading models and saved to /home/urdatorn/stanza_resources
|
| 22 |
-
Traceback (most recent call last):
|
| 23 |
-
File "<frozen runpy>", line 198, in _run_module_as_main
|
| 24 |
-
File "<frozen runpy>", line 88, in _run_code
|
| 25 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 151, in <module>
|
| 26 |
-
main()
|
| 27 |
-
~~~~^^
|
| 28 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 147, in main
|
| 29 |
-
common.main(process_treebank, common.ModelType.DEPPARSE, add_specific_args)
|
| 30 |
-
~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 31 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 335, in main
|
| 32 |
-
process_treebank(treebank, model_type, paths, args)
|
| 33 |
-
~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 34 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 109, in process_treebank
|
| 35 |
-
tagger_model = choose_tagger_model(short_language, dataset, args.tagger_model, args)
|
| 36 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 90, in choose_tagger_model
|
| 37 |
-
raise FileNotFoundError("Could not figure out which model file to use for %s. Just tried to download to %s the models %s" % (short_language, args.save_dir, download_list))
|
| 38 |
-
FileNotFoundError: Could not figure out which model file to use for sv. Just tried to download to saved_models/pos the models []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_conll17.pt_sv_20260218_190956.txt
DELETED
|
@@ -1,38 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_20260218_190956.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Sourcing scripts/config.sh
|
| 6 |
-
Running stanza dataset preparation…
|
| 7 |
-
2026-02-18 19:09:58 INFO: Datasets program called with:
|
| 8 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
|
| 9 |
-
2026-02-18 19:09:58 DEBUG: Looking for tagger for lang |sv| dataset |diachronic|. Suggested model |None|. Looking first in |saved_models/pos|.
|
| 10 |
-
2026-02-18 19:09:58 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
|
| 11 |
-
|
| 12 |
-
2026-02-18 19:09:58 INFO: Downloaded file to /home/urdatorn/stanza_resources/resources.json
|
| 13 |
-
2026-02-18 19:09:58 DEBUG: Processing parameter "processors"...
|
| 14 |
-
2026-02-18 19:09:58 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
|
| 15 |
-
2026-02-18 19:09:58 INFO: Downloading these customized packages for language: sv (Swedish)...
|
| 16 |
-
=======================
|
| 17 |
-
| Processor | Package |
|
| 18 |
-
-----------------------
|
| 19 |
-
=======================
|
| 20 |
-
|
| 21 |
-
2026-02-18 19:09:58 INFO: Finished downloading models and saved to /home/urdatorn/stanza_resources
|
| 22 |
-
Traceback (most recent call last):
|
| 23 |
-
File "<frozen runpy>", line 198, in _run_module_as_main
|
| 24 |
-
File "<frozen runpy>", line 88, in _run_code
|
| 25 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 151, in <module>
|
| 26 |
-
main()
|
| 27 |
-
~~~~^^
|
| 28 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 147, in main
|
| 29 |
-
common.main(process_treebank, common.ModelType.DEPPARSE, add_specific_args)
|
| 30 |
-
~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 31 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 335, in main
|
| 32 |
-
process_treebank(treebank, model_type, paths, args)
|
| 33 |
-
~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 34 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 109, in process_treebank
|
| 35 |
-
tagger_model = choose_tagger_model(short_language, dataset, args.tagger_model, args)
|
| 36 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 90, in choose_tagger_model
|
| 37 |
-
raise FileNotFoundError("Could not figure out which model file to use for %s. Just tried to download to %s the models %s" % (short_language, args.save_dir, download_list))
|
| 38 |
-
FileNotFoundError: Could not figure out which model file to use for sv. Just tried to download to saved_models/pos the models []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_conll17.pt_sv_20260218_191305.txt
DELETED
|
@@ -1,38 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_20260218_191305.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Sourcing scripts/config.sh
|
| 6 |
-
Running stanza dataset preparation…
|
| 7 |
-
2026-02-18 19:13:07 INFO: Datasets program called with:
|
| 8 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
|
| 9 |
-
2026-02-18 19:13:07 DEBUG: Looking for tagger for lang |sv| dataset |diachronic|. Suggested model |None|. Looking first in |saved_models/pos|.
|
| 10 |
-
2026-02-18 19:13:07 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
|
| 11 |
-
|
| 12 |
-
2026-02-18 19:13:07 INFO: Downloaded file to /home/urdatorn/stanza_resources/resources.json
|
| 13 |
-
2026-02-18 19:13:07 DEBUG: Processing parameter "processors"...
|
| 14 |
-
2026-02-18 19:13:07 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
|
| 15 |
-
2026-02-18 19:13:07 INFO: Downloading these customized packages for language: sv (Swedish)...
|
| 16 |
-
=======================
|
| 17 |
-
| Processor | Package |
|
| 18 |
-
-----------------------
|
| 19 |
-
=======================
|
| 20 |
-
|
| 21 |
-
2026-02-18 19:13:07 INFO: Finished downloading models and saved to /home/urdatorn/stanza_resources
|
| 22 |
-
Traceback (most recent call last):
|
| 23 |
-
File "<frozen runpy>", line 198, in _run_module_as_main
|
| 24 |
-
File "<frozen runpy>", line 88, in _run_code
|
| 25 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 151, in <module>
|
| 26 |
-
main()
|
| 27 |
-
~~~~^^
|
| 28 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 147, in main
|
| 29 |
-
common.main(process_treebank, common.ModelType.DEPPARSE, add_specific_args)
|
| 30 |
-
~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 31 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 335, in main
|
| 32 |
-
process_treebank(treebank, model_type, paths, args)
|
| 33 |
-
~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 34 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 109, in process_treebank
|
| 35 |
-
tagger_model = choose_tagger_model(short_language, dataset, args.tagger_model, args)
|
| 36 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 90, in choose_tagger_model
|
| 37 |
-
raise FileNotFoundError("Could not figure out which model file to use for %s. Just tried to download to %s the models %s" % (short_language, args.save_dir, download_list))
|
| 38 |
-
FileNotFoundError: Could not figure out which model file to use for sv. Just tried to download to saved_models/pos the models []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_conll17.pt_sv_20260218_191719.txt
DELETED
|
@@ -1,38 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_20260218_191719.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Sourcing scripts/config.sh
|
| 6 |
-
Running stanza dataset preparation…
|
| 7 |
-
2026-02-18 19:17:21 INFO: Datasets program called with:
|
| 8 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
|
| 9 |
-
2026-02-18 19:17:21 DEBUG: Looking for tagger for lang |sv| dataset |diachronic|. Suggested model |None|. Looking first in |saved_models/pos|.
|
| 10 |
-
2026-02-18 19:17:21 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
|
| 11 |
-
|
| 12 |
-
2026-02-18 19:17:21 INFO: Downloaded file to /home/urdatorn/stanza_resources/resources.json
|
| 13 |
-
2026-02-18 19:17:21 DEBUG: Processing parameter "processors"...
|
| 14 |
-
2026-02-18 19:17:21 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
|
| 15 |
-
2026-02-18 19:17:21 INFO: Downloading these customized packages for language: sv (Swedish)...
|
| 16 |
-
=======================
|
| 17 |
-
| Processor | Package |
|
| 18 |
-
-----------------------
|
| 19 |
-
=======================
|
| 20 |
-
|
| 21 |
-
2026-02-18 19:17:21 INFO: Finished downloading models and saved to /home/urdatorn/stanza_resources
|
| 22 |
-
Traceback (most recent call last):
|
| 23 |
-
File "<frozen runpy>", line 198, in _run_module_as_main
|
| 24 |
-
File "<frozen runpy>", line 88, in _run_code
|
| 25 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 151, in <module>
|
| 26 |
-
main()
|
| 27 |
-
~~~~^^
|
| 28 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 147, in main
|
| 29 |
-
common.main(process_treebank, common.ModelType.DEPPARSE, add_specific_args)
|
| 30 |
-
~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 31 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 335, in main
|
| 32 |
-
process_treebank(treebank, model_type, paths, args)
|
| 33 |
-
~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 34 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 109, in process_treebank
|
| 35 |
-
tagger_model = choose_tagger_model(short_language, dataset, args.tagger_model, args)
|
| 36 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 90, in choose_tagger_model
|
| 37 |
-
raise FileNotFoundError("Could not figure out which model file to use for %s. Just tried to download to %s the models %s" % (short_language, args.save_dir, download_list))
|
| 38 |
-
FileNotFoundError: Could not figure out which model file to use for sv. Just tried to download to saved_models/pos the models []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_conll17.pt_sv_20260219_104841.txt
DELETED
|
@@ -1,44 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_20260219_104841.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Sourcing scripts/config.sh
|
| 6 |
-
Running stanza dataset preparation…
|
| 7 |
-
2026-02-19 10:48:43 INFO: Datasets program called with:
|
| 8 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt --gold
|
| 9 |
-
Augmented 60 quotes: Counter({'《》': 9, '»«': 8, '″″': 8, '““': 7, '„”': 6, '""': 6, '„“': 6, '«»': 4, '””': 4, '「」': 2})
|
| 10 |
-
Traceback (most recent call last):
|
| 11 |
-
File "<frozen runpy>", line 198, in _run_module_as_main
|
| 12 |
-
File "<frozen runpy>", line 88, in _run_code
|
| 13 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 151, in <module>
|
| 14 |
-
main()
|
| 15 |
-
~~~~^^
|
| 16 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 147, in main
|
| 17 |
-
common.main(process_treebank, common.ModelType.DEPPARSE, add_specific_args)
|
| 18 |
-
~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 19 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 335, in main
|
| 20 |
-
process_treebank(treebank, model_type, paths, args)
|
| 21 |
-
~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 22 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py", line 96, in process_treebank
|
| 23 |
-
prepare_tokenizer_treebank.copy_conllu_treebank(treebank, model_type, paths, paths["DEPPARSE_DATA_DIR"])
|
| 24 |
-
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 25 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_tokenizer_treebank.py", line 75, in copy_conllu_treebank
|
| 26 |
-
process_treebank(treebank, model_type, paths, args)
|
| 27 |
-
~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 28 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_tokenizer_treebank.py", line 1499, in process_treebank
|
| 29 |
-
process_ud_treebank(treebank, udbase_dir, tokenizer_dir, short_name, short_language, args.augment)
|
| 30 |
-
~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 31 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_tokenizer_treebank.py", line 1357, in process_ud_treebank
|
| 32 |
-
prepare_ud_dataset(treebank, udbase_dir, tokenizer_dir, short_name, short_language, "test", augment)
|
| 33 |
-
~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 34 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_tokenizer_treebank.py", line 1332, in prepare_ud_dataset
|
| 35 |
-
input_conllu = common.find_treebank_dataset_file(treebank, udbase_dir, dataset, "conllu", fail=True)
|
| 36 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/common.py", line 196, in find_treebank_dataset_file
|
| 37 |
-
raise FileNotFoundError("Could not find any treebank files which matched {}\nIf you have the data elsewhere, you can change the base directory for the search by changing the {} environment variable".format(filename, env_var))
|
| 38 |
-
FileNotFoundError: Could not find any treebank files which matched /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/*-ud-test.conllu
|
| 39 |
-
If you have the data elsewhere, you can change the base directory for the search by changing the UDBASE environment variable
|
| 40 |
-
Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
|
| 41 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /tmp/tmp468rp5x_/sv_diachronic.train.gold.conllu
|
| 42 |
-
Swapped 'w1, w2' for 'w1 ,w2' 93 times
|
| 43 |
-
Added 86 new sentences with asdf, zzzz -> asdf,zzzz
|
| 44 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /tmp/tmp468rp5x_/sv_diachronic.dev.gold.conllu
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_conll17.pt_sv_20260219_105228.txt
DELETED
|
@@ -1,315 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_20260219_105228.txt ===
|
| 2 |
-
Language codes: sv
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-SILVER.py sv
|
| 6 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 7 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 8 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 9 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 10 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 11 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 12 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 13 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 14 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 15 |
-
Reading: /home/urdatorn/git/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 16 |
-
Loading DigPhil MACHINE for SILVER (minus gold overlap)…
|
| 17 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
|
| 18 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 19 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 20 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
|
| 21 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 22 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 23 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 24 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 25 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 26 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 27 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 28 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 29 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 30 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 31 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
|
| 32 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
|
| 33 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 34 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
|
| 35 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
|
| 36 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
|
| 37 |
-
Reading GOLD: /home/urdatorn/git/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
|
| 38 |
-
Adding 74 remaining gold sentences to TRAIN
|
| 39 |
-
Cleaning TRAIN...
|
| 40 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 41 |
-
Cleaning SILVER...
|
| 42 |
-
[REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
|
| 43 |
-
[REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
|
| 44 |
-
[REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
|
| 45 |
-
[REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
|
| 46 |
-
[REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
|
| 47 |
-
[REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
|
| 48 |
-
[REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
|
| 49 |
-
[REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
|
| 50 |
-
[REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
|
| 51 |
-
[REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
|
| 52 |
-
[REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
|
| 53 |
-
[REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
|
| 54 |
-
[REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
|
| 55 |
-
[REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
|
| 56 |
-
[REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
|
| 57 |
-
[REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
|
| 58 |
-
[REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
|
| 59 |
-
[REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
|
| 60 |
-
[REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
|
| 61 |
-
[REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
|
| 62 |
-
[REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
|
| 63 |
-
[REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
|
| 64 |
-
[REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
|
| 65 |
-
[REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
|
| 66 |
-
[REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
|
| 67 |
-
[REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
|
| 68 |
-
[REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
|
| 69 |
-
[REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
|
| 70 |
-
[REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
|
| 71 |
-
[REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
|
| 72 |
-
[REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
|
| 73 |
-
[REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
|
| 74 |
-
[REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
|
| 75 |
-
[REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
|
| 76 |
-
[REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
|
| 77 |
-
[REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
|
| 78 |
-
[REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
|
| 79 |
-
[REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
|
| 80 |
-
[REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
|
| 81 |
-
[REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
|
| 82 |
-
[REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
|
| 83 |
-
[REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
|
| 84 |
-
[REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
|
| 85 |
-
[REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
|
| 86 |
-
[REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
|
| 87 |
-
[REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
|
| 88 |
-
[REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
|
| 89 |
-
[REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
|
| 90 |
-
[REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
|
| 91 |
-
[REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
|
| 92 |
-
[REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
|
| 93 |
-
[REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
|
| 94 |
-
[REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
|
| 95 |
-
[REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
|
| 96 |
-
[REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
|
| 97 |
-
[REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
|
| 98 |
-
[REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
|
| 99 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
|
| 100 |
-
[REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
|
| 101 |
-
[REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
|
| 102 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
|
| 103 |
-
[REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
|
| 104 |
-
[REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
|
| 105 |
-
Cleaning DEV...
|
| 106 |
-
Writing TRAIN → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (13867 valid sentences)
|
| 107 |
-
Writing SILVER → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train-silver.conllu (46432 valid sentences)
|
| 108 |
-
Writing DEV → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (35 valid sentences)
|
| 109 |
-
Writing TEST → /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (copy of dev, required by Stanza)
|
| 110 |
-
Done.
|
| 111 |
-
Sourcing scripts/config.sh
|
| 112 |
-
Running stanza dataset preparation…
|
| 113 |
-
2026-02-19 10:52:31 INFO: Datasets program called with:
|
| 114 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt --gold
|
| 115 |
-
Augmented 60 quotes: Counter({'《》': 9, '»«': 8, '″″': 8, '““': 7, '„”': 6, '""': 6, '„“': 6, '«»': 4, '””': 4, '「」': 2})
|
| 116 |
-
Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
|
| 117 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /tmp/tmpwrb749kh/sv_diachronic.train.gold.conllu
|
| 118 |
-
Swapped 'w1, w2' for 'w1 ,w2' 98 times
|
| 119 |
-
Added 86 new sentences with asdf, zzzz -> asdf,zzzz
|
| 120 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /tmp/tmpwrb749kh/sv_diachronic.dev.gold.conllu
|
| 121 |
-
Reading from /home/urdatorn/git/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /tmp/tmpwrb749kh/sv_diachronic.test.gold.conllu
|
| 122 |
-
Copying from /tmp/tmpwrb749kh/sv_diachronic.train.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 123 |
-
Copying from /tmp/tmpwrb749kh/sv_diachronic.dev.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 124 |
-
Copying from /tmp/tmpwrb749kh/sv_diachronic.test.gold.conllu to /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
|
| 125 |
-
Running stanza dependency parser training…
|
| 126 |
-
2026-02-19 10:52:34 INFO: Training program called with:
|
| 127 |
-
/home/urdatorn/git/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --batch_size 32 --dropout 0.33 --use_bert --bert_model vesteinn/ScandiBERT --silver_file ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu
|
| 128 |
-
2026-02-19 10:52:34 DEBUG: UD_Swedish-diachronic: sv_diachronic
|
| 129 |
-
2026-02-19 10:52:34 INFO: Using model /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 130 |
-
2026-02-19 10:52:34 INFO: Using model /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 131 |
-
2026-02-19 10:52:34 INFO: Using default pretrain for language sv, found in /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt To use a different pretrain, specify --wordvec_pretrain_file
|
| 132 |
-
2026-02-19 10:52:34 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_scandibert_parser.pt does not exist, training new model
|
| 133 |
-
2026-02-19 10:52:34 INFO: Using model /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 134 |
-
2026-02-19 10:52:34 INFO: Using model /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 135 |
-
2026-02-19 10:52:34 INFO: Using default pretrain for language sv, found in /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt To use a different pretrain, specify --wordvec_pretrain_file
|
| 136 |
-
2026-02-19 10:52:34 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/home/urdatorn/stanza_resources/sv/pretrain', '--train_file', '/home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--wordvec_pretrain_file', '/home/urdatorn/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt', '--batch_size', '32', '--dropout', '0.33', '--bert_model', 'vesteinn/ScandiBERT', '--silver_file', 'ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu']
|
| 137 |
-
2026-02-19 10:52:34 INFO: Running parser in train mode
|
| 138 |
-
2026-02-19 10:52:35 DEBUG: Loaded pretrain from /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
|
| 139 |
-
2026-02-19 10:52:35 INFO: Using 7 as the word cutoff based on the size of the pretrain (250000)
|
| 140 |
-
2026-02-19 10:52:35 INFO: Using pretrained contextualized char embedding
|
| 141 |
-
2026-02-19 10:52:35 INFO: ARGS USED AT TRAINING TIME:
|
| 142 |
-
augment_nopunct: None
|
| 143 |
-
batch_size: 32
|
| 144 |
-
bert_finetune: False
|
| 145 |
-
bert_finetune_layers: None
|
| 146 |
-
bert_hidden_layers: 4
|
| 147 |
-
bert_learning_rate: 1.0
|
| 148 |
-
bert_model: vesteinn/ScandiBERT
|
| 149 |
-
bert_start_finetuning: 200
|
| 150 |
-
bert_warmup_steps: 200
|
| 151 |
-
bert_weight_decay: 0.0
|
| 152 |
-
beta2: 0.999
|
| 153 |
-
char: True
|
| 154 |
-
char_emb_dim: 100
|
| 155 |
-
char_hidden_dim: 400
|
| 156 |
-
char_num_layers: 1
|
| 157 |
-
char_rec_dropout: 0
|
| 158 |
-
charlm: True
|
| 159 |
-
charlm_backward_file: /home/urdatorn/stanza_resources/sv/backward_charlm/conll17.pt
|
| 160 |
-
charlm_forward_file: /home/urdatorn/stanza_resources/sv/forward_charlm/conll17.pt
|
| 161 |
-
charlm_save_dir: saved_models/charlm
|
| 162 |
-
charlm_shorthand: sv_conll17
|
| 163 |
-
checkpoint: True
|
| 164 |
-
checkpoint_interval: 500
|
| 165 |
-
checkpoint_save_name: None
|
| 166 |
-
continue_from: None
|
| 167 |
-
data_dir: data/depparse
|
| 168 |
-
deep_biaff_hidden_dim: 400
|
| 169 |
-
deep_biaff_output_dim: 160
|
| 170 |
-
device: cpu
|
| 171 |
-
distance: True
|
| 172 |
-
dropout: 0.33
|
| 173 |
-
eval_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 174 |
-
eval_interval: 100
|
| 175 |
-
gold_labels: True
|
| 176 |
-
hidden_dim: 400
|
| 177 |
-
lang: sv
|
| 178 |
-
linearization: True
|
| 179 |
-
log_norms: False
|
| 180 |
-
log_step: 20
|
| 181 |
-
lora_alpha: 128
|
| 182 |
-
lora_dropout: 0.1
|
| 183 |
-
lora_modules_to_save: []
|
| 184 |
-
lora_rank: 64
|
| 185 |
-
lora_target_modules: ['query', 'value', 'output.dense', 'intermediate.dense']
|
| 186 |
-
lr: 2.0
|
| 187 |
-
max_grad_norm: 1.0
|
| 188 |
-
max_steps: 50000
|
| 189 |
-
max_steps_before_stop: 2000
|
| 190 |
-
mode: train
|
| 191 |
-
model_type: graph
|
| 192 |
-
num_layers: 3
|
| 193 |
-
optim: adadelta
|
| 194 |
-
output_file: None
|
| 195 |
-
output_latex: False
|
| 196 |
-
pretrain: True
|
| 197 |
-
pretrain_max_vocab: 250000
|
| 198 |
-
rec_dropout: 0
|
| 199 |
-
reversed: False
|
| 200 |
-
sample_train: 1.0
|
| 201 |
-
save_dir: saved_models/depparse
|
| 202 |
-
save_name: {shorthand}_{embedding}_parser.pt
|
| 203 |
-
second_batch_size: None
|
| 204 |
-
second_bert_learning_rate: 0.001
|
| 205 |
-
second_lr: 0.0002
|
| 206 |
-
second_optim: adam
|
| 207 |
-
second_optim_start_step: 10000
|
| 208 |
-
second_warmup_steps: 200
|
| 209 |
-
seed: 1234
|
| 210 |
-
shorthand: sv_diachronic
|
| 211 |
-
silver_file: ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu
|
| 212 |
-
silver_weight: 0.5
|
| 213 |
-
tag_emb_dim: 50
|
| 214 |
-
train_file: /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 215 |
-
train_size: None
|
| 216 |
-
transformed_dim: 125
|
| 217 |
-
transition_embedding_dim: 20
|
| 218 |
-
transition_hidden_dim: 20
|
| 219 |
-
transition_merge_hidden_dim: 200
|
| 220 |
-
transition_subtree_combination: SubtreeCombination.NONE
|
| 221 |
-
transition_subtree_nonlinearity: none
|
| 222 |
-
use_arc_embedding: False
|
| 223 |
-
use_peft: False
|
| 224 |
-
use_ufeats: True
|
| 225 |
-
use_upos: True
|
| 226 |
-
use_xpos: True
|
| 227 |
-
wandb: False
|
| 228 |
-
wandb_name: None
|
| 229 |
-
weight_decay: 1e-05
|
| 230 |
-
word_cutoff: 7
|
| 231 |
-
word_dropout: 0.33
|
| 232 |
-
word_emb_dim: 75
|
| 233 |
-
wordvec_dir: /home/urdatorn/stanza_resources/sv/pretrain
|
| 234 |
-
wordvec_file: None
|
| 235 |
-
wordvec_pretrain_file: /home/urdatorn/stanza_resources/sv/pretrain/conll17.pt
|
| 236 |
-
|
| 237 |
-
2026-02-19 10:52:35 INFO: Loading data with batch size 32...
|
| 238 |
-
2026-02-19 10:52:36 INFO: Train File /home/urdatorn/git/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 13953
|
| 239 |
-
2026-02-19 10:52:36 INFO: Original data size: 13953
|
| 240 |
-
2026-02-19 10:52:38 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
|
| 241 |
-
2026-02-19 10:52:38 INFO: Original length = 13953
|
| 242 |
-
2026-02-19 10:52:38 INFO: Filtered length = 13953
|
| 243 |
-
2026-02-19 10:52:40 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='-') for the xpos factory for sv_diachronic
|
| 244 |
-
2026-02-19 10:52:42 DEBUG: 9512 batches created.
|
| 245 |
-
2026-02-19 10:52:42 INFO: Augmented data size: 14711
|
| 246 |
-
2026-02-19 10:52:45 DEBUG: 10020 batches created.
|
| 247 |
-
2026-02-19 10:52:45 DEBUG: 34 batches created.
|
| 248 |
-
2026-02-19 10:52:45 DEBUG: 34 batches created.
|
| 249 |
-
2026-02-19 10:52:45 INFO: Training parser...
|
| 250 |
-
|
| 251 |
-
[1mXLMRobertaModel LOAD REPORT[0m from: vesteinn/ScandiBERT
|
| 252 |
-
Key | Status |
|
| 253 |
-
--------------------------------+------------+-
|
| 254 |
-
lm_head.bias | UNEXPECTED |
|
| 255 |
-
lm_head.decoder.bias | UNEXPECTED |
|
| 256 |
-
lm_head.layer_norm.bias | UNEXPECTED |
|
| 257 |
-
lm_head.dense.weight | UNEXPECTED |
|
| 258 |
-
lm_head.layer_norm.weight | UNEXPECTED |
|
| 259 |
-
roberta.embeddings.position_ids | UNEXPECTED |
|
| 260 |
-
lm_head.dense.bias | UNEXPECTED |
|
| 261 |
-
pooler.dense.bias | MISSING |
|
| 262 |
-
pooler.dense.weight | MISSING |
|
| 263 |
-
|
| 264 |
-
[3mNotes:
|
| 265 |
-
- UNEXPECTED[3m :can be ignored when loading from different task/architecture; not ok if you expect identical arch.
|
| 266 |
-
- MISSING[3m :those params were newly initialized because missing from the checkpoint. Consider training on your downstream task.[0m
|
| 267 |
-
Warning: You are sending unauthenticated requests to the HF Hub. Please set a HF_TOKEN to enable higher rate limits and faster downloads.
|
| 268 |
-
Traceback (most recent call last):
|
| 269 |
-
File "<frozen runpy>", line 198, in _run_module_as_main
|
| 270 |
-
File "<frozen runpy>", line 88, in _run_code
|
| 271 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/training/run_depparse.py", line 145, in <module>
|
| 272 |
-
main()
|
| 273 |
-
~~~~^^
|
| 274 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/training/run_depparse.py", line 142, in main
|
| 275 |
-
common.main(run_treebank, "depparse", "parser", add_depparse_args, sub_argparse=parser.build_argparse(), build_model_filename=build_model_filename, choose_charlm_method=choose_depparse_charlm)
|
| 276 |
-
~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 277 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/training/common.py", line 201, in main
|
| 278 |
-
run_treebank(mode, paths, treebank, short_name, command_args, extra_args + save_name_args)
|
| 279 |
-
~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 280 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/utils/training/run_depparse.py", line 94, in run_treebank
|
| 281 |
-
parser.main(train_args)
|
| 282 |
-
~~~~~~~~~~~^^^^^^^^^^^^
|
| 283 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/models/parser.py", line 266, in main
|
| 284 |
-
return train(args)
|
| 285 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/models/parser.py", line 392, in train
|
| 286 |
-
trainer = model_type(args=args, vocab=vocab, pretrain=pretrain, device=args['device'])
|
| 287 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/models/depparse/trainer.py", line 65, in __init__
|
| 288 |
-
bert_model, bert_tokenizer = load_bert(self.args['bert_model'])
|
| 289 |
-
~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 290 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/models/common/foundation_cache.py", line 115, in load_bert
|
| 291 |
-
return bert_embedding.load_bert(model_name, local_files_only=local_files_only)
|
| 292 |
-
~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 293 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/models/common/bert_embedding.py", line 66, in load_bert
|
| 294 |
-
bert_tokenizer = load_tokenizer(model_name, tokenizer_kwargs=tokenizer_kwargs, local_files_only=local_files_only)
|
| 295 |
-
File "/home/urdatorn/git/stanza-digphil/stanza/models/common/bert_embedding.py", line 50, in load_tokenizer
|
| 296 |
-
bert_tokenizer = AutoTokenizer.from_pretrained(model_name, **bert_args)
|
| 297 |
-
File "/home/urdatorn/.pyenv/versions/venv/lib/python3.14/site-packages/transformers/models/auto/tokenization_auto.py", line 712, in from_pretrained
|
| 298 |
-
return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs)
|
| 299 |
-
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 300 |
-
File "/home/urdatorn/.pyenv/versions/venv/lib/python3.14/site-packages/transformers/tokenization_utils_base.py", line 1712, in from_pretrained
|
| 301 |
-
return cls._from_pretrained(
|
| 302 |
-
~~~~~~~~~~~~~~~~~~~~^
|
| 303 |
-
resolved_vocab_files,
|
| 304 |
-
^^^^^^^^^^^^^^^^^^^^^
|
| 305 |
-
...<9 lines>...
|
| 306 |
-
**kwargs,
|
| 307 |
-
^^^^^^^^^
|
| 308 |
-
)
|
| 309 |
-
^
|
| 310 |
-
File "/home/urdatorn/.pyenv/versions/venv/lib/python3.14/site-packages/transformers/tokenization_utils_base.py", line 1897, in _from_pretrained
|
| 311 |
-
init_kwargs = cls.convert_to_native_format(**init_kwargs)
|
| 312 |
-
File "/home/urdatorn/.pyenv/versions/venv/lib/python3.14/site-packages/transformers/tokenization_utils_tokenizers.py", line 127, in convert_to_native_format
|
| 313 |
-
if vocab and isinstance(vocab[0], (list, tuple)):
|
| 314 |
-
~~~~~^^^
|
| 315 |
-
KeyError: 0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_conll17.pt_sv_20260219_105756.txt
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
logs/log_conll17.pt_sv_20260219_114115.txt
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
logs/log_conll17.pt_sv_diachron_20251203_223822.txt
DELETED
|
@@ -1,731 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_diachron_20251203_223822.txt ===
|
| 2 |
-
Language codes: sv diachron
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-test.py sv diachron
|
| 6 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 7 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 8 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 9 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 10 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 11 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 12 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/ucxn_ud_swedish-talbanken.conllu
|
| 13 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 14 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 15 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 16 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 17 |
-
Including DigPhil MACHINE in TRAIN (minus gold)…
|
| 18 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 19 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 20 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 21 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 22 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
|
| 23 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 24 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 25 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 26 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
|
| 27 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
|
| 28 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 29 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
|
| 30 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
|
| 31 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
|
| 32 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 33 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 34 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 35 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 36 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 37 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
|
| 38 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
|
| 39 |
-
Cleaning TRAIN...
|
| 40 |
-
[REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
|
| 41 |
-
[REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
|
| 42 |
-
[REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
|
| 43 |
-
[REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
|
| 44 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
|
| 45 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
|
| 46 |
-
[REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
|
| 47 |
-
[REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
|
| 48 |
-
[REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
|
| 49 |
-
[REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
|
| 50 |
-
[REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
|
| 51 |
-
[REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
|
| 52 |
-
[REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
|
| 53 |
-
[REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
|
| 54 |
-
[REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
|
| 55 |
-
[REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
|
| 56 |
-
[REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
|
| 57 |
-
[REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
|
| 58 |
-
[REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
|
| 59 |
-
[REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
|
| 60 |
-
[REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
|
| 61 |
-
[REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
|
| 62 |
-
[REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
|
| 63 |
-
[REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
|
| 64 |
-
[REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
|
| 65 |
-
[REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
|
| 66 |
-
[REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
|
| 67 |
-
[REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
|
| 68 |
-
[REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
|
| 69 |
-
[REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
|
| 70 |
-
[REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
|
| 71 |
-
[REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
|
| 72 |
-
[REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
|
| 73 |
-
[REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
|
| 74 |
-
[REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
|
| 75 |
-
[REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
|
| 76 |
-
[REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
|
| 77 |
-
[REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
|
| 78 |
-
[REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
|
| 79 |
-
[REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
|
| 80 |
-
[REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
|
| 81 |
-
[REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
|
| 82 |
-
[REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
|
| 83 |
-
[REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
|
| 84 |
-
[REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
|
| 85 |
-
[REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
|
| 86 |
-
[REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
|
| 87 |
-
[REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
|
| 88 |
-
[REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
|
| 89 |
-
[REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
|
| 90 |
-
[REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
|
| 91 |
-
[REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
|
| 92 |
-
[REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
|
| 93 |
-
[REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
|
| 94 |
-
[REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
|
| 95 |
-
[REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
|
| 96 |
-
[REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
|
| 97 |
-
[REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
|
| 98 |
-
[REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
|
| 99 |
-
[REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
|
| 100 |
-
[REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
|
| 101 |
-
[REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
|
| 102 |
-
[REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
|
| 103 |
-
Cleaning DEV...
|
| 104 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 105 |
-
Cleaning TEST...
|
| 106 |
-
Writing TRAIN → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (66252 valid sentences)
|
| 107 |
-
Writing DEV → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (9 valid sentences)
|
| 108 |
-
Writing TEST → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (99 valid sentences)
|
| 109 |
-
Done.
|
| 110 |
-
Sourcing scripts/config_alvis.sh
|
| 111 |
-
Running stanza dataset preparation…
|
| 112 |
-
2025-12-03 22:38:31 INFO: Datasets program called with:
|
| 113 |
-
/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 114 |
-
2025-12-03 22:38:31 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
|
| 115 |
-
|
| 116 |
-
2025-12-03 22:38:31 INFO: Downloaded file to /cephyr/users/cleland/Alvis/stanza_resources/resources.json
|
| 117 |
-
2025-12-03 22:38:31 DEBUG: Processing parameter "processors"...
|
| 118 |
-
2025-12-03 22:38:31 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
|
| 119 |
-
2025-12-03 22:38:31 INFO: Downloading these customized packages for language: sv (Swedish)...
|
| 120 |
-
=======================
|
| 121 |
-
| Processor | Package |
|
| 122 |
-
-----------------------
|
| 123 |
-
=======================
|
| 124 |
-
|
| 125 |
-
2025-12-03 22:38:31 INFO: Finished downloading models and saved to /cephyr/users/cleland/Alvis/stanza_resources
|
| 126 |
-
2025-12-03 22:38:31 INFO: Using tagger model in /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt for sv_diachronic
|
| 127 |
-
2025-12-03 22:38:31 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 128 |
-
2025-12-03 22:38:31 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 129 |
-
Augmented 192 quotes: Counter({'„”': 28, '""': 27, '「」': 20, '″″': 20, '»«': 18, '«»': 18, '《》': 17, '””': 17, '„“': 15, '““': 12})
|
| 130 |
-
2025-12-03 22:38:34 INFO: Running tagger to retag /local/tmp.5441282/tmpi3ipyceb/sv_diachronic.train.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 131 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5441282/tmpi3ipyceb/sv_diachronic.train.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu']
|
| 132 |
-
2025-12-03 22:38:34 INFO: Running tagger in predict mode
|
| 133 |
-
2025-12-03 22:38:34 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 134 |
-
2025-12-03 22:38:36 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 135 |
-
2025-12-03 22:38:36 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 136 |
-
2025-12-03 22:38:36 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 137 |
-
2025-12-03 22:38:36 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 138 |
-
2025-12-03 22:38:37 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 139 |
-
2025-12-03 22:38:39 INFO: Loading data with batch size 250...
|
| 140 |
-
2025-12-03 22:39:18 INFO: Start evaluation...
|
| 141 |
-
2025-12-03 22:42:53 INFO: UPOS XPOS UFeats AllTags
|
| 142 |
-
2025-12-03 22:42:53 INFO: 99.37 88.39 98.13 87.63
|
| 143 |
-
2025-12-03 22:42:53 INFO: POS Tagger score: sv_diachronic 87.63
|
| 144 |
-
2025-12-03 22:42:53 INFO: Running tagger to retag /local/tmp.5441282/tmpi3ipyceb/sv_diachronic.dev.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 145 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5441282/tmpi3ipyceb/sv_diachronic.dev.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu']
|
| 146 |
-
2025-12-03 22:42:53 INFO: Running tagger in predict mode
|
| 147 |
-
2025-12-03 22:42:53 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 148 |
-
2025-12-03 22:42:55 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 149 |
-
2025-12-03 22:42:55 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 150 |
-
2025-12-03 22:42:55 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 151 |
-
2025-12-03 22:42:55 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 152 |
-
2025-12-03 22:42:55 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 153 |
-
2025-12-03 22:42:55 INFO: Loading data with batch size 250...
|
| 154 |
-
2025-12-03 22:42:55 INFO: Start evaluation...
|
| 155 |
-
2025-12-03 22:42:56 INFO: UPOS XPOS UFeats AllTags
|
| 156 |
-
2025-12-03 22:42:56 INFO: 93.32 90.84 93.32 85.64
|
| 157 |
-
2025-12-03 22:42:56 INFO: POS Tagger score: sv_diachronic 85.64
|
| 158 |
-
2025-12-03 22:42:56 INFO: Running tagger to retag /local/tmp.5441282/tmpi3ipyceb/sv_diachronic.test.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
|
| 159 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5441282/tmpi3ipyceb/sv_diachronic.test.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu']
|
| 160 |
-
2025-12-03 22:42:56 INFO: Running tagger in predict mode
|
| 161 |
-
2025-12-03 22:42:56 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 162 |
-
2025-12-03 22:42:57 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 163 |
-
2025-12-03 22:42:57 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 164 |
-
2025-12-03 22:42:57 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 165 |
-
2025-12-03 22:42:57 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 166 |
-
2025-12-03 22:42:58 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 167 |
-
2025-12-03 22:42:58 INFO: Loading data with batch size 250...
|
| 168 |
-
2025-12-03 22:42:58 INFO: Start evaluation...
|
| 169 |
-
2025-12-03 22:42:58 INFO: UPOS XPOS UFeats AllTags
|
| 170 |
-
2025-12-03 22:42:58 INFO: 93.14 96.78 95.32 90.28
|
| 171 |
-
2025-12-03 22:42:58 INFO: POS Tagger score: sv_diachronic 90.28
|
| 172 |
-
Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
|
| 173 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /local/tmp.5441282/tmpi3ipyceb/sv_diachronic.train.gold.conllu
|
| 174 |
-
Swapped 'w1, w2' for 'w1 ,w2' 132 times
|
| 175 |
-
Added 506 new sentences with asdf, zzzz -> asdf,zzzz
|
| 176 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /local/tmp.5441282/tmpi3ipyceb/sv_diachronic.dev.gold.conllu
|
| 177 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /local/tmp.5441282/tmpi3ipyceb/sv_diachronic.test.gold.conllu
|
| 178 |
-
Running stanza dependency parser training…
|
| 179 |
-
2025-12-03 22:43:12 INFO: Training program called with:
|
| 180 |
-
/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt --batch_size 32 --dropout 0.33
|
| 181 |
-
2025-12-03 22:43:12 DEBUG: UD_Swedish-diachronic: sv_diachronic
|
| 182 |
-
2025-12-03 22:43:12 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 183 |
-
2025-12-03 22:43:12 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 184 |
-
2025-12-03 22:43:12 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_charlm_parser.pt does not exist, training new model
|
| 185 |
-
2025-12-03 22:43:12 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 186 |
-
2025-12-03 22:43:12 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 187 |
-
2025-12-03 22:43:12 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--train_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
|
| 188 |
-
2025-12-03 22:43:12 INFO: Running parser in train mode
|
| 189 |
-
2025-12-03 22:43:12 INFO: Using pretrained contextualized char embedding
|
| 190 |
-
2025-12-03 22:43:12 INFO: Loading data with batch size 32...
|
| 191 |
-
2025-12-03 22:43:20 INFO: Train File /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 66758
|
| 192 |
-
2025-12-03 22:43:20 INFO: Original data size: 66758
|
| 193 |
-
2025-12-03 22:43:21 INFO: Augmented data size: 66875
|
| 194 |
-
2025-12-03 22:43:38 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
|
| 195 |
-
2025-12-03 22:43:38 INFO: Original length = 66875
|
| 196 |
-
2025-12-03 22:43:38 INFO: Filtered length = 66875
|
| 197 |
-
2025-12-03 22:43:55 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='|') for the xpos factory for sv_diachronic
|
| 198 |
-
2025-12-03 22:44:01 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 199 |
-
2025-12-03 22:44:15 DEBUG: 38839 batches created.
|
| 200 |
-
2025-12-03 22:44:15 DEBUG: 9 batches created.
|
| 201 |
-
2025-12-03 22:44:15 INFO: Training parser...
|
| 202 |
-
2025-12-03 22:44:15 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 203 |
-
2025-12-03 22:44:15 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 204 |
-
2025-12-03 22:44:15 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 205 |
-
2025-12-03 22:44:16 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 206 |
-
2025-12-03 22:44:24 INFO: Finished STEP 20/50000, loss = 6.296603 (0.201 sec/batch), lr: 0.003000
|
| 207 |
-
2025-12-03 22:44:27 INFO: Finished STEP 40/50000, loss = 5.081994 (0.181 sec/batch), lr: 0.003000
|
| 208 |
-
2025-12-03 22:44:31 INFO: Finished STEP 60/50000, loss = 4.866196 (0.144 sec/batch), lr: 0.003000
|
| 209 |
-
2025-12-03 22:44:34 INFO: Finished STEP 80/50000, loss = 4.271401 (0.138 sec/batch), lr: 0.003000
|
| 210 |
-
2025-12-03 22:44:36 INFO: Finished STEP 100/50000, loss = 4.687382 (0.143 sec/batch), lr: 0.003000
|
| 211 |
-
2025-12-03 22:44:36 INFO: Evaluating on dev set...
|
| 212 |
-
2025-12-03 22:44:37 INFO: LAS MLAS BLEX
|
| 213 |
-
2025-12-03 22:44:37 INFO: 43.56 33.47 35.15
|
| 214 |
-
2025-12-03 22:44:37 INFO: step 100: train_loss = 7.396650, dev_score = 0.4356
|
| 215 |
-
2025-12-03 22:44:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 216 |
-
2025-12-03 22:44:37 INFO: new best model saved.
|
| 217 |
-
2025-12-03 22:44:38 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 218 |
-
2025-12-03 22:44:38 INFO: new model checkpoint saved.
|
| 219 |
-
2025-12-03 22:44:41 INFO: Finished STEP 120/50000, loss = 3.784610 (0.125 sec/batch), lr: 0.003000
|
| 220 |
-
2025-12-03 22:44:43 INFO: Finished STEP 140/50000, loss = 4.674234 (0.131 sec/batch), lr: 0.003000
|
| 221 |
-
2025-12-03 22:44:46 INFO: Finished STEP 160/50000, loss = 4.012815 (0.137 sec/batch), lr: 0.003000
|
| 222 |
-
2025-12-03 22:44:48 INFO: Finished STEP 180/50000, loss = 4.966728 (0.122 sec/batch), lr: 0.003000
|
| 223 |
-
2025-12-03 22:44:51 INFO: Finished STEP 200/50000, loss = 4.140534 (0.121 sec/batch), lr: 0.003000
|
| 224 |
-
2025-12-03 22:44:51 INFO: Evaluating on dev set...
|
| 225 |
-
2025-12-03 22:44:51 INFO: LAS MLAS BLEX
|
| 226 |
-
2025-12-03 22:44:51 INFO: 51.98 38.30 43.83
|
| 227 |
-
2025-12-03 22:44:51 INFO: step 200: train_loss = 4.683159, dev_score = 0.5198
|
| 228 |
-
2025-12-03 22:44:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 229 |
-
2025-12-03 22:44:51 INFO: new best model saved.
|
| 230 |
-
2025-12-03 22:44:52 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 231 |
-
2025-12-03 22:44:52 INFO: new model checkpoint saved.
|
| 232 |
-
2025-12-03 22:44:55 INFO: Finished STEP 220/50000, loss = 3.998342 (0.126 sec/batch), lr: 0.003000
|
| 233 |
-
2025-12-03 22:44:57 INFO: Finished STEP 240/50000, loss = 5.531576 (0.124 sec/batch), lr: 0.003000
|
| 234 |
-
2025-12-03 22:44:59 INFO: Finished STEP 260/50000, loss = 4.806288 (0.116 sec/batch), lr: 0.003000
|
| 235 |
-
2025-12-03 22:45:02 INFO: Finished STEP 280/50000, loss = 4.965835 (0.111 sec/batch), lr: 0.003000
|
| 236 |
-
2025-12-03 22:45:04 INFO: Finished STEP 300/50000, loss = 3.812420 (0.115 sec/batch), lr: 0.003000
|
| 237 |
-
2025-12-03 22:45:04 INFO: Evaluating on dev set...
|
| 238 |
-
2025-12-03 22:45:04 INFO: LAS MLAS BLEX
|
| 239 |
-
2025-12-03 22:45:04 INFO: 55.45 44.40 49.47
|
| 240 |
-
2025-12-03 22:45:04 INFO: step 300: train_loss = 4.276576, dev_score = 0.5545
|
| 241 |
-
2025-12-03 22:45:05 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 242 |
-
2025-12-03 22:45:05 INFO: new best model saved.
|
| 243 |
-
2025-12-03 22:45:05 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 244 |
-
2025-12-03 22:45:05 INFO: new model checkpoint saved.
|
| 245 |
-
2025-12-03 22:45:07 INFO: Finished STEP 320/50000, loss = 3.429461 (0.111 sec/batch), lr: 0.003000
|
| 246 |
-
2025-12-03 22:45:10 INFO: Finished STEP 340/50000, loss = 3.642277 (0.118 sec/batch), lr: 0.003000
|
| 247 |
-
2025-12-03 22:45:12 INFO: Finished STEP 360/50000, loss = 3.760803 (0.116 sec/batch), lr: 0.003000
|
| 248 |
-
2025-12-03 22:45:14 INFO: Finished STEP 380/50000, loss = 4.786219 (0.105 sec/batch), lr: 0.003000
|
| 249 |
-
2025-12-03 22:45:16 INFO: Finished STEP 400/50000, loss = 2.972147 (0.108 sec/batch), lr: 0.003000
|
| 250 |
-
2025-12-03 22:45:16 INFO: Evaluating on dev set...
|
| 251 |
-
2025-12-03 22:45:17 INFO: LAS MLAS BLEX
|
| 252 |
-
2025-12-03 22:45:17 INFO: 57.43 46.61 50.85
|
| 253 |
-
2025-12-03 22:45:17 INFO: step 400: train_loss = 3.942294, dev_score = 0.5743
|
| 254 |
-
2025-12-03 22:45:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 255 |
-
2025-12-03 22:45:17 INFO: new best model saved.
|
| 256 |
-
2025-12-03 22:45:18 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 257 |
-
2025-12-03 22:45:18 INFO: new model checkpoint saved.
|
| 258 |
-
2025-12-03 22:45:20 INFO: Finished STEP 420/50000, loss = 4.565236 (0.105 sec/batch), lr: 0.003000
|
| 259 |
-
2025-12-03 22:45:22 INFO: Finished STEP 440/50000, loss = 4.531408 (0.098 sec/batch), lr: 0.003000
|
| 260 |
-
2025-12-03 22:45:24 INFO: Finished STEP 460/50000, loss = 5.082150 (0.090 sec/batch), lr: 0.003000
|
| 261 |
-
2025-12-03 22:45:26 INFO: Finished STEP 480/50000, loss = 2.923217 (0.110 sec/batch), lr: 0.003000
|
| 262 |
-
2025-12-03 22:45:28 INFO: Finished STEP 500/50000, loss = 3.408716 (0.100 sec/batch), lr: 0.003000
|
| 263 |
-
2025-12-03 22:45:28 INFO: Evaluating on dev set...
|
| 264 |
-
2025-12-03 22:45:28 INFO: LAS MLAS BLEX
|
| 265 |
-
2025-12-03 22:45:28 INFO: 60.89 51.48 55.27
|
| 266 |
-
2025-12-03 22:45:28 INFO: step 500: train_loss = 4.009553, dev_score = 0.6089
|
| 267 |
-
2025-12-03 22:45:29 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 268 |
-
2025-12-03 22:45:29 INFO: new best model saved.
|
| 269 |
-
2025-12-03 22:45:29 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 270 |
-
2025-12-03 22:45:29 INFO: new model checkpoint saved.
|
| 271 |
-
2025-12-03 22:45:31 INFO: Finished STEP 520/50000, loss = 5.306584 (0.101 sec/batch), lr: 0.003000
|
| 272 |
-
2025-12-03 22:45:33 INFO: Finished STEP 540/50000, loss = 3.657916 (0.100 sec/batch), lr: 0.003000
|
| 273 |
-
2025-12-03 22:45:35 INFO: Finished STEP 560/50000, loss = 4.001961 (0.086 sec/batch), lr: 0.003000
|
| 274 |
-
2025-12-03 22:45:37 INFO: Finished STEP 580/50000, loss = 4.052956 (0.085 sec/batch), lr: 0.003000
|
| 275 |
-
2025-12-03 22:45:39 INFO: Finished STEP 600/50000, loss = 3.852856 (0.099 sec/batch), lr: 0.003000
|
| 276 |
-
2025-12-03 22:45:39 INFO: Evaluating on dev set...
|
| 277 |
-
2025-12-03 22:45:40 INFO: LAS MLAS BLEX
|
| 278 |
-
2025-12-03 22:45:40 INFO: 59.90 50.00 55.51
|
| 279 |
-
2025-12-03 22:45:40 INFO: step 600: train_loss = 3.724004, dev_score = 0.5990
|
| 280 |
-
2025-12-03 22:45:40 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 281 |
-
2025-12-03 22:45:40 INFO: new model checkpoint saved.
|
| 282 |
-
2025-12-03 22:45:42 INFO: Finished STEP 620/50000, loss = 3.484413 (0.101 sec/batch), lr: 0.003000
|
| 283 |
-
2025-12-03 22:45:44 INFO: Finished STEP 640/50000, loss = 3.220206 (0.102 sec/batch), lr: 0.003000
|
| 284 |
-
2025-12-03 22:45:46 INFO: Finished STEP 660/50000, loss = 3.800846 (0.094 sec/batch), lr: 0.003000
|
| 285 |
-
2025-12-03 22:45:48 INFO: Finished STEP 680/50000, loss = 3.769488 (0.093 sec/batch), lr: 0.003000
|
| 286 |
-
2025-12-03 22:45:50 INFO: Finished STEP 700/50000, loss = 3.285978 (0.097 sec/batch), lr: 0.003000
|
| 287 |
-
2025-12-03 22:45:50 INFO: Evaluating on dev set...
|
| 288 |
-
2025-12-03 22:45:50 INFO: LAS MLAS BLEX
|
| 289 |
-
2025-12-03 22:45:50 INFO: 57.92 46.93 53.70
|
| 290 |
-
2025-12-03 22:45:50 INFO: step 700: train_loss = 3.829402, dev_score = 0.5792
|
| 291 |
-
2025-12-03 22:45:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 292 |
-
2025-12-03 22:45:51 INFO: new model checkpoint saved.
|
| 293 |
-
2025-12-03 22:45:53 INFO: Finished STEP 720/50000, loss = 4.905636 (0.100 sec/batch), lr: 0.003000
|
| 294 |
-
2025-12-03 22:45:55 INFO: Finished STEP 740/50000, loss = 3.337056 (0.102 sec/batch), lr: 0.003000
|
| 295 |
-
2025-12-03 22:45:57 INFO: Finished STEP 760/50000, loss = 5.074053 (0.092 sec/batch), lr: 0.003000
|
| 296 |
-
2025-12-03 22:45:59 INFO: Finished STEP 780/50000, loss = 4.024767 (0.089 sec/batch), lr: 0.003000
|
| 297 |
-
2025-12-03 22:46:00 INFO: Finished STEP 800/50000, loss = 4.552472 (0.091 sec/batch), lr: 0.003000
|
| 298 |
-
2025-12-03 22:46:00 INFO: Evaluating on dev set...
|
| 299 |
-
2025-12-03 22:46:01 INFO: LAS MLAS BLEX
|
| 300 |
-
2025-12-03 22:46:01 INFO: 61.39 49.05 53.70
|
| 301 |
-
2025-12-03 22:46:01 INFO: step 800: train_loss = 3.612613, dev_score = 0.6139
|
| 302 |
-
2025-12-03 22:46:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 303 |
-
2025-12-03 22:46:01 INFO: new best model saved.
|
| 304 |
-
2025-12-03 22:46:02 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 305 |
-
2025-12-03 22:46:02 INFO: new model checkpoint saved.
|
| 306 |
-
2025-12-03 22:46:04 INFO: Finished STEP 820/50000, loss = 6.260066 (0.090 sec/batch), lr: 0.003000
|
| 307 |
-
2025-12-03 22:46:05 INFO: Finished STEP 840/50000, loss = 4.288107 (0.092 sec/batch), lr: 0.003000
|
| 308 |
-
2025-12-03 22:46:07 INFO: Finished STEP 860/50000, loss = 3.849105 (0.097 sec/batch), lr: 0.003000
|
| 309 |
-
2025-12-03 22:46:09 INFO: Finished STEP 880/50000, loss = 2.492743 (0.088 sec/batch), lr: 0.003000
|
| 310 |
-
2025-12-03 22:46:11 INFO: Finished STEP 900/50000, loss = 4.095746 (0.090 sec/batch), lr: 0.003000
|
| 311 |
-
2025-12-03 22:46:11 INFO: Evaluating on dev set...
|
| 312 |
-
2025-12-03 22:46:11 INFO: LAS MLAS BLEX
|
| 313 |
-
2025-12-03 22:46:11 INFO: 60.40 51.68 55.04
|
| 314 |
-
2025-12-03 22:46:11 INFO: step 900: train_loss = 3.946219, dev_score = 0.6040
|
| 315 |
-
2025-12-03 22:46:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 316 |
-
2025-12-03 22:46:12 INFO: new model checkpoint saved.
|
| 317 |
-
2025-12-03 22:46:14 INFO: Finished STEP 920/50000, loss = 2.770604 (0.084 sec/batch), lr: 0.003000
|
| 318 |
-
2025-12-03 22:46:16 INFO: Finished STEP 940/50000, loss = 3.612128 (0.100 sec/batch), lr: 0.003000
|
| 319 |
-
2025-12-03 22:46:17 INFO: Finished STEP 960/50000, loss = 4.249920 (0.089 sec/batch), lr: 0.003000
|
| 320 |
-
2025-12-03 22:46:19 INFO: Finished STEP 980/50000, loss = 2.418294 (0.091 sec/batch), lr: 0.003000
|
| 321 |
-
2025-12-03 22:46:21 INFO: Finished STEP 1000/50000, loss = 4.842584 (0.094 sec/batch), lr: 0.003000
|
| 322 |
-
2025-12-03 22:46:21 INFO: Evaluating on dev set...
|
| 323 |
-
2025-12-03 22:46:21 INFO: LAS MLAS BLEX
|
| 324 |
-
2025-12-03 22:46:21 INFO: 61.39 53.70 56.66
|
| 325 |
-
2025-12-03 22:46:21 INFO: step 1000: train_loss = 3.696227, dev_score = 0.6139
|
| 326 |
-
2025-12-03 22:46:22 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 327 |
-
2025-12-03 22:46:22 INFO: new best model saved.
|
| 328 |
-
2025-12-03 22:46:22 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 329 |
-
2025-12-03 22:46:22 INFO: new model checkpoint saved.
|
| 330 |
-
2025-12-03 22:46:24 INFO: Finished STEP 1020/50000, loss = 3.346077 (0.088 sec/batch), lr: 0.003000
|
| 331 |
-
2025-12-03 22:46:26 INFO: Finished STEP 1040/50000, loss = 3.296835 (0.087 sec/batch), lr: 0.003000
|
| 332 |
-
2025-12-03 22:46:28 INFO: Finished STEP 1060/50000, loss = 2.515590 (0.088 sec/batch), lr: 0.003000
|
| 333 |
-
2025-12-03 22:46:29 INFO: Finished STEP 1080/50000, loss = 3.913376 (0.075 sec/batch), lr: 0.003000
|
| 334 |
-
2025-12-03 22:46:31 INFO: Finished STEP 1100/50000, loss = 5.241524 (0.077 sec/batch), lr: 0.003000
|
| 335 |
-
2025-12-03 22:46:31 INFO: Evaluating on dev set...
|
| 336 |
-
2025-12-03 22:46:31 INFO: LAS MLAS BLEX
|
| 337 |
-
2025-12-03 22:46:31 INFO: 59.65 49.47 53.28
|
| 338 |
-
2025-12-03 22:46:31 INFO: step 1100: train_loss = 3.721023, dev_score = 0.5965
|
| 339 |
-
2025-12-03 22:46:32 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 340 |
-
2025-12-03 22:46:32 INFO: new model checkpoint saved.
|
| 341 |
-
2025-12-03 22:46:34 INFO: Finished STEP 1120/50000, loss = 2.731287 (0.086 sec/batch), lr: 0.003000
|
| 342 |
-
2025-12-03 22:46:36 INFO: Finished STEP 1140/50000, loss = 3.536034 (0.080 sec/batch), lr: 0.003000
|
| 343 |
-
2025-12-03 22:46:37 INFO: Finished STEP 1160/50000, loss = 3.398331 (0.079 sec/batch), lr: 0.003000
|
| 344 |
-
2025-12-03 22:46:39 INFO: Finished STEP 1180/50000, loss = 5.028436 (0.078 sec/batch), lr: 0.003000
|
| 345 |
-
2025-12-03 22:46:41 INFO: Finished STEP 1200/50000, loss = 3.061586 (0.077 sec/batch), lr: 0.003000
|
| 346 |
-
2025-12-03 22:46:41 INFO: Evaluating on dev set...
|
| 347 |
-
2025-12-03 22:46:41 INFO: LAS MLAS BLEX
|
| 348 |
-
2025-12-03 22:46:41 INFO: 61.14 53.28 56.66
|
| 349 |
-
2025-12-03 22:46:41 INFO: step 1200: train_loss = 3.522418, dev_score = 0.6114
|
| 350 |
-
2025-12-03 22:46:42 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 351 |
-
2025-12-03 22:46:42 INFO: new model checkpoint saved.
|
| 352 |
-
2025-12-03 22:46:43 INFO: Finished STEP 1220/50000, loss = 2.513215 (0.089 sec/batch), lr: 0.003000
|
| 353 |
-
2025-12-03 22:46:45 INFO: Finished STEP 1240/50000, loss = 4.864305 (0.085 sec/batch), lr: 0.003000
|
| 354 |
-
2025-12-03 22:46:47 INFO: Finished STEP 1260/50000, loss = 7.466803 (0.082 sec/batch), lr: 0.003000
|
| 355 |
-
2025-12-03 22:46:49 INFO: Finished STEP 1280/50000, loss = 3.586653 (0.086 sec/batch), lr: 0.003000
|
| 356 |
-
2025-12-03 22:46:50 INFO: Finished STEP 1300/50000, loss = 3.791464 (0.078 sec/batch), lr: 0.003000
|
| 357 |
-
2025-12-03 22:46:50 INFO: Evaluating on dev set...
|
| 358 |
-
2025-12-03 22:46:51 INFO: LAS MLAS BLEX
|
| 359 |
-
2025-12-03 22:46:51 INFO: 61.63 52.01 55.81
|
| 360 |
-
2025-12-03 22:46:51 INFO: step 1300: train_loss = 3.743181, dev_score = 0.6163
|
| 361 |
-
2025-12-03 22:46:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 362 |
-
2025-12-03 22:46:51 INFO: new best model saved.
|
| 363 |
-
2025-12-03 22:46:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 364 |
-
2025-12-03 22:46:51 INFO: new model checkpoint saved.
|
| 365 |
-
2025-12-03 22:46:53 INFO: Finished STEP 1320/50000, loss = 2.740164 (0.083 sec/batch), lr: 0.003000
|
| 366 |
-
2025-12-03 22:46:55 INFO: Finished STEP 1340/50000, loss = 4.084477 (0.088 sec/batch), lr: 0.003000
|
| 367 |
-
2025-12-03 22:46:56 INFO: Finished STEP 1360/50000, loss = 4.818987 (0.070 sec/batch), lr: 0.003000
|
| 368 |
-
2025-12-03 22:46:58 INFO: Finished STEP 1380/50000, loss = 2.790164 (0.076 sec/batch), lr: 0.003000
|
| 369 |
-
2025-12-03 22:47:00 INFO: Finished STEP 1400/50000, loss = 4.602959 (0.081 sec/batch), lr: 0.003000
|
| 370 |
-
2025-12-03 22:47:00 INFO: Evaluating on dev set...
|
| 371 |
-
2025-12-03 22:47:00 INFO: LAS MLAS BLEX
|
| 372 |
-
2025-12-03 22:47:00 INFO: 62.13 55.08 58.05
|
| 373 |
-
2025-12-03 22:47:00 INFO: step 1400: train_loss = 3.724900, dev_score = 0.6213
|
| 374 |
-
2025-12-03 22:47:00 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 375 |
-
2025-12-03 22:47:00 INFO: new best model saved.
|
| 376 |
-
2025-12-03 22:47:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 377 |
-
2025-12-03 22:47:01 INFO: new model checkpoint saved.
|
| 378 |
-
2025-12-03 22:47:03 INFO: Finished STEP 1420/50000, loss = 4.015442 (0.079 sec/batch), lr: 0.003000
|
| 379 |
-
2025-12-03 22:47:04 INFO: Finished STEP 1440/50000, loss = 3.268615 (0.082 sec/batch), lr: 0.003000
|
| 380 |
-
2025-12-03 22:47:06 INFO: Finished STEP 1460/50000, loss = 3.735999 (0.080 sec/batch), lr: 0.003000
|
| 381 |
-
2025-12-03 22:47:07 INFO: Finished STEP 1480/50000, loss = 3.804015 (0.073 sec/batch), lr: 0.003000
|
| 382 |
-
2025-12-03 22:47:09 INFO: Finished STEP 1500/50000, loss = 4.517982 (0.087 sec/batch), lr: 0.003000
|
| 383 |
-
2025-12-03 22:47:09 INFO: Evaluating on dev set...
|
| 384 |
-
2025-12-03 22:47:10 INFO: LAS MLAS BLEX
|
| 385 |
-
2025-12-03 22:47:10 INFO: 61.63 52.54 56.36
|
| 386 |
-
2025-12-03 22:47:10 INFO: step 1500: train_loss = 3.783271, dev_score = 0.6163
|
| 387 |
-
2025-12-03 22:47:10 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 388 |
-
2025-12-03 22:47:10 INFO: new model checkpoint saved.
|
| 389 |
-
2025-12-03 22:47:12 INFO: Finished STEP 1520/50000, loss = 3.094217 (0.085 sec/batch), lr: 0.003000
|
| 390 |
-
2025-12-03 22:47:13 INFO: Finished STEP 1540/50000, loss = 3.673062 (0.073 sec/batch), lr: 0.003000
|
| 391 |
-
2025-12-03 22:47:15 INFO: Finished STEP 1560/50000, loss = 3.584330 (0.078 sec/batch), lr: 0.003000
|
| 392 |
-
2025-12-03 22:47:16 INFO: Finished STEP 1580/50000, loss = 3.714322 (0.078 sec/batch), lr: 0.003000
|
| 393 |
-
2025-12-03 22:47:18 INFO: Finished STEP 1600/50000, loss = 3.680001 (0.070 sec/batch), lr: 0.003000
|
| 394 |
-
2025-12-03 22:47:18 INFO: Evaluating on dev set...
|
| 395 |
-
2025-12-03 22:47:18 INFO: LAS MLAS BLEX
|
| 396 |
-
2025-12-03 22:47:18 INFO: 60.89 49.47 54.55
|
| 397 |
-
2025-12-03 22:47:18 INFO: step 1600: train_loss = 3.629141, dev_score = 0.6089
|
| 398 |
-
2025-12-03 22:47:19 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 399 |
-
2025-12-03 22:47:19 INFO: new model checkpoint saved.
|
| 400 |
-
2025-12-03 22:47:21 INFO: Finished STEP 1620/50000, loss = 4.405625 (0.076 sec/batch), lr: 0.003000
|
| 401 |
-
2025-12-03 22:47:22 INFO: Finished STEP 1640/50000, loss = 6.583941 (0.084 sec/batch), lr: 0.003000
|
| 402 |
-
2025-12-03 22:47:24 INFO: Finished STEP 1660/50000, loss = 2.624713 (0.082 sec/batch), lr: 0.003000
|
| 403 |
-
2025-12-03 22:47:25 INFO: Finished STEP 1680/50000, loss = 3.746646 (0.081 sec/batch), lr: 0.003000
|
| 404 |
-
2025-12-03 22:47:27 INFO: Finished STEP 1700/50000, loss = 3.647429 (0.078 sec/batch), lr: 0.003000
|
| 405 |
-
2025-12-03 22:47:27 INFO: Evaluating on dev set...
|
| 406 |
-
2025-12-03 22:47:27 INFO: LAS MLAS BLEX
|
| 407 |
-
2025-12-03 22:47:27 INFO: 62.38 53.59 56.96
|
| 408 |
-
2025-12-03 22:47:27 INFO: step 1700: train_loss = 3.711031, dev_score = 0.6238
|
| 409 |
-
2025-12-03 22:47:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 410 |
-
2025-12-03 22:47:28 INFO: new best model saved.
|
| 411 |
-
2025-12-03 22:47:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 412 |
-
2025-12-03 22:47:28 INFO: new model checkpoint saved.
|
| 413 |
-
2025-12-03 22:47:30 INFO: Finished STEP 1720/50000, loss = 3.788727 (0.082 sec/batch), lr: 0.003000
|
| 414 |
-
2025-12-03 22:47:31 INFO: Finished STEP 1740/50000, loss = 4.010241 (0.075 sec/batch), lr: 0.003000
|
| 415 |
-
2025-12-03 22:47:33 INFO: Finished STEP 1760/50000, loss = 3.686716 (0.081 sec/batch), lr: 0.003000
|
| 416 |
-
2025-12-03 22:47:35 INFO: Finished STEP 1780/50000, loss = 2.884626 (0.079 sec/batch), lr: 0.003000
|
| 417 |
-
2025-12-03 22:47:36 INFO: Finished STEP 1800/50000, loss = 3.530944 (0.073 sec/batch), lr: 0.003000
|
| 418 |
-
2025-12-03 22:47:36 INFO: Evaluating on dev set...
|
| 419 |
-
2025-12-03 22:47:37 INFO: LAS MLAS BLEX
|
| 420 |
-
2025-12-03 22:47:37 INFO: 63.37 53.81 58.47
|
| 421 |
-
2025-12-03 22:47:37 INFO: step 1800: train_loss = 3.627865, dev_score = 0.6337
|
| 422 |
-
2025-12-03 22:47:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 423 |
-
2025-12-03 22:47:37 INFO: new best model saved.
|
| 424 |
-
2025-12-03 22:47:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 425 |
-
2025-12-03 22:47:37 INFO: new model checkpoint saved.
|
| 426 |
-
2025-12-03 22:47:39 INFO: Finished STEP 1820/50000, loss = 3.387048 (0.078 sec/batch), lr: 0.003000
|
| 427 |
-
2025-12-03 22:47:40 INFO: Finished STEP 1840/50000, loss = 3.075635 (0.081 sec/batch), lr: 0.003000
|
| 428 |
-
2025-12-03 22:47:42 INFO: Finished STEP 1860/50000, loss = 3.544627 (0.074 sec/batch), lr: 0.003000
|
| 429 |
-
2025-12-03 22:47:44 INFO: Finished STEP 1880/50000, loss = 2.697122 (0.073 sec/batch), lr: 0.003000
|
| 430 |
-
2025-12-03 22:47:45 INFO: Finished STEP 1900/50000, loss = 4.582170 (0.074 sec/batch), lr: 0.003000
|
| 431 |
-
2025-12-03 22:47:45 INFO: Evaluating on dev set...
|
| 432 |
-
2025-12-03 22:47:46 INFO: LAS MLAS BLEX
|
| 433 |
-
2025-12-03 22:47:46 INFO: 63.12 55.81 57.93
|
| 434 |
-
2025-12-03 22:47:46 INFO: step 1900: train_loss = 3.833838, dev_score = 0.6312
|
| 435 |
-
2025-12-03 22:47:46 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 436 |
-
2025-12-03 22:47:46 INFO: new model checkpoint saved.
|
| 437 |
-
2025-12-03 22:47:48 INFO: Finished STEP 1920/50000, loss = 4.797754 (0.071 sec/batch), lr: 0.003000
|
| 438 |
-
2025-12-03 22:47:49 INFO: Finished STEP 1940/50000, loss = 3.999227 (0.076 sec/batch), lr: 0.003000
|
| 439 |
-
2025-12-03 22:47:51 INFO: Finished STEP 1960/50000, loss = 2.863396 (0.081 sec/batch), lr: 0.003000
|
| 440 |
-
2025-12-03 22:47:52 INFO: Finished STEP 1980/50000, loss = 3.603798 (0.074 sec/batch), lr: 0.003000
|
| 441 |
-
2025-12-03 22:47:54 INFO: Finished STEP 2000/50000, loss = 2.377973 (0.075 sec/batch), lr: 0.003000
|
| 442 |
-
2025-12-03 22:47:54 INFO: Evaluating on dev set...
|
| 443 |
-
2025-12-03 22:47:54 INFO: LAS MLAS BLEX
|
| 444 |
-
2025-12-03 22:47:54 INFO: 58.91 50.63 54.85
|
| 445 |
-
2025-12-03 22:47:54 INFO: step 2000: train_loss = 3.578752, dev_score = 0.5891
|
| 446 |
-
2025-12-03 22:47:55 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 447 |
-
2025-12-03 22:47:55 INFO: new model checkpoint saved.
|
| 448 |
-
2025-12-03 22:47:56 INFO: Finished STEP 2020/50000, loss = 3.455811 (0.069 sec/batch), lr: 0.003000
|
| 449 |
-
2025-12-03 22:47:58 INFO: Finished STEP 2040/50000, loss = 4.113457 (0.067 sec/batch), lr: 0.003000
|
| 450 |
-
2025-12-03 22:47:59 INFO: Finished STEP 2060/50000, loss = 3.953318 (0.070 sec/batch), lr: 0.003000
|
| 451 |
-
2025-12-03 22:48:01 INFO: Finished STEP 2080/50000, loss = 2.824056 (0.075 sec/batch), lr: 0.003000
|
| 452 |
-
2025-12-03 22:48:02 INFO: Finished STEP 2100/50000, loss = 4.579782 (0.079 sec/batch), lr: 0.003000
|
| 453 |
-
2025-12-03 22:48:02 INFO: Evaluating on dev set...
|
| 454 |
-
2025-12-03 22:48:03 INFO: LAS MLAS BLEX
|
| 455 |
-
2025-12-03 22:48:03 INFO: 61.63 52.85 57.51
|
| 456 |
-
2025-12-03 22:48:03 INFO: step 2100: train_loss = 3.729740, dev_score = 0.6163
|
| 457 |
-
2025-12-03 22:48:03 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 458 |
-
2025-12-03 22:48:03 INFO: new model checkpoint saved.
|
| 459 |
-
2025-12-03 22:48:05 INFO: Finished STEP 2120/50000, loss = 3.769734 (0.077 sec/batch), lr: 0.003000
|
| 460 |
-
2025-12-03 22:48:06 INFO: Finished STEP 2140/50000, loss = 3.904819 (0.071 sec/batch), lr: 0.003000
|
| 461 |
-
2025-12-03 22:48:08 INFO: Finished STEP 2160/50000, loss = 3.850474 (0.068 sec/batch), lr: 0.003000
|
| 462 |
-
2025-12-03 22:48:09 INFO: Finished STEP 2180/50000, loss = 3.001653 (0.072 sec/batch), lr: 0.003000
|
| 463 |
-
2025-12-03 22:48:11 INFO: Finished STEP 2200/50000, loss = 1.957051 (0.074 sec/batch), lr: 0.003000
|
| 464 |
-
2025-12-03 22:48:11 INFO: Evaluating on dev set...
|
| 465 |
-
2025-12-03 22:48:11 INFO: LAS MLAS BLEX
|
| 466 |
-
2025-12-03 22:48:11 INFO: 60.64 51.16 55.39
|
| 467 |
-
2025-12-03 22:48:11 INFO: step 2200: train_loss = 3.531071, dev_score = 0.6064
|
| 468 |
-
2025-12-03 22:48:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 469 |
-
2025-12-03 22:48:12 INFO: new model checkpoint saved.
|
| 470 |
-
2025-12-03 22:48:13 INFO: Finished STEP 2220/50000, loss = 3.429658 (0.077 sec/batch), lr: 0.003000
|
| 471 |
-
2025-12-03 22:48:15 INFO: Finished STEP 2240/50000, loss = 3.791333 (0.073 sec/batch), lr: 0.003000
|
| 472 |
-
2025-12-03 22:48:16 INFO: Finished STEP 2260/50000, loss = 2.263699 (0.072 sec/batch), lr: 0.003000
|
| 473 |
-
2025-12-03 22:48:18 INFO: Finished STEP 2280/50000, loss = 3.650295 (0.067 sec/batch), lr: 0.003000
|
| 474 |
-
2025-12-03 22:48:19 INFO: Finished STEP 2300/50000, loss = 4.082314 (0.071 sec/batch), lr: 0.003000
|
| 475 |
-
2025-12-03 22:48:19 INFO: Evaluating on dev set...
|
| 476 |
-
2025-12-03 22:48:20 INFO: LAS MLAS BLEX
|
| 477 |
-
2025-12-03 22:48:20 INFO: 62.13 54.55 57.93
|
| 478 |
-
2025-12-03 22:48:20 INFO: step 2300: train_loss = 3.777273, dev_score = 0.6213
|
| 479 |
-
2025-12-03 22:48:20 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 480 |
-
2025-12-03 22:48:20 INFO: new model checkpoint saved.
|
| 481 |
-
2025-12-03 22:48:22 INFO: Finished STEP 2320/50000, loss = 3.704285 (0.071 sec/batch), lr: 0.003000
|
| 482 |
-
2025-12-03 22:48:23 INFO: Finished STEP 2340/50000, loss = 2.753342 (0.069 sec/batch), lr: 0.003000
|
| 483 |
-
2025-12-03 22:48:25 INFO: Finished STEP 2360/50000, loss = 3.819938 (0.072 sec/batch), lr: 0.003000
|
| 484 |
-
2025-12-03 22:48:26 INFO: Finished STEP 2380/50000, loss = 4.015243 (0.075 sec/batch), lr: 0.003000
|
| 485 |
-
2025-12-03 22:48:27 INFO: Finished STEP 2400/50000, loss = 4.291789 (0.077 sec/batch), lr: 0.003000
|
| 486 |
-
2025-12-03 22:48:27 INFO: Evaluating on dev set...
|
| 487 |
-
2025-12-03 22:48:28 INFO: LAS MLAS BLEX
|
| 488 |
-
2025-12-03 22:48:28 INFO: 60.64 53.28 57.08
|
| 489 |
-
2025-12-03 22:48:28 INFO: step 2400: train_loss = 3.659402, dev_score = 0.6064
|
| 490 |
-
2025-12-03 22:48:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 491 |
-
2025-12-03 22:48:28 INFO: new model checkpoint saved.
|
| 492 |
-
2025-12-03 22:48:30 INFO: Finished STEP 2420/50000, loss = 3.808647 (0.070 sec/batch), lr: 0.003000
|
| 493 |
-
2025-12-03 22:48:31 INFO: Finished STEP 2440/50000, loss = 3.675776 (0.070 sec/batch), lr: 0.003000
|
| 494 |
-
2025-12-03 22:48:33 INFO: Finished STEP 2460/50000, loss = 4.045568 (0.071 sec/batch), lr: 0.003000
|
| 495 |
-
2025-12-03 22:48:34 INFO: Finished STEP 2480/50000, loss = 2.708247 (0.071 sec/batch), lr: 0.003000
|
| 496 |
-
2025-12-03 22:48:36 INFO: Finished STEP 2500/50000, loss = 3.840647 (0.072 sec/batch), lr: 0.003000
|
| 497 |
-
2025-12-03 22:48:36 INFO: Evaluating on dev set...
|
| 498 |
-
2025-12-03 22:48:36 INFO: LAS MLAS BLEX
|
| 499 |
-
2025-12-03 22:48:36 INFO: 63.61 54.97 57.93
|
| 500 |
-
2025-12-03 22:48:36 INFO: step 2500: train_loss = 3.616523, dev_score = 0.6361
|
| 501 |
-
2025-12-03 22:48:36 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 502 |
-
2025-12-03 22:48:36 INFO: new best model saved.
|
| 503 |
-
2025-12-03 22:48:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 504 |
-
2025-12-03 22:48:37 INFO: new model checkpoint saved.
|
| 505 |
-
2025-12-03 22:48:38 INFO: Finished STEP 2520/50000, loss = 3.329048 (0.075 sec/batch), lr: 0.003000
|
| 506 |
-
2025-12-03 22:48:40 INFO: Finished STEP 2540/50000, loss = 3.042371 (0.066 sec/batch), lr: 0.003000
|
| 507 |
-
2025-12-03 22:48:41 INFO: Finished STEP 2560/50000, loss = 3.319040 (0.070 sec/batch), lr: 0.003000
|
| 508 |
-
2025-12-03 22:48:43 INFO: Finished STEP 2580/50000, loss = 4.341519 (0.056 sec/batch), lr: 0.003000
|
| 509 |
-
2025-12-03 22:48:44 INFO: Finished STEP 2600/50000, loss = 5.309865 (0.070 sec/batch), lr: 0.003000
|
| 510 |
-
2025-12-03 22:48:44 INFO: Evaluating on dev set...
|
| 511 |
-
2025-12-03 22:48:44 INFO: LAS MLAS BLEX
|
| 512 |
-
2025-12-03 22:48:44 INFO: 61.14 52.01 56.24
|
| 513 |
-
2025-12-03 22:48:44 INFO: step 2600: train_loss = 3.623874, dev_score = 0.6114
|
| 514 |
-
2025-12-03 22:48:45 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 515 |
-
2025-12-03 22:48:45 INFO: new model checkpoint saved.
|
| 516 |
-
2025-12-03 22:48:46 INFO: Finished STEP 2620/50000, loss = 2.560167 (0.066 sec/batch), lr: 0.003000
|
| 517 |
-
2025-12-03 22:48:48 INFO: Finished STEP 2640/50000, loss = 3.262659 (0.066 sec/batch), lr: 0.003000
|
| 518 |
-
2025-12-03 22:48:49 INFO: Finished STEP 2660/50000, loss = 2.611564 (0.072 sec/batch), lr: 0.003000
|
| 519 |
-
2025-12-03 22:48:51 INFO: Finished STEP 2680/50000, loss = 3.953672 (0.070 sec/batch), lr: 0.003000
|
| 520 |
-
2025-12-03 22:48:52 INFO: Finished STEP 2700/50000, loss = 3.726388 (0.072 sec/batch), lr: 0.003000
|
| 521 |
-
2025-12-03 22:48:52 INFO: Evaluating on dev set...
|
| 522 |
-
2025-12-03 22:48:53 INFO: LAS MLAS BLEX
|
| 523 |
-
2025-12-03 22:48:53 INFO: 61.88 53.39 57.20
|
| 524 |
-
2025-12-03 22:48:53 INFO: step 2700: train_loss = 3.740555, dev_score = 0.6188
|
| 525 |
-
2025-12-03 22:48:53 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 526 |
-
2025-12-03 22:48:53 INFO: new model checkpoint saved.
|
| 527 |
-
2025-12-03 22:48:54 INFO: Finished STEP 2720/50000, loss = 3.536426 (0.066 sec/batch), lr: 0.003000
|
| 528 |
-
2025-12-03 22:48:56 INFO: Finished STEP 2740/50000, loss = 4.492881 (0.072 sec/batch), lr: 0.003000
|
| 529 |
-
2025-12-03 22:48:57 INFO: Finished STEP 2760/50000, loss = 3.437390 (0.073 sec/batch), lr: 0.003000
|
| 530 |
-
2025-12-03 22:48:59 INFO: Finished STEP 2780/50000, loss = 3.811538 (0.065 sec/batch), lr: 0.003000
|
| 531 |
-
2025-12-03 22:49:00 INFO: Finished STEP 2800/50000, loss = 2.615445 (0.069 sec/batch), lr: 0.003000
|
| 532 |
-
2025-12-03 22:49:00 INFO: Evaluating on dev set...
|
| 533 |
-
2025-12-03 22:49:01 INFO: LAS MLAS BLEX
|
| 534 |
-
2025-12-03 22:49:01 INFO: 61.14 51.59 56.66
|
| 535 |
-
2025-12-03 22:49:01 INFO: step 2800: train_loss = 3.751917, dev_score = 0.6114
|
| 536 |
-
2025-12-03 22:49:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 537 |
-
2025-12-03 22:49:01 INFO: new model checkpoint saved.
|
| 538 |
-
2025-12-03 22:49:03 INFO: Finished STEP 2820/50000, loss = 3.525189 (0.076 sec/batch), lr: 0.003000
|
| 539 |
-
2025-12-03 22:49:04 INFO: Finished STEP 2840/50000, loss = 3.113450 (0.066 sec/batch), lr: 0.003000
|
| 540 |
-
2025-12-03 22:49:05 INFO: Finished STEP 2860/50000, loss = 2.488067 (0.066 sec/batch), lr: 0.003000
|
| 541 |
-
2025-12-03 22:49:07 INFO: Finished STEP 2880/50000, loss = 2.702721 (0.071 sec/batch), lr: 0.003000
|
| 542 |
-
2025-12-03 22:49:08 INFO: Finished STEP 2900/50000, loss = 3.248651 (0.068 sec/batch), lr: 0.003000
|
| 543 |
-
2025-12-03 22:49:08 INFO: Evaluating on dev set...
|
| 544 |
-
2025-12-03 22:49:09 INFO: LAS MLAS BLEX
|
| 545 |
-
2025-12-03 22:49:09 INFO: 62.13 52.43 56.24
|
| 546 |
-
2025-12-03 22:49:09 INFO: step 2900: train_loss = 3.761651, dev_score = 0.6213
|
| 547 |
-
2025-12-03 22:49:09 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 548 |
-
2025-12-03 22:49:09 INFO: new model checkpoint saved.
|
| 549 |
-
2025-12-03 22:49:11 INFO: Finished STEP 2920/50000, loss = 3.129019 (0.072 sec/batch), lr: 0.003000
|
| 550 |
-
2025-12-03 22:49:12 INFO: Finished STEP 2940/50000, loss = 3.579517 (0.067 sec/batch), lr: 0.003000
|
| 551 |
-
2025-12-03 22:49:13 INFO: Finished STEP 2960/50000, loss = 3.505895 (0.067 sec/batch), lr: 0.003000
|
| 552 |
-
2025-12-03 22:49:15 INFO: Finished STEP 2980/50000, loss = 2.193599 (0.073 sec/batch), lr: 0.003000
|
| 553 |
-
2025-12-03 22:49:16 INFO: Finished STEP 3000/50000, loss = 1.882619 (0.064 sec/batch), lr: 0.003000
|
| 554 |
-
2025-12-03 22:49:16 INFO: Evaluating on dev set...
|
| 555 |
-
2025-12-03 22:49:17 INFO: LAS MLAS BLEX
|
| 556 |
-
2025-12-03 22:49:17 INFO: 61.63 51.16 55.81
|
| 557 |
-
2025-12-03 22:49:17 INFO: step 3000: train_loss = 4.014063, dev_score = 0.6163
|
| 558 |
-
2025-12-03 22:49:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 559 |
-
2025-12-03 22:49:17 INFO: new model checkpoint saved.
|
| 560 |
-
2025-12-03 22:49:19 INFO: Finished STEP 3020/50000, loss = 5.977424 (0.058 sec/batch), lr: 0.003000
|
| 561 |
-
2025-12-03 22:49:20 INFO: Finished STEP 3040/50000, loss = 3.319470 (0.063 sec/batch), lr: 0.003000
|
| 562 |
-
2025-12-03 22:49:21 INFO: Finished STEP 3060/50000, loss = 3.211635 (0.069 sec/batch), lr: 0.003000
|
| 563 |
-
2025-12-03 22:49:23 INFO: Finished STEP 3080/50000, loss = 3.183325 (0.065 sec/batch), lr: 0.003000
|
| 564 |
-
2025-12-03 22:49:24 INFO: Finished STEP 3100/50000, loss = 2.788619 (0.069 sec/batch), lr: 0.003000
|
| 565 |
-
2025-12-03 22:49:24 INFO: Evaluating on dev set...
|
| 566 |
-
2025-12-03 22:49:25 INFO: LAS MLAS BLEX
|
| 567 |
-
2025-12-03 22:49:25 INFO: 60.89 51.05 56.54
|
| 568 |
-
2025-12-03 22:49:25 INFO: step 3100: train_loss = 3.763229, dev_score = 0.6089
|
| 569 |
-
2025-12-03 22:49:25 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 570 |
-
2025-12-03 22:49:25 INFO: new model checkpoint saved.
|
| 571 |
-
2025-12-03 22:49:27 INFO: Finished STEP 3120/50000, loss = 4.927651 (0.063 sec/batch), lr: 0.003000
|
| 572 |
-
2025-12-03 22:49:28 INFO: Finished STEP 3140/50000, loss = 3.578332 (0.068 sec/batch), lr: 0.003000
|
| 573 |
-
2025-12-03 22:49:29 INFO: Finished STEP 3160/50000, loss = 3.173575 (0.076 sec/batch), lr: 0.003000
|
| 574 |
-
2025-12-03 22:49:31 INFO: Finished STEP 3180/50000, loss = 4.860454 (0.075 sec/batch), lr: 0.003000
|
| 575 |
-
2025-12-03 22:49:32 INFO: Finished STEP 3200/50000, loss = 1.778136 (0.065 sec/batch), lr: 0.003000
|
| 576 |
-
2025-12-03 22:49:32 INFO: Evaluating on dev set...
|
| 577 |
-
2025-12-03 22:49:33 INFO: LAS MLAS BLEX
|
| 578 |
-
2025-12-03 22:49:33 INFO: 60.15 50.74 54.12
|
| 579 |
-
2025-12-03 22:49:33 INFO: step 3200: train_loss = 3.791774, dev_score = 0.6015
|
| 580 |
-
2025-12-03 22:49:33 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 581 |
-
2025-12-03 22:49:33 INFO: new model checkpoint saved.
|
| 582 |
-
2025-12-03 22:49:35 INFO: Finished STEP 3220/50000, loss = 8.492014 (0.070 sec/batch), lr: 0.003000
|
| 583 |
-
2025-12-03 22:49:36 INFO: Finished STEP 3240/50000, loss = 3.045755 (0.064 sec/batch), lr: 0.003000
|
| 584 |
-
2025-12-03 22:49:37 INFO: Finished STEP 3260/50000, loss = 4.122291 (0.076 sec/batch), lr: 0.003000
|
| 585 |
-
2025-12-03 22:49:39 INFO: Finished STEP 3280/50000, loss = 3.458145 (0.063 sec/batch), lr: 0.003000
|
| 586 |
-
2025-12-03 22:49:40 INFO: Finished STEP 3300/50000, loss = 2.681028 (0.064 sec/batch), lr: 0.003000
|
| 587 |
-
2025-12-03 22:49:40 INFO: Evaluating on dev set...
|
| 588 |
-
2025-12-03 22:49:41 INFO: LAS MLAS BLEX
|
| 589 |
-
2025-12-03 22:49:41 INFO: 61.88 51.59 55.39
|
| 590 |
-
2025-12-03 22:49:41 INFO: step 3300: train_loss = 3.750064, dev_score = 0.6188
|
| 591 |
-
2025-12-03 22:49:41 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 592 |
-
2025-12-03 22:49:41 INFO: new model checkpoint saved.
|
| 593 |
-
2025-12-03 22:49:42 INFO: Finished STEP 3320/50000, loss = 3.499715 (0.066 sec/batch), lr: 0.003000
|
| 594 |
-
2025-12-03 22:49:44 INFO: Finished STEP 3340/50000, loss = 3.765631 (0.064 sec/batch), lr: 0.003000
|
| 595 |
-
2025-12-03 22:49:45 INFO: Finished STEP 3360/50000, loss = 2.908885 (0.064 sec/batch), lr: 0.003000
|
| 596 |
-
2025-12-03 22:49:47 INFO: Finished STEP 3380/50000, loss = 4.147782 (0.067 sec/batch), lr: 0.003000
|
| 597 |
-
2025-12-03 22:49:48 INFO: Finished STEP 3400/50000, loss = 4.854831 (0.073 sec/batch), lr: 0.003000
|
| 598 |
-
2025-12-03 22:49:48 INFO: Evaluating on dev set...
|
| 599 |
-
2025-12-03 22:49:48 INFO: LAS MLAS BLEX
|
| 600 |
-
2025-12-03 22:49:48 INFO: 61.39 51.59 54.55
|
| 601 |
-
2025-12-03 22:49:48 INFO: step 3400: train_loss = 3.827870, dev_score = 0.6139
|
| 602 |
-
2025-12-03 22:49:49 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 603 |
-
2025-12-03 22:49:49 INFO: new model checkpoint saved.
|
| 604 |
-
2025-12-03 22:49:50 INFO: Finished STEP 3420/50000, loss = 3.672020 (0.066 sec/batch), lr: 0.003000
|
| 605 |
-
2025-12-03 22:49:52 INFO: Finished STEP 3440/50000, loss = 3.315593 (0.064 sec/batch), lr: 0.003000
|
| 606 |
-
2025-12-03 22:49:53 INFO: Finished STEP 3460/50000, loss = 2.334443 (0.064 sec/batch), lr: 0.003000
|
| 607 |
-
2025-12-03 22:49:54 INFO: Finished STEP 3480/50000, loss = 4.692723 (0.061 sec/batch), lr: 0.003000
|
| 608 |
-
2025-12-03 22:49:56 INFO: Finished STEP 3500/50000, loss = 2.920178 (0.064 sec/batch), lr: 0.003000
|
| 609 |
-
2025-12-03 22:49:56 INFO: Evaluating on dev set...
|
| 610 |
-
2025-12-03 22:49:56 INFO: LAS MLAS BLEX
|
| 611 |
-
2025-12-03 22:49:56 INFO: 60.40 50.74 53.70
|
| 612 |
-
2025-12-03 22:49:56 INFO: step 3500: train_loss = 3.750002, dev_score = 0.6040
|
| 613 |
-
2025-12-03 22:49:56 INFO: Training ended with 3500 steps.
|
| 614 |
-
2025-12-03 22:49:56 INFO: Best dev F1 = 63.61, at iteration = 2500
|
| 615 |
-
2025-12-03 22:49:57 INFO: Running dev depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
|
| 616 |
-
2025-12-03 22:49:57 INFO: Running parser in predict mode
|
| 617 |
-
2025-12-03 22:49:57 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 618 |
-
2025-12-03 22:49:59 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 619 |
-
2025-12-03 22:49:59 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 620 |
-
2025-12-03 22:49:59 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 621 |
-
2025-12-03 22:49:59 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 622 |
-
2025-12-03 22:50:00 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 623 |
-
2025-12-03 22:50:00 INFO: Loading data with batch size 32...
|
| 624 |
-
2025-12-03 22:50:00 DEBUG: 9 batches created.
|
| 625 |
-
2025-12-03 22:50:00 INFO: F1 scores for each dependency:
|
| 626 |
-
Note that unlabeled attachment errors hurt the labeled attachment scores
|
| 627 |
-
acl: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
|
| 628 |
-
acl:relcl: p 0.2222 r 0.2857 f1 0.2500 (7 actual)
|
| 629 |
-
advcl: p 0.1667 r 0.2000 f1 0.1818 (5 actual)
|
| 630 |
-
advmod: p 0.4828 r 0.5600 f1 0.5185 (25 actual)
|
| 631 |
-
amod: p 0.8889 r 0.7742 f1 0.8276 (31 actual)
|
| 632 |
-
appos: p 1.0000 r 0.5000 f1 0.6667 (4 actual)
|
| 633 |
-
aux: p 0.8182 r 0.8182 f1 0.8182 (11 actual)
|
| 634 |
-
case: p 0.9444 r 0.9107 f1 0.9273 (56 actual)
|
| 635 |
-
cc: p 0.6923 r 0.6923 f1 0.6923 (13 actual)
|
| 636 |
-
ccomp: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 637 |
-
compound:prt: p 0.0000 r 0.0000 f1 0.0000 (0 actual)
|
| 638 |
-
conj: p 0.3750 r 0.5000 f1 0.4286 (12 actual)
|
| 639 |
-
cop: p 0.5000 r 0.3333 f1 0.4000 (3 actual)
|
| 640 |
-
csubj: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 641 |
-
det: p 0.8696 r 0.9091 f1 0.8889 (22 actual)
|
| 642 |
-
expl: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 643 |
-
iobj: p 0.3333 r 0.5000 f1 0.4000 (2 actual)
|
| 644 |
-
mark: p 0.6364 r 0.5833 f1 0.6087 (12 actual)
|
| 645 |
-
nmod: p 0.4118 r 0.4667 f1 0.4375 (15 actual)
|
| 646 |
-
nmod:poss: p 1.0000 r 0.8947 f1 0.9444 (19 actual)
|
| 647 |
-
nsubj: p 0.5200 r 0.7647 f1 0.6190 (17 actual)
|
| 648 |
-
nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
|
| 649 |
-
obj: p 0.5769 r 0.6818 f1 0.6250 (22 actual)
|
| 650 |
-
obl: p 0.5714 r 0.5854 f1 0.5783 (41 actual)
|
| 651 |
-
obl:agent: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 652 |
-
orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 653 |
-
parataxis: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
|
| 654 |
-
punct: p 0.4423 r 0.4423 f1 0.4423 (52 actual)
|
| 655 |
-
root: p 0.5556 r 0.5556 f1 0.5556 (9 actual)
|
| 656 |
-
xcomp: p 0.6667 r 0.2500 f1 0.3636 (8 actual)
|
| 657 |
-
2025-12-03 22:50:00 INFO: LAS MLAS BLEX
|
| 658 |
-
2025-12-03 22:50:00 INFO: 63.61 54.97 57.93
|
| 659 |
-
2025-12-03 22:50:00 INFO: Parser score:
|
| 660 |
-
2025-12-03 22:50:00 INFO: sv_diachronic 63.61
|
| 661 |
-
2025-12-03 22:50:00 INFO: Finished running dev set on
|
| 662 |
-
UD_Swedish-diachronic
|
| 663 |
-
UAS LAS CLAS MLAS BLEX
|
| 664 |
-
70.54 63.61 57.93 54.97 57.93
|
| 665 |
-
2025-12-03 22:50:00 INFO: Running test depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
|
| 666 |
-
2025-12-03 22:50:00 INFO: Running parser in predict mode
|
| 667 |
-
2025-12-03 22:50:00 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 668 |
-
2025-12-03 22:50:02 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 669 |
-
2025-12-03 22:50:02 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 670 |
-
2025-12-03 22:50:02 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 671 |
-
2025-12-03 22:50:02 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 672 |
-
2025-12-03 22:50:02 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 673 |
-
2025-12-03 22:50:02 INFO: Loading data with batch size 32...
|
| 674 |
-
2025-12-03 22:50:02 DEBUG: 93 batches created.
|
| 675 |
-
2025-12-03 22:50:07 INFO: F1 scores for each dependency:
|
| 676 |
-
Note that unlabeled attachment errors hurt the labeled attachment scores
|
| 677 |
-
acl: p 0.3333 r 0.0312 f1 0.0571 (32 actual)
|
| 678 |
-
acl:cleft: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 679 |
-
acl:relcl: p 0.3061 r 0.2000 f1 0.2419 (75 actual)
|
| 680 |
-
advcl: p 0.0893 r 0.1667 f1 0.1163 (60 actual)
|
| 681 |
-
advcl:relcl: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 682 |
-
advmod: p 0.5745 r 0.5896 f1 0.5820 (268 actual)
|
| 683 |
-
amod: p 0.8139 r 0.8174 f1 0.8156 (230 actual)
|
| 684 |
-
appos: p 0.0000 r 0.0000 f1 0.0000 (13 actual)
|
| 685 |
-
aux: p 0.8554 r 0.8452 f1 0.8503 (84 actual)
|
| 686 |
-
aux:pass: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 687 |
-
case: p 0.8661 r 0.8150 f1 0.8398 (373 actual)
|
| 688 |
-
cc: p 0.6474 r 0.6516 f1 0.6495 (155 actual)
|
| 689 |
-
ccomp: p 0.0000 r 0.0000 f1 0.0000 (35 actual)
|
| 690 |
-
compound:prt: p 0.6800 r 0.8095 f1 0.7391 (21 actual)
|
| 691 |
-
conj: p 0.2938 r 0.2975 f1 0.2956 (158 actual)
|
| 692 |
-
cop: p 0.7188 r 0.5000 f1 0.5897 (46 actual)
|
| 693 |
-
csubj: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 694 |
-
dep: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 695 |
-
det: p 0.8308 r 0.8029 f1 0.8166 (208 actual)
|
| 696 |
-
discourse: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
|
| 697 |
-
dislocated: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 698 |
-
expl: p 0.5000 r 0.0909 f1 0.1538 (11 actual)
|
| 699 |
-
expl:pv: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 700 |
-
fixed: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
|
| 701 |
-
flat: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 702 |
-
flat:name: p 0.0000 r 0.0000 f1 0.0000 (12 actual)
|
| 703 |
-
goeswith: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 704 |
-
iobj: p 0.1935 r 0.4286 f1 0.2667 (14 actual)
|
| 705 |
-
mark: p 0.6624 r 0.6797 f1 0.6710 (153 actual)
|
| 706 |
-
nmod: p 0.2661 r 0.2843 f1 0.2749 (102 actual)
|
| 707 |
-
nmod:poss: p 0.8865 r 0.8803 f1 0.8834 (142 actual)
|
| 708 |
-
nsubj: p 0.5385 r 0.6750 f1 0.5990 (280 actual)
|
| 709 |
-
nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (25 actual)
|
| 710 |
-
nummod: p 0.8000 r 0.8000 f1 0.8000 (10 actual)
|
| 711 |
-
obj: p 0.5556 r 0.6011 f1 0.5774 (183 actual)
|
| 712 |
-
obl: p 0.5030 r 0.5935 f1 0.5446 (278 actual)
|
| 713 |
-
obl:agent: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 714 |
-
orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 715 |
-
parataxis: p 0.1000 r 0.2222 f1 0.1379 (18 actual)
|
| 716 |
-
punct: p 0.4685 r 0.4729 f1 0.4707 (425 actual)
|
| 717 |
-
reparandum: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 718 |
-
root: p 0.5657 r 0.5657 f1 0.5657 (99 actual)
|
| 719 |
-
vocative: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
|
| 720 |
-
xcomp: p 0.5000 r 0.3467 f1 0.4094 (75 actual)
|
| 721 |
-
2025-12-03 22:50:07 INFO: LAS MLAS BLEX
|
| 722 |
-
2025-12-03 22:50:07 INFO: 59.06 49.85 53.26
|
| 723 |
-
2025-12-03 22:50:07 INFO: Parser score:
|
| 724 |
-
2025-12-03 22:50:07 INFO: sv_diachronic 59.06
|
| 725 |
-
2025-12-03 22:50:07 INFO: Finished running test set on
|
| 726 |
-
UD_Swedish-diachronic
|
| 727 |
-
UAS LAS CLAS MLAS BLEX
|
| 728 |
-
68.07 59.06 53.26 49.85 53.26
|
| 729 |
-
DONE.
|
| 730 |
-
Full log saved to: logs/log_conll17.pt_sv_diachron_20251203_223822.txt
|
| 731 |
-
Symlink updated: logs/latest.txt → log_conll17.pt_sv_diachron_20251203_223822.txt
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_conll17.pt_sv_diachron_20251212_145854.txt
DELETED
|
@@ -1,161 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_diachron_20251212_145854.txt ===
|
| 2 |
-
Language codes: sv diachron
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-test.py sv diachron
|
| 6 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec991-spf148.conllu
|
| 7 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec988-spf145.conllu
|
| 8 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 9 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 10 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 11 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 12 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 13 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 14 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 15 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 16 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 17 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1102-spf259.conllu
|
| 18 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-letter141673-Stalhammar.conllu
|
| 19 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1033-spf190.conllu
|
| 20 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 21 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 22 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 23 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 24 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 25 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 26 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 27 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 28 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 29 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 30 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 31 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec25-Runius.conllu
|
| 32 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 33 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 34 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec987-spf144.conllu
|
| 35 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1063-spf220.conllu
|
| 36 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 37 |
-
Including DigPhil MACHINE in TRAIN (minus gold)…
|
| 38 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 39 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 40 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 41 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 42 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
|
| 43 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 44 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 45 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 46 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
|
| 47 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
|
| 48 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 49 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
|
| 50 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
|
| 51 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
|
| 52 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 53 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 54 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 55 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 56 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 57 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
|
| 58 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
|
| 59 |
-
Cleaning TRAIN...
|
| 60 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 61 |
-
[REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
|
| 62 |
-
[REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
|
| 63 |
-
[REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
|
| 64 |
-
[REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
|
| 65 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
|
| 66 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
|
| 67 |
-
[REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
|
| 68 |
-
[REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
|
| 69 |
-
[REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
|
| 70 |
-
[REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
|
| 71 |
-
[REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
|
| 72 |
-
[REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
|
| 73 |
-
[REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
|
| 74 |
-
[REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
|
| 75 |
-
[REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
|
| 76 |
-
[REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
|
| 77 |
-
[REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
|
| 78 |
-
[REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
|
| 79 |
-
[REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
|
| 80 |
-
[REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
|
| 81 |
-
[REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
|
| 82 |
-
[REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
|
| 83 |
-
[REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
|
| 84 |
-
[REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
|
| 85 |
-
[REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
|
| 86 |
-
[REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
|
| 87 |
-
[REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
|
| 88 |
-
[REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
|
| 89 |
-
[REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
|
| 90 |
-
[REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
|
| 91 |
-
[REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
|
| 92 |
-
[REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
|
| 93 |
-
[REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
|
| 94 |
-
[REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
|
| 95 |
-
[REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
|
| 96 |
-
[REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
|
| 97 |
-
[REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
|
| 98 |
-
[REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
|
| 99 |
-
[REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
|
| 100 |
-
[REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
|
| 101 |
-
[REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
|
| 102 |
-
[REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
|
| 103 |
-
[REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
|
| 104 |
-
[REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
|
| 105 |
-
[REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
|
| 106 |
-
[REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
|
| 107 |
-
[REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
|
| 108 |
-
[REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
|
| 109 |
-
[REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
|
| 110 |
-
[REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
|
| 111 |
-
[REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
|
| 112 |
-
[REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
|
| 113 |
-
[REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
|
| 114 |
-
[REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
|
| 115 |
-
[REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
|
| 116 |
-
[REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
|
| 117 |
-
[REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
|
| 118 |
-
[REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
|
| 119 |
-
[REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
|
| 120 |
-
[REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
|
| 121 |
-
[REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
|
| 122 |
-
[REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
|
| 123 |
-
[REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
|
| 124 |
-
Cleaning DEV...
|
| 125 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 126 |
-
Cleaning TEST...
|
| 127 |
-
Writing TRAIN → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (60334 valid sentences)
|
| 128 |
-
Writing DEV → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (9 valid sentences)
|
| 129 |
-
Writing TEST → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (99 valid sentences)
|
| 130 |
-
Done.
|
| 131 |
-
Sourcing scripts/config_alvis.sh
|
| 132 |
-
Running stanza dataset preparation…
|
| 133 |
-
2025-12-12 14:59:02 INFO: Datasets program called with:
|
| 134 |
-
/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 135 |
-
2025-12-12 14:59:02 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
|
| 136 |
-
|
| 137 |
-
2025-12-12 14:59:02 INFO: Downloaded file to /cephyr/users/cleland/Alvis/stanza_resources/resources.json
|
| 138 |
-
2025-12-12 14:59:02 DEBUG: Processing parameter "processors"...
|
| 139 |
-
2025-12-12 14:59:02 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
|
| 140 |
-
2025-12-12 14:59:02 INFO: Downloading these customized packages for language: sv (Swedish)...
|
| 141 |
-
=======================
|
| 142 |
-
| Processor | Package |
|
| 143 |
-
-----------------------
|
| 144 |
-
=======================
|
| 145 |
-
|
| 146 |
-
2025-12-12 14:59:02 INFO: Finished downloading models and saved to /cephyr/users/cleland/Alvis/stanza_resources
|
| 147 |
-
2025-12-12 14:59:02 INFO: Using tagger model in /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt for sv_diachronic
|
| 148 |
-
2025-12-12 14:59:02 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 149 |
-
2025-12-12 14:59:02 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 150 |
-
Augmented 189 quotes: Counter({'""': 26, '「」': 23, '„”': 23, '»«': 22, '″″': 18, '””': 18, '““': 17, '《》': 17, '„“': 13, '«»': 12})
|
| 151 |
-
2025-12-12 14:59:05 INFO: Running tagger to retag /local/tmp.5491708/tmpc2soyxjt/sv_diachronic.train.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 152 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5491708/tmpc2soyxjt/sv_diachronic.train.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu']
|
| 153 |
-
2025-12-12 14:59:05 INFO: Running tagger in predict mode
|
| 154 |
-
2025-12-12 14:59:05 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 155 |
-
2025-12-12 14:59:07 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 156 |
-
2025-12-12 14:59:07 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 157 |
-
2025-12-12 14:59:07 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 158 |
-
2025-12-12 14:59:07 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 159 |
-
2025-12-12 14:59:08 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 160 |
-
2025-12-12 14:59:11 INFO: Loading data with batch size 250...
|
| 161 |
-
./make_new_model.sh: line 58: 3492354 Terminated python -m stanza.utils.datasets.prepare_depparse_treebank UD_Swedish-diachronic --wordvec_pretrain_file "/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/${PRETRAINED_MODEL}"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_conll17.pt_sv_diachron_20251212_150001.txt
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
logs/log_conll17.pt_sv_diachron_20260116_065352.txt
DELETED
|
@@ -1,830 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_diachron_20260116_065352.txt ===
|
| 2 |
-
Language codes: sv diachron
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-test.py sv diachron
|
| 6 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec991-spf148.conllu
|
| 7 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec988-spf145.conllu
|
| 8 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 9 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 10 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 11 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 12 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 13 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 14 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 15 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 16 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 17 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1102-spf259.conllu
|
| 18 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-letter141673-Stalhammar.conllu
|
| 19 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1033-spf190.conllu
|
| 20 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 21 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 22 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 23 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 24 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 25 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 26 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 27 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 28 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 29 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 30 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 31 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec25-Runius.conllu
|
| 32 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 33 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 34 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec987-spf144.conllu
|
| 35 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1063-spf220.conllu
|
| 36 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 37 |
-
Including DigPhil MACHINE in TRAIN (minus gold)…
|
| 38 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 39 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 40 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 41 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 42 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
|
| 43 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 44 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 45 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 46 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
|
| 47 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
|
| 48 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 49 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
|
| 50 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
|
| 51 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
|
| 52 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 53 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 54 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 55 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 56 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 57 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
|
| 58 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
|
| 59 |
-
Cleaning TRAIN...
|
| 60 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 61 |
-
[REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
|
| 62 |
-
[REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
|
| 63 |
-
[REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
|
| 64 |
-
[REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
|
| 65 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
|
| 66 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
|
| 67 |
-
[REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
|
| 68 |
-
[REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
|
| 69 |
-
[REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
|
| 70 |
-
[REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
|
| 71 |
-
[REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
|
| 72 |
-
[REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
|
| 73 |
-
[REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
|
| 74 |
-
[REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
|
| 75 |
-
[REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
|
| 76 |
-
[REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
|
| 77 |
-
[REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
|
| 78 |
-
[REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
|
| 79 |
-
[REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
|
| 80 |
-
[REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
|
| 81 |
-
[REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
|
| 82 |
-
[REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
|
| 83 |
-
[REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
|
| 84 |
-
[REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
|
| 85 |
-
[REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
|
| 86 |
-
[REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
|
| 87 |
-
[REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
|
| 88 |
-
[REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
|
| 89 |
-
[REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
|
| 90 |
-
[REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
|
| 91 |
-
[REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
|
| 92 |
-
[REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
|
| 93 |
-
[REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
|
| 94 |
-
[REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
|
| 95 |
-
[REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
|
| 96 |
-
[REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
|
| 97 |
-
[REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
|
| 98 |
-
[REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
|
| 99 |
-
[REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
|
| 100 |
-
[REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
|
| 101 |
-
[REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
|
| 102 |
-
[REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
|
| 103 |
-
[REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
|
| 104 |
-
[REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
|
| 105 |
-
[REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
|
| 106 |
-
[REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
|
| 107 |
-
[REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
|
| 108 |
-
[REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
|
| 109 |
-
[REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
|
| 110 |
-
[REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
|
| 111 |
-
[REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
|
| 112 |
-
[REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
|
| 113 |
-
[REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
|
| 114 |
-
[REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
|
| 115 |
-
[REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
|
| 116 |
-
[REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
|
| 117 |
-
[REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
|
| 118 |
-
[REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
|
| 119 |
-
[REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
|
| 120 |
-
[REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
|
| 121 |
-
[REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
|
| 122 |
-
[REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
|
| 123 |
-
[REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
|
| 124 |
-
Cleaning DEV...
|
| 125 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 126 |
-
Cleaning TEST...
|
| 127 |
-
Writing TRAIN → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (60334 valid sentences)
|
| 128 |
-
Writing DEV → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (9 valid sentences)
|
| 129 |
-
Writing TEST → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (99 valid sentences)
|
| 130 |
-
Done.
|
| 131 |
-
Sourcing scripts/config_alvis.sh
|
| 132 |
-
Running stanza dataset preparation…
|
| 133 |
-
2026-01-16 06:54:03 INFO: Datasets program called with:
|
| 134 |
-
/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 135 |
-
2026-01-16 06:54:03 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
|
| 136 |
-
|
| 137 |
-
2026-01-16 06:54:03 INFO: Downloaded file to /cephyr/users/cleland/Alvis/stanza_resources/resources.json
|
| 138 |
-
2026-01-16 06:54:03 DEBUG: Processing parameter "processors"...
|
| 139 |
-
2026-01-16 06:54:03 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
|
| 140 |
-
2026-01-16 06:54:03 INFO: Downloading these customized packages for language: sv (Swedish)...
|
| 141 |
-
=======================
|
| 142 |
-
| Processor | Package |
|
| 143 |
-
-----------------------
|
| 144 |
-
=======================
|
| 145 |
-
|
| 146 |
-
2026-01-16 06:54:03 INFO: Finished downloading models and saved to /cephyr/users/cleland/Alvis/stanza_resources
|
| 147 |
-
2026-01-16 06:54:03 INFO: Using tagger model in /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt for sv_diachronic
|
| 148 |
-
2026-01-16 06:54:03 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 149 |
-
2026-01-16 06:54:03 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 150 |
-
Augmented 189 quotes: Counter({'""': 26, '「」': 23, '„”': 23, '»«': 22, '″″': 18, '””': 18, '““': 17, '《》': 17, '„“': 13, '«»': 12})
|
| 151 |
-
2026-01-16 06:54:06 INFO: Running tagger to retag /local/tmp.5658622/tmp9_y5_489/sv_diachronic.train.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 152 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5658622/tmp9_y5_489/sv_diachronic.train.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu']
|
| 153 |
-
2026-01-16 06:54:06 INFO: Running tagger in predict mode
|
| 154 |
-
2026-01-16 06:54:06 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 155 |
-
2026-01-16 06:54:08 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 156 |
-
2026-01-16 06:54:08 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 157 |
-
2026-01-16 06:54:08 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 158 |
-
2026-01-16 06:54:08 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 159 |
-
2026-01-16 06:54:08 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 160 |
-
2026-01-16 06:54:12 INFO: Loading data with batch size 250...
|
| 161 |
-
2026-01-16 06:54:44 INFO: Start evaluation...
|
| 162 |
-
2026-01-16 06:57:49 INFO: UPOS XPOS UFeats AllTags
|
| 163 |
-
2026-01-16 06:57:49 INFO: 99.37 87.38 98.56 87.09
|
| 164 |
-
2026-01-16 06:57:49 INFO: POS Tagger score: sv_diachronic 87.09
|
| 165 |
-
2026-01-16 06:57:50 INFO: Running tagger to retag /local/tmp.5658622/tmp9_y5_489/sv_diachronic.dev.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 166 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5658622/tmp9_y5_489/sv_diachronic.dev.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu']
|
| 167 |
-
2026-01-16 06:57:50 INFO: Running tagger in predict mode
|
| 168 |
-
2026-01-16 06:57:50 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 169 |
-
2026-01-16 06:57:51 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 170 |
-
2026-01-16 06:57:51 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 171 |
-
2026-01-16 06:57:51 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 172 |
-
2026-01-16 06:57:52 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 173 |
-
2026-01-16 06:57:52 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 174 |
-
2026-01-16 06:57:52 INFO: Loading data with batch size 250...
|
| 175 |
-
2026-01-16 06:57:52 INFO: Start evaluation...
|
| 176 |
-
2026-01-16 06:57:52 INFO: UPOS XPOS UFeats AllTags
|
| 177 |
-
2026-01-16 06:57:52 INFO: 93.32 90.84 93.32 85.64
|
| 178 |
-
2026-01-16 06:57:52 INFO: POS Tagger score: sv_diachronic 85.64
|
| 179 |
-
2026-01-16 06:57:52 INFO: Running tagger to retag /local/tmp.5658622/tmp9_y5_489/sv_diachronic.test.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
|
| 180 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5658622/tmp9_y5_489/sv_diachronic.test.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu']
|
| 181 |
-
2026-01-16 06:57:52 INFO: Running tagger in predict mode
|
| 182 |
-
2026-01-16 06:57:52 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 183 |
-
2026-01-16 06:57:54 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 184 |
-
2026-01-16 06:57:54 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 185 |
-
2026-01-16 06:57:54 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 186 |
-
2026-01-16 06:57:54 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 187 |
-
2026-01-16 06:57:54 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 188 |
-
2026-01-16 06:57:54 INFO: Loading data with batch size 250...
|
| 189 |
-
2026-01-16 06:57:54 INFO: Start evaluation...
|
| 190 |
-
2026-01-16 06:57:55 INFO: UPOS XPOS UFeats AllTags
|
| 191 |
-
2026-01-16 06:57:55 INFO: 93.14 96.78 95.32 90.28
|
| 192 |
-
2026-01-16 06:57:55 INFO: POS Tagger score: sv_diachronic 90.28
|
| 193 |
-
Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
|
| 194 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /local/tmp.5658622/tmp9_y5_489/sv_diachronic.train.gold.conllu
|
| 195 |
-
Swapped 'w1, w2' for 'w1 ,w2' 106 times
|
| 196 |
-
Added 520 new sentences with asdf, zzzz -> asdf,zzzz
|
| 197 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /local/tmp.5658622/tmp9_y5_489/sv_diachronic.dev.gold.conllu
|
| 198 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /local/tmp.5658622/tmp9_y5_489/sv_diachronic.test.gold.conllu
|
| 199 |
-
Running stanza dependency parser training…
|
| 200 |
-
2026-01-16 06:58:08 INFO: Training program called with:
|
| 201 |
-
/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt --batch_size 8 --dropout 0.33
|
| 202 |
-
2026-01-16 06:58:08 DEBUG: UD_Swedish-diachronic: sv_diachronic
|
| 203 |
-
2026-01-16 06:58:08 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 204 |
-
2026-01-16 06:58:08 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 205 |
-
2026-01-16 06:58:08 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_charlm_parser.pt does not exist, training new model
|
| 206 |
-
2026-01-16 06:58:08 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 207 |
-
2026-01-16 06:58:08 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 208 |
-
2026-01-16 06:58:08 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--train_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '8', '--dropout', '0.33']
|
| 209 |
-
2026-01-16 06:58:08 INFO: Running parser in train mode
|
| 210 |
-
2026-01-16 06:58:08 INFO: Using pretrained contextualized char embedding
|
| 211 |
-
2026-01-16 06:58:08 INFO: Loading data with batch size 8...
|
| 212 |
-
2026-01-16 06:58:14 INFO: Train File /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 60854
|
| 213 |
-
2026-01-16 06:58:14 INFO: Original data size: 60854
|
| 214 |
-
2026-01-16 06:58:14 INFO: Augmented data size: 60854
|
| 215 |
-
2026-01-16 06:58:29 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
|
| 216 |
-
2026-01-16 06:58:29 INFO: Original length = 60854
|
| 217 |
-
2026-01-16 06:58:29 INFO: Filtered length = 60854
|
| 218 |
-
2026-01-16 06:58:43 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='|') for the xpos factory for sv_diachronic
|
| 219 |
-
2026-01-16 06:58:48 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 220 |
-
2026-01-16 06:58:59 DEBUG: 58067 batches created.
|
| 221 |
-
2026-01-16 06:58:59 DEBUG: 9 batches created.
|
| 222 |
-
2026-01-16 06:58:59 INFO: Training parser...
|
| 223 |
-
2026-01-16 06:58:59 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 224 |
-
2026-01-16 06:58:59 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 225 |
-
2026-01-16 06:59:00 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 226 |
-
2026-01-16 06:59:00 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 227 |
-
2026-01-16 06:59:04 INFO: Finished STEP 20/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
|
| 228 |
-
2026-01-16 06:59:05 INFO: Finished STEP 40/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
|
| 229 |
-
2026-01-16 06:59:05 INFO: Finished STEP 60/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
|
| 230 |
-
2026-01-16 06:59:06 INFO: Finished STEP 80/50000, loss = 0.000000 (0.032 sec/batch), lr: 0.003000
|
| 231 |
-
2026-01-16 06:59:07 INFO: Finished STEP 100/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
|
| 232 |
-
2026-01-16 06:59:07 INFO: Evaluating on dev set...
|
| 233 |
-
2026-01-16 06:59:07 INFO: LAS MLAS BLEX
|
| 234 |
-
2026-01-16 06:59:07 INFO: 0.50 0.31 0.63
|
| 235 |
-
2026-01-16 06:59:07 INFO: step 100: train_loss = 192.073832, dev_score = 0.0050
|
| 236 |
-
2026-01-16 06:59:07 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 237 |
-
2026-01-16 06:59:07 INFO: new best model saved.
|
| 238 |
-
2026-01-16 06:59:08 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 239 |
-
2026-01-16 06:59:08 INFO: new model checkpoint saved.
|
| 240 |
-
2026-01-16 06:59:08 INFO: Finished STEP 120/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
|
| 241 |
-
2026-01-16 06:59:09 INFO: Finished STEP 140/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
|
| 242 |
-
2026-01-16 06:59:10 INFO: Finished STEP 160/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
|
| 243 |
-
2026-01-16 06:59:10 INFO: Finished STEP 180/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
|
| 244 |
-
2026-01-16 06:59:11 INFO: Finished STEP 200/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
|
| 245 |
-
2026-01-16 06:59:11 INFO: Evaluating on dev set...
|
| 246 |
-
2026-01-16 06:59:11 INFO: LAS MLAS BLEX
|
| 247 |
-
2026-01-16 06:59:11 INFO: 0.50 0.31 0.63
|
| 248 |
-
2026-01-16 06:59:11 INFO: step 200: train_loss = 0.000000, dev_score = 0.0050
|
| 249 |
-
2026-01-16 06:59:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 250 |
-
2026-01-16 06:59:12 INFO: new best model saved.
|
| 251 |
-
2026-01-16 06:59:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 252 |
-
2026-01-16 06:59:12 INFO: new model checkpoint saved.
|
| 253 |
-
2026-01-16 06:59:13 INFO: Finished STEP 220/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
|
| 254 |
-
2026-01-16 06:59:13 INFO: Finished STEP 240/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
|
| 255 |
-
2026-01-16 06:59:14 INFO: Finished STEP 260/50000, loss = 0.000000 (0.031 sec/batch), lr: 0.003000
|
| 256 |
-
2026-01-16 06:59:15 INFO: Finished STEP 280/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
|
| 257 |
-
2026-01-16 06:59:15 INFO: Finished STEP 300/50000, loss = 2.877054 (0.032 sec/batch), lr: 0.003000
|
| 258 |
-
2026-01-16 06:59:15 INFO: Evaluating on dev set...
|
| 259 |
-
2026-01-16 06:59:16 INFO: LAS MLAS BLEX
|
| 260 |
-
2026-01-16 06:59:16 INFO: 1.49 0.00 0.00
|
| 261 |
-
2026-01-16 06:59:16 INFO: step 300: train_loss = 32.398629, dev_score = 0.0149
|
| 262 |
-
2026-01-16 06:59:16 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 263 |
-
2026-01-16 06:59:16 INFO: new best model saved.
|
| 264 |
-
2026-01-16 06:59:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 265 |
-
2026-01-16 06:59:17 INFO: new model checkpoint saved.
|
| 266 |
-
2026-01-16 06:59:17 INFO: Finished STEP 320/50000, loss = 4.388504 (0.032 sec/batch), lr: 0.003000
|
| 267 |
-
2026-01-16 06:59:18 INFO: Finished STEP 340/50000, loss = 1.865119 (0.031 sec/batch), lr: 0.003000
|
| 268 |
-
2026-01-16 06:59:18 INFO: Finished STEP 360/50000, loss = 2.029490 (0.031 sec/batch), lr: 0.003000
|
| 269 |
-
2026-01-16 06:59:19 INFO: Finished STEP 380/50000, loss = 3.120337 (0.033 sec/batch), lr: 0.003000
|
| 270 |
-
2026-01-16 06:59:20 INFO: Finished STEP 400/50000, loss = 0.527415 (0.031 sec/batch), lr: 0.003000
|
| 271 |
-
2026-01-16 06:59:20 INFO: Evaluating on dev set...
|
| 272 |
-
2026-01-16 06:59:20 INFO: LAS MLAS BLEX
|
| 273 |
-
2026-01-16 06:59:20 INFO: 2.97 2.74 2.74
|
| 274 |
-
2026-01-16 06:59:20 INFO: step 400: train_loss = 2.460383, dev_score = 0.0297
|
| 275 |
-
2026-01-16 06:59:21 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 276 |
-
2026-01-16 06:59:21 INFO: new best model saved.
|
| 277 |
-
2026-01-16 06:59:21 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 278 |
-
2026-01-16 06:59:21 INFO: new model checkpoint saved.
|
| 279 |
-
2026-01-16 06:59:22 INFO: Finished STEP 420/50000, loss = 0.000224 (0.030 sec/batch), lr: 0.003000
|
| 280 |
-
2026-01-16 06:59:22 INFO: Finished STEP 440/50000, loss = 1.204638 (0.031 sec/batch), lr: 0.003000
|
| 281 |
-
2026-01-16 06:59:23 INFO: Finished STEP 460/50000, loss = 1.052635 (0.031 sec/batch), lr: 0.003000
|
| 282 |
-
2026-01-16 06:59:24 INFO: Finished STEP 480/50000, loss = 2.311400 (0.032 sec/batch), lr: 0.003000
|
| 283 |
-
2026-01-16 06:59:24 INFO: Finished STEP 500/50000, loss = 1.092946 (0.031 sec/batch), lr: 0.003000
|
| 284 |
-
2026-01-16 06:59:24 INFO: Evaluating on dev set...
|
| 285 |
-
2026-01-16 06:59:25 INFO: LAS MLAS BLEX
|
| 286 |
-
2026-01-16 06:59:25 INFO: 2.48 0.70 1.41
|
| 287 |
-
2026-01-16 06:59:25 INFO: step 500: train_loss = 1.338739, dev_score = 0.0248
|
| 288 |
-
2026-01-16 06:59:25 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 289 |
-
2026-01-16 06:59:25 INFO: new model checkpoint saved.
|
| 290 |
-
2026-01-16 06:59:26 INFO: Finished STEP 520/50000, loss = 0.652456 (0.030 sec/batch), lr: 0.003000
|
| 291 |
-
2026-01-16 06:59:27 INFO: Finished STEP 540/50000, loss = 0.147004 (0.031 sec/batch), lr: 0.003000
|
| 292 |
-
2026-01-16 06:59:27 INFO: Finished STEP 560/50000, loss = 0.782623 (0.030 sec/batch), lr: 0.003000
|
| 293 |
-
2026-01-16 06:59:28 INFO: Finished STEP 580/50000, loss = 0.076848 (0.030 sec/batch), lr: 0.003000
|
| 294 |
-
2026-01-16 06:59:28 INFO: Finished STEP 600/50000, loss = 0.429073 (0.031 sec/batch), lr: 0.003000
|
| 295 |
-
2026-01-16 06:59:28 INFO: Evaluating on dev set...
|
| 296 |
-
2026-01-16 06:59:29 INFO: LAS MLAS BLEX
|
| 297 |
-
2026-01-16 06:59:29 INFO: 3.71 2.21 2.65
|
| 298 |
-
2026-01-16 06:59:29 INFO: step 600: train_loss = 1.109608, dev_score = 0.0371
|
| 299 |
-
2026-01-16 06:59:29 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 300 |
-
2026-01-16 06:59:29 INFO: new best model saved.
|
| 301 |
-
2026-01-16 06:59:30 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 302 |
-
2026-01-16 06:59:30 INFO: new model checkpoint saved.
|
| 303 |
-
2026-01-16 06:59:30 INFO: Finished STEP 620/50000, loss = 0.636850 (0.033 sec/batch), lr: 0.003000
|
| 304 |
-
2026-01-16 06:59:31 INFO: Finished STEP 640/50000, loss = 0.097528 (0.032 sec/batch), lr: 0.003000
|
| 305 |
-
2026-01-16 06:59:32 INFO: Finished STEP 660/50000, loss = 0.328014 (0.031 sec/batch), lr: 0.003000
|
| 306 |
-
2026-01-16 06:59:32 INFO: Finished STEP 680/50000, loss = 1.418195 (0.031 sec/batch), lr: 0.003000
|
| 307 |
-
2026-01-16 06:59:33 INFO: Finished STEP 700/50000, loss = 0.000002 (0.031 sec/batch), lr: 0.003000
|
| 308 |
-
2026-01-16 06:59:33 INFO: Evaluating on dev set...
|
| 309 |
-
2026-01-16 06:59:34 INFO: LAS MLAS BLEX
|
| 310 |
-
2026-01-16 06:59:34 INFO: 3.47 2.50 2.50
|
| 311 |
-
2026-01-16 06:59:34 INFO: step 700: train_loss = 1.152763, dev_score = 0.0347
|
| 312 |
-
2026-01-16 06:59:34 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 313 |
-
2026-01-16 06:59:34 INFO: new model checkpoint saved.
|
| 314 |
-
2026-01-16 06:59:35 INFO: Finished STEP 720/50000, loss = 3.007647 (0.031 sec/batch), lr: 0.003000
|
| 315 |
-
2026-01-16 06:59:35 INFO: Finished STEP 740/50000, loss = 1.066953 (0.031 sec/batch), lr: 0.003000
|
| 316 |
-
2026-01-16 06:59:36 INFO: Finished STEP 760/50000, loss = 1.151768 (0.030 sec/batch), lr: 0.003000
|
| 317 |
-
2026-01-16 06:59:37 INFO: Finished STEP 780/50000, loss = 1.043154 (0.032 sec/batch), lr: 0.003000
|
| 318 |
-
2026-01-16 06:59:37 INFO: Finished STEP 800/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
|
| 319 |
-
2026-01-16 06:59:37 INFO: Evaluating on dev set...
|
| 320 |
-
2026-01-16 06:59:38 INFO: LAS MLAS BLEX
|
| 321 |
-
2026-01-16 06:59:38 INFO: 3.71 2.21 2.58
|
| 322 |
-
2026-01-16 06:59:38 INFO: step 800: train_loss = 1.042035, dev_score = 0.0371
|
| 323 |
-
2026-01-16 06:59:38 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 324 |
-
2026-01-16 06:59:38 INFO: new best model saved.
|
| 325 |
-
2026-01-16 06:59:38 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 326 |
-
2026-01-16 06:59:38 INFO: new model checkpoint saved.
|
| 327 |
-
2026-01-16 06:59:39 INFO: Finished STEP 820/50000, loss = 0.309987 (0.031 sec/batch), lr: 0.003000
|
| 328 |
-
2026-01-16 06:59:40 INFO: Finished STEP 840/50000, loss = 0.046532 (0.030 sec/batch), lr: 0.003000
|
| 329 |
-
2026-01-16 06:59:40 INFO: Finished STEP 860/50000, loss = 2.719592 (0.031 sec/batch), lr: 0.003000
|
| 330 |
-
2026-01-16 06:59:41 INFO: Finished STEP 880/50000, loss = 1.216663 (0.030 sec/batch), lr: 0.003000
|
| 331 |
-
2026-01-16 06:59:42 INFO: Finished STEP 900/50000, loss = 1.323724 (0.032 sec/batch), lr: 0.003000
|
| 332 |
-
2026-01-16 06:59:42 INFO: Evaluating on dev set...
|
| 333 |
-
2026-01-16 06:59:42 INFO: LAS MLAS BLEX
|
| 334 |
-
2026-01-16 06:59:42 INFO: 3.71 2.71 3.05
|
| 335 |
-
2026-01-16 06:59:42 INFO: step 900: train_loss = 1.123192, dev_score = 0.0371
|
| 336 |
-
2026-01-16 06:59:42 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 337 |
-
2026-01-16 06:59:42 INFO: new best model saved.
|
| 338 |
-
2026-01-16 06:59:43 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 339 |
-
2026-01-16 06:59:43 INFO: new model checkpoint saved.
|
| 340 |
-
2026-01-16 06:59:43 INFO: Finished STEP 920/50000, loss = 0.152121 (0.031 sec/batch), lr: 0.003000
|
| 341 |
-
2026-01-16 06:59:44 INFO: Finished STEP 940/50000, loss = 0.897826 (0.031 sec/batch), lr: 0.003000
|
| 342 |
-
2026-01-16 06:59:45 INFO: Finished STEP 960/50000, loss = 0.000258 (0.030 sec/batch), lr: 0.003000
|
| 343 |
-
2026-01-16 06:59:45 INFO: Finished STEP 980/50000, loss = 2.134546 (0.030 sec/batch), lr: 0.003000
|
| 344 |
-
2026-01-16 06:59:46 INFO: Finished STEP 1000/50000, loss = 0.035686 (0.030 sec/batch), lr: 0.003000
|
| 345 |
-
2026-01-16 06:59:46 INFO: Evaluating on dev set...
|
| 346 |
-
2026-01-16 06:59:46 INFO: LAS MLAS BLEX
|
| 347 |
-
2026-01-16 06:59:46 INFO: 3.71 2.49 2.85
|
| 348 |
-
2026-01-16 06:59:46 INFO: step 1000: train_loss = 0.984753, dev_score = 0.0371
|
| 349 |
-
2026-01-16 06:59:47 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 350 |
-
2026-01-16 06:59:47 INFO: new best model saved.
|
| 351 |
-
2026-01-16 06:59:47 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 352 |
-
2026-01-16 06:59:47 INFO: new model checkpoint saved.
|
| 353 |
-
2026-01-16 06:59:48 INFO: Finished STEP 1020/50000, loss = 1.837077 (0.031 sec/batch), lr: 0.003000
|
| 354 |
-
2026-01-16 06:59:48 INFO: Finished STEP 1040/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
|
| 355 |
-
2026-01-16 06:59:49 INFO: Finished STEP 1060/50000, loss = 2.368280 (0.030 sec/batch), lr: 0.003000
|
| 356 |
-
2026-01-16 06:59:50 INFO: Finished STEP 1080/50000, loss = 0.469828 (0.031 sec/batch), lr: 0.003000
|
| 357 |
-
2026-01-16 06:59:50 INFO: Finished STEP 1100/50000, loss = 0.000491 (0.031 sec/batch), lr: 0.003000
|
| 358 |
-
2026-01-16 06:59:50 INFO: Evaluating on dev set...
|
| 359 |
-
2026-01-16 06:59:51 INFO: LAS MLAS BLEX
|
| 360 |
-
2026-01-16 06:59:51 INFO: 4.95 4.11 5.48
|
| 361 |
-
2026-01-16 06:59:51 INFO: step 1100: train_loss = 1.174486, dev_score = 0.0495
|
| 362 |
-
2026-01-16 06:59:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 363 |
-
2026-01-16 06:59:51 INFO: new best model saved.
|
| 364 |
-
2026-01-16 06:59:52 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 365 |
-
2026-01-16 06:59:52 INFO: new model checkpoint saved.
|
| 366 |
-
2026-01-16 06:59:52 INFO: Finished STEP 1120/50000, loss = 5.331606 (0.033 sec/batch), lr: 0.003000
|
| 367 |
-
2026-01-16 06:59:53 INFO: Finished STEP 1140/50000, loss = 4.487067 (0.032 sec/batch), lr: 0.003000
|
| 368 |
-
2026-01-16 06:59:54 INFO: Finished STEP 1160/50000, loss = 1.940725 (0.031 sec/batch), lr: 0.003000
|
| 369 |
-
2026-01-16 06:59:54 INFO: Finished STEP 1180/50000, loss = 2.472967 (0.032 sec/batch), lr: 0.003000
|
| 370 |
-
2026-01-16 06:59:55 INFO: Finished STEP 1200/50000, loss = 0.686162 (0.032 sec/batch), lr: 0.003000
|
| 371 |
-
2026-01-16 06:59:55 INFO: Evaluating on dev set...
|
| 372 |
-
2026-01-16 06:59:56 INFO: LAS MLAS BLEX
|
| 373 |
-
2026-01-16 06:59:56 INFO: 6.68 4.08 5.10
|
| 374 |
-
2026-01-16 06:59:56 INFO: step 1200: train_loss = 2.657063, dev_score = 0.0668
|
| 375 |
-
2026-01-16 06:59:56 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 376 |
-
2026-01-16 06:59:56 INFO: new best model saved.
|
| 377 |
-
2026-01-16 06:59:56 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 378 |
-
2026-01-16 06:59:56 INFO: new model checkpoint saved.
|
| 379 |
-
2026-01-16 06:59:57 INFO: Finished STEP 1220/50000, loss = 0.828940 (0.031 sec/batch), lr: 0.003000
|
| 380 |
-
2026-01-16 06:59:58 INFO: Finished STEP 1240/50000, loss = 1.596623 (0.032 sec/batch), lr: 0.003000
|
| 381 |
-
2026-01-16 06:59:58 INFO: Finished STEP 1260/50000, loss = 6.153142 (0.032 sec/batch), lr: 0.003000
|
| 382 |
-
2026-01-16 06:59:59 INFO: Finished STEP 1280/50000, loss = 1.988011 (0.031 sec/batch), lr: 0.003000
|
| 383 |
-
2026-01-16 07:00:00 INFO: Finished STEP 1300/50000, loss = 3.000342 (0.032 sec/batch), lr: 0.003000
|
| 384 |
-
2026-01-16 07:00:00 INFO: Evaluating on dev set...
|
| 385 |
-
2026-01-16 07:00:00 INFO: LAS MLAS BLEX
|
| 386 |
-
2026-01-16 07:00:00 INFO: 7.67 5.80 6.48
|
| 387 |
-
2026-01-16 07:00:00 INFO: step 1300: train_loss = 1.683935, dev_score = 0.0767
|
| 388 |
-
2026-01-16 07:00:00 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 389 |
-
2026-01-16 07:00:00 INFO: new best model saved.
|
| 390 |
-
2026-01-16 07:00:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 391 |
-
2026-01-16 07:00:01 INFO: new model checkpoint saved.
|
| 392 |
-
2026-01-16 07:00:02 INFO: Finished STEP 1320/50000, loss = 2.859746 (0.032 sec/batch), lr: 0.003000
|
| 393 |
-
2026-01-16 07:00:02 INFO: Finished STEP 1340/50000, loss = 1.214685 (0.032 sec/batch), lr: 0.003000
|
| 394 |
-
2026-01-16 07:00:03 INFO: Finished STEP 1360/50000, loss = 0.369348 (0.032 sec/batch), lr: 0.003000
|
| 395 |
-
2026-01-16 07:00:04 INFO: Finished STEP 1380/50000, loss = 1.401275 (0.032 sec/batch), lr: 0.003000
|
| 396 |
-
2026-01-16 07:00:04 INFO: Finished STEP 1400/50000, loss = 1.433550 (0.032 sec/batch), lr: 0.003000
|
| 397 |
-
2026-01-16 07:00:04 INFO: Evaluating on dev set...
|
| 398 |
-
2026-01-16 07:00:05 INFO: LAS MLAS BLEX
|
| 399 |
-
2026-01-16 07:00:05 INFO: 9.65 4.51 5.64
|
| 400 |
-
2026-01-16 07:00:05 INFO: step 1400: train_loss = 1.665584, dev_score = 0.0965
|
| 401 |
-
2026-01-16 07:00:05 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 402 |
-
2026-01-16 07:00:05 INFO: new best model saved.
|
| 403 |
-
2026-01-16 07:00:05 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 404 |
-
2026-01-16 07:00:05 INFO: new model checkpoint saved.
|
| 405 |
-
2026-01-16 07:00:06 INFO: Finished STEP 1420/50000, loss = 1.668130 (0.032 sec/batch), lr: 0.003000
|
| 406 |
-
2026-01-16 07:00:07 INFO: Finished STEP 1440/50000, loss = 0.647845 (0.031 sec/batch), lr: 0.003000
|
| 407 |
-
2026-01-16 07:00:07 INFO: Finished STEP 1460/50000, loss = 1.816998 (0.034 sec/batch), lr: 0.003000
|
| 408 |
-
2026-01-16 07:00:08 INFO: Finished STEP 1480/50000, loss = 1.645519 (0.032 sec/batch), lr: 0.003000
|
| 409 |
-
2026-01-16 07:00:09 INFO: Finished STEP 1500/50000, loss = 3.368581 (0.032 sec/batch), lr: 0.003000
|
| 410 |
-
2026-01-16 07:00:09 INFO: Evaluating on dev set...
|
| 411 |
-
2026-01-16 07:00:10 INFO: LAS MLAS BLEX
|
| 412 |
-
2026-01-16 07:00:10 INFO: 11.14 6.09 6.81
|
| 413 |
-
2026-01-16 07:00:10 INFO: step 1500: train_loss = 1.689450, dev_score = 0.1114
|
| 414 |
-
2026-01-16 07:00:10 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 415 |
-
2026-01-16 07:00:10 INFO: new best model saved.
|
| 416 |
-
2026-01-16 07:00:10 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 417 |
-
2026-01-16 07:00:10 INFO: new model checkpoint saved.
|
| 418 |
-
2026-01-16 07:00:11 INFO: Finished STEP 1520/50000, loss = 0.696790 (0.033 sec/batch), lr: 0.003000
|
| 419 |
-
2026-01-16 07:00:12 INFO: Finished STEP 1540/50000, loss = 0.979858 (0.032 sec/batch), lr: 0.003000
|
| 420 |
-
2026-01-16 07:00:12 INFO: Finished STEP 1560/50000, loss = 1.371460 (0.032 sec/batch), lr: 0.003000
|
| 421 |
-
2026-01-16 07:00:13 INFO: Finished STEP 1580/50000, loss = 1.076014 (0.031 sec/batch), lr: 0.003000
|
| 422 |
-
2026-01-16 07:00:14 INFO: Finished STEP 1600/50000, loss = 0.347749 (0.031 sec/batch), lr: 0.003000
|
| 423 |
-
2026-01-16 07:00:14 INFO: Evaluating on dev set...
|
| 424 |
-
2026-01-16 07:00:14 INFO: LAS MLAS BLEX
|
| 425 |
-
2026-01-16 07:00:14 INFO: 18.07 4.21 5.35
|
| 426 |
-
2026-01-16 07:00:14 INFO: step 1600: train_loss = 1.549157, dev_score = 0.1807
|
| 427 |
-
2026-01-16 07:00:14 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 428 |
-
2026-01-16 07:00:14 INFO: new best model saved.
|
| 429 |
-
2026-01-16 07:00:15 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 430 |
-
2026-01-16 07:00:15 INFO: new model checkpoint saved.
|
| 431 |
-
2026-01-16 07:00:16 INFO: Finished STEP 1620/50000, loss = 0.799195 (0.032 sec/batch), lr: 0.003000
|
| 432 |
-
2026-01-16 07:00:16 INFO: Finished STEP 1640/50000, loss = 1.727682 (0.032 sec/batch), lr: 0.003000
|
| 433 |
-
2026-01-16 07:00:17 INFO: Finished STEP 1660/50000, loss = 2.875053 (0.033 sec/batch), lr: 0.003000
|
| 434 |
-
2026-01-16 07:00:18 INFO: Finished STEP 1680/50000, loss = 2.254138 (0.031 sec/batch), lr: 0.003000
|
| 435 |
-
2026-01-16 07:00:18 INFO: Finished STEP 1700/50000, loss = 0.754430 (0.031 sec/batch), lr: 0.003000
|
| 436 |
-
2026-01-16 07:00:18 INFO: Evaluating on dev set...
|
| 437 |
-
2026-01-16 07:00:19 INFO: LAS MLAS BLEX
|
| 438 |
-
2026-01-16 07:00:19 INFO: 11.39 4.62 5.01
|
| 439 |
-
2026-01-16 07:00:19 INFO: step 1700: train_loss = 1.452840, dev_score = 0.1139
|
| 440 |
-
2026-01-16 07:00:19 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 441 |
-
2026-01-16 07:00:19 INFO: new model checkpoint saved.
|
| 442 |
-
2026-01-16 07:00:20 INFO: Finished STEP 1720/50000, loss = 0.672960 (0.032 sec/batch), lr: 0.003000
|
| 443 |
-
2026-01-16 07:00:21 INFO: Finished STEP 1740/50000, loss = 1.758359 (0.031 sec/batch), lr: 0.003000
|
| 444 |
-
2026-01-16 07:00:21 INFO: Finished STEP 1760/50000, loss = 0.112267 (0.032 sec/batch), lr: 0.003000
|
| 445 |
-
2026-01-16 07:00:22 INFO: Finished STEP 1780/50000, loss = 2.130425 (0.035 sec/batch), lr: 0.003000
|
| 446 |
-
2026-01-16 07:00:23 INFO: Finished STEP 1800/50000, loss = 0.867664 (0.032 sec/batch), lr: 0.003000
|
| 447 |
-
2026-01-16 07:00:23 INFO: Evaluating on dev set...
|
| 448 |
-
2026-01-16 07:00:23 INFO: LAS MLAS BLEX
|
| 449 |
-
2026-01-16 07:00:23 INFO: 8.17 5.17 5.52
|
| 450 |
-
2026-01-16 07:00:23 INFO: step 1800: train_loss = 1.566285, dev_score = 0.0817
|
| 451 |
-
2026-01-16 07:00:24 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 452 |
-
2026-01-16 07:00:24 INFO: new model checkpoint saved.
|
| 453 |
-
2026-01-16 07:00:24 INFO: Finished STEP 1820/50000, loss = 0.918402 (0.032 sec/batch), lr: 0.003000
|
| 454 |
-
2026-01-16 07:00:25 INFO: Finished STEP 1840/50000, loss = 0.602067 (0.031 sec/batch), lr: 0.003000
|
| 455 |
-
2026-01-16 07:00:26 INFO: Finished STEP 1860/50000, loss = 4.889768 (0.031 sec/batch), lr: 0.003000
|
| 456 |
-
2026-01-16 07:00:26 INFO: Finished STEP 1880/50000, loss = 3.249488 (0.037 sec/batch), lr: 0.003000
|
| 457 |
-
2026-01-16 07:00:27 INFO: Finished STEP 1900/50000, loss = 1.829388 (0.032 sec/batch), lr: 0.003000
|
| 458 |
-
2026-01-16 07:00:27 INFO: Evaluating on dev set...
|
| 459 |
-
2026-01-16 07:00:28 INFO: LAS MLAS BLEX
|
| 460 |
-
2026-01-16 07:00:28 INFO: 9.16 3.77 5.28
|
| 461 |
-
2026-01-16 07:00:28 INFO: step 1900: train_loss = 1.531276, dev_score = 0.0916
|
| 462 |
-
2026-01-16 07:00:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 463 |
-
2026-01-16 07:00:28 INFO: new model checkpoint saved.
|
| 464 |
-
2026-01-16 07:00:29 INFO: Finished STEP 1920/50000, loss = 1.377406 (0.033 sec/batch), lr: 0.003000
|
| 465 |
-
2026-01-16 07:00:29 INFO: Finished STEP 1940/50000, loss = 0.690662 (0.032 sec/batch), lr: 0.003000
|
| 466 |
-
2026-01-16 07:00:30 INFO: Finished STEP 1960/50000, loss = 0.488795 (0.032 sec/batch), lr: 0.003000
|
| 467 |
-
2026-01-16 07:00:31 INFO: Finished STEP 1980/50000, loss = 1.489328 (0.031 sec/batch), lr: 0.003000
|
| 468 |
-
2026-01-16 07:00:31 INFO: Finished STEP 2000/50000, loss = 0.265036 (0.032 sec/batch), lr: 0.003000
|
| 469 |
-
2026-01-16 07:00:31 INFO: Evaluating on dev set...
|
| 470 |
-
2026-01-16 07:00:32 INFO: LAS MLAS BLEX
|
| 471 |
-
2026-01-16 07:00:32 INFO: 3.71 2.34 3.90
|
| 472 |
-
2026-01-16 07:00:32 INFO: step 2000: train_loss = 1.577713, dev_score = 0.0371
|
| 473 |
-
2026-01-16 07:00:32 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 474 |
-
2026-01-16 07:00:32 INFO: new model checkpoint saved.
|
| 475 |
-
2026-01-16 07:00:33 INFO: Finished STEP 2020/50000, loss = 1.118794 (0.031 sec/batch), lr: 0.003000
|
| 476 |
-
2026-01-16 07:00:34 INFO: Finished STEP 2040/50000, loss = 1.678119 (0.032 sec/batch), lr: 0.003000
|
| 477 |
-
2026-01-16 07:00:34 INFO: Finished STEP 2060/50000, loss = 0.566802 (0.032 sec/batch), lr: 0.003000
|
| 478 |
-
2026-01-16 07:00:35 INFO: Finished STEP 2080/50000, loss = 1.566789 (0.032 sec/batch), lr: 0.003000
|
| 479 |
-
2026-01-16 07:00:36 INFO: Finished STEP 2100/50000, loss = 1.108176 (0.031 sec/batch), lr: 0.003000
|
| 480 |
-
2026-01-16 07:00:36 INFO: Evaluating on dev set...
|
| 481 |
-
2026-01-16 07:00:36 INFO: LAS MLAS BLEX
|
| 482 |
-
2026-01-16 07:00:36 INFO: 8.17 5.37 6.91
|
| 483 |
-
2026-01-16 07:00:36 INFO: step 2100: train_loss = 1.353945, dev_score = 0.0817
|
| 484 |
-
2026-01-16 07:00:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 485 |
-
2026-01-16 07:00:37 INFO: new model checkpoint saved.
|
| 486 |
-
2026-01-16 07:00:37 INFO: Finished STEP 2120/50000, loss = 0.896808 (0.033 sec/batch), lr: 0.003000
|
| 487 |
-
2026-01-16 07:00:38 INFO: Finished STEP 2140/50000, loss = 3.207397 (0.032 sec/batch), lr: 0.003000
|
| 488 |
-
2026-01-16 07:00:39 INFO: Finished STEP 2160/50000, loss = 0.004755 (0.037 sec/batch), lr: 0.003000
|
| 489 |
-
2026-01-16 07:00:39 INFO: Finished STEP 2180/50000, loss = 0.760959 (0.031 sec/batch), lr: 0.003000
|
| 490 |
-
2026-01-16 07:00:40 INFO: Finished STEP 2200/50000, loss = 2.696491 (0.032 sec/batch), lr: 0.003000
|
| 491 |
-
2026-01-16 07:00:40 INFO: Evaluating on dev set...
|
| 492 |
-
2026-01-16 07:00:41 INFO: LAS MLAS BLEX
|
| 493 |
-
2026-01-16 07:00:41 INFO: 7.67 1.83 2.19
|
| 494 |
-
2026-01-16 07:00:41 INFO: step 2200: train_loss = 1.574935, dev_score = 0.0767
|
| 495 |
-
2026-01-16 07:00:41 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 496 |
-
2026-01-16 07:00:41 INFO: new model checkpoint saved.
|
| 497 |
-
2026-01-16 07:00:42 INFO: Finished STEP 2220/50000, loss = 2.683285 (0.031 sec/batch), lr: 0.003000
|
| 498 |
-
2026-01-16 07:00:42 INFO: Finished STEP 2240/50000, loss = 2.951740 (0.031 sec/batch), lr: 0.003000
|
| 499 |
-
2026-01-16 07:00:43 INFO: Finished STEP 2260/50000, loss = 4.218589 (0.031 sec/batch), lr: 0.003000
|
| 500 |
-
2026-01-16 07:00:44 INFO: Finished STEP 2280/50000, loss = 1.695766 (0.031 sec/batch), lr: 0.003000
|
| 501 |
-
2026-01-16 07:00:44 INFO: Finished STEP 2300/50000, loss = 2.362932 (0.031 sec/batch), lr: 0.003000
|
| 502 |
-
2026-01-16 07:00:44 INFO: Evaluating on dev set...
|
| 503 |
-
2026-01-16 07:00:45 INFO: LAS MLAS BLEX
|
| 504 |
-
2026-01-16 07:00:45 INFO: 12.62 3.67 4.48
|
| 505 |
-
2026-01-16 07:00:45 INFO: step 2300: train_loss = 3.007224, dev_score = 0.1262
|
| 506 |
-
2026-01-16 07:00:46 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 507 |
-
2026-01-16 07:00:46 INFO: new model checkpoint saved.
|
| 508 |
-
2026-01-16 07:00:46 INFO: Finished STEP 2320/50000, loss = 2.829835 (0.031 sec/batch), lr: 0.003000
|
| 509 |
-
2026-01-16 07:00:47 INFO: Finished STEP 2340/50000, loss = 1.246457 (0.031 sec/batch), lr: 0.003000
|
| 510 |
-
2026-01-16 07:00:47 INFO: Finished STEP 2360/50000, loss = 1.362900 (0.032 sec/batch), lr: 0.003000
|
| 511 |
-
2026-01-16 07:00:48 INFO: Finished STEP 2380/50000, loss = 1.726964 (0.031 sec/batch), lr: 0.003000
|
| 512 |
-
2026-01-16 07:00:49 INFO: Finished STEP 2400/50000, loss = 1.644795 (0.031 sec/batch), lr: 0.003000
|
| 513 |
-
2026-01-16 07:00:49 INFO: Evaluating on dev set...
|
| 514 |
-
2026-01-16 07:00:49 INFO: LAS MLAS BLEX
|
| 515 |
-
2026-01-16 07:00:49 INFO: 11.14 6.71 7.77
|
| 516 |
-
2026-01-16 07:00:49 INFO: step 2400: train_loss = 2.391629, dev_score = 0.1114
|
| 517 |
-
2026-01-16 07:00:50 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 518 |
-
2026-01-16 07:00:50 INFO: new model checkpoint saved.
|
| 519 |
-
2026-01-16 07:00:50 INFO: Finished STEP 2420/50000, loss = 1.448694 (0.030 sec/batch), lr: 0.003000
|
| 520 |
-
2026-01-16 07:00:51 INFO: Finished STEP 2440/50000, loss = 0.224588 (0.030 sec/batch), lr: 0.003000
|
| 521 |
-
2026-01-16 07:00:52 INFO: Finished STEP 2460/50000, loss = 1.348104 (0.031 sec/batch), lr: 0.003000
|
| 522 |
-
2026-01-16 07:00:52 INFO: Finished STEP 2480/50000, loss = 1.120497 (0.031 sec/batch), lr: 0.003000
|
| 523 |
-
2026-01-16 07:00:53 INFO: Finished STEP 2500/50000, loss = 2.224450 (0.032 sec/batch), lr: 0.003000
|
| 524 |
-
2026-01-16 07:00:53 INFO: Evaluating on dev set...
|
| 525 |
-
2026-01-16 07:00:53 INFO: LAS MLAS BLEX
|
| 526 |
-
2026-01-16 07:00:53 INFO: 18.56 6.11 7.63
|
| 527 |
-
2026-01-16 07:00:53 INFO: step 2500: train_loss = 2.279046, dev_score = 0.1856
|
| 528 |
-
2026-01-16 07:00:54 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 529 |
-
2026-01-16 07:00:54 INFO: new best model saved.
|
| 530 |
-
2026-01-16 07:00:54 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 531 |
-
2026-01-16 07:00:54 INFO: new model checkpoint saved.
|
| 532 |
-
2026-01-16 07:00:55 INFO: Finished STEP 2520/50000, loss = 4.519064 (0.031 sec/batch), lr: 0.003000
|
| 533 |
-
2026-01-16 07:00:55 INFO: Finished STEP 2540/50000, loss = 3.034460 (0.034 sec/batch), lr: 0.003000
|
| 534 |
-
2026-01-16 07:00:56 INFO: Finished STEP 2560/50000, loss = 2.673581 (0.032 sec/batch), lr: 0.003000
|
| 535 |
-
2026-01-16 07:00:57 INFO: Finished STEP 2580/50000, loss = 0.801564 (0.030 sec/batch), lr: 0.003000
|
| 536 |
-
2026-01-16 07:00:57 INFO: Finished STEP 2600/50000, loss = 0.521267 (0.032 sec/batch), lr: 0.003000
|
| 537 |
-
2026-01-16 07:00:57 INFO: Evaluating on dev set...
|
| 538 |
-
2026-01-16 07:00:58 INFO: LAS MLAS BLEX
|
| 539 |
-
2026-01-16 07:00:58 INFO: 18.56 6.01 7.21
|
| 540 |
-
2026-01-16 07:00:58 INFO: step 2600: train_loss = 2.528788, dev_score = 0.1856
|
| 541 |
-
2026-01-16 07:00:58 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 542 |
-
2026-01-16 07:00:58 INFO: new best model saved.
|
| 543 |
-
2026-01-16 07:00:59 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 544 |
-
2026-01-16 07:00:59 INFO: new model checkpoint saved.
|
| 545 |
-
2026-01-16 07:00:59 INFO: Finished STEP 2620/50000, loss = 1.897941 (0.031 sec/batch), lr: 0.003000
|
| 546 |
-
2026-01-16 07:01:00 INFO: Finished STEP 2640/50000, loss = 2.104889 (0.031 sec/batch), lr: 0.003000
|
| 547 |
-
2026-01-16 07:01:00 INFO: Finished STEP 2660/50000, loss = 2.285084 (0.031 sec/batch), lr: 0.003000
|
| 548 |
-
2026-01-16 07:01:01 INFO: Finished STEP 2680/50000, loss = 2.218475 (0.031 sec/batch), lr: 0.003000
|
| 549 |
-
2026-01-16 07:01:02 INFO: Finished STEP 2700/50000, loss = 6.859745 (0.031 sec/batch), lr: 0.003000
|
| 550 |
-
2026-01-16 07:01:02 INFO: Evaluating on dev set...
|
| 551 |
-
2026-01-16 07:01:02 INFO: LAS MLAS BLEX
|
| 552 |
-
2026-01-16 07:01:02 INFO: 19.55 6.19 8.12
|
| 553 |
-
2026-01-16 07:01:02 INFO: step 2700: train_loss = 2.664627, dev_score = 0.1955
|
| 554 |
-
2026-01-16 07:01:03 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 555 |
-
2026-01-16 07:01:03 INFO: new best model saved.
|
| 556 |
-
2026-01-16 07:01:03 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 557 |
-
2026-01-16 07:01:03 INFO: new model checkpoint saved.
|
| 558 |
-
2026-01-16 07:01:04 INFO: Finished STEP 2720/50000, loss = 2.067003 (0.031 sec/batch), lr: 0.003000
|
| 559 |
-
2026-01-16 07:01:04 INFO: Finished STEP 2740/50000, loss = 1.080532 (0.031 sec/batch), lr: 0.003000
|
| 560 |
-
2026-01-16 07:01:05 INFO: Finished STEP 2760/50000, loss = 4.616343 (0.032 sec/batch), lr: 0.003000
|
| 561 |
-
2026-01-16 07:01:06 INFO: Finished STEP 2780/50000, loss = 1.479079 (0.032 sec/batch), lr: 0.003000
|
| 562 |
-
2026-01-16 07:01:06 INFO: Finished STEP 2800/50000, loss = 1.063658 (0.032 sec/batch), lr: 0.003000
|
| 563 |
-
2026-01-16 07:01:06 INFO: Evaluating on dev set...
|
| 564 |
-
2026-01-16 07:01:07 INFO: LAS MLAS BLEX
|
| 565 |
-
2026-01-16 07:01:07 INFO: 17.82 4.66 6.21
|
| 566 |
-
2026-01-16 07:01:07 INFO: step 2800: train_loss = 2.274723, dev_score = 0.1782
|
| 567 |
-
2026-01-16 07:01:08 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 568 |
-
2026-01-16 07:01:08 INFO: new model checkpoint saved.
|
| 569 |
-
2026-01-16 07:01:08 INFO: Finished STEP 2820/50000, loss = 2.248970 (0.032 sec/batch), lr: 0.003000
|
| 570 |
-
2026-01-16 07:01:09 INFO: Finished STEP 2840/50000, loss = 3.309697 (0.031 sec/batch), lr: 0.003000
|
| 571 |
-
2026-01-16 07:01:10 INFO: Finished STEP 2860/50000, loss = 0.586270 (0.032 sec/batch), lr: 0.003000
|
| 572 |
-
2026-01-16 07:01:10 INFO: Finished STEP 2880/50000, loss = 1.731944 (0.031 sec/batch), lr: 0.003000
|
| 573 |
-
2026-01-16 07:01:11 INFO: Finished STEP 2900/50000, loss = 2.552610 (0.030 sec/batch), lr: 0.003000
|
| 574 |
-
2026-01-16 07:01:11 INFO: Evaluating on dev set...
|
| 575 |
-
2026-01-16 07:01:11 INFO: LAS MLAS BLEX
|
| 576 |
-
2026-01-16 07:01:11 INFO: 17.33 5.02 6.95
|
| 577 |
-
2026-01-16 07:01:11 INFO: step 2900: train_loss = 2.368524, dev_score = 0.1733
|
| 578 |
-
2026-01-16 07:01:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 579 |
-
2026-01-16 07:01:12 INFO: new model checkpoint saved.
|
| 580 |
-
2026-01-16 07:01:13 INFO: Finished STEP 2920/50000, loss = 3.126576 (0.031 sec/batch), lr: 0.003000
|
| 581 |
-
2026-01-16 07:01:13 INFO: Finished STEP 2940/50000, loss = 1.741402 (0.031 sec/batch), lr: 0.003000
|
| 582 |
-
2026-01-16 07:01:14 INFO: Finished STEP 2960/50000, loss = 3.777675 (0.031 sec/batch), lr: 0.003000
|
| 583 |
-
2026-01-16 07:01:15 INFO: Finished STEP 2980/50000, loss = 1.370117 (0.031 sec/batch), lr: 0.003000
|
| 584 |
-
2026-01-16 07:01:15 INFO: Finished STEP 3000/50000, loss = 3.574947 (0.031 sec/batch), lr: 0.003000
|
| 585 |
-
2026-01-16 07:01:15 INFO: Evaluating on dev set...
|
| 586 |
-
2026-01-16 07:01:16 INFO: LAS MLAS BLEX
|
| 587 |
-
2026-01-16 07:01:16 INFO: 17.82 6.94 8.48
|
| 588 |
-
2026-01-16 07:01:16 INFO: step 3000: train_loss = 2.539027, dev_score = 0.1782
|
| 589 |
-
2026-01-16 07:01:16 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 590 |
-
2026-01-16 07:01:16 INFO: new model checkpoint saved.
|
| 591 |
-
2026-01-16 07:01:17 INFO: Finished STEP 3020/50000, loss = 1.453705 (0.031 sec/batch), lr: 0.003000
|
| 592 |
-
2026-01-16 07:01:18 INFO: Finished STEP 3040/50000, loss = 6.184045 (0.032 sec/batch), lr: 0.003000
|
| 593 |
-
2026-01-16 07:01:18 INFO: Finished STEP 3060/50000, loss = 4.618065 (0.030 sec/batch), lr: 0.003000
|
| 594 |
-
2026-01-16 07:01:19 INFO: Finished STEP 3080/50000, loss = 2.472624 (0.032 sec/batch), lr: 0.003000
|
| 595 |
-
2026-01-16 07:01:20 INFO: Finished STEP 3100/50000, loss = 2.051110 (0.032 sec/batch), lr: 0.003000
|
| 596 |
-
2026-01-16 07:01:20 INFO: Evaluating on dev set...
|
| 597 |
-
2026-01-16 07:01:20 INFO: LAS MLAS BLEX
|
| 598 |
-
2026-01-16 07:01:20 INFO: 23.51 9.39 13.88
|
| 599 |
-
2026-01-16 07:01:20 INFO: step 3100: train_loss = 2.557801, dev_score = 0.2351
|
| 600 |
-
2026-01-16 07:01:20 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 601 |
-
2026-01-16 07:01:20 INFO: new best model saved.
|
| 602 |
-
2026-01-16 07:01:21 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 603 |
-
2026-01-16 07:01:21 INFO: new model checkpoint saved.
|
| 604 |
-
2026-01-16 07:01:21 INFO: Finished STEP 3120/50000, loss = 1.929233 (0.031 sec/batch), lr: 0.003000
|
| 605 |
-
2026-01-16 07:01:22 INFO: Finished STEP 3140/50000, loss = 1.142895 (0.032 sec/batch), lr: 0.003000
|
| 606 |
-
2026-01-16 07:01:23 INFO: Finished STEP 3160/50000, loss = 2.250463 (0.031 sec/batch), lr: 0.003000
|
| 607 |
-
2026-01-16 07:01:23 INFO: Finished STEP 3180/50000, loss = 1.848980 (0.031 sec/batch), lr: 0.003000
|
| 608 |
-
2026-01-16 07:01:24 INFO: Finished STEP 3200/50000, loss = 1.592728 (0.031 sec/batch), lr: 0.003000
|
| 609 |
-
2026-01-16 07:01:24 INFO: Evaluating on dev set...
|
| 610 |
-
2026-01-16 07:01:24 INFO: LAS MLAS BLEX
|
| 611 |
-
2026-01-16 07:01:24 INFO: 14.11 7.38 10.10
|
| 612 |
-
2026-01-16 07:01:24 INFO: step 3200: train_loss = 2.549193, dev_score = 0.1411
|
| 613 |
-
2026-01-16 07:01:25 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 614 |
-
2026-01-16 07:01:25 INFO: new model checkpoint saved.
|
| 615 |
-
2026-01-16 07:01:26 INFO: Finished STEP 3220/50000, loss = 1.161050 (0.031 sec/batch), lr: 0.003000
|
| 616 |
-
2026-01-16 07:01:26 INFO: Finished STEP 3240/50000, loss = 2.996624 (0.031 sec/batch), lr: 0.003000
|
| 617 |
-
2026-01-16 07:01:27 INFO: Finished STEP 3260/50000, loss = 1.401396 (0.031 sec/batch), lr: 0.003000
|
| 618 |
-
2026-01-16 07:01:28 INFO: Finished STEP 3280/50000, loss = 3.537535 (0.032 sec/batch), lr: 0.003000
|
| 619 |
-
2026-01-16 07:01:28 INFO: Finished STEP 3300/50000, loss = 2.966022 (0.031 sec/batch), lr: 0.003000
|
| 620 |
-
2026-01-16 07:01:28 INFO: Evaluating on dev set...
|
| 621 |
-
2026-01-16 07:01:29 INFO: LAS MLAS BLEX
|
| 622 |
-
2026-01-16 07:01:29 INFO: 21.78 6.75 7.54
|
| 623 |
-
2026-01-16 07:01:29 INFO: step 3300: train_loss = 2.125264, dev_score = 0.2178
|
| 624 |
-
2026-01-16 07:01:29 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 625 |
-
2026-01-16 07:01:29 INFO: new model checkpoint saved.
|
| 626 |
-
2026-01-16 07:01:30 INFO: Finished STEP 3320/50000, loss = 1.093041 (0.032 sec/batch), lr: 0.003000
|
| 627 |
-
2026-01-16 07:01:30 INFO: Finished STEP 3340/50000, loss = 4.219791 (0.031 sec/batch), lr: 0.003000
|
| 628 |
-
2026-01-16 07:01:31 INFO: Finished STEP 3360/50000, loss = 1.543054 (0.032 sec/batch), lr: 0.003000
|
| 629 |
-
2026-01-16 07:01:32 INFO: Finished STEP 3380/50000, loss = 1.405899 (0.031 sec/batch), lr: 0.003000
|
| 630 |
-
2026-01-16 07:01:32 INFO: Finished STEP 3400/50000, loss = 1.838370 (0.032 sec/batch), lr: 0.003000
|
| 631 |
-
2026-01-16 07:01:32 INFO: Evaluating on dev set...
|
| 632 |
-
2026-01-16 07:01:33 INFO: LAS MLAS BLEX
|
| 633 |
-
2026-01-16 07:01:33 INFO: 11.39 5.43 9.69
|
| 634 |
-
2026-01-16 07:01:33 INFO: step 3400: train_loss = 2.268760, dev_score = 0.1139
|
| 635 |
-
2026-01-16 07:01:33 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 636 |
-
2026-01-16 07:01:33 INFO: new model checkpoint saved.
|
| 637 |
-
2026-01-16 07:01:34 INFO: Finished STEP 3420/50000, loss = 2.768859 (0.032 sec/batch), lr: 0.003000
|
| 638 |
-
2026-01-16 07:01:35 INFO: Finished STEP 3440/50000, loss = 1.197934 (0.031 sec/batch), lr: 0.003000
|
| 639 |
-
2026-01-16 07:01:35 INFO: Finished STEP 3460/50000, loss = 1.991076 (0.034 sec/batch), lr: 0.003000
|
| 640 |
-
2026-01-16 07:01:36 INFO: Finished STEP 3480/50000, loss = 1.895437 (0.032 sec/batch), lr: 0.003000
|
| 641 |
-
2026-01-16 07:01:37 INFO: Finished STEP 3500/50000, loss = 1.187534 (0.030 sec/batch), lr: 0.003000
|
| 642 |
-
2026-01-16 07:01:37 INFO: Evaluating on dev set...
|
| 643 |
-
2026-01-16 07:01:37 INFO: LAS MLAS BLEX
|
| 644 |
-
2026-01-16 07:01:37 INFO: 19.06 10.51 15.76
|
| 645 |
-
2026-01-16 07:01:37 INFO: step 3500: train_loss = 2.106533, dev_score = 0.1906
|
| 646 |
-
2026-01-16 07:01:38 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 647 |
-
2026-01-16 07:01:38 INFO: new model checkpoint saved.
|
| 648 |
-
2026-01-16 07:01:38 INFO: Finished STEP 3520/50000, loss = 1.585278 (0.032 sec/batch), lr: 0.003000
|
| 649 |
-
2026-01-16 07:01:39 INFO: Finished STEP 3540/50000, loss = 1.423405 (0.031 sec/batch), lr: 0.003000
|
| 650 |
-
2026-01-16 07:01:39 INFO: Finished STEP 3560/50000, loss = 0.421177 (0.030 sec/batch), lr: 0.003000
|
| 651 |
-
2026-01-16 07:01:40 INFO: Finished STEP 3580/50000, loss = 1.731428 (0.031 sec/batch), lr: 0.003000
|
| 652 |
-
2026-01-16 07:01:41 INFO: Finished STEP 3600/50000, loss = 0.526926 (0.031 sec/batch), lr: 0.003000
|
| 653 |
-
2026-01-16 07:01:41 INFO: Evaluating on dev set...
|
| 654 |
-
2026-01-16 07:01:41 INFO: LAS MLAS BLEX
|
| 655 |
-
2026-01-16 07:01:41 INFO: 19.55 8.20 10.55
|
| 656 |
-
2026-01-16 07:01:41 INFO: step 3600: train_loss = 2.619668, dev_score = 0.1955
|
| 657 |
-
2026-01-16 07:01:42 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 658 |
-
2026-01-16 07:01:42 INFO: new model checkpoint saved.
|
| 659 |
-
2026-01-16 07:01:42 INFO: Finished STEP 3620/50000, loss = 0.616525 (0.030 sec/batch), lr: 0.003000
|
| 660 |
-
2026-01-16 07:01:43 INFO: Finished STEP 3640/50000, loss = 1.884155 (0.030 sec/batch), lr: 0.003000
|
| 661 |
-
2026-01-16 07:01:44 INFO: Finished STEP 3660/50000, loss = 2.431090 (0.030 sec/batch), lr: 0.003000
|
| 662 |
-
2026-01-16 07:01:44 INFO: Finished STEP 3680/50000, loss = 4.406909 (0.029 sec/batch), lr: 0.003000
|
| 663 |
-
2026-01-16 07:01:45 INFO: Finished STEP 3700/50000, loss = 1.068943 (0.030 sec/batch), lr: 0.003000
|
| 664 |
-
2026-01-16 07:01:45 INFO: Evaluating on dev set...
|
| 665 |
-
2026-01-16 07:01:45 INFO: LAS MLAS BLEX
|
| 666 |
-
2026-01-16 07:01:45 INFO: 10.15 6.90 9.20
|
| 667 |
-
2026-01-16 07:01:45 INFO: step 3700: train_loss = 2.132062, dev_score = 0.1015
|
| 668 |
-
2026-01-16 07:01:46 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 669 |
-
2026-01-16 07:01:46 INFO: new model checkpoint saved.
|
| 670 |
-
2026-01-16 07:01:46 INFO: Finished STEP 3720/50000, loss = 1.305085 (0.030 sec/batch), lr: 0.003000
|
| 671 |
-
2026-01-16 07:01:47 INFO: Finished STEP 3740/50000, loss = 0.134945 (0.031 sec/batch), lr: 0.003000
|
| 672 |
-
2026-01-16 07:01:48 INFO: Finished STEP 3760/50000, loss = 0.634447 (0.030 sec/batch), lr: 0.003000
|
| 673 |
-
2026-01-16 07:01:48 INFO: Finished STEP 3780/50000, loss = 1.520501 (0.030 sec/batch), lr: 0.003000
|
| 674 |
-
2026-01-16 07:01:49 INFO: Finished STEP 3800/50000, loss = 2.499804 (0.030 sec/batch), lr: 0.003000
|
| 675 |
-
2026-01-16 07:01:49 INFO: Evaluating on dev set...
|
| 676 |
-
2026-01-16 07:01:49 INFO: LAS MLAS BLEX
|
| 677 |
-
2026-01-16 07:01:49 INFO: 12.87 8.68 11.44
|
| 678 |
-
2026-01-16 07:01:49 INFO: step 3800: train_loss = 2.251492, dev_score = 0.1287
|
| 679 |
-
2026-01-16 07:01:50 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 680 |
-
2026-01-16 07:01:50 INFO: new model checkpoint saved.
|
| 681 |
-
2026-01-16 07:01:50 INFO: Finished STEP 3820/50000, loss = 4.027611 (0.030 sec/batch), lr: 0.003000
|
| 682 |
-
2026-01-16 07:01:51 INFO: Finished STEP 3840/50000, loss = 0.331204 (0.029 sec/batch), lr: 0.003000
|
| 683 |
-
2026-01-16 07:01:52 INFO: Finished STEP 3860/50000, loss = 1.255922 (0.029 sec/batch), lr: 0.003000
|
| 684 |
-
2026-01-16 07:01:52 INFO: Finished STEP 3880/50000, loss = 1.898022 (0.030 sec/batch), lr: 0.003000
|
| 685 |
-
2026-01-16 07:01:53 INFO: Finished STEP 3900/50000, loss = 1.984990 (0.029 sec/batch), lr: 0.003000
|
| 686 |
-
2026-01-16 07:01:53 INFO: Evaluating on dev set...
|
| 687 |
-
2026-01-16 07:01:53 INFO: LAS MLAS BLEX
|
| 688 |
-
2026-01-16 07:01:53 INFO: 14.36 8.71 9.90
|
| 689 |
-
2026-01-16 07:01:53 INFO: step 3900: train_loss = 2.995180, dev_score = 0.1436
|
| 690 |
-
2026-01-16 07:01:54 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 691 |
-
2026-01-16 07:01:54 INFO: new model checkpoint saved.
|
| 692 |
-
2026-01-16 07:01:54 INFO: Finished STEP 3920/50000, loss = 2.342984 (0.031 sec/batch), lr: 0.003000
|
| 693 |
-
2026-01-16 07:01:55 INFO: Finished STEP 3940/50000, loss = 0.587340 (0.030 sec/batch), lr: 0.003000
|
| 694 |
-
2026-01-16 07:01:56 INFO: Finished STEP 3960/50000, loss = 1.429672 (0.029 sec/batch), lr: 0.003000
|
| 695 |
-
2026-01-16 07:01:56 INFO: Finished STEP 3980/50000, loss = 1.932409 (0.030 sec/batch), lr: 0.003000
|
| 696 |
-
2026-01-16 07:01:57 INFO: Finished STEP 4000/50000, loss = 2.074286 (0.031 sec/batch), lr: 0.003000
|
| 697 |
-
2026-01-16 07:01:57 INFO: Evaluating on dev set...
|
| 698 |
-
2026-01-16 07:01:57 INFO: LAS MLAS BLEX
|
| 699 |
-
2026-01-16 07:01:57 INFO: 13.61 7.11 9.49
|
| 700 |
-
2026-01-16 07:01:57 INFO: step 4000: train_loss = 2.342431, dev_score = 0.1361
|
| 701 |
-
2026-01-16 07:01:58 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 702 |
-
2026-01-16 07:01:58 INFO: new model checkpoint saved.
|
| 703 |
-
2026-01-16 07:01:58 INFO: Finished STEP 4020/50000, loss = 0.798264 (0.029 sec/batch), lr: 0.003000
|
| 704 |
-
2026-01-16 07:01:59 INFO: Finished STEP 4040/50000, loss = 0.128710 (0.029 sec/batch), lr: 0.003000
|
| 705 |
-
2026-01-16 07:02:00 INFO: Finished STEP 4060/50000, loss = 0.836676 (0.029 sec/batch), lr: 0.003000
|
| 706 |
-
2026-01-16 07:02:00 INFO: Finished STEP 4080/50000, loss = 1.385599 (0.030 sec/batch), lr: 0.003000
|
| 707 |
-
2026-01-16 07:02:01 INFO: Finished STEP 4100/50000, loss = 1.616785 (0.032 sec/batch), lr: 0.003000
|
| 708 |
-
2026-01-16 07:02:01 INFO: Evaluating on dev set...
|
| 709 |
-
2026-01-16 07:02:01 INFO: LAS MLAS BLEX
|
| 710 |
-
2026-01-16 07:02:01 INFO: 17.33 6.88 8.91
|
| 711 |
-
2026-01-16 07:02:01 INFO: step 4100: train_loss = 2.344978, dev_score = 0.1733
|
| 712 |
-
2026-01-16 07:02:01 INFO: Training ended with 4100 steps.
|
| 713 |
-
2026-01-16 07:02:01 INFO: Best dev F1 = 23.51, at iteration = 3100
|
| 714 |
-
2026-01-16 07:02:02 INFO: Running dev depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '8', '--dropout', '0.33']
|
| 715 |
-
2026-01-16 07:02:02 INFO: Running parser in predict mode
|
| 716 |
-
2026-01-16 07:02:02 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 717 |
-
2026-01-16 07:02:04 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 718 |
-
2026-01-16 07:02:04 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 719 |
-
2026-01-16 07:02:04 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 720 |
-
2026-01-16 07:02:04 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 721 |
-
2026-01-16 07:02:04 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 722 |
-
2026-01-16 07:02:04 INFO: Loading data with batch size 8...
|
| 723 |
-
2026-01-16 07:02:04 DEBUG: 9 batches created.
|
| 724 |
-
2026-01-16 07:02:05 INFO: F1 scores for each dependency:
|
| 725 |
-
Note that unlabeled attachment errors hurt the labeled attachment scores
|
| 726 |
-
acl: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
|
| 727 |
-
acl:relcl: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
|
| 728 |
-
advcl: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
|
| 729 |
-
advmod: p 0.1000 r 0.1200 f1 0.1091 (25 actual)
|
| 730 |
-
amod: p 0.0000 r 0.0000 f1 0.0000 (31 actual)
|
| 731 |
-
appos: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 732 |
-
aux: p 0.0000 r 0.0000 f1 0.0000 (11 actual)
|
| 733 |
-
case: p 0.4557 r 0.6429 f1 0.5333 (56 actual)
|
| 734 |
-
cc: p 0.0000 r 0.0000 f1 0.0000 (13 actual)
|
| 735 |
-
ccomp: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 736 |
-
conj: p 0.0000 r 0.0000 f1 0.0000 (12 actual)
|
| 737 |
-
cop: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
|
| 738 |
-
csubj: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 739 |
-
det: p 0.5000 r 0.4091 f1 0.4500 (22 actual)
|
| 740 |
-
expl: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 741 |
-
flat:name: p 0.0000 r 0.0000 f1 0.0000 (0 actual)
|
| 742 |
-
iobj: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 743 |
-
mark: p 0.0000 r 0.0000 f1 0.0000 (12 actual)
|
| 744 |
-
nmod: p 0.0000 r 0.0000 f1 0.0000 (15 actual)
|
| 745 |
-
nmod:poss: p 0.0000 r 0.0000 f1 0.0000 (19 actual)
|
| 746 |
-
nsubj: p 0.0641 r 0.5882 f1 0.1156 (17 actual)
|
| 747 |
-
nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
|
| 748 |
-
obj: p 0.0000 r 0.0000 f1 0.0000 (22 actual)
|
| 749 |
-
obl: p 0.2373 r 0.3415 f1 0.2800 (41 actual)
|
| 750 |
-
obl:agent: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 751 |
-
orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 752 |
-
parataxis: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
|
| 753 |
-
punct: p 0.3077 r 0.3077 f1 0.3077 (52 actual)
|
| 754 |
-
root: p 0.5556 r 0.5556 f1 0.5556 (9 actual)
|
| 755 |
-
xcomp: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
|
| 756 |
-
2026-01-16 07:02:05 INFO: LAS MLAS BLEX
|
| 757 |
-
2026-01-16 07:02:05 INFO: 23.51 9.39 13.88
|
| 758 |
-
2026-01-16 07:02:05 INFO: Parser score:
|
| 759 |
-
2026-01-16 07:02:05 INFO: sv_diachronic 23.51
|
| 760 |
-
2026-01-16 07:02:05 INFO: Finished running dev set on
|
| 761 |
-
UD_Swedish-diachronic
|
| 762 |
-
UAS LAS CLAS MLAS BLEX
|
| 763 |
-
48.27 23.51 13.88 9.39 13.88
|
| 764 |
-
2026-01-16 07:02:05 INFO: Running test depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '8', '--dropout', '0.33']
|
| 765 |
-
2026-01-16 07:02:05 INFO: Running parser in predict mode
|
| 766 |
-
2026-01-16 07:02:05 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 767 |
-
2026-01-16 07:02:07 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 768 |
-
2026-01-16 07:02:07 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 769 |
-
2026-01-16 07:02:07 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 770 |
-
2026-01-16 07:02:07 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 771 |
-
2026-01-16 07:02:07 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 772 |
-
2026-01-16 07:02:07 INFO: Loading data with batch size 8...
|
| 773 |
-
2026-01-16 07:02:07 DEBUG: 99 batches created.
|
| 774 |
-
2026-01-16 07:02:11 INFO: F1 scores for each dependency:
|
| 775 |
-
Note that unlabeled attachment errors hurt the labeled attachment scores
|
| 776 |
-
acl: p 0.0000 r 0.0000 f1 0.0000 (32 actual)
|
| 777 |
-
acl:cleft: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 778 |
-
acl:relcl: p 0.0000 r 0.0000 f1 0.0000 (75 actual)
|
| 779 |
-
advcl: p 0.0000 r 0.0000 f1 0.0000 (60 actual)
|
| 780 |
-
advcl:relcl: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 781 |
-
advmod: p 0.0662 r 0.1007 f1 0.0799 (268 actual)
|
| 782 |
-
amod: p 0.0000 r 0.0000 f1 0.0000 (230 actual)
|
| 783 |
-
appos: p 0.0000 r 0.0000 f1 0.0000 (13 actual)
|
| 784 |
-
aux: p 0.0000 r 0.0000 f1 0.0000 (84 actual)
|
| 785 |
-
aux:pass: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 786 |
-
case: p 0.3323 r 0.5818 f1 0.4230 (373 actual)
|
| 787 |
-
cc: p 0.0000 r 0.0000 f1 0.0000 (155 actual)
|
| 788 |
-
ccomp: p 0.0000 r 0.0000 f1 0.0000 (35 actual)
|
| 789 |
-
compound:prt: p 0.0000 r 0.0000 f1 0.0000 (21 actual)
|
| 790 |
-
conj: p 0.0000 r 0.0000 f1 0.0000 (158 actual)
|
| 791 |
-
cop: p 0.0000 r 0.0000 f1 0.0000 (46 actual)
|
| 792 |
-
csubj: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 793 |
-
dep: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 794 |
-
det: p 0.2518 r 0.1683 f1 0.2017 (208 actual)
|
| 795 |
-
discourse: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
|
| 796 |
-
dislocated: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 797 |
-
expl: p 0.0000 r 0.0000 f1 0.0000 (11 actual)
|
| 798 |
-
expl:pv: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 799 |
-
fixed: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
|
| 800 |
-
flat: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 801 |
-
flat:name: p 0.4000 r 0.1667 f1 0.2353 (12 actual)
|
| 802 |
-
goeswith: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 803 |
-
iobj: p 0.0000 r 0.0000 f1 0.0000 (14 actual)
|
| 804 |
-
mark: p 0.0000 r 0.0000 f1 0.0000 (153 actual)
|
| 805 |
-
nmod: p 0.0000 r 0.0000 f1 0.0000 (102 actual)
|
| 806 |
-
nmod:poss: p 0.0000 r 0.0000 f1 0.0000 (142 actual)
|
| 807 |
-
nsubj: p 0.1027 r 0.5464 f1 0.1729 (280 actual)
|
| 808 |
-
nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (25 actual)
|
| 809 |
-
nummod: p 0.0000 r 0.0000 f1 0.0000 (10 actual)
|
| 810 |
-
obj: p 0.0000 r 0.0000 f1 0.0000 (183 actual)
|
| 811 |
-
obl: p 0.2272 r 0.3309 f1 0.2694 (278 actual)
|
| 812 |
-
obl:agent: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 813 |
-
orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 814 |
-
parataxis: p 0.0000 r 0.0000 f1 0.0000 (18 actual)
|
| 815 |
-
punct: p 0.3411 r 0.3459 f1 0.3435 (425 actual)
|
| 816 |
-
reparandum: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 817 |
-
root: p 0.4545 r 0.4545 f1 0.4545 (99 actual)
|
| 818 |
-
vocative: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
|
| 819 |
-
xcomp: p 0.0000 r 0.0000 f1 0.0000 (75 actual)
|
| 820 |
-
2026-01-16 07:02:11 INFO: LAS MLAS BLEX
|
| 821 |
-
2026-01-16 07:02:11 INFO: 20.06 11.15 14.33
|
| 822 |
-
2026-01-16 07:02:11 INFO: Parser score:
|
| 823 |
-
2026-01-16 07:02:11 INFO: sv_diachronic 20.06
|
| 824 |
-
2026-01-16 07:02:11 INFO: Finished running test set on
|
| 825 |
-
UD_Swedish-diachronic
|
| 826 |
-
UAS LAS CLAS MLAS BLEX
|
| 827 |
-
48.02 20.06 14.33 11.15 14.33
|
| 828 |
-
DONE.
|
| 829 |
-
Full log saved to: logs/log_conll17.pt_sv_diachron_20260116_065352.txt
|
| 830 |
-
Symlink updated: logs/latest.txt → log_conll17.pt_sv_diachron_20260116_065352.txt
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_conll17.pt_sv_diachron_20260116_073035.txt
DELETED
|
@@ -1,868 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_diachron_20260116_073035.txt ===
|
| 2 |
-
Language codes: sv diachron
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-test.py sv diachron
|
| 6 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec991-spf148.conllu
|
| 7 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec988-spf145.conllu
|
| 8 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 9 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 10 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 11 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 12 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 13 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 14 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 15 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 16 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 17 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1102-spf259.conllu
|
| 18 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-letter141673-Stalhammar.conllu
|
| 19 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1033-spf190.conllu
|
| 20 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 21 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 22 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 23 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 24 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 25 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 26 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 27 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 28 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 29 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 30 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 31 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec25-Runius.conllu
|
| 32 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 33 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 34 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec987-spf144.conllu
|
| 35 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec1063-spf220.conllu
|
| 36 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 37 |
-
Including DigPhil MACHINE in TRAIN (minus gold)…
|
| 38 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 39 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 40 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 41 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 42 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
|
| 43 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 44 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 45 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 46 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
|
| 47 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
|
| 48 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 49 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
|
| 50 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
|
| 51 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
|
| 52 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 53 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 54 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 55 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 56 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 57 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
|
| 58 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
|
| 59 |
-
Cleaning TRAIN...
|
| 60 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 61 |
-
[REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
|
| 62 |
-
[REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
|
| 63 |
-
[REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
|
| 64 |
-
[REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
|
| 65 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
|
| 66 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
|
| 67 |
-
[REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
|
| 68 |
-
[REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
|
| 69 |
-
[REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
|
| 70 |
-
[REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
|
| 71 |
-
[REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
|
| 72 |
-
[REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
|
| 73 |
-
[REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
|
| 74 |
-
[REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
|
| 75 |
-
[REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
|
| 76 |
-
[REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
|
| 77 |
-
[REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
|
| 78 |
-
[REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
|
| 79 |
-
[REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
|
| 80 |
-
[REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
|
| 81 |
-
[REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
|
| 82 |
-
[REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
|
| 83 |
-
[REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
|
| 84 |
-
[REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
|
| 85 |
-
[REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
|
| 86 |
-
[REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
|
| 87 |
-
[REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
|
| 88 |
-
[REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
|
| 89 |
-
[REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
|
| 90 |
-
[REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
|
| 91 |
-
[REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
|
| 92 |
-
[REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
|
| 93 |
-
[REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
|
| 94 |
-
[REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
|
| 95 |
-
[REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
|
| 96 |
-
[REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
|
| 97 |
-
[REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
|
| 98 |
-
[REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
|
| 99 |
-
[REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
|
| 100 |
-
[REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
|
| 101 |
-
[REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
|
| 102 |
-
[REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
|
| 103 |
-
[REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
|
| 104 |
-
[REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
|
| 105 |
-
[REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
|
| 106 |
-
[REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
|
| 107 |
-
[REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
|
| 108 |
-
[REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
|
| 109 |
-
[REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
|
| 110 |
-
[REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
|
| 111 |
-
[REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
|
| 112 |
-
[REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
|
| 113 |
-
[REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
|
| 114 |
-
[REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
|
| 115 |
-
[REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
|
| 116 |
-
[REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
|
| 117 |
-
[REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
|
| 118 |
-
[REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
|
| 119 |
-
[REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
|
| 120 |
-
[REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
|
| 121 |
-
[REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
|
| 122 |
-
[REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
|
| 123 |
-
[REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
|
| 124 |
-
Cleaning DEV...
|
| 125 |
-
Cleaning TEST...
|
| 126 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 127 |
-
Writing TRAIN → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (60334 valid sentences)
|
| 128 |
-
Writing DEV → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (34 valid sentences)
|
| 129 |
-
Writing TEST → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (74 valid sentences)
|
| 130 |
-
Done.
|
| 131 |
-
Sourcing scripts/config_alvis.sh
|
| 132 |
-
Running stanza dataset preparation…
|
| 133 |
-
2026-01-16 07:30:43 INFO: Datasets program called with:
|
| 134 |
-
/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 135 |
-
2026-01-16 07:30:43 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
|
| 136 |
-
|
| 137 |
-
2026-01-16 07:30:44 INFO: Downloaded file to /cephyr/users/cleland/Alvis/stanza_resources/resources.json
|
| 138 |
-
2026-01-16 07:30:44 DEBUG: Processing parameter "processors"...
|
| 139 |
-
2026-01-16 07:30:44 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
|
| 140 |
-
2026-01-16 07:30:44 INFO: Downloading these customized packages for language: sv (Swedish)...
|
| 141 |
-
=======================
|
| 142 |
-
| Processor | Package |
|
| 143 |
-
-----------------------
|
| 144 |
-
=======================
|
| 145 |
-
|
| 146 |
-
2026-01-16 07:30:44 INFO: Finished downloading models and saved to /cephyr/users/cleland/Alvis/stanza_resources
|
| 147 |
-
2026-01-16 07:30:44 INFO: Using tagger model in /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt for sv_diachronic
|
| 148 |
-
2026-01-16 07:30:44 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 149 |
-
2026-01-16 07:30:44 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 150 |
-
Augmented 189 quotes: Counter({'""': 26, '「」': 23, '„”': 23, '»«': 22, '″″': 18, '””': 18, '““': 17, '《》': 17, '„“': 13, '«»': 12})
|
| 151 |
-
2026-01-16 07:30:46 INFO: Running tagger to retag /local/tmp.5658622/tmp0y7bjycl/sv_diachronic.train.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 152 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5658622/tmp0y7bjycl/sv_diachronic.train.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu']
|
| 153 |
-
2026-01-16 07:30:47 INFO: Running tagger in predict mode
|
| 154 |
-
2026-01-16 07:30:47 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 155 |
-
2026-01-16 07:30:48 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 156 |
-
2026-01-16 07:30:48 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 157 |
-
2026-01-16 07:30:48 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 158 |
-
2026-01-16 07:30:48 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 159 |
-
2026-01-16 07:30:49 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 160 |
-
2026-01-16 07:30:51 INFO: Loading data with batch size 250...
|
| 161 |
-
2026-01-16 07:31:23 INFO: Start evaluation...
|
| 162 |
-
2026-01-16 07:34:24 INFO: UPOS XPOS UFeats AllTags
|
| 163 |
-
2026-01-16 07:34:24 INFO: 99.37 87.38 98.56 87.09
|
| 164 |
-
2026-01-16 07:34:24 INFO: POS Tagger score: sv_diachronic 87.09
|
| 165 |
-
2026-01-16 07:34:25 INFO: Running tagger to retag /local/tmp.5658622/tmp0y7bjycl/sv_diachronic.dev.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 166 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5658622/tmp0y7bjycl/sv_diachronic.dev.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu']
|
| 167 |
-
2026-01-16 07:34:25 INFO: Running tagger in predict mode
|
| 168 |
-
2026-01-16 07:34:25 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 169 |
-
2026-01-16 07:34:26 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 170 |
-
2026-01-16 07:34:26 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 171 |
-
2026-01-16 07:34:26 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 172 |
-
2026-01-16 07:34:26 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 173 |
-
2026-01-16 07:34:26 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 174 |
-
2026-01-16 07:34:26 INFO: Loading data with batch size 250...
|
| 175 |
-
2026-01-16 07:34:26 INFO: Start evaluation...
|
| 176 |
-
2026-01-16 07:34:27 INFO: UPOS XPOS UFeats AllTags
|
| 177 |
-
2026-01-16 07:34:27 INFO: 90.62 97.55 94.98 89.40
|
| 178 |
-
2026-01-16 07:34:27 INFO: POS Tagger score: sv_diachronic 89.40
|
| 179 |
-
2026-01-16 07:34:27 INFO: Running tagger to retag /local/tmp.5658622/tmp0y7bjycl/sv_diachronic.test.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
|
| 180 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5658622/tmp0y7bjycl/sv_diachronic.test.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu']
|
| 181 |
-
2026-01-16 07:34:27 INFO: Running tagger in predict mode
|
| 182 |
-
2026-01-16 07:34:27 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 183 |
-
2026-01-16 07:34:29 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 184 |
-
2026-01-16 07:34:29 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 185 |
-
2026-01-16 07:34:29 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 186 |
-
2026-01-16 07:34:29 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 187 |
-
2026-01-16 07:34:29 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 188 |
-
2026-01-16 07:34:29 INFO: Loading data with batch size 250...
|
| 189 |
-
2026-01-16 07:34:29 INFO: Start evaluation...
|
| 190 |
-
2026-01-16 07:34:29 INFO: UPOS XPOS UFeats AllTags
|
| 191 |
-
2026-01-16 07:34:29 INFO: 94.88 95.25 95.21 90.09
|
| 192 |
-
2026-01-16 07:34:29 INFO: POS Tagger score: sv_diachronic 90.09
|
| 193 |
-
Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
|
| 194 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /local/tmp.5658622/tmp0y7bjycl/sv_diachronic.train.gold.conllu
|
| 195 |
-
Swapped 'w1, w2' for 'w1 ,w2' 106 times
|
| 196 |
-
Added 520 new sentences with asdf, zzzz -> asdf,zzzz
|
| 197 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /local/tmp.5658622/tmp0y7bjycl/sv_diachronic.dev.gold.conllu
|
| 198 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /local/tmp.5658622/tmp0y7bjycl/sv_diachronic.test.gold.conllu
|
| 199 |
-
Running stanza dependency parser training…
|
| 200 |
-
2026-01-16 07:34:42 INFO: Training program called with:
|
| 201 |
-
/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt --batch_size 8 --dropout 0.15
|
| 202 |
-
2026-01-16 07:34:42 DEBUG: UD_Swedish-diachronic: sv_diachronic
|
| 203 |
-
2026-01-16 07:34:42 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 204 |
-
2026-01-16 07:34:42 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 205 |
-
2026-01-16 07:34:42 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_charlm_parser.pt does not exist, training new model
|
| 206 |
-
2026-01-16 07:34:42 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 207 |
-
2026-01-16 07:34:42 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 208 |
-
2026-01-16 07:34:42 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--train_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '8', '--dropout', '0.15']
|
| 209 |
-
2026-01-16 07:34:42 INFO: Running parser in train mode
|
| 210 |
-
2026-01-16 07:34:42 INFO: Using pretrained contextualized char embedding
|
| 211 |
-
2026-01-16 07:34:42 INFO: Loading data with batch size 8...
|
| 212 |
-
2026-01-16 07:34:48 INFO: Train File /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 60854
|
| 213 |
-
2026-01-16 07:34:48 INFO: Original data size: 60854
|
| 214 |
-
2026-01-16 07:34:48 INFO: Augmented data size: 60854
|
| 215 |
-
2026-01-16 07:35:03 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
|
| 216 |
-
2026-01-16 07:35:03 INFO: Original length = 60854
|
| 217 |
-
2026-01-16 07:35:03 INFO: Filtered length = 60854
|
| 218 |
-
2026-01-16 07:35:17 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='|') for the xpos factory for sv_diachronic
|
| 219 |
-
2026-01-16 07:35:22 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 220 |
-
2026-01-16 07:35:33 DEBUG: 58067 batches created.
|
| 221 |
-
2026-01-16 07:35:33 DEBUG: 34 batches created.
|
| 222 |
-
2026-01-16 07:35:33 INFO: Training parser...
|
| 223 |
-
2026-01-16 07:35:33 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 224 |
-
2026-01-16 07:35:33 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 225 |
-
2026-01-16 07:35:34 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 226 |
-
2026-01-16 07:35:34 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 227 |
-
2026-01-16 07:35:37 INFO: Finished STEP 20/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
|
| 228 |
-
2026-01-16 07:35:38 INFO: Finished STEP 40/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
|
| 229 |
-
2026-01-16 07:35:38 INFO: Finished STEP 60/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
|
| 230 |
-
2026-01-16 07:35:39 INFO: Finished STEP 80/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
|
| 231 |
-
2026-01-16 07:35:39 INFO: Finished STEP 100/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
|
| 232 |
-
2026-01-16 07:35:39 INFO: Evaluating on dev set...
|
| 233 |
-
2026-01-16 07:35:41 INFO: LAS MLAS BLEX
|
| 234 |
-
2026-01-16 07:35:41 INFO: 0.06 0.08 0.08
|
| 235 |
-
2026-01-16 07:35:41 INFO: step 100: train_loss = 379.523051, dev_score = 0.0006
|
| 236 |
-
2026-01-16 07:35:41 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 237 |
-
2026-01-16 07:35:41 INFO: new best model saved.
|
| 238 |
-
2026-01-16 07:35:42 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 239 |
-
2026-01-16 07:35:42 INFO: new model checkpoint saved.
|
| 240 |
-
2026-01-16 07:35:42 INFO: Finished STEP 120/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
|
| 241 |
-
2026-01-16 07:35:43 INFO: Finished STEP 140/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
|
| 242 |
-
2026-01-16 07:35:44 INFO: Finished STEP 160/50000, loss = 0.000000 (0.028 sec/batch), lr: 0.003000
|
| 243 |
-
2026-01-16 07:35:44 INFO: Finished STEP 180/50000, loss = 0.000000 (0.030 sec/batch), lr: 0.003000
|
| 244 |
-
2026-01-16 07:35:45 INFO: Finished STEP 200/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
|
| 245 |
-
2026-01-16 07:35:45 INFO: Evaluating on dev set...
|
| 246 |
-
2026-01-16 07:35:46 INFO: LAS MLAS BLEX
|
| 247 |
-
2026-01-16 07:35:46 INFO: 0.06 0.08 0.08
|
| 248 |
-
2026-01-16 07:35:46 INFO: step 200: train_loss = 0.000000, dev_score = 0.0006
|
| 249 |
-
2026-01-16 07:35:47 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 250 |
-
2026-01-16 07:35:47 INFO: new best model saved.
|
| 251 |
-
2026-01-16 07:35:47 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 252 |
-
2026-01-16 07:35:47 INFO: new model checkpoint saved.
|
| 253 |
-
2026-01-16 07:35:48 INFO: Finished STEP 220/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
|
| 254 |
-
2026-01-16 07:35:48 INFO: Finished STEP 240/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
|
| 255 |
-
2026-01-16 07:35:49 INFO: Finished STEP 260/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
|
| 256 |
-
2026-01-16 07:35:50 INFO: Finished STEP 280/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
|
| 257 |
-
2026-01-16 07:35:50 INFO: Finished STEP 300/50000, loss = 2.424209 (0.030 sec/batch), lr: 0.003000
|
| 258 |
-
2026-01-16 07:35:50 INFO: Evaluating on dev set...
|
| 259 |
-
2026-01-16 07:35:52 INFO: LAS MLAS BLEX
|
| 260 |
-
2026-01-16 07:35:52 INFO: 1.35 0.00 0.00
|
| 261 |
-
2026-01-16 07:35:52 INFO: step 300: train_loss = 158.649473, dev_score = 0.0135
|
| 262 |
-
2026-01-16 07:35:52 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 263 |
-
2026-01-16 07:35:52 INFO: new best model saved.
|
| 264 |
-
2026-01-16 07:35:53 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 265 |
-
2026-01-16 07:35:53 INFO: new model checkpoint saved.
|
| 266 |
-
2026-01-16 07:35:53 INFO: Finished STEP 320/50000, loss = 1.969488 (0.031 sec/batch), lr: 0.003000
|
| 267 |
-
2026-01-16 07:35:54 INFO: Finished STEP 340/50000, loss = 1.649248 (0.029 sec/batch), lr: 0.003000
|
| 268 |
-
2026-01-16 07:35:54 INFO: Finished STEP 360/50000, loss = 1.934138 (0.030 sec/batch), lr: 0.003000
|
| 269 |
-
2026-01-16 07:35:55 INFO: Finished STEP 380/50000, loss = 2.360754 (0.031 sec/batch), lr: 0.003000
|
| 270 |
-
2026-01-16 07:35:56 INFO: Finished STEP 400/50000, loss = 0.562345 (0.030 sec/batch), lr: 0.003000
|
| 271 |
-
2026-01-16 07:35:56 INFO: Evaluating on dev set...
|
| 272 |
-
2026-01-16 07:35:58 INFO: LAS MLAS BLEX
|
| 273 |
-
2026-01-16 07:35:58 INFO: 4.72 4.16 4.25
|
| 274 |
-
2026-01-16 07:35:58 INFO: step 400: train_loss = 2.121128, dev_score = 0.0472
|
| 275 |
-
2026-01-16 07:35:58 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 276 |
-
2026-01-16 07:35:58 INFO: new best model saved.
|
| 277 |
-
2026-01-16 07:35:59 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 278 |
-
2026-01-16 07:35:59 INFO: new model checkpoint saved.
|
| 279 |
-
2026-01-16 07:35:59 INFO: Finished STEP 420/50000, loss = 0.002566 (0.029 sec/batch), lr: 0.003000
|
| 280 |
-
2026-01-16 07:36:00 INFO: Finished STEP 440/50000, loss = 0.734853 (0.030 sec/batch), lr: 0.003000
|
| 281 |
-
2026-01-16 07:36:01 INFO: Finished STEP 460/50000, loss = 0.494820 (0.029 sec/batch), lr: 0.003000
|
| 282 |
-
2026-01-16 07:36:01 INFO: Finished STEP 480/50000, loss = 3.201444 (0.030 sec/batch), lr: 0.003000
|
| 283 |
-
2026-01-16 07:36:02 INFO: Finished STEP 500/50000, loss = 0.795342 (0.030 sec/batch), lr: 0.003000
|
| 284 |
-
2026-01-16 07:36:02 INFO: Evaluating on dev set...
|
| 285 |
-
2026-01-16 07:36:04 INFO: LAS MLAS BLEX
|
| 286 |
-
2026-01-16 07:36:04 INFO: 5.88 5.61 6.11
|
| 287 |
-
2026-01-16 07:36:04 INFO: step 500: train_loss = 1.037986, dev_score = 0.0588
|
| 288 |
-
2026-01-16 07:36:04 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 289 |
-
2026-01-16 07:36:04 INFO: new best model saved.
|
| 290 |
-
2026-01-16 07:36:05 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 291 |
-
2026-01-16 07:36:05 INFO: new model checkpoint saved.
|
| 292 |
-
2026-01-16 07:36:05 INFO: Finished STEP 520/50000, loss = 0.778292 (0.029 sec/batch), lr: 0.003000
|
| 293 |
-
2026-01-16 07:36:06 INFO: Finished STEP 540/50000, loss = 0.409450 (0.030 sec/batch), lr: 0.003000
|
| 294 |
-
2026-01-16 07:36:06 INFO: Finished STEP 560/50000, loss = 0.671611 (0.029 sec/batch), lr: 0.003000
|
| 295 |
-
2026-01-16 07:36:07 INFO: Finished STEP 580/50000, loss = 0.081582 (0.029 sec/batch), lr: 0.003000
|
| 296 |
-
2026-01-16 07:36:08 INFO: Finished STEP 600/50000, loss = 0.436447 (0.030 sec/batch), lr: 0.003000
|
| 297 |
-
2026-01-16 07:36:08 INFO: Evaluating on dev set...
|
| 298 |
-
2026-01-16 07:36:10 INFO: LAS MLAS BLEX
|
| 299 |
-
2026-01-16 07:36:10 INFO: 5.76 6.09 6.18
|
| 300 |
-
2026-01-16 07:36:10 INFO: step 600: train_loss = 1.044479, dev_score = 0.0576
|
| 301 |
-
2026-01-16 07:36:11 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 302 |
-
2026-01-16 07:36:11 INFO: new model checkpoint saved.
|
| 303 |
-
2026-01-16 07:36:11 INFO: Finished STEP 620/50000, loss = 0.887764 (0.031 sec/batch), lr: 0.003000
|
| 304 |
-
2026-01-16 07:36:12 INFO: Finished STEP 640/50000, loss = 0.527841 (0.030 sec/batch), lr: 0.003000
|
| 305 |
-
2026-01-16 07:36:12 INFO: Finished STEP 660/50000, loss = 0.269212 (0.030 sec/batch), lr: 0.003000
|
| 306 |
-
2026-01-16 07:36:13 INFO: Finished STEP 680/50000, loss = 2.426815 (0.029 sec/batch), lr: 0.003000
|
| 307 |
-
2026-01-16 07:36:14 INFO: Finished STEP 700/50000, loss = 0.000517 (0.029 sec/batch), lr: 0.003000
|
| 308 |
-
2026-01-16 07:36:14 INFO: Evaluating on dev set...
|
| 309 |
-
2026-01-16 07:36:15 INFO: LAS MLAS BLEX
|
| 310 |
-
2026-01-16 07:36:15 INFO: 5.21 4.95 5.37
|
| 311 |
-
2026-01-16 07:36:15 INFO: step 700: train_loss = 0.978077, dev_score = 0.0521
|
| 312 |
-
2026-01-16 07:36:16 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 313 |
-
2026-01-16 07:36:16 INFO: new model checkpoint saved.
|
| 314 |
-
2026-01-16 07:36:16 INFO: Finished STEP 720/50000, loss = 2.358844 (0.029 sec/batch), lr: 0.003000
|
| 315 |
-
2026-01-16 07:36:17 INFO: Finished STEP 740/50000, loss = 0.791893 (0.030 sec/batch), lr: 0.003000
|
| 316 |
-
2026-01-16 07:36:18 INFO: Finished STEP 760/50000, loss = 0.906499 (0.029 sec/batch), lr: 0.003000
|
| 317 |
-
2026-01-16 07:36:18 INFO: Finished STEP 780/50000, loss = 1.075008 (0.030 sec/batch), lr: 0.003000
|
| 318 |
-
2026-01-16 07:36:19 INFO: Finished STEP 800/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
|
| 319 |
-
2026-01-16 07:36:19 INFO: Evaluating on dev set...
|
| 320 |
-
2026-01-16 07:36:21 INFO: LAS MLAS BLEX
|
| 321 |
-
2026-01-16 07:36:21 INFO: 2.14 1.13 1.13
|
| 322 |
-
2026-01-16 07:36:21 INFO: step 800: train_loss = 0.975283, dev_score = 0.0214
|
| 323 |
-
2026-01-16 07:36:21 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 324 |
-
2026-01-16 07:36:21 INFO: new model checkpoint saved.
|
| 325 |
-
2026-01-16 07:36:22 INFO: Finished STEP 820/50000, loss = 0.678568 (0.030 sec/batch), lr: 0.003000
|
| 326 |
-
2026-01-16 07:36:22 INFO: Finished STEP 840/50000, loss = 0.000088 (0.029 sec/batch), lr: 0.003000
|
| 327 |
-
2026-01-16 07:36:23 INFO: Finished STEP 860/50000, loss = 1.472390 (0.029 sec/batch), lr: 0.003000
|
| 328 |
-
2026-01-16 07:36:24 INFO: Finished STEP 880/50000, loss = 0.818992 (0.029 sec/batch), lr: 0.003000
|
| 329 |
-
2026-01-16 07:36:24 INFO: Finished STEP 900/50000, loss = 1.486508 (0.031 sec/batch), lr: 0.003000
|
| 330 |
-
2026-01-16 07:36:24 INFO: Evaluating on dev set...
|
| 331 |
-
2026-01-16 07:36:26 INFO: LAS MLAS BLEX
|
| 332 |
-
2026-01-16 07:36:26 INFO: 3.98 1.89 3.04
|
| 333 |
-
2026-01-16 07:36:26 INFO: step 900: train_loss = 0.911072, dev_score = 0.0398
|
| 334 |
-
2026-01-16 07:36:27 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 335 |
-
2026-01-16 07:36:27 INFO: new model checkpoint saved.
|
| 336 |
-
2026-01-16 07:36:27 INFO: Finished STEP 920/50000, loss = 0.001453 (0.030 sec/batch), lr: 0.003000
|
| 337 |
-
2026-01-16 07:36:28 INFO: Finished STEP 940/50000, loss = 0.934357 (0.029 sec/batch), lr: 0.003000
|
| 338 |
-
2026-01-16 07:36:28 INFO: Finished STEP 960/50000, loss = 0.010020 (0.029 sec/batch), lr: 0.003000
|
| 339 |
-
2026-01-16 07:36:29 INFO: Finished STEP 980/50000, loss = 0.627853 (0.029 sec/batch), lr: 0.003000
|
| 340 |
-
2026-01-16 07:36:30 INFO: Finished STEP 1000/50000, loss = 0.240197 (0.029 sec/batch), lr: 0.003000
|
| 341 |
-
2026-01-16 07:36:30 INFO: Evaluating on dev set...
|
| 342 |
-
2026-01-16 07:36:31 INFO: LAS MLAS BLEX
|
| 343 |
-
2026-01-16 07:36:31 INFO: 2.76 1.78 1.78
|
| 344 |
-
2026-01-16 07:36:31 INFO: step 1000: train_loss = 0.752214, dev_score = 0.0276
|
| 345 |
-
2026-01-16 07:36:32 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 346 |
-
2026-01-16 07:36:32 INFO: new model checkpoint saved.
|
| 347 |
-
2026-01-16 07:36:32 INFO: Finished STEP 1020/50000, loss = 1.534254 (0.029 sec/batch), lr: 0.003000
|
| 348 |
-
2026-01-16 07:36:33 INFO: Finished STEP 1040/50000, loss = 0.000038 (0.029 sec/batch), lr: 0.003000
|
| 349 |
-
2026-01-16 07:36:34 INFO: Finished STEP 1060/50000, loss = 1.659585 (0.029 sec/batch), lr: 0.003000
|
| 350 |
-
2026-01-16 07:36:34 INFO: Finished STEP 1080/50000, loss = 0.403725 (0.029 sec/batch), lr: 0.003000
|
| 351 |
-
2026-01-16 07:36:35 INFO: Finished STEP 1100/50000, loss = 0.000000 (0.029 sec/batch), lr: 0.003000
|
| 352 |
-
2026-01-16 07:36:35 INFO: Evaluating on dev set...
|
| 353 |
-
2026-01-16 07:36:37 INFO: LAS MLAS BLEX
|
| 354 |
-
2026-01-16 07:36:37 INFO: 3.12 2.80 2.89
|
| 355 |
-
2026-01-16 07:36:37 INFO: step 1100: train_loss = 0.824820, dev_score = 0.0312
|
| 356 |
-
2026-01-16 07:36:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 357 |
-
2026-01-16 07:36:37 INFO: new model checkpoint saved.
|
| 358 |
-
2026-01-16 07:36:38 INFO: Finished STEP 1120/50000, loss = 6.452912 (0.031 sec/batch), lr: 0.003000
|
| 359 |
-
2026-01-16 07:36:38 INFO: Finished STEP 1140/50000, loss = 0.749853 (0.031 sec/batch), lr: 0.003000
|
| 360 |
-
2026-01-16 07:36:39 INFO: Finished STEP 1160/50000, loss = 4.511549 (0.030 sec/batch), lr: 0.003000
|
| 361 |
-
2026-01-16 07:36:40 INFO: Finished STEP 1180/50000, loss = 2.717284 (0.030 sec/batch), lr: 0.003000
|
| 362 |
-
2026-01-16 07:36:40 INFO: Finished STEP 1200/50000, loss = 0.523814 (0.031 sec/batch), lr: 0.003000
|
| 363 |
-
2026-01-16 07:36:40 INFO: Evaluating on dev set...
|
| 364 |
-
2026-01-16 07:36:42 INFO: LAS MLAS BLEX
|
| 365 |
-
2026-01-16 07:36:42 INFO: 4.53 1.41 1.50
|
| 366 |
-
2026-01-16 07:36:42 INFO: step 1200: train_loss = 2.303929, dev_score = 0.0453
|
| 367 |
-
2026-01-16 07:36:42 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 368 |
-
2026-01-16 07:36:42 INFO: new model checkpoint saved.
|
| 369 |
-
2026-01-16 07:36:43 INFO: Finished STEP 1220/50000, loss = 0.628675 (0.030 sec/batch), lr: 0.003000
|
| 370 |
-
2026-01-16 07:36:44 INFO: Finished STEP 1240/50000, loss = 1.167853 (0.031 sec/batch), lr: 0.003000
|
| 371 |
-
2026-01-16 07:36:44 INFO: Finished STEP 1260/50000, loss = 2.458635 (0.031 sec/batch), lr: 0.003000
|
| 372 |
-
2026-01-16 07:36:45 INFO: Finished STEP 1280/50000, loss = 1.727715 (0.030 sec/batch), lr: 0.003000
|
| 373 |
-
2026-01-16 07:36:46 INFO: Finished STEP 1300/50000, loss = 3.715376 (0.031 sec/batch), lr: 0.003000
|
| 374 |
-
2026-01-16 07:36:46 INFO: Evaluating on dev set...
|
| 375 |
-
2026-01-16 07:36:47 INFO: LAS MLAS BLEX
|
| 376 |
-
2026-01-16 07:36:47 INFO: 3.06 1.24 1.32
|
| 377 |
-
2026-01-16 07:36:47 INFO: step 1300: train_loss = 1.477564, dev_score = 0.0306
|
| 378 |
-
2026-01-16 07:36:48 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 379 |
-
2026-01-16 07:36:48 INFO: new model checkpoint saved.
|
| 380 |
-
2026-01-16 07:36:48 INFO: Finished STEP 1320/50000, loss = 1.768359 (0.030 sec/batch), lr: 0.003000
|
| 381 |
-
2026-01-16 07:36:49 INFO: Finished STEP 1340/50000, loss = 1.245610 (0.030 sec/batch), lr: 0.003000
|
| 382 |
-
2026-01-16 07:36:50 INFO: Finished STEP 1360/50000, loss = 0.451835 (0.031 sec/batch), lr: 0.003000
|
| 383 |
-
2026-01-16 07:36:50 INFO: Finished STEP 1380/50000, loss = 1.443755 (0.031 sec/batch), lr: 0.003000
|
| 384 |
-
2026-01-16 07:36:51 INFO: Finished STEP 1400/50000, loss = 1.238335 (0.030 sec/batch), lr: 0.003000
|
| 385 |
-
2026-01-16 07:36:51 INFO: Evaluating on dev set...
|
| 386 |
-
2026-01-16 07:36:53 INFO: LAS MLAS BLEX
|
| 387 |
-
2026-01-16 07:36:53 INFO: 7.11 6.09 6.96
|
| 388 |
-
2026-01-16 07:36:53 INFO: step 1400: train_loss = 1.361486, dev_score = 0.0711
|
| 389 |
-
2026-01-16 07:36:53 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 390 |
-
2026-01-16 07:36:53 INFO: new best model saved.
|
| 391 |
-
2026-01-16 07:36:53 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 392 |
-
2026-01-16 07:36:53 INFO: new model checkpoint saved.
|
| 393 |
-
2026-01-16 07:36:54 INFO: Finished STEP 1420/50000, loss = 1.591924 (0.031 sec/batch), lr: 0.003000
|
| 394 |
-
2026-01-16 07:36:55 INFO: Finished STEP 1440/50000, loss = 0.735520 (0.030 sec/batch), lr: 0.003000
|
| 395 |
-
2026-01-16 07:36:55 INFO: Finished STEP 1460/50000, loss = 1.212776 (0.032 sec/batch), lr: 0.003000
|
| 396 |
-
2026-01-16 07:36:56 INFO: Finished STEP 1480/50000, loss = 1.210722 (0.031 sec/batch), lr: 0.003000
|
| 397 |
-
2026-01-16 07:36:57 INFO: Finished STEP 1500/50000, loss = 3.332783 (0.031 sec/batch), lr: 0.003000
|
| 398 |
-
2026-01-16 07:36:57 INFO: Evaluating on dev set...
|
| 399 |
-
2026-01-16 07:36:58 INFO: LAS MLAS BLEX
|
| 400 |
-
2026-01-16 07:36:58 INFO: 6.74 3.03 3.88
|
| 401 |
-
2026-01-16 07:36:58 INFO: step 1500: train_loss = 1.381426, dev_score = 0.0674
|
| 402 |
-
2026-01-16 07:36:59 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 403 |
-
2026-01-16 07:36:59 INFO: new model checkpoint saved.
|
| 404 |
-
2026-01-16 07:37:00 INFO: Finished STEP 1520/50000, loss = 0.796412 (0.031 sec/batch), lr: 0.003000
|
| 405 |
-
2026-01-16 07:37:00 INFO: Finished STEP 1540/50000, loss = 0.730972 (0.031 sec/batch), lr: 0.003000
|
| 406 |
-
2026-01-16 07:37:01 INFO: Finished STEP 1560/50000, loss = 1.262556 (0.031 sec/batch), lr: 0.003000
|
| 407 |
-
2026-01-16 07:37:01 INFO: Finished STEP 1580/50000, loss = 0.734725 (0.030 sec/batch), lr: 0.003000
|
| 408 |
-
2026-01-16 07:37:02 INFO: Finished STEP 1600/50000, loss = 0.073490 (0.030 sec/batch), lr: 0.003000
|
| 409 |
-
2026-01-16 07:37:02 INFO: Evaluating on dev set...
|
| 410 |
-
2026-01-16 07:37:05 INFO: LAS MLAS BLEX
|
| 411 |
-
2026-01-16 07:37:05 INFO: 12.87 10.38 11.31
|
| 412 |
-
2026-01-16 07:37:05 INFO: step 1600: train_loss = 1.284919, dev_score = 0.1287
|
| 413 |
-
2026-01-16 07:37:06 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 414 |
-
2026-01-16 07:37:06 INFO: new best model saved.
|
| 415 |
-
2026-01-16 07:37:06 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 416 |
-
2026-01-16 07:37:06 INFO: new model checkpoint saved.
|
| 417 |
-
2026-01-16 07:37:07 INFO: Finished STEP 1620/50000, loss = 0.640151 (0.030 sec/batch), lr: 0.003000
|
| 418 |
-
2026-01-16 07:37:07 INFO: Finished STEP 1640/50000, loss = 1.584237 (0.030 sec/batch), lr: 0.003000
|
| 419 |
-
2026-01-16 07:37:08 INFO: Finished STEP 1660/50000, loss = 2.824079 (0.031 sec/batch), lr: 0.003000
|
| 420 |
-
2026-01-16 07:37:09 INFO: Finished STEP 1680/50000, loss = 1.456016 (0.030 sec/batch), lr: 0.003000
|
| 421 |
-
2026-01-16 07:37:09 INFO: Finished STEP 1700/50000, loss = 1.605577 (0.030 sec/batch), lr: 0.003000
|
| 422 |
-
2026-01-16 07:37:09 INFO: Evaluating on dev set...
|
| 423 |
-
2026-01-16 07:37:11 INFO: LAS MLAS BLEX
|
| 424 |
-
2026-01-16 07:37:11 INFO: 13.36 9.21 10.85
|
| 425 |
-
2026-01-16 07:37:11 INFO: step 1700: train_loss = 1.177195, dev_score = 0.1336
|
| 426 |
-
2026-01-16 07:37:11 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 427 |
-
2026-01-16 07:37:11 INFO: new best model saved.
|
| 428 |
-
2026-01-16 07:37:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 429 |
-
2026-01-16 07:37:12 INFO: new model checkpoint saved.
|
| 430 |
-
2026-01-16 07:37:12 INFO: Finished STEP 1720/50000, loss = 1.005753 (0.030 sec/batch), lr: 0.003000
|
| 431 |
-
2026-01-16 07:37:13 INFO: Finished STEP 1740/50000, loss = 1.587673 (0.029 sec/batch), lr: 0.003000
|
| 432 |
-
2026-01-16 07:37:14 INFO: Finished STEP 1760/50000, loss = 1.467617 (0.030 sec/batch), lr: 0.003000
|
| 433 |
-
2026-01-16 07:37:14 INFO: Finished STEP 1780/50000, loss = 2.158596 (0.034 sec/batch), lr: 0.003000
|
| 434 |
-
2026-01-16 07:37:15 INFO: Finished STEP 1800/50000, loss = 0.786507 (0.031 sec/batch), lr: 0.003000
|
| 435 |
-
2026-01-16 07:37:15 INFO: Evaluating on dev set...
|
| 436 |
-
2026-01-16 07:37:17 INFO: LAS MLAS BLEX
|
| 437 |
-
2026-01-16 07:37:17 INFO: 10.11 6.65 7.70
|
| 438 |
-
2026-01-16 07:37:17 INFO: step 1800: train_loss = 1.175222, dev_score = 0.1011
|
| 439 |
-
2026-01-16 07:37:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 440 |
-
2026-01-16 07:37:17 INFO: new model checkpoint saved.
|
| 441 |
-
2026-01-16 07:37:18 INFO: Finished STEP 1820/50000, loss = 0.861782 (0.031 sec/batch), lr: 0.003000
|
| 442 |
-
2026-01-16 07:37:19 INFO: Finished STEP 1840/50000, loss = 1.011274 (0.030 sec/batch), lr: 0.003000
|
| 443 |
-
2026-01-16 07:37:19 INFO: Finished STEP 1860/50000, loss = 0.641151 (0.030 sec/batch), lr: 0.003000
|
| 444 |
-
2026-01-16 07:37:20 INFO: Finished STEP 1880/50000, loss = 0.764767 (0.036 sec/batch), lr: 0.003000
|
| 445 |
-
2026-01-16 07:37:21 INFO: Finished STEP 1900/50000, loss = 1.876217 (0.030 sec/batch), lr: 0.003000
|
| 446 |
-
2026-01-16 07:37:21 INFO: Evaluating on dev set...
|
| 447 |
-
2026-01-16 07:37:25 INFO: LAS MLAS BLEX
|
| 448 |
-
2026-01-16 07:37:25 INFO: 7.66 4.98 5.64
|
| 449 |
-
2026-01-16 07:37:25 INFO: step 1900: train_loss = 1.227179, dev_score = 0.0766
|
| 450 |
-
2026-01-16 07:37:26 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 451 |
-
2026-01-16 07:37:26 INFO: new model checkpoint saved.
|
| 452 |
-
2026-01-16 07:37:27 INFO: Finished STEP 1920/50000, loss = 1.518084 (0.031 sec/batch), lr: 0.003000
|
| 453 |
-
2026-01-16 07:37:27 INFO: Finished STEP 1940/50000, loss = 0.689516 (0.031 sec/batch), lr: 0.003000
|
| 454 |
-
2026-01-16 07:37:28 INFO: Finished STEP 1960/50000, loss = 0.541874 (0.031 sec/batch), lr: 0.003000
|
| 455 |
-
2026-01-16 07:37:29 INFO: Finished STEP 1980/50000, loss = 1.434824 (0.030 sec/batch), lr: 0.003000
|
| 456 |
-
2026-01-16 07:37:29 INFO: Finished STEP 2000/50000, loss = 0.343702 (0.031 sec/batch), lr: 0.003000
|
| 457 |
-
2026-01-16 07:37:29 INFO: Evaluating on dev set...
|
| 458 |
-
2026-01-16 07:37:31 INFO: LAS MLAS BLEX
|
| 459 |
-
2026-01-16 07:37:31 INFO: 11.52 4.14 5.58
|
| 460 |
-
2026-01-16 07:37:31 INFO: step 2000: train_loss = 1.288903, dev_score = 0.1152
|
| 461 |
-
2026-01-16 07:37:32 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 462 |
-
2026-01-16 07:37:32 INFO: new model checkpoint saved.
|
| 463 |
-
2026-01-16 07:37:33 INFO: Finished STEP 2020/50000, loss = 1.551595 (0.030 sec/batch), lr: 0.003000
|
| 464 |
-
2026-01-16 07:37:33 INFO: Finished STEP 2040/50000, loss = 0.806195 (0.031 sec/batch), lr: 0.003000
|
| 465 |
-
2026-01-16 07:37:34 INFO: Finished STEP 2060/50000, loss = 0.586268 (0.031 sec/batch), lr: 0.003000
|
| 466 |
-
2026-01-16 07:37:34 INFO: Finished STEP 2080/50000, loss = 1.326350 (0.031 sec/batch), lr: 0.003000
|
| 467 |
-
2026-01-16 07:37:35 INFO: Finished STEP 2100/50000, loss = 1.068810 (0.031 sec/batch), lr: 0.003000
|
| 468 |
-
2026-01-16 07:37:35 INFO: Evaluating on dev set...
|
| 469 |
-
2026-01-16 07:37:37 INFO: LAS MLAS BLEX
|
| 470 |
-
2026-01-16 07:37:37 INFO: 3.49 1.97 2.55
|
| 471 |
-
2026-01-16 07:37:37 INFO: step 2100: train_loss = 1.064336, dev_score = 0.0349
|
| 472 |
-
2026-01-16 07:37:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 473 |
-
2026-01-16 07:37:37 INFO: new model checkpoint saved.
|
| 474 |
-
2026-01-16 07:37:38 INFO: Finished STEP 2120/50000, loss = 0.723737 (0.033 sec/batch), lr: 0.003000
|
| 475 |
-
2026-01-16 07:37:39 INFO: Finished STEP 2140/50000, loss = 1.762422 (0.031 sec/batch), lr: 0.003000
|
| 476 |
-
2026-01-16 07:37:39 INFO: Finished STEP 2160/50000, loss = 0.039095 (0.036 sec/batch), lr: 0.003000
|
| 477 |
-
2026-01-16 07:37:40 INFO: Finished STEP 2180/50000, loss = 0.746787 (0.030 sec/batch), lr: 0.003000
|
| 478 |
-
2026-01-16 07:37:41 INFO: Finished STEP 2200/50000, loss = 0.145349 (0.031 sec/batch), lr: 0.003000
|
| 479 |
-
2026-01-16 07:37:41 INFO: Evaluating on dev set...
|
| 480 |
-
2026-01-16 07:37:43 INFO: LAS MLAS BLEX
|
| 481 |
-
2026-01-16 07:37:43 INFO: 13.30 9.26 9.89
|
| 482 |
-
2026-01-16 07:37:43 INFO: step 2200: train_loss = 1.202942, dev_score = 0.1330
|
| 483 |
-
2026-01-16 07:37:43 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 484 |
-
2026-01-16 07:37:43 INFO: new model checkpoint saved.
|
| 485 |
-
2026-01-16 07:37:44 INFO: Finished STEP 2220/50000, loss = 2.421659 (0.030 sec/batch), lr: 0.003000
|
| 486 |
-
2026-01-16 07:37:44 INFO: Finished STEP 2240/50000, loss = 2.592595 (0.030 sec/batch), lr: 0.003000
|
| 487 |
-
2026-01-16 07:37:45 INFO: Finished STEP 2260/50000, loss = 3.920859 (0.030 sec/batch), lr: 0.003000
|
| 488 |
-
2026-01-16 07:37:46 INFO: Finished STEP 2280/50000, loss = 1.928826 (0.030 sec/batch), lr: 0.003000
|
| 489 |
-
2026-01-16 07:37:46 INFO: Finished STEP 2300/50000, loss = 5.176385 (0.030 sec/batch), lr: 0.003000
|
| 490 |
-
2026-01-16 07:37:46 INFO: Evaluating on dev set...
|
| 491 |
-
2026-01-16 07:37:48 INFO: LAS MLAS BLEX
|
| 492 |
-
2026-01-16 07:37:48 INFO: 17.83 9.07 10.24
|
| 493 |
-
2026-01-16 07:37:48 INFO: step 2300: train_loss = 2.391255, dev_score = 0.1783
|
| 494 |
-
2026-01-16 07:37:48 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 495 |
-
2026-01-16 07:37:48 INFO: new best model saved.
|
| 496 |
-
2026-01-16 07:37:49 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 497 |
-
2026-01-16 07:37:49 INFO: new model checkpoint saved.
|
| 498 |
-
2026-01-16 07:37:49 INFO: Finished STEP 2320/50000, loss = 2.267345 (0.030 sec/batch), lr: 0.003000
|
| 499 |
-
2026-01-16 07:37:50 INFO: Finished STEP 2340/50000, loss = 0.719973 (0.030 sec/batch), lr: 0.003000
|
| 500 |
-
2026-01-16 07:37:51 INFO: Finished STEP 2360/50000, loss = 1.036772 (0.031 sec/batch), lr: 0.003000
|
| 501 |
-
2026-01-16 07:37:51 INFO: Finished STEP 2380/50000, loss = 2.774978 (0.030 sec/batch), lr: 0.003000
|
| 502 |
-
2026-01-16 07:37:52 INFO: Finished STEP 2400/50000, loss = 1.541276 (0.030 sec/batch), lr: 0.003000
|
| 503 |
-
2026-01-16 07:37:52 INFO: Evaluating on dev set...
|
| 504 |
-
2026-01-16 07:37:54 INFO: LAS MLAS BLEX
|
| 505 |
-
2026-01-16 07:37:54 INFO: 5.76 3.46 4.14
|
| 506 |
-
2026-01-16 07:37:54 INFO: step 2400: train_loss = 2.107856, dev_score = 0.0576
|
| 507 |
-
2026-01-16 07:37:54 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 508 |
-
2026-01-16 07:37:54 INFO: new model checkpoint saved.
|
| 509 |
-
2026-01-16 07:37:55 INFO: Finished STEP 2420/50000, loss = 0.558537 (0.029 sec/batch), lr: 0.003000
|
| 510 |
-
2026-01-16 07:37:55 INFO: Finished STEP 2440/50000, loss = 0.417549 (0.029 sec/batch), lr: 0.003000
|
| 511 |
-
2026-01-16 07:37:56 INFO: Finished STEP 2460/50000, loss = 1.211123 (0.029 sec/batch), lr: 0.003000
|
| 512 |
-
2026-01-16 07:37:57 INFO: Finished STEP 2480/50000, loss = 1.066298 (0.030 sec/batch), lr: 0.003000
|
| 513 |
-
2026-01-16 07:37:57 INFO: Finished STEP 2500/50000, loss = 1.788547 (0.031 sec/batch), lr: 0.003000
|
| 514 |
-
2026-01-16 07:37:57 INFO: Evaluating on dev set...
|
| 515 |
-
2026-01-16 07:37:59 INFO: LAS MLAS BLEX
|
| 516 |
-
2026-01-16 07:37:59 INFO: 12.68 5.16 6.11
|
| 517 |
-
2026-01-16 07:37:59 INFO: step 2500: train_loss = 2.112291, dev_score = 0.1268
|
| 518 |
-
2026-01-16 07:38:00 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 519 |
-
2026-01-16 07:38:00 INFO: new model checkpoint saved.
|
| 520 |
-
2026-01-16 07:38:00 INFO: Finished STEP 2520/50000, loss = 1.717021 (0.029 sec/batch), lr: 0.003000
|
| 521 |
-
2026-01-16 07:38:01 INFO: Finished STEP 2540/50000, loss = 1.660125 (0.032 sec/batch), lr: 0.003000
|
| 522 |
-
2026-01-16 07:38:01 INFO: Finished STEP 2560/50000, loss = 1.768633 (0.030 sec/batch), lr: 0.003000
|
| 523 |
-
2026-01-16 07:38:02 INFO: Finished STEP 2580/50000, loss = 0.593540 (0.029 sec/batch), lr: 0.003000
|
| 524 |
-
2026-01-16 07:38:03 INFO: Finished STEP 2600/50000, loss = 0.938626 (0.030 sec/batch), lr: 0.003000
|
| 525 |
-
2026-01-16 07:38:03 INFO: Evaluating on dev set...
|
| 526 |
-
2026-01-16 07:38:05 INFO: LAS MLAS BLEX
|
| 527 |
-
2026-01-16 07:38:05 INFO: 19.12 9.02 10.39
|
| 528 |
-
2026-01-16 07:38:05 INFO: step 2600: train_loss = 1.933433, dev_score = 0.1912
|
| 529 |
-
2026-01-16 07:38:05 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 530 |
-
2026-01-16 07:38:05 INFO: new best model saved.
|
| 531 |
-
2026-01-16 07:38:06 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 532 |
-
2026-01-16 07:38:06 INFO: new model checkpoint saved.
|
| 533 |
-
2026-01-16 07:38:06 INFO: Finished STEP 2620/50000, loss = 2.193763 (0.029 sec/batch), lr: 0.003000
|
| 534 |
-
2026-01-16 07:38:07 INFO: Finished STEP 2640/50000, loss = 1.114236 (0.030 sec/batch), lr: 0.003000
|
| 535 |
-
2026-01-16 07:38:07 INFO: Finished STEP 2660/50000, loss = 1.077905 (0.030 sec/batch), lr: 0.003000
|
| 536 |
-
2026-01-16 07:38:08 INFO: Finished STEP 2680/50000, loss = 2.145391 (0.029 sec/batch), lr: 0.003000
|
| 537 |
-
2026-01-16 07:38:09 INFO: Finished STEP 2700/50000, loss = 4.426371 (0.030 sec/batch), lr: 0.003000
|
| 538 |
-
2026-01-16 07:38:09 INFO: Evaluating on dev set...
|
| 539 |
-
2026-01-16 07:38:11 INFO: LAS MLAS BLEX
|
| 540 |
-
2026-01-16 07:38:11 INFO: 11.76 7.72 9.78
|
| 541 |
-
2026-01-16 07:38:11 INFO: step 2700: train_loss = 1.947624, dev_score = 0.1176
|
| 542 |
-
2026-01-16 07:38:11 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 543 |
-
2026-01-16 07:38:11 INFO: new model checkpoint saved.
|
| 544 |
-
2026-01-16 07:38:12 INFO: Finished STEP 2720/50000, loss = 1.433867 (0.030 sec/batch), lr: 0.003000
|
| 545 |
-
2026-01-16 07:38:12 INFO: Finished STEP 2740/50000, loss = 0.081805 (0.030 sec/batch), lr: 0.003000
|
| 546 |
-
2026-01-16 07:38:13 INFO: Finished STEP 2760/50000, loss = 2.694524 (0.031 sec/batch), lr: 0.003000
|
| 547 |
-
2026-01-16 07:38:14 INFO: Finished STEP 2780/50000, loss = 2.323376 (0.030 sec/batch), lr: 0.003000
|
| 548 |
-
2026-01-16 07:38:14 INFO: Finished STEP 2800/50000, loss = 0.507773 (0.031 sec/batch), lr: 0.003000
|
| 549 |
-
2026-01-16 07:38:14 INFO: Evaluating on dev set...
|
| 550 |
-
2026-01-16 07:38:16 INFO: LAS MLAS BLEX
|
| 551 |
-
2026-01-16 07:38:16 INFO: 16.05 7.79 9.64
|
| 552 |
-
2026-01-16 07:38:16 INFO: step 2800: train_loss = 2.021911, dev_score = 0.1605
|
| 553 |
-
2026-01-16 07:38:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 554 |
-
2026-01-16 07:38:17 INFO: new model checkpoint saved.
|
| 555 |
-
2026-01-16 07:38:17 INFO: Finished STEP 2820/50000, loss = 2.562084 (0.030 sec/batch), lr: 0.003000
|
| 556 |
-
2026-01-16 07:38:18 INFO: Finished STEP 2840/50000, loss = 2.842648 (0.029 sec/batch), lr: 0.003000
|
| 557 |
-
2026-01-16 07:38:18 INFO: Finished STEP 2860/50000, loss = 0.741765 (0.031 sec/batch), lr: 0.003000
|
| 558 |
-
2026-01-16 07:38:19 INFO: Finished STEP 2880/50000, loss = 1.145487 (0.030 sec/batch), lr: 0.003000
|
| 559 |
-
2026-01-16 07:38:20 INFO: Finished STEP 2900/50000, loss = 1.986015 (0.029 sec/batch), lr: 0.003000
|
| 560 |
-
2026-01-16 07:38:20 INFO: Evaluating on dev set...
|
| 561 |
-
2026-01-16 07:38:22 INFO: LAS MLAS BLEX
|
| 562 |
-
2026-01-16 07:38:22 INFO: 19.42 11.02 14.05
|
| 563 |
-
2026-01-16 07:38:22 INFO: step 2900: train_loss = 1.980637, dev_score = 0.1942
|
| 564 |
-
2026-01-16 07:38:22 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 565 |
-
2026-01-16 07:38:22 INFO: new best model saved.
|
| 566 |
-
2026-01-16 07:38:23 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 567 |
-
2026-01-16 07:38:23 INFO: new model checkpoint saved.
|
| 568 |
-
2026-01-16 07:38:23 INFO: Finished STEP 2920/50000, loss = 2.065670 (0.029 sec/batch), lr: 0.003000
|
| 569 |
-
2026-01-16 07:38:24 INFO: Finished STEP 2940/50000, loss = 1.709635 (0.030 sec/batch), lr: 0.003000
|
| 570 |
-
2026-01-16 07:38:24 INFO: Finished STEP 2960/50000, loss = 3.801945 (0.030 sec/batch), lr: 0.003000
|
| 571 |
-
2026-01-16 07:38:25 INFO: Finished STEP 2980/50000, loss = 1.160876 (0.030 sec/batch), lr: 0.003000
|
| 572 |
-
2026-01-16 07:38:26 INFO: Finished STEP 3000/50000, loss = 3.921757 (0.030 sec/batch), lr: 0.003000
|
| 573 |
-
2026-01-16 07:38:26 INFO: Evaluating on dev set...
|
| 574 |
-
2026-01-16 07:38:28 INFO: LAS MLAS BLEX
|
| 575 |
-
2026-01-16 07:38:28 INFO: 12.19 4.47 6.57
|
| 576 |
-
2026-01-16 07:38:28 INFO: step 3000: train_loss = 1.918748, dev_score = 0.1219
|
| 577 |
-
2026-01-16 07:38:29 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 578 |
-
2026-01-16 07:38:29 INFO: new model checkpoint saved.
|
| 579 |
-
2026-01-16 07:38:29 INFO: Finished STEP 3020/50000, loss = 1.314914 (0.030 sec/batch), lr: 0.003000
|
| 580 |
-
2026-01-16 07:38:30 INFO: Finished STEP 3040/50000, loss = 3.862856 (0.031 sec/batch), lr: 0.003000
|
| 581 |
-
2026-01-16 07:38:31 INFO: Finished STEP 3060/50000, loss = 2.133690 (0.029 sec/batch), lr: 0.003000
|
| 582 |
-
2026-01-16 07:38:31 INFO: Finished STEP 3080/50000, loss = 2.644999 (0.031 sec/batch), lr: 0.003000
|
| 583 |
-
2026-01-16 07:38:32 INFO: Finished STEP 3100/50000, loss = 2.620658 (0.031 sec/batch), lr: 0.003000
|
| 584 |
-
2026-01-16 07:38:32 INFO: Evaluating on dev set...
|
| 585 |
-
2026-01-16 07:38:34 INFO: LAS MLAS BLEX
|
| 586 |
-
2026-01-16 07:38:34 INFO: 9.56 5.70 7.32
|
| 587 |
-
2026-01-16 07:38:34 INFO: step 3100: train_loss = 1.990910, dev_score = 0.0956
|
| 588 |
-
2026-01-16 07:38:34 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 589 |
-
2026-01-16 07:38:34 INFO: new model checkpoint saved.
|
| 590 |
-
2026-01-16 07:38:35 INFO: Finished STEP 3120/50000, loss = 1.526267 (0.029 sec/batch), lr: 0.003000
|
| 591 |
-
2026-01-16 07:38:35 INFO: Finished STEP 3140/50000, loss = 1.378218 (0.031 sec/batch), lr: 0.003000
|
| 592 |
-
2026-01-16 07:38:36 INFO: Finished STEP 3160/50000, loss = 1.390367 (0.030 sec/batch), lr: 0.003000
|
| 593 |
-
2026-01-16 07:38:37 INFO: Finished STEP 3180/50000, loss = 1.705234 (0.030 sec/batch), lr: 0.003000
|
| 594 |
-
2026-01-16 07:38:37 INFO: Finished STEP 3200/50000, loss = 1.402759 (0.030 sec/batch), lr: 0.003000
|
| 595 |
-
2026-01-16 07:38:37 INFO: Evaluating on dev set...
|
| 596 |
-
2026-01-16 07:38:40 INFO: LAS MLAS BLEX
|
| 597 |
-
2026-01-16 07:38:40 INFO: 17.65 10.33 12.02
|
| 598 |
-
2026-01-16 07:38:40 INFO: step 3200: train_loss = 2.082199, dev_score = 0.1765
|
| 599 |
-
2026-01-16 07:38:40 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 600 |
-
2026-01-16 07:38:40 INFO: new model checkpoint saved.
|
| 601 |
-
2026-01-16 07:38:41 INFO: Finished STEP 3220/50000, loss = 1.167270 (0.030 sec/batch), lr: 0.003000
|
| 602 |
-
2026-01-16 07:38:42 INFO: Finished STEP 3240/50000, loss = 3.461316 (0.031 sec/batch), lr: 0.003000
|
| 603 |
-
2026-01-16 07:38:42 INFO: Finished STEP 3260/50000, loss = 2.926430 (0.030 sec/batch), lr: 0.003000
|
| 604 |
-
2026-01-16 07:38:43 INFO: Finished STEP 3280/50000, loss = 3.029675 (0.030 sec/batch), lr: 0.003000
|
| 605 |
-
2026-01-16 07:38:44 INFO: Finished STEP 3300/50000, loss = 2.962304 (0.030 sec/batch), lr: 0.003000
|
| 606 |
-
2026-01-16 07:38:44 INFO: Evaluating on dev set...
|
| 607 |
-
2026-01-16 07:38:46 INFO: LAS MLAS BLEX
|
| 608 |
-
2026-01-16 07:38:46 INFO: 21.14 14.06 15.91
|
| 609 |
-
2026-01-16 07:38:46 INFO: step 3300: train_loss = 2.159422, dev_score = 0.2114
|
| 610 |
-
2026-01-16 07:38:46 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 611 |
-
2026-01-16 07:38:46 INFO: new best model saved.
|
| 612 |
-
2026-01-16 07:38:47 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 613 |
-
2026-01-16 07:38:47 INFO: new model checkpoint saved.
|
| 614 |
-
2026-01-16 07:38:47 INFO: Finished STEP 3320/50000, loss = 1.153689 (0.031 sec/batch), lr: 0.003000
|
| 615 |
-
2026-01-16 07:38:48 INFO: Finished STEP 3340/50000, loss = 6.680528 (0.030 sec/batch), lr: 0.003000
|
| 616 |
-
2026-01-16 07:38:49 INFO: Finished STEP 3360/50000, loss = 5.256881 (0.031 sec/batch), lr: 0.003000
|
| 617 |
-
2026-01-16 07:38:49 INFO: Finished STEP 3380/50000, loss = 2.249988 (0.030 sec/batch), lr: 0.003000
|
| 618 |
-
2026-01-16 07:38:50 INFO: Finished STEP 3400/50000, loss = 2.187397 (0.031 sec/batch), lr: 0.003000
|
| 619 |
-
2026-01-16 07:38:50 INFO: Evaluating on dev set...
|
| 620 |
-
2026-01-16 07:38:52 INFO: LAS MLAS BLEX
|
| 621 |
-
2026-01-16 07:38:52 INFO: 10.85 9.14 11.20
|
| 622 |
-
2026-01-16 07:38:52 INFO: step 3400: train_loss = 1.977379, dev_score = 0.1085
|
| 623 |
-
2026-01-16 07:38:52 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 624 |
-
2026-01-16 07:38:52 INFO: new model checkpoint saved.
|
| 625 |
-
2026-01-16 07:38:53 INFO: Finished STEP 3420/50000, loss = 1.144123 (0.031 sec/batch), lr: 0.003000
|
| 626 |
-
2026-01-16 07:38:53 INFO: Finished STEP 3440/50000, loss = 1.048997 (0.030 sec/batch), lr: 0.003000
|
| 627 |
-
2026-01-16 07:38:54 INFO: Finished STEP 3460/50000, loss = 2.239283 (0.033 sec/batch), lr: 0.003000
|
| 628 |
-
2026-01-16 07:38:55 INFO: Finished STEP 3480/50000, loss = 1.407322 (0.030 sec/batch), lr: 0.003000
|
| 629 |
-
2026-01-16 07:38:55 INFO: Finished STEP 3500/50000, loss = 1.793783 (0.029 sec/batch), lr: 0.003000
|
| 630 |
-
2026-01-16 07:38:55 INFO: Evaluating on dev set...
|
| 631 |
-
2026-01-16 07:38:57 INFO: LAS MLAS BLEX
|
| 632 |
-
2026-01-16 07:38:57 INFO: 25.55 14.33 15.18
|
| 633 |
-
2026-01-16 07:38:57 INFO: step 3500: train_loss = 1.776203, dev_score = 0.2555
|
| 634 |
-
2026-01-16 07:38:58 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 635 |
-
2026-01-16 07:38:58 INFO: new best model saved.
|
| 636 |
-
2026-01-16 07:38:58 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 637 |
-
2026-01-16 07:38:58 INFO: new model checkpoint saved.
|
| 638 |
-
2026-01-16 07:38:59 INFO: Finished STEP 3520/50000, loss = 1.173083 (0.031 sec/batch), lr: 0.003000
|
| 639 |
-
2026-01-16 07:38:59 INFO: Finished STEP 3540/50000, loss = 2.557373 (0.029 sec/batch), lr: 0.003000
|
| 640 |
-
2026-01-16 07:39:00 INFO: Finished STEP 3560/50000, loss = 0.362348 (0.029 sec/batch), lr: 0.003000
|
| 641 |
-
2026-01-16 07:39:01 INFO: Finished STEP 3580/50000, loss = 0.959906 (0.030 sec/batch), lr: 0.003000
|
| 642 |
-
2026-01-16 07:39:01 INFO: Finished STEP 3600/50000, loss = 0.376063 (0.029 sec/batch), lr: 0.003000
|
| 643 |
-
2026-01-16 07:39:01 INFO: Evaluating on dev set...
|
| 644 |
-
2026-01-16 07:39:03 INFO: LAS MLAS BLEX
|
| 645 |
-
2026-01-16 07:39:03 INFO: 10.60 5.75 9.91
|
| 646 |
-
2026-01-16 07:39:03 INFO: step 3600: train_loss = 1.918432, dev_score = 0.1060
|
| 647 |
-
2026-01-16 07:39:04 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 648 |
-
2026-01-16 07:39:04 INFO: new model checkpoint saved.
|
| 649 |
-
2026-01-16 07:39:04 INFO: Finished STEP 3620/50000, loss = 0.217418 (0.030 sec/batch), lr: 0.003000
|
| 650 |
-
2026-01-16 07:39:05 INFO: Finished STEP 3640/50000, loss = 1.392662 (0.030 sec/batch), lr: 0.003000
|
| 651 |
-
2026-01-16 07:39:05 INFO: Finished STEP 3660/50000, loss = 2.835163 (0.030 sec/batch), lr: 0.003000
|
| 652 |
-
2026-01-16 07:39:06 INFO: Finished STEP 3680/50000, loss = 2.799474 (0.030 sec/batch), lr: 0.003000
|
| 653 |
-
2026-01-16 07:39:07 INFO: Finished STEP 3700/50000, loss = 0.710460 (0.030 sec/batch), lr: 0.003000
|
| 654 |
-
2026-01-16 07:39:07 INFO: Evaluating on dev set...
|
| 655 |
-
2026-01-16 07:39:08 INFO: LAS MLAS BLEX
|
| 656 |
-
2026-01-16 07:39:08 INFO: 13.11 10.84 12.64
|
| 657 |
-
2026-01-16 07:39:08 INFO: step 3700: train_loss = 1.913150, dev_score = 0.1311
|
| 658 |
-
2026-01-16 07:39:09 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 659 |
-
2026-01-16 07:39:09 INFO: new model checkpoint saved.
|
| 660 |
-
2026-01-16 07:39:10 INFO: Finished STEP 3720/50000, loss = 2.065038 (0.030 sec/batch), lr: 0.003000
|
| 661 |
-
2026-01-16 07:39:10 INFO: Finished STEP 3740/50000, loss = 1.423430 (0.031 sec/batch), lr: 0.003000
|
| 662 |
-
2026-01-16 07:39:11 INFO: Finished STEP 3760/50000, loss = 0.174135 (0.031 sec/batch), lr: 0.003000
|
| 663 |
-
2026-01-16 07:39:11 INFO: Finished STEP 3780/50000, loss = 0.994005 (0.030 sec/batch), lr: 0.003000
|
| 664 |
-
2026-01-16 07:39:12 INFO: Finished STEP 3800/50000, loss = 2.113207 (0.030 sec/batch), lr: 0.003000
|
| 665 |
-
2026-01-16 07:39:12 INFO: Evaluating on dev set...
|
| 666 |
-
2026-01-16 07:39:14 INFO: LAS MLAS BLEX
|
| 667 |
-
2026-01-16 07:39:14 INFO: 11.70 9.55 12.36
|
| 668 |
-
2026-01-16 07:39:14 INFO: step 3800: train_loss = 1.940419, dev_score = 0.1170
|
| 669 |
-
2026-01-16 07:39:14 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 670 |
-
2026-01-16 07:39:14 INFO: new model checkpoint saved.
|
| 671 |
-
2026-01-16 07:39:15 INFO: Finished STEP 3820/50000, loss = 1.886823 (0.030 sec/batch), lr: 0.003000
|
| 672 |
-
2026-01-16 07:39:16 INFO: Finished STEP 3840/50000, loss = 0.537330 (0.029 sec/batch), lr: 0.003000
|
| 673 |
-
2026-01-16 07:39:16 INFO: Finished STEP 3860/50000, loss = 1.127199 (0.030 sec/batch), lr: 0.003000
|
| 674 |
-
2026-01-16 07:39:17 INFO: Finished STEP 3880/50000, loss = 1.348202 (0.030 sec/batch), lr: 0.003000
|
| 675 |
-
2026-01-16 07:39:17 INFO: Finished STEP 3900/50000, loss = 2.103697 (0.029 sec/batch), lr: 0.003000
|
| 676 |
-
2026-01-16 07:39:17 INFO: Evaluating on dev set...
|
| 677 |
-
2026-01-16 07:39:19 INFO: LAS MLAS BLEX
|
| 678 |
-
2026-01-16 07:39:19 INFO: 19.36 13.16 14.37
|
| 679 |
-
2026-01-16 07:39:19 INFO: step 3900: train_loss = 2.127213, dev_score = 0.1936
|
| 680 |
-
2026-01-16 07:39:20 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 681 |
-
2026-01-16 07:39:20 INFO: new model checkpoint saved.
|
| 682 |
-
2026-01-16 07:39:20 INFO: Finished STEP 3920/50000, loss = 2.552736 (0.031 sec/batch), lr: 0.003000
|
| 683 |
-
2026-01-16 07:39:21 INFO: Finished STEP 3940/50000, loss = 0.713954 (0.030 sec/batch), lr: 0.003000
|
| 684 |
-
2026-01-16 07:39:22 INFO: Finished STEP 3960/50000, loss = 0.829729 (0.030 sec/batch), lr: 0.003000
|
| 685 |
-
2026-01-16 07:39:22 INFO: Finished STEP 3980/50000, loss = 2.967856 (0.030 sec/batch), lr: 0.003000
|
| 686 |
-
2026-01-16 07:39:23 INFO: Finished STEP 4000/50000, loss = 2.044243 (0.031 sec/batch), lr: 0.003000
|
| 687 |
-
2026-01-16 07:39:23 INFO: Evaluating on dev set...
|
| 688 |
-
2026-01-16 07:39:25 INFO: LAS MLAS BLEX
|
| 689 |
-
2026-01-16 07:39:25 INFO: 13.91 7.27 8.25
|
| 690 |
-
2026-01-16 07:39:25 INFO: step 4000: train_loss = 1.918335, dev_score = 0.1391
|
| 691 |
-
2026-01-16 07:39:25 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 692 |
-
2026-01-16 07:39:25 INFO: new model checkpoint saved.
|
| 693 |
-
2026-01-16 07:39:26 INFO: Finished STEP 4020/50000, loss = 0.570849 (0.030 sec/batch), lr: 0.003000
|
| 694 |
-
2026-01-16 07:39:26 INFO: Finished STEP 4040/50000, loss = 0.004094 (0.030 sec/batch), lr: 0.003000
|
| 695 |
-
2026-01-16 07:39:27 INFO: Finished STEP 4060/50000, loss = 0.615223 (0.029 sec/batch), lr: 0.003000
|
| 696 |
-
2026-01-16 07:39:28 INFO: Finished STEP 4080/50000, loss = 1.598375 (0.031 sec/batch), lr: 0.003000
|
| 697 |
-
2026-01-16 07:39:28 INFO: Finished STEP 4100/50000, loss = 1.763149 (0.033 sec/batch), lr: 0.003000
|
| 698 |
-
2026-01-16 07:39:28 INFO: Evaluating on dev set...
|
| 699 |
-
2026-01-16 07:39:30 INFO: LAS MLAS BLEX
|
| 700 |
-
2026-01-16 07:39:30 INFO: 17.22 6.34 9.61
|
| 701 |
-
2026-01-16 07:39:30 INFO: step 4100: train_loss = 1.769413, dev_score = 0.1722
|
| 702 |
-
2026-01-16 07:39:31 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 703 |
-
2026-01-16 07:39:31 INFO: new model checkpoint saved.
|
| 704 |
-
2026-01-16 07:39:31 INFO: Finished STEP 4120/50000, loss = 4.137807 (0.031 sec/batch), lr: 0.003000
|
| 705 |
-
2026-01-16 07:39:32 INFO: Finished STEP 4140/50000, loss = 3.163526 (0.030 sec/batch), lr: 0.003000
|
| 706 |
-
2026-01-16 07:39:32 INFO: Finished STEP 4160/50000, loss = 7.217985 (0.030 sec/batch), lr: 0.003000
|
| 707 |
-
2026-01-16 07:39:33 INFO: Finished STEP 4180/50000, loss = 1.032177 (0.031 sec/batch), lr: 0.003000
|
| 708 |
-
2026-01-16 07:39:34 INFO: Finished STEP 4200/50000, loss = 0.475087 (0.030 sec/batch), lr: 0.003000
|
| 709 |
-
2026-01-16 07:39:34 INFO: Evaluating on dev set...
|
| 710 |
-
2026-01-16 07:39:36 INFO: LAS MLAS BLEX
|
| 711 |
-
2026-01-16 07:39:36 INFO: 10.11 6.28 11.64
|
| 712 |
-
2026-01-16 07:39:36 INFO: step 4200: train_loss = 1.876384, dev_score = 0.1011
|
| 713 |
-
2026-01-16 07:39:36 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 714 |
-
2026-01-16 07:39:36 INFO: new model checkpoint saved.
|
| 715 |
-
2026-01-16 07:39:37 INFO: Finished STEP 4220/50000, loss = 1.246820 (0.030 sec/batch), lr: 0.003000
|
| 716 |
-
2026-01-16 07:39:37 INFO: Finished STEP 4240/50000, loss = 1.793286 (0.030 sec/batch), lr: 0.003000
|
| 717 |
-
2026-01-16 07:39:38 INFO: Finished STEP 4260/50000, loss = 2.043184 (0.031 sec/batch), lr: 0.003000
|
| 718 |
-
2026-01-16 07:39:39 INFO: Finished STEP 4280/50000, loss = 2.374679 (0.030 sec/batch), lr: 0.003000
|
| 719 |
-
2026-01-16 07:39:39 INFO: Finished STEP 4300/50000, loss = 1.198715 (0.031 sec/batch), lr: 0.003000
|
| 720 |
-
2026-01-16 07:39:39 INFO: Evaluating on dev set...
|
| 721 |
-
2026-01-16 07:39:41 INFO: LAS MLAS BLEX
|
| 722 |
-
2026-01-16 07:39:41 INFO: 12.13 4.51 8.45
|
| 723 |
-
2026-01-16 07:39:41 INFO: step 4300: train_loss = 1.807878, dev_score = 0.1213
|
| 724 |
-
2026-01-16 07:39:42 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 725 |
-
2026-01-16 07:39:42 INFO: new model checkpoint saved.
|
| 726 |
-
2026-01-16 07:39:42 INFO: Finished STEP 4320/50000, loss = 1.510360 (0.030 sec/batch), lr: 0.003000
|
| 727 |
-
2026-01-16 07:39:43 INFO: Finished STEP 4340/50000, loss = 2.409623 (0.030 sec/batch), lr: 0.003000
|
| 728 |
-
2026-01-16 07:39:43 INFO: Finished STEP 4360/50000, loss = 1.254725 (0.031 sec/batch), lr: 0.003000
|
| 729 |
-
2026-01-16 07:39:44 INFO: Finished STEP 4380/50000, loss = 7.248045 (0.030 sec/batch), lr: 0.003000
|
| 730 |
-
2026-01-16 07:39:45 INFO: Finished STEP 4400/50000, loss = 2.985596 (0.030 sec/batch), lr: 0.003000
|
| 731 |
-
2026-01-16 07:39:45 INFO: Evaluating on dev set...
|
| 732 |
-
2026-01-16 07:39:46 INFO: LAS MLAS BLEX
|
| 733 |
-
2026-01-16 07:39:46 INFO: 11.95 7.40 9.70
|
| 734 |
-
2026-01-16 07:39:46 INFO: step 4400: train_loss = 1.787297, dev_score = 0.1195
|
| 735 |
-
2026-01-16 07:39:47 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 736 |
-
2026-01-16 07:39:47 INFO: new model checkpoint saved.
|
| 737 |
-
2026-01-16 07:39:47 INFO: Finished STEP 4420/50000, loss = 1.253725 (0.030 sec/batch), lr: 0.003000
|
| 738 |
-
2026-01-16 07:39:48 INFO: Finished STEP 4440/50000, loss = 2.283563 (0.029 sec/batch), lr: 0.003000
|
| 739 |
-
2026-01-16 07:39:49 INFO: Finished STEP 4460/50000, loss = 2.580783 (0.030 sec/batch), lr: 0.003000
|
| 740 |
-
2026-01-16 07:39:49 INFO: Finished STEP 4480/50000, loss = 1.721368 (0.032 sec/batch), lr: 0.003000
|
| 741 |
-
2026-01-16 07:39:50 INFO: Finished STEP 4500/50000, loss = 1.226920 (0.034 sec/batch), lr: 0.003000
|
| 742 |
-
2026-01-16 07:39:50 INFO: Evaluating on dev set...
|
| 743 |
-
2026-01-16 07:39:52 INFO: LAS MLAS BLEX
|
| 744 |
-
2026-01-16 07:39:52 INFO: 18.81 11.17 13.67
|
| 745 |
-
2026-01-16 07:39:52 INFO: step 4500: train_loss = 2.092089, dev_score = 0.1881
|
| 746 |
-
2026-01-16 07:39:52 INFO: Training ended with 4500 steps.
|
| 747 |
-
2026-01-16 07:39:52 INFO: Best dev F1 = 25.55, at iteration = 3500
|
| 748 |
-
2026-01-16 07:39:53 INFO: Running dev depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '8', '--dropout', '0.15']
|
| 749 |
-
2026-01-16 07:39:53 INFO: Running parser in predict mode
|
| 750 |
-
2026-01-16 07:39:53 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 751 |
-
2026-01-16 07:39:54 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 752 |
-
2026-01-16 07:39:54 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 753 |
-
2026-01-16 07:39:54 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 754 |
-
2026-01-16 07:39:54 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 755 |
-
2026-01-16 07:39:55 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 756 |
-
2026-01-16 07:39:55 INFO: Loading data with batch size 8...
|
| 757 |
-
2026-01-16 07:39:55 DEBUG: 34 batches created.
|
| 758 |
-
2026-01-16 07:39:57 INFO: F1 scores for each dependency:
|
| 759 |
-
Note that unlabeled attachment errors hurt the labeled attachment scores
|
| 760 |
-
acl: p 0.0000 r 0.0000 f1 0.0000 (14 actual)
|
| 761 |
-
acl:relcl: p 0.0000 r 0.0000 f1 0.0000 (32 actual)
|
| 762 |
-
advcl: p 0.0000 r 0.0000 f1 0.0000 (26 actual)
|
| 763 |
-
advcl:relcl: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 764 |
-
advmod: p 0.1391 r 0.5081 f1 0.2184 (124 actual)
|
| 765 |
-
amod: p 0.1538 r 0.0194 f1 0.0345 (103 actual)
|
| 766 |
-
appos: p 0.0000 r 0.0000 f1 0.0000 (12 actual)
|
| 767 |
-
aux: p 0.0000 r 0.0000 f1 0.0000 (42 actual)
|
| 768 |
-
aux:pass: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 769 |
-
case: p 0.6946 r 0.6784 f1 0.6864 (171 actual)
|
| 770 |
-
cc: p 0.0000 r 0.0000 f1 0.0000 (67 actual)
|
| 771 |
-
ccomp: p 0.0000 r 0.0000 f1 0.0000 (17 actual)
|
| 772 |
-
compound:prt: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
|
| 773 |
-
conj: p 0.0000 r 0.0000 f1 0.0000 (64 actual)
|
| 774 |
-
cop: p 0.0000 r 0.0000 f1 0.0000 (16 actual)
|
| 775 |
-
csubj: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 776 |
-
dep: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 777 |
-
det: p 0.7222 r 0.7800 f1 0.7500 (100 actual)
|
| 778 |
-
discourse: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
|
| 779 |
-
expl: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 780 |
-
fixed: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
|
| 781 |
-
flat: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 782 |
-
flat:name: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
|
| 783 |
-
goeswith: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 784 |
-
iobj: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
|
| 785 |
-
mark: p 0.0000 r 0.0000 f1 0.0000 (68 actual)
|
| 786 |
-
nmod: p 0.0000 r 0.0000 f1 0.0000 (55 actual)
|
| 787 |
-
nmod:poss: p 0.3750 r 0.0484 f1 0.0857 (62 actual)
|
| 788 |
-
nsubj: p 0.1684 r 0.5294 f1 0.2556 (119 actual)
|
| 789 |
-
nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (10 actual)
|
| 790 |
-
nummod: p 0.0000 r 0.0000 f1 0.0000 (10 actual)
|
| 791 |
-
obj: p 0.0517 r 0.1481 f1 0.0767 (81 actual)
|
| 792 |
-
obl: p 0.0270 r 0.0080 f1 0.0123 (125 actual)
|
| 793 |
-
obl:agent: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 794 |
-
parataxis: p 0.0000 r 0.0000 f1 0.0000 (13 actual)
|
| 795 |
-
punct: p 0.3245 r 0.3333 f1 0.3288 (183 actual)
|
| 796 |
-
reparandum: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 797 |
-
root: p 0.4412 r 0.4412 f1 0.4412 (34 actual)
|
| 798 |
-
vocative: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
|
| 799 |
-
xcomp: p 0.0000 r 0.0000 f1 0.0000 (28 actual)
|
| 800 |
-
2026-01-16 07:39:57 INFO: LAS MLAS BLEX
|
| 801 |
-
2026-01-16 07:39:57 INFO: 25.55 14.33 15.18
|
| 802 |
-
2026-01-16 07:39:57 INFO: Parser score:
|
| 803 |
-
2026-01-16 07:39:57 INFO: sv_diachronic 25.55
|
| 804 |
-
2026-01-16 07:39:57 INFO: Finished running dev set on
|
| 805 |
-
UD_Swedish-diachronic
|
| 806 |
-
UAS LAS CLAS MLAS BLEX
|
| 807 |
-
49.08 25.55 15.18 14.33 15.18
|
| 808 |
-
2026-01-16 07:39:57 INFO: Running test depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '8', '--dropout', '0.15']
|
| 809 |
-
2026-01-16 07:39:57 INFO: Running parser in predict mode
|
| 810 |
-
2026-01-16 07:39:57 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 811 |
-
2026-01-16 07:39:59 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 812 |
-
2026-01-16 07:39:59 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 813 |
-
2026-01-16 07:39:59 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 814 |
-
2026-01-16 07:39:59 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 815 |
-
2026-01-16 07:39:59 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 816 |
-
2026-01-16 07:39:59 INFO: Loading data with batch size 8...
|
| 817 |
-
2026-01-16 07:39:59 DEBUG: 74 batches created.
|
| 818 |
-
2026-01-16 07:40:02 INFO: F1 scores for each dependency:
|
| 819 |
-
Note that unlabeled attachment errors hurt the labeled attachment scores
|
| 820 |
-
acl: p 0.0000 r 0.0000 f1 0.0000 (21 actual)
|
| 821 |
-
acl:cleft: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 822 |
-
acl:relcl: p 0.0000 r 0.0000 f1 0.0000 (50 actual)
|
| 823 |
-
advcl: p 0.0000 r 0.0000 f1 0.0000 (39 actual)
|
| 824 |
-
advmod: p 0.1618 r 0.5858 f1 0.2535 (169 actual)
|
| 825 |
-
amod: p 0.1698 r 0.0570 f1 0.0853 (158 actual)
|
| 826 |
-
appos: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
|
| 827 |
-
aux: p 0.0000 r 0.0000 f1 0.0000 (53 actual)
|
| 828 |
-
aux:pass: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 829 |
-
case: p 0.7148 r 0.7868 f1 0.7491 (258 actual)
|
| 830 |
-
cc: p 0.0000 r 0.0000 f1 0.0000 (101 actual)
|
| 831 |
-
ccomp: p 0.0000 r 0.0000 f1 0.0000 (20 actual)
|
| 832 |
-
compound:prt: p 0.0000 r 0.0000 f1 0.0000 (14 actual)
|
| 833 |
-
conj: p 0.0000 r 0.0000 f1 0.0000 (106 actual)
|
| 834 |
-
cop: p 0.0000 r 0.0000 f1 0.0000 (33 actual)
|
| 835 |
-
csubj: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 836 |
-
det: p 0.6382 r 0.7462 f1 0.6879 (130 actual)
|
| 837 |
-
discourse: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 838 |
-
dislocated: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 839 |
-
expl: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
|
| 840 |
-
expl:pv: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 841 |
-
fixed: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 842 |
-
flat:name: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
|
| 843 |
-
iobj: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
|
| 844 |
-
mark: p 0.0000 r 0.0000 f1 0.0000 (97 actual)
|
| 845 |
-
nmod: p 0.0000 r 0.0000 f1 0.0000 (62 actual)
|
| 846 |
-
nmod:poss: p 0.0588 r 0.0101 f1 0.0172 (99 actual)
|
| 847 |
-
nsubj: p 0.1690 r 0.4719 f1 0.2489 (178 actual)
|
| 848 |
-
nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (20 actual)
|
| 849 |
-
obj: p 0.0508 r 0.1532 f1 0.0763 (124 actual)
|
| 850 |
-
obl: p 0.0000 r 0.0000 f1 0.0000 (194 actual)
|
| 851 |
-
obl:agent: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
|
| 852 |
-
orphan: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 853 |
-
parataxis: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
|
| 854 |
-
punct: p 0.3435 r 0.3435 f1 0.3435 (294 actual)
|
| 855 |
-
root: p 0.4189 r 0.4189 f1 0.4189 (74 actual)
|
| 856 |
-
vocative: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 857 |
-
xcomp: p 0.0000 r 0.0000 f1 0.0000 (55 actual)
|
| 858 |
-
2026-01-16 07:40:02 INFO: LAS MLAS BLEX
|
| 859 |
-
2026-01-16 07:40:02 INFO: 27.02 15.21 16.05
|
| 860 |
-
2026-01-16 07:40:02 INFO: Parser score:
|
| 861 |
-
2026-01-16 07:40:02 INFO: sv_diachronic 27.02
|
| 862 |
-
2026-01-16 07:40:03 INFO: Finished running test set on
|
| 863 |
-
UD_Swedish-diachronic
|
| 864 |
-
UAS LAS CLAS MLAS BLEX
|
| 865 |
-
49.63 27.02 16.05 15.21 16.05
|
| 866 |
-
DONE.
|
| 867 |
-
Full log saved to: logs/log_conll17.pt_sv_diachron_20260116_073035.txt
|
| 868 |
-
Symlink updated: logs/latest.txt → log_conll17.pt_sv_diachron_20260116_073035.txt
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
logs/log_conll17.pt_sv_diachron_de_lit_20251203_154509.txt
DELETED
|
@@ -1,740 +0,0 @@
|
|
| 1 |
-
=== LOGFILE: logs/log_conll17.pt_sv_diachron_de_lit_20251203_154509.txt ===
|
| 2 |
-
Language codes: sv diachron de_lit
|
| 3 |
-
Using pretrained model: conll17.pt
|
| 4 |
-
|
| 5 |
-
Running: python prepare-train-val-test.py sv diachron de_lit
|
| 6 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-dev.conllu
|
| 7 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test.conllu
|
| 8 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_pud-ud-test.conllu
|
| 9 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-test.conllu
|
| 10 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_swell-ud-test-trg.conllu
|
| 11 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-dev.conllu
|
| 12 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/ucxn_ud_swedish-talbanken.conllu
|
| 13 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_talbanken-ud-train.conllu
|
| 14 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_old-ud-test.conllu
|
| 15 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-train.conllu
|
| 16 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-sv/sv_lines-ud-test.conllu
|
| 17 |
-
Reading: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud-treebanks-de_lit/de_lit-ud-test.conllu
|
| 18 |
-
Including DigPhil MACHINE in TRAIN (minus gold)…
|
| 19 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec330-GyllenborgC_SwenskaSpratthoken.conllu
|
| 20 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec254-CederborghF_BerattelseOmJohnHall.conllu
|
| 21 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec277-EnbomPU_MedborgeligtSkalde.conllu
|
| 22 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec268-DulciU_VitterhetsNojen3.conllu
|
| 23 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1063-spf220.conllu
|
| 24 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec397-AngeredStrandbergH_UnderSodernsSol.conllu
|
| 25 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec324-GranbergPA_Enslighetsalskaren.conllu
|
| 26 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec252-BremerF_Teckningar1.conllu
|
| 27 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec988-spf145.conllu
|
| 28 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec987-spf144.conllu
|
| 29 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec631-HasselskogN_HallaHallaGronkoping.conllu
|
| 30 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-letter141673-Stalhammar.conllu
|
| 31 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1033-spf190.conllu
|
| 32 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec25-Runius.conllu
|
| 33 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec486-SchwartzMS_BellmansSkor.conllu
|
| 34 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec452-NyblomH_FantasierFyra.conllu
|
| 35 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec613-EngstromA_StrindbergOchJag.conllu
|
| 36 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec208-Anonym_DetGrasligaMordet.conllu
|
| 37 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec639-HeidenstamV_Proletarfilosofiens.conllu
|
| 38 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec1102-spf259.conllu
|
| 39 |
-
Reading GOLD: /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/alanev_raw_files/diachron-validated/svediakorp-sec991-spf148.conllu
|
| 40 |
-
Cleaning TRAIN...
|
| 41 |
-
[REMOVED] sent_id=6 ERRORS=['Line 24: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 29: Invalid token ID or head', 'Token 30 has invalid head 24']
|
| 42 |
-
[REMOVED] sent_id=7_8 ERRORS=['Multiple roots found: [5, 10]']
|
| 43 |
-
[REMOVED] sent_id=30_31 ERRORS=['Multiple roots found: [3, 18]']
|
| 44 |
-
[REMOVED] sent_id=35 ERRORS=['Line 36: Invalid token ID or head']
|
| 45 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [1, 5]']
|
| 46 |
-
[REMOVED] sent_id=2_3 ERRORS=['Multiple roots found: [7, 20]']
|
| 47 |
-
[REMOVED] sent_id=8_9 ERRORS=['Multiple roots found: [24, 57]']
|
| 48 |
-
[REMOVED] sent_id=12_13 ERRORS=['Multiple roots found: [11, 16]']
|
| 49 |
-
[REMOVED] sent_id=124_split2 ERRORS=['Line 4: Invalid token ID or head', 'No root found', 'Token 1 has invalid head 4', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 6 has invalid head 4', 'Token 11 has invalid head 4', 'Token 15 has invalid head 4']
|
| 50 |
-
[REMOVED] sent_id=396 ERRORS=['Token 2: Missing form']
|
| 51 |
-
[REMOVED] sent_id=416 ERRORS=['Token 2: Missing form']
|
| 52 |
-
[REMOVED] sent_id=589 ERRORS=['Token 2: Missing form']
|
| 53 |
-
[REMOVED] sent_id=909 ERRORS=['Token 2: Missing form']
|
| 54 |
-
[REMOVED] sent_id=912 ERRORS=['Token 2: Missing form']
|
| 55 |
-
[REMOVED] sent_id=3_split1 ERRORS=['Multiple roots found: [4, 15, 17]']
|
| 56 |
-
[REMOVED] sent_id=3_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 8: Invalid token ID or head', 'Line 15: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1', 'Token 3 has invalid head 8', 'Token 4 has invalid head 8', 'Token 5 has invalid head 8', 'Token 7 has invalid head 8', 'Token 10 has invalid head 8', 'Token 13 has invalid head 8', 'Token 14 has invalid head 8']
|
| 57 |
-
[REMOVED] sent_id=3_4 ERRORS=['Multiple roots found: [1, 5]']
|
| 58 |
-
[REMOVED] sent_id=5_6 ERRORS=['Multiple roots found: [3, 24]']
|
| 59 |
-
[REMOVED] sent_id=11_12_13 ERRORS=['Multiple roots found: [5, 17, 25]']
|
| 60 |
-
[REMOVED] sent_id=119 ERRORS=['Token 2: Missing form']
|
| 61 |
-
[REMOVED] sent_id=179 ERRORS=['Token 2: Missing form']
|
| 62 |
-
[REMOVED] sent_id=188 ERRORS=['Token 2: Missing form']
|
| 63 |
-
[REMOVED] sent_id=223 ERRORS=['Token 2: Missing form']
|
| 64 |
-
[REMOVED] sent_id=268 ERRORS=['Token 2: Missing form']
|
| 65 |
-
[REMOVED] sent_id=325 ERRORS=['Token 2: Missing form']
|
| 66 |
-
[REMOVED] sent_id=388 ERRORS=['Token 2: Missing form']
|
| 67 |
-
[REMOVED] sent_id=399 ERRORS=['Token 2: Missing form']
|
| 68 |
-
[REMOVED] sent_id=475 ERRORS=['Token 2: Missing form']
|
| 69 |
-
[REMOVED] sent_id=505 ERRORS=['Token 2: Missing form']
|
| 70 |
-
[REMOVED] sent_id=520 ERRORS=['Token 2: Missing form']
|
| 71 |
-
[REMOVED] sent_id=562 ERRORS=['Token 2: Missing form']
|
| 72 |
-
[REMOVED] sent_id=669 ERRORS=['Token 2: Missing form']
|
| 73 |
-
[REMOVED] sent_id=711 ERRORS=['Token 2: Missing form']
|
| 74 |
-
[REMOVED] sent_id=731 ERRORS=['Token 2: Missing form']
|
| 75 |
-
[REMOVED] sent_id=867 ERRORS=['Token 2: Missing form']
|
| 76 |
-
[REMOVED] sent_id=884 ERRORS=['Token 2: Missing form']
|
| 77 |
-
[REMOVED] sent_id=923 ERRORS=['Token 2: Missing form']
|
| 78 |
-
[REMOVED] sent_id=939 ERRORS=['Token 2: Missing form']
|
| 79 |
-
[REMOVED] sent_id=1086 ERRORS=['Token 2: Missing form']
|
| 80 |
-
[REMOVED] sent_id=1179 ERRORS=['Token 2: Missing form']
|
| 81 |
-
[REMOVED] sent_id=1251 ERRORS=['Token 2: Missing form']
|
| 82 |
-
[REMOVED] sent_id=1345 ERRORS=['Token 2: Missing form']
|
| 83 |
-
[REMOVED] sent_id=1459 ERRORS=['Token 2: Missing form']
|
| 84 |
-
[REMOVED] sent_id=1656 ERRORS=['Token 2: Missing form']
|
| 85 |
-
[REMOVED] sent_id=1669 ERRORS=['Token 2: Missing form']
|
| 86 |
-
[REMOVED] sent_id=87_88 ERRORS=['Multiple roots found: [3, 6]']
|
| 87 |
-
[REMOVED] sent_id=65_split2_66_split2 ERRORS=['Line 4: Invalid token ID or head', 'Token 2 has invalid head 4', 'Token 3 has invalid head 4', 'Token 5 has invalid head 4']
|
| 88 |
-
[REMOVED] sent_id=25 ERRORS=['Token 2: Missing form']
|
| 89 |
-
[REMOVED] sent_id=136 ERRORS=['Token 2: Missing form']
|
| 90 |
-
[REMOVED] sent_id=208 ERRORS=['Token 2: Missing form']
|
| 91 |
-
[REMOVED] sent_id=230 ERRORS=['Token 2: Missing form']
|
| 92 |
-
[REMOVED] sent_id=245 ERRORS=['Token 2: Missing form']
|
| 93 |
-
[REMOVED] sent_id=276 ERRORS=['Token 2: Missing form']
|
| 94 |
-
[REMOVED] sent_id=320 ERRORS=['Token 2: Missing form']
|
| 95 |
-
[REMOVED] sent_id=366 ERRORS=['Token 2: Missing form']
|
| 96 |
-
[REMOVED] sent_id=519 ERRORS=['Token 2: Missing form']
|
| 97 |
-
[REMOVED] sent_id=569 ERRORS=['Token 2: Missing form']
|
| 98 |
-
[REMOVED] sent_id=50_split2 ERRORS=['Line 1: Invalid token ID or head', 'Line 6: Invalid token ID or head', 'No root found', 'Token 2 has invalid head 1']
|
| 99 |
-
[REMOVED] sent_id=53_54 ERRORS=['Multiple roots found: [27, 91]']
|
| 100 |
-
[REMOVED] sent_id=55_56_57 ERRORS=['Multiple roots found: [2, 4, 13]']
|
| 101 |
-
[REMOVED] sent_id=17_split1 ERRORS=['Multiple roots found: [2, 14, 17]']
|
| 102 |
-
[REMOVED] sent_id=17_split2 ERRORS=['Line 8: Invalid token ID or head', 'Line 25: Invalid token ID or head', 'Line 38: Invalid token ID or head', 'No root found', 'Token 3 has invalid head 8', 'Token 7 has invalid head 8', 'Token 9 has invalid head 8', 'Token 10 has invalid head 8', 'Token 17 has invalid head 8', 'Token 22 has invalid head 25', 'Token 23 has invalid head 25', 'Token 24 has invalid head 25', 'Token 26 has invalid head 25', 'Token 27 has invalid head 25', 'Token 28 has invalid head 25']
|
| 103 |
-
[REMOVED] sent_id=19_split1 ERRORS=['Multiple roots found: [3, 31]']
|
| 104 |
-
Cleaning DEV...
|
| 105 |
-
[REMOVED] sent_id=33 ERRORS=['Token 15: Missing deprel']
|
| 106 |
-
Cleaning TEST...
|
| 107 |
-
Writing TRAIN → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu (68172 valid sentences)
|
| 108 |
-
Writing DEV → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu (9 valid sentences)
|
| 109 |
-
Writing TEST → /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu (99 valid sentences)
|
| 110 |
-
Done.
|
| 111 |
-
Sourcing scripts/config_alvis.sh
|
| 112 |
-
Running stanza dataset preparation…
|
| 113 |
-
2025-12-03 15:45:18 INFO: Datasets program called with:
|
| 114 |
-
/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/datasets/prepare_depparse_treebank.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 115 |
-
2025-12-03 15:45:18 DEBUG: Downloading resource file from https://raw.githubusercontent.com/stanfordnlp/stanza-resources/main/resources_1.11.0.json
|
| 116 |
-
|
| 117 |
-
2025-12-03 15:45:18 INFO: Downloaded file to /cephyr/users/cleland/Alvis/stanza_resources/resources.json
|
| 118 |
-
2025-12-03 15:45:18 DEBUG: Processing parameter "processors"...
|
| 119 |
-
2025-12-03 15:45:18 WARNING: Can not find pos: diachronic from official model list. Ignoring it.
|
| 120 |
-
2025-12-03 15:45:18 INFO: Downloading these customized packages for language: sv (Swedish)...
|
| 121 |
-
=======================
|
| 122 |
-
| Processor | Package |
|
| 123 |
-
-----------------------
|
| 124 |
-
=======================
|
| 125 |
-
|
| 126 |
-
2025-12-03 15:45:18 INFO: Finished downloading models and saved to /cephyr/users/cleland/Alvis/stanza_resources
|
| 127 |
-
2025-12-03 15:45:18 INFO: Using tagger model in /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt for sv_diachronic
|
| 128 |
-
2025-12-03 15:45:18 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 129 |
-
2025-12-03 15:45:18 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 130 |
-
Augmented 188 quotes: Counter({'«»': 23, '»«': 22, '″″': 22, '""': 20, '““': 20, '„”': 19, '「」': 18, '《》': 16, '„“': 15, '””': 13})
|
| 131 |
-
2025-12-03 15:45:21 INFO: Running tagger to retag /local/tmp.5440223/tmplttwhhzq/sv_diachronic.train.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu
|
| 132 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5440223/tmplttwhhzq/sv_diachronic.train.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu']
|
| 133 |
-
2025-12-03 15:45:22 INFO: Running tagger in predict mode
|
| 134 |
-
2025-12-03 15:45:22 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 135 |
-
2025-12-03 15:45:23 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 136 |
-
2025-12-03 15:45:23 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 137 |
-
2025-12-03 15:45:23 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 138 |
-
2025-12-03 15:45:23 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 139 |
-
2025-12-03 15:45:24 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 140 |
-
2025-12-03 15:45:26 INFO: Loading data with batch size 250...
|
| 141 |
-
2025-12-03 15:46:06 INFO: Start evaluation...
|
| 142 |
-
2025-12-03 15:49:44 INFO: UPOS XPOS UFeats AllTags
|
| 143 |
-
2025-12-03 15:49:44 INFO: 96.83 85.29 95.46 84.47
|
| 144 |
-
2025-12-03 15:49:44 INFO: POS Tagger score: sv_diachronic 84.47
|
| 145 |
-
2025-12-03 15:49:45 INFO: Running tagger to retag /local/tmp.5440223/tmplttwhhzq/sv_diachronic.dev.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu
|
| 146 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5440223/tmplttwhhzq/sv_diachronic.dev.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu']
|
| 147 |
-
2025-12-03 15:49:45 INFO: Running tagger in predict mode
|
| 148 |
-
2025-12-03 15:49:45 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 149 |
-
2025-12-03 15:49:47 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 150 |
-
2025-12-03 15:49:47 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 151 |
-
2025-12-03 15:49:47 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 152 |
-
2025-12-03 15:49:47 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 153 |
-
2025-12-03 15:49:47 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 154 |
-
2025-12-03 15:49:47 INFO: Loading data with batch size 250...
|
| 155 |
-
2025-12-03 15:49:47 INFO: Start evaluation...
|
| 156 |
-
2025-12-03 15:49:47 INFO: UPOS XPOS UFeats AllTags
|
| 157 |
-
2025-12-03 15:49:47 INFO: 93.32 90.84 93.32 85.64
|
| 158 |
-
2025-12-03 15:49:47 INFO: POS Tagger score: sv_diachronic 85.64
|
| 159 |
-
2025-12-03 15:49:47 INFO: Running tagger to retag /local/tmp.5440223/tmplttwhhzq/sv_diachronic.test.gold.conllu to /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu
|
| 160 |
-
Args: ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--save_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pos', '--save_name', 'diachronic.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--eval_file', '/local/tmp.5440223/tmplttwhhzq/sv_diachronic.test.gold.conllu', '--output_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu']
|
| 161 |
-
2025-12-03 15:49:47 INFO: Running tagger in predict mode
|
| 162 |
-
2025-12-03 15:49:47 INFO: Loading model from: /cephyr/users/cleland/Alvis/stanza_resources/sv/pos/diachronic.pt
|
| 163 |
-
2025-12-03 15:49:49 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 164 |
-
2025-12-03 15:49:49 DEBUG: POS model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 165 |
-
2025-12-03 15:49:49 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 166 |
-
2025-12-03 15:49:49 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 167 |
-
2025-12-03 15:49:49 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 168 |
-
2025-12-03 15:49:49 INFO: Loading data with batch size 250...
|
| 169 |
-
2025-12-03 15:49:49 INFO: Start evaluation...
|
| 170 |
-
2025-12-03 15:49:50 INFO: UPOS XPOS UFeats AllTags
|
| 171 |
-
2025-12-03 15:49:50 INFO: 93.14 96.78 95.32 90.28
|
| 172 |
-
2025-12-03 15:49:50 INFO: POS Tagger score: sv_diachronic 90.28
|
| 173 |
-
Preparing data for UD_Swedish-diachronic: sv_diachronic, sv
|
| 174 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-train.conllu and writing to /local/tmp.5440223/tmplttwhhzq/sv_diachronic.train.gold.conllu
|
| 175 |
-
Swapped 'w1, w2' for 'w1 ,w2' 141 times
|
| 176 |
-
Added 573 new sentences with asdf, zzzz -> asdf,zzzz
|
| 177 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-dev.conllu and writing to /local/tmp.5440223/tmplttwhhzq/sv_diachronic.dev.gold.conllu
|
| 178 |
-
Reading from /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/ud/UD_Swedish-diachronic/sv_diachronic-ud-test.conllu and writing to /local/tmp.5440223/tmplttwhhzq/sv_diachronic.test.gold.conllu
|
| 179 |
-
Running stanza dependency parser training…
|
| 180 |
-
2025-12-03 15:50:04 INFO: Training program called with:
|
| 181 |
-
/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/stanza/utils/training/run_depparse.py UD_Swedish-diachronic --wordvec_pretrain_file /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt --batch_size 32 --dropout 0.33
|
| 182 |
-
2025-12-03 15:50:04 DEBUG: UD_Swedish-diachronic: sv_diachronic
|
| 183 |
-
2025-12-03 15:50:04 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 184 |
-
2025-12-03 15:50:04 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 185 |
-
2025-12-03 15:50:04 INFO: UD_Swedish-diachronic: saved_models/depparse/sv_diachronic_charlm_parser.pt does not exist, training new model
|
| 186 |
-
2025-12-03 15:50:04 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt for forward charlm
|
| 187 |
-
2025-12-03 15:50:04 INFO: Using model /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt for backward charlm
|
| 188 |
-
2025-12-03 15:50:04 INFO: Running train depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--train_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--batch_size', '5000', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'train', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
|
| 189 |
-
2025-12-03 15:50:04 INFO: Running parser in train mode
|
| 190 |
-
2025-12-03 15:50:04 INFO: Using pretrained contextualized char embedding
|
| 191 |
-
2025-12-03 15:50:04 INFO: Loading data with batch size 32...
|
| 192 |
-
2025-12-03 15:50:11 INFO: Train File /mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.train.in.conllu, Data Size: 68745
|
| 193 |
-
2025-12-03 15:50:11 INFO: Original data size: 68745
|
| 194 |
-
2025-12-03 15:50:12 INFO: Augmented data size: 69046
|
| 195 |
-
2025-12-03 15:50:31 WARNING: sv_diachronic is not a known dataset. Examining the data to choose which xpos vocab to use
|
| 196 |
-
2025-12-03 15:50:31 INFO: Original length = 69046
|
| 197 |
-
2025-12-03 15:50:31 INFO: Filtered length = 69046
|
| 198 |
-
2025-12-03 15:50:47 WARNING: Chose XPOSDescription(xpos_type=<XPOSType.XPOS: 1>, sep='|') for the xpos factory for sv_diachronic
|
| 199 |
-
2025-12-03 15:50:55 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 200 |
-
2025-12-03 15:51:09 DEBUG: 40457 batches created.
|
| 201 |
-
2025-12-03 15:51:09 DEBUG: 9 batches created.
|
| 202 |
-
2025-12-03 15:51:09 INFO: Training parser...
|
| 203 |
-
2025-12-03 15:51:09 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 204 |
-
2025-12-03 15:51:09 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 205 |
-
2025-12-03 15:51:09 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 206 |
-
2025-12-03 15:51:10 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 207 |
-
2025-12-03 15:51:17 INFO: Finished STEP 20/50000, loss = 6.791992 (0.200 sec/batch), lr: 0.003000
|
| 208 |
-
2025-12-03 15:51:21 INFO: Finished STEP 40/50000, loss = 5.211706 (0.167 sec/batch), lr: 0.003000
|
| 209 |
-
2025-12-03 15:51:24 INFO: Finished STEP 60/50000, loss = 5.668643 (0.147 sec/batch), lr: 0.003000
|
| 210 |
-
2025-12-03 15:51:27 INFO: Finished STEP 80/50000, loss = 4.197357 (0.151 sec/batch), lr: 0.003000
|
| 211 |
-
2025-12-03 15:51:30 INFO: Finished STEP 100/50000, loss = 5.356755 (0.129 sec/batch), lr: 0.003000
|
| 212 |
-
2025-12-03 15:51:30 INFO: Evaluating on dev set...
|
| 213 |
-
2025-12-03 15:51:31 INFO: LAS MLAS BLEX
|
| 214 |
-
2025-12-03 15:51:31 INFO: 39.85 26.44 33.26
|
| 215 |
-
2025-12-03 15:51:31 INFO: step 100: train_loss = 8.587716, dev_score = 0.3985
|
| 216 |
-
2025-12-03 15:51:31 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 217 |
-
2025-12-03 15:51:31 INFO: new best model saved.
|
| 218 |
-
2025-12-03 15:51:32 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 219 |
-
2025-12-03 15:51:32 INFO: new model checkpoint saved.
|
| 220 |
-
2025-12-03 15:51:34 INFO: Finished STEP 120/50000, loss = 4.996027 (0.120 sec/batch), lr: 0.003000
|
| 221 |
-
2025-12-03 15:51:37 INFO: Finished STEP 140/50000, loss = 5.125143 (0.131 sec/batch), lr: 0.003000
|
| 222 |
-
2025-12-03 15:51:40 INFO: Finished STEP 160/50000, loss = 4.139769 (0.138 sec/batch), lr: 0.003000
|
| 223 |
-
2025-12-03 15:51:42 INFO: Finished STEP 180/50000, loss = 4.071211 (0.131 sec/batch), lr: 0.003000
|
| 224 |
-
2025-12-03 15:51:45 INFO: Finished STEP 200/50000, loss = 4.302421 (0.121 sec/batch), lr: 0.003000
|
| 225 |
-
2025-12-03 15:51:45 INFO: Evaluating on dev set...
|
| 226 |
-
2025-12-03 15:51:45 INFO: LAS MLAS BLEX
|
| 227 |
-
2025-12-03 15:51:45 INFO: 50.00 35.44 40.93
|
| 228 |
-
2025-12-03 15:51:45 INFO: step 200: train_loss = 4.776229, dev_score = 0.5000
|
| 229 |
-
2025-12-03 15:51:45 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 230 |
-
2025-12-03 15:51:45 INFO: new best model saved.
|
| 231 |
-
2025-12-03 15:51:46 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 232 |
-
2025-12-03 15:51:46 INFO: new model checkpoint saved.
|
| 233 |
-
2025-12-03 15:51:48 INFO: Finished STEP 220/50000, loss = 4.349181 (0.124 sec/batch), lr: 0.003000
|
| 234 |
-
2025-12-03 15:51:51 INFO: Finished STEP 240/50000, loss = 5.862440 (0.125 sec/batch), lr: 0.003000
|
| 235 |
-
2025-12-03 15:51:53 INFO: Finished STEP 260/50000, loss = 4.765276 (0.118 sec/batch), lr: 0.003000
|
| 236 |
-
2025-12-03 15:51:56 INFO: Finished STEP 280/50000, loss = 4.617913 (0.116 sec/batch), lr: 0.003000
|
| 237 |
-
2025-12-03 15:51:58 INFO: Finished STEP 300/50000, loss = 5.353960 (0.113 sec/batch), lr: 0.003000
|
| 238 |
-
2025-12-03 15:51:58 INFO: Evaluating on dev set...
|
| 239 |
-
2025-12-03 15:51:58 INFO: LAS MLAS BLEX
|
| 240 |
-
2025-12-03 15:51:58 INFO: 51.49 38.56 43.22
|
| 241 |
-
2025-12-03 15:51:58 INFO: step 300: train_loss = 4.421819, dev_score = 0.5149
|
| 242 |
-
2025-12-03 15:51:59 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 243 |
-
2025-12-03 15:51:59 INFO: new best model saved.
|
| 244 |
-
2025-12-03 15:51:59 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 245 |
-
2025-12-03 15:51:59 INFO: new model checkpoint saved.
|
| 246 |
-
2025-12-03 15:52:02 INFO: Finished STEP 320/50000, loss = 5.069269 (0.121 sec/batch), lr: 0.003000
|
| 247 |
-
2025-12-03 15:52:04 INFO: Finished STEP 340/50000, loss = 3.614627 (0.112 sec/batch), lr: 0.003000
|
| 248 |
-
2025-12-03 15:52:06 INFO: Finished STEP 360/50000, loss = 3.878788 (0.115 sec/batch), lr: 0.003000
|
| 249 |
-
2025-12-03 15:52:08 INFO: Finished STEP 380/50000, loss = 4.367438 (0.100 sec/batch), lr: 0.003000
|
| 250 |
-
2025-12-03 15:52:10 INFO: Finished STEP 400/50000, loss = 4.081972 (0.107 sec/batch), lr: 0.003000
|
| 251 |
-
2025-12-03 15:52:10 INFO: Evaluating on dev set...
|
| 252 |
-
2025-12-03 15:52:11 INFO: LAS MLAS BLEX
|
| 253 |
-
2025-12-03 15:52:11 INFO: 55.69 43.55 46.51
|
| 254 |
-
2025-12-03 15:52:11 INFO: step 400: train_loss = 4.280971, dev_score = 0.5569
|
| 255 |
-
2025-12-03 15:52:11 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 256 |
-
2025-12-03 15:52:11 INFO: new best model saved.
|
| 257 |
-
2025-12-03 15:52:12 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 258 |
-
2025-12-03 15:52:12 INFO: new model checkpoint saved.
|
| 259 |
-
2025-12-03 15:52:14 INFO: Finished STEP 420/50000, loss = 3.892664 (0.110 sec/batch), lr: 0.003000
|
| 260 |
-
2025-12-03 15:52:16 INFO: Finished STEP 440/50000, loss = 3.998952 (0.100 sec/batch), lr: 0.003000
|
| 261 |
-
2025-12-03 15:52:18 INFO: Finished STEP 460/50000, loss = 5.833274 (0.112 sec/batch), lr: 0.003000
|
| 262 |
-
2025-12-03 15:52:20 INFO: Finished STEP 480/50000, loss = 3.561453 (0.106 sec/batch), lr: 0.003000
|
| 263 |
-
2025-12-03 15:52:22 INFO: Finished STEP 500/50000, loss = 3.267099 (0.106 sec/batch), lr: 0.003000
|
| 264 |
-
2025-12-03 15:52:22 INFO: Evaluating on dev set...
|
| 265 |
-
2025-12-03 15:52:23 INFO: LAS MLAS BLEX
|
| 266 |
-
2025-12-03 15:52:23 INFO: 55.45 44.30 48.95
|
| 267 |
-
2025-12-03 15:52:23 INFO: step 500: train_loss = 4.074671, dev_score = 0.5545
|
| 268 |
-
2025-12-03 15:52:24 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 269 |
-
2025-12-03 15:52:24 INFO: new model checkpoint saved.
|
| 270 |
-
2025-12-03 15:52:26 INFO: Finished STEP 520/50000, loss = 3.987905 (0.101 sec/batch), lr: 0.003000
|
| 271 |
-
2025-12-03 15:52:28 INFO: Finished STEP 540/50000, loss = 4.040174 (0.105 sec/batch), lr: 0.003000
|
| 272 |
-
2025-12-03 15:52:30 INFO: Finished STEP 560/50000, loss = 5.425183 (0.096 sec/batch), lr: 0.003000
|
| 273 |
-
2025-12-03 15:52:32 INFO: Finished STEP 580/50000, loss = 3.171549 (0.098 sec/batch), lr: 0.003000
|
| 274 |
-
2025-12-03 15:52:34 INFO: Finished STEP 600/50000, loss = 5.054599 (0.087 sec/batch), lr: 0.003000
|
| 275 |
-
2025-12-03 15:52:34 INFO: Evaluating on dev set...
|
| 276 |
-
2025-12-03 15:52:34 INFO: LAS MLAS BLEX
|
| 277 |
-
2025-12-03 15:52:34 INFO: 59.41 50.85 53.39
|
| 278 |
-
2025-12-03 15:52:34 INFO: step 600: train_loss = 3.989031, dev_score = 0.5941
|
| 279 |
-
2025-12-03 15:52:34 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 280 |
-
2025-12-03 15:52:34 INFO: new best model saved.
|
| 281 |
-
2025-12-03 15:52:35 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 282 |
-
2025-12-03 15:52:35 INFO: new model checkpoint saved.
|
| 283 |
-
2025-12-03 15:52:37 INFO: Finished STEP 620/50000, loss = 5.185298 (0.107 sec/batch), lr: 0.003000
|
| 284 |
-
2025-12-03 15:52:39 INFO: Finished STEP 640/50000, loss = 3.282787 (0.098 sec/batch), lr: 0.003000
|
| 285 |
-
2025-12-03 15:52:41 INFO: Finished STEP 660/50000, loss = 7.106239 (0.095 sec/batch), lr: 0.003000
|
| 286 |
-
2025-12-03 15:52:43 INFO: Finished STEP 680/50000, loss = 3.663255 (0.099 sec/batch), lr: 0.003000
|
| 287 |
-
2025-12-03 15:52:45 INFO: Finished STEP 700/50000, loss = 4.011485 (0.098 sec/batch), lr: 0.003000
|
| 288 |
-
2025-12-03 15:52:45 INFO: Evaluating on dev set...
|
| 289 |
-
2025-12-03 15:52:46 INFO: LAS MLAS BLEX
|
| 290 |
-
2025-12-03 15:52:46 INFO: 59.16 46.74 52.21
|
| 291 |
-
2025-12-03 15:52:46 INFO: step 700: train_loss = 3.973323, dev_score = 0.5916
|
| 292 |
-
2025-12-03 15:52:46 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 293 |
-
2025-12-03 15:52:46 INFO: new model checkpoint saved.
|
| 294 |
-
2025-12-03 15:52:48 INFO: Finished STEP 720/50000, loss = 4.034673 (0.094 sec/batch), lr: 0.003000
|
| 295 |
-
2025-12-03 15:52:50 INFO: Finished STEP 740/50000, loss = 4.719307 (0.099 sec/batch), lr: 0.003000
|
| 296 |
-
2025-12-03 15:52:52 INFO: Finished STEP 760/50000, loss = 3.175733 (0.089 sec/batch), lr: 0.003000
|
| 297 |
-
2025-12-03 15:52:54 INFO: Finished STEP 780/50000, loss = 3.158538 (0.098 sec/batch), lr: 0.003000
|
| 298 |
-
2025-12-03 15:52:56 INFO: Finished STEP 800/50000, loss = 2.881503 (0.099 sec/batch), lr: 0.003000
|
| 299 |
-
2025-12-03 15:52:56 INFO: Evaluating on dev set...
|
| 300 |
-
2025-12-03 15:52:56 INFO: LAS MLAS BLEX
|
| 301 |
-
2025-12-03 15:52:56 INFO: 57.92 46.19 49.15
|
| 302 |
-
2025-12-03 15:52:56 INFO: step 800: train_loss = 3.763542, dev_score = 0.5792
|
| 303 |
-
2025-12-03 15:52:57 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 304 |
-
2025-12-03 15:52:57 INFO: new model checkpoint saved.
|
| 305 |
-
2025-12-03 15:52:59 INFO: Finished STEP 820/50000, loss = 4.045841 (0.087 sec/batch), lr: 0.003000
|
| 306 |
-
2025-12-03 15:53:01 INFO: Finished STEP 840/50000, loss = 3.735970 (0.088 sec/batch), lr: 0.003000
|
| 307 |
-
2025-12-03 15:53:02 INFO: Finished STEP 860/50000, loss = 4.020472 (0.080 sec/batch), lr: 0.003000
|
| 308 |
-
2025-12-03 15:53:04 INFO: Finished STEP 880/50000, loss = 5.290604 (0.091 sec/batch), lr: 0.003000
|
| 309 |
-
2025-12-03 15:53:06 INFO: Finished STEP 900/50000, loss = 3.524802 (0.089 sec/batch), lr: 0.003000
|
| 310 |
-
2025-12-03 15:53:06 INFO: Evaluating on dev set...
|
| 311 |
-
2025-12-03 15:53:07 INFO: LAS MLAS BLEX
|
| 312 |
-
2025-12-03 15:53:07 INFO: 56.93 45.67 49.47
|
| 313 |
-
2025-12-03 15:53:07 INFO: step 900: train_loss = 3.948068, dev_score = 0.5693
|
| 314 |
-
2025-12-03 15:53:07 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 315 |
-
2025-12-03 15:53:07 INFO: new model checkpoint saved.
|
| 316 |
-
2025-12-03 15:53:09 INFO: Finished STEP 920/50000, loss = 3.348033 (0.088 sec/batch), lr: 0.003000
|
| 317 |
-
2025-12-03 15:53:11 INFO: Finished STEP 940/50000, loss = 3.860394 (0.091 sec/batch), lr: 0.003000
|
| 318 |
-
2025-12-03 15:53:13 INFO: Finished STEP 960/50000, loss = 3.459932 (0.082 sec/batch), lr: 0.003000
|
| 319 |
-
2025-12-03 15:53:15 INFO: Finished STEP 980/50000, loss = 4.335596 (0.085 sec/batch), lr: 0.003000
|
| 320 |
-
2025-12-03 15:53:17 INFO: Finished STEP 1000/50000, loss = 3.108686 (0.089 sec/batch), lr: 0.003000
|
| 321 |
-
2025-12-03 15:53:17 INFO: Evaluating on dev set...
|
| 322 |
-
2025-12-03 15:53:17 INFO: LAS MLAS BLEX
|
| 323 |
-
2025-12-03 15:53:17 INFO: 60.40 51.27 53.81
|
| 324 |
-
2025-12-03 15:53:17 INFO: step 1000: train_loss = 3.976391, dev_score = 0.6040
|
| 325 |
-
2025-12-03 15:53:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 326 |
-
2025-12-03 15:53:17 INFO: new best model saved.
|
| 327 |
-
2025-12-03 15:53:18 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 328 |
-
2025-12-03 15:53:18 INFO: new model checkpoint saved.
|
| 329 |
-
2025-12-03 15:53:20 INFO: Finished STEP 1020/50000, loss = 6.081649 (0.089 sec/batch), lr: 0.003000
|
| 330 |
-
2025-12-03 15:53:21 INFO: Finished STEP 1040/50000, loss = 3.454382 (0.089 sec/batch), lr: 0.003000
|
| 331 |
-
2025-12-03 15:53:23 INFO: Finished STEP 1060/50000, loss = 3.093398 (0.085 sec/batch), lr: 0.003000
|
| 332 |
-
2025-12-03 15:53:25 INFO: Finished STEP 1080/50000, loss = 3.780082 (0.088 sec/batch), lr: 0.003000
|
| 333 |
-
2025-12-03 15:53:27 INFO: Finished STEP 1100/50000, loss = 3.758047 (0.075 sec/batch), lr: 0.003000
|
| 334 |
-
2025-12-03 15:53:27 INFO: Evaluating on dev set...
|
| 335 |
-
2025-12-03 15:53:27 INFO: LAS MLAS BLEX
|
| 336 |
-
2025-12-03 15:53:27 INFO: 59.41 49.05 52.85
|
| 337 |
-
2025-12-03 15:53:27 INFO: step 1100: train_loss = 3.779099, dev_score = 0.5941
|
| 338 |
-
2025-12-03 15:53:28 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 339 |
-
2025-12-03 15:53:28 INFO: new model checkpoint saved.
|
| 340 |
-
2025-12-03 15:53:30 INFO: Finished STEP 1120/50000, loss = 3.774802 (0.088 sec/batch), lr: 0.003000
|
| 341 |
-
2025-12-03 15:53:31 INFO: Finished STEP 1140/50000, loss = 2.977976 (0.089 sec/batch), lr: 0.003000
|
| 342 |
-
2025-12-03 15:53:33 INFO: Finished STEP 1160/50000, loss = 3.838738 (0.078 sec/batch), lr: 0.003000
|
| 343 |
-
2025-12-03 15:53:35 INFO: Finished STEP 1180/50000, loss = 4.660880 (0.080 sec/batch), lr: 0.003000
|
| 344 |
-
2025-12-03 15:53:36 INFO: Finished STEP 1200/50000, loss = 2.489745 (0.088 sec/batch), lr: 0.003000
|
| 345 |
-
2025-12-03 15:53:36 INFO: Evaluating on dev set...
|
| 346 |
-
2025-12-03 15:53:37 INFO: LAS MLAS BLEX
|
| 347 |
-
2025-12-03 15:53:37 INFO: 58.66 48.31 51.69
|
| 348 |
-
2025-12-03 15:53:37 INFO: step 1200: train_loss = 3.720397, dev_score = 0.5866
|
| 349 |
-
2025-12-03 15:53:37 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 350 |
-
2025-12-03 15:53:37 INFO: new model checkpoint saved.
|
| 351 |
-
2025-12-03 15:53:39 INFO: Finished STEP 1220/50000, loss = 3.247226 (0.090 sec/batch), lr: 0.003000
|
| 352 |
-
2025-12-03 15:53:41 INFO: Finished STEP 1240/50000, loss = 4.093720 (0.079 sec/batch), lr: 0.003000
|
| 353 |
-
2025-12-03 15:53:43 INFO: Finished STEP 1260/50000, loss = 2.970687 (0.080 sec/batch), lr: 0.003000
|
| 354 |
-
2025-12-03 15:53:44 INFO: Finished STEP 1280/50000, loss = 3.959898 (0.087 sec/batch), lr: 0.003000
|
| 355 |
-
2025-12-03 15:53:46 INFO: Finished STEP 1300/50000, loss = 3.890834 (0.081 sec/batch), lr: 0.003000
|
| 356 |
-
2025-12-03 15:53:46 INFO: Evaluating on dev set...
|
| 357 |
-
2025-12-03 15:53:47 INFO: LAS MLAS BLEX
|
| 358 |
-
2025-12-03 15:53:47 INFO: 57.67 50.00 52.97
|
| 359 |
-
2025-12-03 15:53:47 INFO: step 1300: train_loss = 3.945667, dev_score = 0.5767
|
| 360 |
-
2025-12-03 15:53:47 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 361 |
-
2025-12-03 15:53:47 INFO: new model checkpoint saved.
|
| 362 |
-
2025-12-03 15:53:49 INFO: Finished STEP 1320/50000, loss = 4.299366 (0.076 sec/batch), lr: 0.003000
|
| 363 |
-
2025-12-03 15:53:51 INFO: Finished STEP 1340/50000, loss = 3.080094 (0.082 sec/batch), lr: 0.003000
|
| 364 |
-
2025-12-03 15:53:52 INFO: Finished STEP 1360/50000, loss = 5.326971 (0.087 sec/batch), lr: 0.003000
|
| 365 |
-
2025-12-03 15:53:54 INFO: Finished STEP 1380/50000, loss = 4.148820 (0.083 sec/batch), lr: 0.003000
|
| 366 |
-
2025-12-03 15:53:56 INFO: Finished STEP 1400/50000, loss = 3.131835 (0.085 sec/batch), lr: 0.003000
|
| 367 |
-
2025-12-03 15:53:56 INFO: Evaluating on dev set...
|
| 368 |
-
2025-12-03 15:53:56 INFO: LAS MLAS BLEX
|
| 369 |
-
2025-12-03 15:53:56 INFO: 59.16 49.47 52.85
|
| 370 |
-
2025-12-03 15:53:56 INFO: step 1400: train_loss = 3.828261, dev_score = 0.5916
|
| 371 |
-
2025-12-03 15:53:57 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 372 |
-
2025-12-03 15:53:57 INFO: new model checkpoint saved.
|
| 373 |
-
2025-12-03 15:53:58 INFO: Finished STEP 1420/50000, loss = 3.877927 (0.079 sec/batch), lr: 0.003000
|
| 374 |
-
2025-12-03 15:54:00 INFO: Finished STEP 1440/50000, loss = 2.828221 (0.077 sec/batch), lr: 0.003000
|
| 375 |
-
2025-12-03 15:54:02 INFO: Finished STEP 1460/50000, loss = 2.771763 (0.088 sec/batch), lr: 0.003000
|
| 376 |
-
2025-12-03 15:54:03 INFO: Finished STEP 1480/50000, loss = 3.887072 (0.082 sec/batch), lr: 0.003000
|
| 377 |
-
2025-12-03 15:54:05 INFO: Finished STEP 1500/50000, loss = 3.089223 (0.080 sec/batch), lr: 0.003000
|
| 378 |
-
2025-12-03 15:54:05 INFO: Evaluating on dev set...
|
| 379 |
-
2025-12-03 15:54:05 INFO: LAS MLAS BLEX
|
| 380 |
-
2025-12-03 15:54:05 INFO: 60.89 49.47 54.12
|
| 381 |
-
2025-12-03 15:54:05 INFO: step 1500: train_loss = 3.662534, dev_score = 0.6089
|
| 382 |
-
2025-12-03 15:54:06 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 383 |
-
2025-12-03 15:54:06 INFO: new best model saved.
|
| 384 |
-
2025-12-03 15:54:06 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 385 |
-
2025-12-03 15:54:06 INFO: new model checkpoint saved.
|
| 386 |
-
2025-12-03 15:54:08 INFO: Finished STEP 1520/50000, loss = 3.424534 (0.079 sec/batch), lr: 0.003000
|
| 387 |
-
2025-12-03 15:54:10 INFO: Finished STEP 1540/50000, loss = 4.257630 (0.081 sec/batch), lr: 0.003000
|
| 388 |
-
2025-12-03 15:54:11 INFO: Finished STEP 1560/50000, loss = 4.004986 (0.078 sec/batch), lr: 0.003000
|
| 389 |
-
2025-12-03 15:54:13 INFO: Finished STEP 1580/50000, loss = 4.259016 (0.076 sec/batch), lr: 0.003000
|
| 390 |
-
2025-12-03 15:54:14 INFO: Finished STEP 1600/50000, loss = 4.886831 (0.071 sec/batch), lr: 0.003000
|
| 391 |
-
2025-12-03 15:54:14 INFO: Evaluating on dev set...
|
| 392 |
-
2025-12-03 15:54:15 INFO: LAS MLAS BLEX
|
| 393 |
-
2025-12-03 15:54:15 INFO: 58.66 49.47 52.01
|
| 394 |
-
2025-12-03 15:54:15 INFO: step 1600: train_loss = 3.940190, dev_score = 0.5866
|
| 395 |
-
2025-12-03 15:54:15 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 396 |
-
2025-12-03 15:54:15 INFO: new model checkpoint saved.
|
| 397 |
-
2025-12-03 15:54:17 INFO: Finished STEP 1620/50000, loss = 3.486598 (0.072 sec/batch), lr: 0.003000
|
| 398 |
-
2025-12-03 15:54:19 INFO: Finished STEP 1640/50000, loss = 3.426835 (0.076 sec/batch), lr: 0.003000
|
| 399 |
-
2025-12-03 15:54:20 INFO: Finished STEP 1660/50000, loss = 3.082500 (0.077 sec/batch), lr: 0.003000
|
| 400 |
-
2025-12-03 15:54:22 INFO: Finished STEP 1680/50000, loss = 2.587858 (0.080 sec/batch), lr: 0.003000
|
| 401 |
-
2025-12-03 15:54:23 INFO: Finished STEP 1700/50000, loss = 4.049259 (0.081 sec/batch), lr: 0.003000
|
| 402 |
-
2025-12-03 15:54:23 INFO: Evaluating on dev set...
|
| 403 |
-
2025-12-03 15:54:24 INFO: LAS MLAS BLEX
|
| 404 |
-
2025-12-03 15:54:24 INFO: 59.41 50.63 54.43
|
| 405 |
-
2025-12-03 15:54:24 INFO: step 1700: train_loss = 3.731144, dev_score = 0.5941
|
| 406 |
-
2025-12-03 15:54:24 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 407 |
-
2025-12-03 15:54:24 INFO: new model checkpoint saved.
|
| 408 |
-
2025-12-03 15:54:26 INFO: Finished STEP 1720/50000, loss = 2.259671 (0.076 sec/batch), lr: 0.003000
|
| 409 |
-
2025-12-03 15:54:28 INFO: Finished STEP 1740/50000, loss = 5.103413 (0.079 sec/batch), lr: 0.003000
|
| 410 |
-
2025-12-03 15:54:29 INFO: Finished STEP 1760/50000, loss = 3.651329 (0.072 sec/batch), lr: 0.003000
|
| 411 |
-
2025-12-03 15:54:31 INFO: Finished STEP 1780/50000, loss = 3.879523 (0.078 sec/batch), lr: 0.003000
|
| 412 |
-
2025-12-03 15:54:32 INFO: Finished STEP 1800/50000, loss = 3.246203 (0.078 sec/batch), lr: 0.003000
|
| 413 |
-
2025-12-03 15:54:32 INFO: Evaluating on dev set...
|
| 414 |
-
2025-12-03 15:54:33 INFO: LAS MLAS BLEX
|
| 415 |
-
2025-12-03 15:54:33 INFO: 60.89 53.39 55.93
|
| 416 |
-
2025-12-03 15:54:33 INFO: step 1800: train_loss = 3.826308, dev_score = 0.6089
|
| 417 |
-
2025-12-03 15:54:33 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 418 |
-
2025-12-03 15:54:33 INFO: new best model saved.
|
| 419 |
-
2025-12-03 15:54:34 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 420 |
-
2025-12-03 15:54:34 INFO: new model checkpoint saved.
|
| 421 |
-
2025-12-03 15:54:35 INFO: Finished STEP 1820/50000, loss = 4.498244 (0.072 sec/batch), lr: 0.003000
|
| 422 |
-
2025-12-03 15:54:37 INFO: Finished STEP 1840/50000, loss = 3.589321 (0.078 sec/batch), lr: 0.003000
|
| 423 |
-
2025-12-03 15:54:39 INFO: Finished STEP 1860/50000, loss = 2.884829 (0.077 sec/batch), lr: 0.003000
|
| 424 |
-
2025-12-03 15:54:40 INFO: Finished STEP 1880/50000, loss = 3.823880 (0.082 sec/batch), lr: 0.003000
|
| 425 |
-
2025-12-03 15:54:42 INFO: Finished STEP 1900/50000, loss = 2.696285 (0.082 sec/batch), lr: 0.003000
|
| 426 |
-
2025-12-03 15:54:42 INFO: Evaluating on dev set...
|
| 427 |
-
2025-12-03 15:54:42 INFO: LAS MLAS BLEX
|
| 428 |
-
2025-12-03 15:54:42 INFO: 60.64 51.16 55.39
|
| 429 |
-
2025-12-03 15:54:42 INFO: step 1900: train_loss = 4.029873, dev_score = 0.6064
|
| 430 |
-
2025-12-03 15:54:43 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 431 |
-
2025-12-03 15:54:43 INFO: new model checkpoint saved.
|
| 432 |
-
2025-12-03 15:54:44 INFO: Finished STEP 1920/50000, loss = 5.128090 (0.068 sec/batch), lr: 0.003000
|
| 433 |
-
2025-12-03 15:54:46 INFO: Finished STEP 1940/50000, loss = 3.145783 (0.078 sec/batch), lr: 0.003000
|
| 434 |
-
2025-12-03 15:54:47 INFO: Finished STEP 1960/50000, loss = 3.195472 (0.075 sec/batch), lr: 0.003000
|
| 435 |
-
2025-12-03 15:54:49 INFO: Finished STEP 1980/50000, loss = 3.154953 (0.072 sec/batch), lr: 0.003000
|
| 436 |
-
2025-12-03 15:54:51 INFO: Finished STEP 2000/50000, loss = 3.763106 (0.080 sec/batch), lr: 0.003000
|
| 437 |
-
2025-12-03 15:54:51 INFO: Evaluating on dev set...
|
| 438 |
-
2025-12-03 15:54:51 INFO: LAS MLAS BLEX
|
| 439 |
-
2025-12-03 15:54:51 INFO: 61.39 50.85 54.66
|
| 440 |
-
2025-12-03 15:54:51 INFO: step 2000: train_loss = 3.764299, dev_score = 0.6139
|
| 441 |
-
2025-12-03 15:54:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 442 |
-
2025-12-03 15:54:51 INFO: new best model saved.
|
| 443 |
-
2025-12-03 15:54:52 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 444 |
-
2025-12-03 15:54:52 INFO: new model checkpoint saved.
|
| 445 |
-
2025-12-03 15:54:53 INFO: Finished STEP 2020/50000, loss = 4.129125 (0.077 sec/batch), lr: 0.003000
|
| 446 |
-
2025-12-03 15:54:55 INFO: Finished STEP 2040/50000, loss = 4.000008 (0.076 sec/batch), lr: 0.003000
|
| 447 |
-
2025-12-03 15:54:56 INFO: Finished STEP 2060/50000, loss = 3.594474 (0.076 sec/batch), lr: 0.003000
|
| 448 |
-
2025-12-03 15:54:58 INFO: Finished STEP 2080/50000, loss = 5.156559 (0.070 sec/batch), lr: 0.003000
|
| 449 |
-
2025-12-03 15:54:59 INFO: Finished STEP 2100/50000, loss = 2.780583 (0.076 sec/batch), lr: 0.003000
|
| 450 |
-
2025-12-03 15:54:59 INFO: Evaluating on dev set...
|
| 451 |
-
2025-12-03 15:55:00 INFO: LAS MLAS BLEX
|
| 452 |
-
2025-12-03 15:55:00 INFO: 58.91 50.32 53.28
|
| 453 |
-
2025-12-03 15:55:00 INFO: step 2100: train_loss = 3.970306, dev_score = 0.5891
|
| 454 |
-
2025-12-03 15:55:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 455 |
-
2025-12-03 15:55:01 INFO: new model checkpoint saved.
|
| 456 |
-
2025-12-03 15:55:02 INFO: Finished STEP 2120/50000, loss = 3.461082 (0.078 sec/batch), lr: 0.003000
|
| 457 |
-
2025-12-03 15:55:04 INFO: Finished STEP 2140/50000, loss = 4.371090 (0.068 sec/batch), lr: 0.003000
|
| 458 |
-
2025-12-03 15:55:05 INFO: Finished STEP 2160/50000, loss = 3.734883 (0.074 sec/batch), lr: 0.003000
|
| 459 |
-
2025-12-03 15:55:07 INFO: Finished STEP 2180/50000, loss = 2.186298 (0.077 sec/batch), lr: 0.003000
|
| 460 |
-
2025-12-03 15:55:08 INFO: Finished STEP 2200/50000, loss = 3.871332 (0.078 sec/batch), lr: 0.003000
|
| 461 |
-
2025-12-03 15:55:08 INFO: Evaluating on dev set...
|
| 462 |
-
2025-12-03 15:55:09 INFO: LAS MLAS BLEX
|
| 463 |
-
2025-12-03 15:55:09 INFO: 58.66 48.73 52.97
|
| 464 |
-
2025-12-03 15:55:09 INFO: step 2200: train_loss = 3.926589, dev_score = 0.5866
|
| 465 |
-
2025-12-03 15:55:09 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 466 |
-
2025-12-03 15:55:09 INFO: new model checkpoint saved.
|
| 467 |
-
2025-12-03 15:55:11 INFO: Finished STEP 2220/50000, loss = 2.865967 (0.073 sec/batch), lr: 0.003000
|
| 468 |
-
2025-12-03 15:55:12 INFO: Finished STEP 2240/50000, loss = 2.578629 (0.069 sec/batch), lr: 0.003000
|
| 469 |
-
2025-12-03 15:55:14 INFO: Finished STEP 2260/50000, loss = 3.295215 (0.068 sec/batch), lr: 0.003000
|
| 470 |
-
2025-12-03 15:55:15 INFO: Finished STEP 2280/50000, loss = 3.903890 (0.074 sec/batch), lr: 0.003000
|
| 471 |
-
2025-12-03 15:55:17 INFO: Finished STEP 2300/50000, loss = 3.614969 (0.071 sec/batch), lr: 0.003000
|
| 472 |
-
2025-12-03 15:55:17 INFO: Evaluating on dev set...
|
| 473 |
-
2025-12-03 15:55:17 INFO: LAS MLAS BLEX
|
| 474 |
-
2025-12-03 15:55:17 INFO: 60.40 52.43 56.66
|
| 475 |
-
2025-12-03 15:55:17 INFO: step 2300: train_loss = 3.953848, dev_score = 0.6040
|
| 476 |
-
2025-12-03 15:55:18 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 477 |
-
2025-12-03 15:55:18 INFO: new model checkpoint saved.
|
| 478 |
-
2025-12-03 15:55:19 INFO: Finished STEP 2320/50000, loss = 3.836884 (0.076 sec/batch), lr: 0.003000
|
| 479 |
-
2025-12-03 15:55:21 INFO: Finished STEP 2340/50000, loss = 2.887274 (0.079 sec/batch), lr: 0.003000
|
| 480 |
-
2025-12-03 15:55:22 INFO: Finished STEP 2360/50000, loss = 4.795152 (0.079 sec/batch), lr: 0.003000
|
| 481 |
-
2025-12-03 15:55:24 INFO: Finished STEP 2380/50000, loss = 4.483939 (0.074 sec/batch), lr: 0.003000
|
| 482 |
-
2025-12-03 15:55:25 INFO: Finished STEP 2400/50000, loss = 4.851534 (0.075 sec/batch), lr: 0.003000
|
| 483 |
-
2025-12-03 15:55:25 INFO: Evaluating on dev set...
|
| 484 |
-
2025-12-03 15:55:26 INFO: LAS MLAS BLEX
|
| 485 |
-
2025-12-03 15:55:26 INFO: 58.91 50.21 54.85
|
| 486 |
-
2025-12-03 15:55:26 INFO: step 2400: train_loss = 3.932371, dev_score = 0.5891
|
| 487 |
-
2025-12-03 15:55:27 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 488 |
-
2025-12-03 15:55:27 INFO: new model checkpoint saved.
|
| 489 |
-
2025-12-03 15:55:28 INFO: Finished STEP 2420/50000, loss = 4.692697 (0.067 sec/batch), lr: 0.003000
|
| 490 |
-
2025-12-03 15:55:30 INFO: Finished STEP 2440/50000, loss = 2.805047 (0.068 sec/batch), lr: 0.003000
|
| 491 |
-
2025-12-03 15:55:31 INFO: Finished STEP 2460/50000, loss = 3.715670 (0.077 sec/batch), lr: 0.003000
|
| 492 |
-
2025-12-03 15:55:32 INFO: Finished STEP 2480/50000, loss = 5.003852 (0.072 sec/batch), lr: 0.003000
|
| 493 |
-
2025-12-03 15:55:34 INFO: Finished STEP 2500/50000, loss = 5.362862 (0.071 sec/batch), lr: 0.003000
|
| 494 |
-
2025-12-03 15:55:34 INFO: Evaluating on dev set...
|
| 495 |
-
2025-12-03 15:55:34 INFO: LAS MLAS BLEX
|
| 496 |
-
2025-12-03 15:55:34 INFO: 62.87 54.12 57.51
|
| 497 |
-
2025-12-03 15:55:34 INFO: step 2500: train_loss = 3.716296, dev_score = 0.6287
|
| 498 |
-
2025-12-03 15:55:35 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 499 |
-
2025-12-03 15:55:35 INFO: new best model saved.
|
| 500 |
-
2025-12-03 15:55:35 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 501 |
-
2025-12-03 15:55:35 INFO: new model checkpoint saved.
|
| 502 |
-
2025-12-03 15:55:37 INFO: Finished STEP 2520/50000, loss = 4.285573 (0.073 sec/batch), lr: 0.003000
|
| 503 |
-
2025-12-03 15:55:38 INFO: Finished STEP 2540/50000, loss = 3.220155 (0.071 sec/batch), lr: 0.003000
|
| 504 |
-
2025-12-03 15:55:40 INFO: Finished STEP 2560/50000, loss = 4.155645 (0.077 sec/batch), lr: 0.003000
|
| 505 |
-
2025-12-03 15:55:41 INFO: Finished STEP 2580/50000, loss = 3.434030 (0.073 sec/batch), lr: 0.003000
|
| 506 |
-
2025-12-03 15:55:43 INFO: Finished STEP 2600/50000, loss = 2.680579 (0.076 sec/batch), lr: 0.003000
|
| 507 |
-
2025-12-03 15:55:43 INFO: Evaluating on dev set...
|
| 508 |
-
2025-12-03 15:55:43 INFO: LAS MLAS BLEX
|
| 509 |
-
2025-12-03 15:55:43 INFO: 62.87 54.66 58.05
|
| 510 |
-
2025-12-03 15:55:43 INFO: step 2600: train_loss = 4.018973, dev_score = 0.6287
|
| 511 |
-
2025-12-03 15:55:43 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 512 |
-
2025-12-03 15:55:43 INFO: new best model saved.
|
| 513 |
-
2025-12-03 15:55:44 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 514 |
-
2025-12-03 15:55:44 INFO: new model checkpoint saved.
|
| 515 |
-
2025-12-03 15:55:45 INFO: Finished STEP 2620/50000, loss = 2.524696 (0.072 sec/batch), lr: 0.003000
|
| 516 |
-
2025-12-03 15:55:47 INFO: Finished STEP 2640/50000, loss = 3.829545 (0.065 sec/batch), lr: 0.003000
|
| 517 |
-
2025-12-03 15:55:48 INFO: Finished STEP 2660/50000, loss = 3.246467 (0.069 sec/batch), lr: 0.003000
|
| 518 |
-
2025-12-03 15:55:50 INFO: Finished STEP 2680/50000, loss = 2.396587 (0.068 sec/batch), lr: 0.003000
|
| 519 |
-
2025-12-03 15:55:51 INFO: Finished STEP 2700/50000, loss = 3.570580 (0.071 sec/batch), lr: 0.003000
|
| 520 |
-
2025-12-03 15:55:51 INFO: Evaluating on dev set...
|
| 521 |
-
2025-12-03 15:55:52 INFO: LAS MLAS BLEX
|
| 522 |
-
2025-12-03 15:55:52 INFO: 58.91 49.58 54.24
|
| 523 |
-
2025-12-03 15:55:52 INFO: step 2700: train_loss = 3.918322, dev_score = 0.5891
|
| 524 |
-
2025-12-03 15:55:52 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 525 |
-
2025-12-03 15:55:52 INFO: new model checkpoint saved.
|
| 526 |
-
2025-12-03 15:55:54 INFO: Finished STEP 2720/50000, loss = 2.712030 (0.072 sec/batch), lr: 0.003000
|
| 527 |
-
2025-12-03 15:55:55 INFO: Finished STEP 2740/50000, loss = 5.811014 (0.077 sec/batch), lr: 0.003000
|
| 528 |
-
2025-12-03 15:55:57 INFO: Finished STEP 2760/50000, loss = 3.773255 (0.076 sec/batch), lr: 0.003000
|
| 529 |
-
2025-12-03 15:55:58 INFO: Finished STEP 2780/50000, loss = 2.087427 (0.075 sec/batch), lr: 0.003000
|
| 530 |
-
2025-12-03 15:56:00 INFO: Finished STEP 2800/50000, loss = 3.811720 (0.063 sec/batch), lr: 0.003000
|
| 531 |
-
2025-12-03 15:56:00 INFO: Evaluating on dev set...
|
| 532 |
-
2025-12-03 15:56:00 INFO: LAS MLAS BLEX
|
| 533 |
-
2025-12-03 15:56:00 INFO: 61.63 51.59 55.39
|
| 534 |
-
2025-12-03 15:56:00 INFO: step 2800: train_loss = 3.902705, dev_score = 0.6163
|
| 535 |
-
2025-12-03 15:56:01 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 536 |
-
2025-12-03 15:56:01 INFO: new model checkpoint saved.
|
| 537 |
-
2025-12-03 15:56:02 INFO: Finished STEP 2820/50000, loss = 3.443505 (0.070 sec/batch), lr: 0.003000
|
| 538 |
-
2025-12-03 15:56:04 INFO: Finished STEP 2840/50000, loss = 3.684120 (0.064 sec/batch), lr: 0.003000
|
| 539 |
-
2025-12-03 15:56:05 INFO: Finished STEP 2860/50000, loss = 5.325919 (0.071 sec/batch), lr: 0.003000
|
| 540 |
-
2025-12-03 15:56:06 INFO: Finished STEP 2880/50000, loss = 3.516696 (0.068 sec/batch), lr: 0.003000
|
| 541 |
-
2025-12-03 15:56:08 INFO: Finished STEP 2900/50000, loss = 3.573216 (0.072 sec/batch), lr: 0.003000
|
| 542 |
-
2025-12-03 15:56:08 INFO: Evaluating on dev set...
|
| 543 |
-
2025-12-03 15:56:09 INFO: LAS MLAS BLEX
|
| 544 |
-
2025-12-03 15:56:09 INFO: 58.17 48.20 53.28
|
| 545 |
-
2025-12-03 15:56:09 INFO: step 2900: train_loss = 4.154457, dev_score = 0.5817
|
| 546 |
-
2025-12-03 15:56:09 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 547 |
-
2025-12-03 15:56:09 INFO: new model checkpoint saved.
|
| 548 |
-
2025-12-03 15:56:11 INFO: Finished STEP 2920/50000, loss = 2.522206 (0.077 sec/batch), lr: 0.003000
|
| 549 |
-
2025-12-03 15:56:12 INFO: Finished STEP 2940/50000, loss = 2.687997 (0.082 sec/batch), lr: 0.003000
|
| 550 |
-
2025-12-03 15:56:13 INFO: Finished STEP 2960/50000, loss = 3.291323 (0.078 sec/batch), lr: 0.003000
|
| 551 |
-
2025-12-03 15:56:15 INFO: Finished STEP 2980/50000, loss = 4.553081 (0.069 sec/batch), lr: 0.003000
|
| 552 |
-
2025-12-03 15:56:16 INFO: Finished STEP 3000/50000, loss = 3.136598 (0.071 sec/batch), lr: 0.003000
|
| 553 |
-
2025-12-03 15:56:16 INFO: Evaluating on dev set...
|
| 554 |
-
2025-12-03 15:56:17 INFO: LAS MLAS BLEX
|
| 555 |
-
2025-12-03 15:56:17 INFO: 61.14 52.43 55.39
|
| 556 |
-
2025-12-03 15:56:17 INFO: step 3000: train_loss = 3.949049, dev_score = 0.6114
|
| 557 |
-
2025-12-03 15:56:17 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 558 |
-
2025-12-03 15:56:17 INFO: new model checkpoint saved.
|
| 559 |
-
2025-12-03 15:56:19 INFO: Finished STEP 3020/50000, loss = 3.020242 (0.077 sec/batch), lr: 0.003000
|
| 560 |
-
2025-12-03 15:56:20 INFO: Finished STEP 3040/50000, loss = 4.876935 (0.069 sec/batch), lr: 0.003000
|
| 561 |
-
2025-12-03 15:56:22 INFO: Finished STEP 3060/50000, loss = 3.937078 (0.080 sec/batch), lr: 0.003000
|
| 562 |
-
2025-12-03 15:56:23 INFO: Finished STEP 3080/50000, loss = 4.811732 (0.064 sec/batch), lr: 0.003000
|
| 563 |
-
2025-12-03 15:56:25 INFO: Finished STEP 3100/50000, loss = 3.313049 (0.073 sec/batch), lr: 0.003000
|
| 564 |
-
2025-12-03 15:56:25 INFO: Evaluating on dev set...
|
| 565 |
-
2025-12-03 15:56:25 INFO: LAS MLAS BLEX
|
| 566 |
-
2025-12-03 15:56:25 INFO: 62.13 54.12 58.35
|
| 567 |
-
2025-12-03 15:56:25 INFO: step 3100: train_loss = 4.106587, dev_score = 0.6213
|
| 568 |
-
2025-12-03 15:56:26 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 569 |
-
2025-12-03 15:56:26 INFO: new model checkpoint saved.
|
| 570 |
-
2025-12-03 15:56:27 INFO: Finished STEP 3120/50000, loss = 3.817453 (0.067 sec/batch), lr: 0.003000
|
| 571 |
-
2025-12-03 15:56:29 INFO: Finished STEP 3140/50000, loss = 5.620703 (0.070 sec/batch), lr: 0.003000
|
| 572 |
-
2025-12-03 15:56:30 INFO: Finished STEP 3160/50000, loss = 4.408039 (0.078 sec/batch), lr: 0.003000
|
| 573 |
-
2025-12-03 15:56:32 INFO: Finished STEP 3180/50000, loss = 2.225734 (0.074 sec/batch), lr: 0.003000
|
| 574 |
-
2025-12-03 15:56:33 INFO: Finished STEP 3200/50000, loss = 6.085883 (0.082 sec/batch), lr: 0.003000
|
| 575 |
-
2025-12-03 15:56:33 INFO: Evaluating on dev set...
|
| 576 |
-
2025-12-03 15:56:34 INFO: LAS MLAS BLEX
|
| 577 |
-
2025-12-03 15:56:34 INFO: 58.91 51.59 53.70
|
| 578 |
-
2025-12-03 15:56:34 INFO: step 3200: train_loss = 4.045106, dev_score = 0.5891
|
| 579 |
-
2025-12-03 15:56:34 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 580 |
-
2025-12-03 15:56:34 INFO: new model checkpoint saved.
|
| 581 |
-
2025-12-03 15:56:36 INFO: Finished STEP 3220/50000, loss = 3.643058 (0.064 sec/batch), lr: 0.003000
|
| 582 |
-
2025-12-03 15:56:37 INFO: Finished STEP 3240/50000, loss = 3.506227 (0.068 sec/batch), lr: 0.003000
|
| 583 |
-
2025-12-03 15:56:38 INFO: Finished STEP 3260/50000, loss = 2.930928 (0.072 sec/batch), lr: 0.003000
|
| 584 |
-
2025-12-03 15:56:40 INFO: Finished STEP 3280/50000, loss = 4.896526 (0.071 sec/batch), lr: 0.003000
|
| 585 |
-
2025-12-03 15:56:41 INFO: Finished STEP 3300/50000, loss = 4.648397 (0.077 sec/batch), lr: 0.003000
|
| 586 |
-
2025-12-03 15:56:41 INFO: Evaluating on dev set...
|
| 587 |
-
2025-12-03 15:56:42 INFO: LAS MLAS BLEX
|
| 588 |
-
2025-12-03 15:56:42 INFO: 59.16 52.01 52.43
|
| 589 |
-
2025-12-03 15:56:42 INFO: step 3300: train_loss = 3.880126, dev_score = 0.5916
|
| 590 |
-
2025-12-03 15:56:42 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 591 |
-
2025-12-03 15:56:42 INFO: new model checkpoint saved.
|
| 592 |
-
2025-12-03 15:56:44 INFO: Finished STEP 3320/50000, loss = 4.656130 (0.065 sec/batch), lr: 0.003000
|
| 593 |
-
2025-12-03 15:56:45 INFO: Finished STEP 3340/50000, loss = 4.644033 (0.070 sec/batch), lr: 0.003000
|
| 594 |
-
2025-12-03 15:56:47 INFO: Finished STEP 3360/50000, loss = 4.342103 (0.070 sec/batch), lr: 0.003000
|
| 595 |
-
2025-12-03 15:56:48 INFO: Finished STEP 3380/50000, loss = 1.757741 (0.069 sec/batch), lr: 0.003000
|
| 596 |
-
2025-12-03 15:56:49 INFO: Finished STEP 3400/50000, loss = 3.296278 (0.077 sec/batch), lr: 0.003000
|
| 597 |
-
2025-12-03 15:56:49 INFO: Evaluating on dev set...
|
| 598 |
-
2025-12-03 15:56:50 INFO: LAS MLAS BLEX
|
| 599 |
-
2025-12-03 15:56:50 INFO: 57.92 48.63 52.43
|
| 600 |
-
2025-12-03 15:56:50 INFO: step 3400: train_loss = 3.761222, dev_score = 0.5792
|
| 601 |
-
2025-12-03 15:56:51 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 602 |
-
2025-12-03 15:56:51 INFO: new model checkpoint saved.
|
| 603 |
-
2025-12-03 15:56:52 INFO: Finished STEP 3420/50000, loss = 4.175681 (0.063 sec/batch), lr: 0.003000
|
| 604 |
-
2025-12-03 15:56:53 INFO: Finished STEP 3440/50000, loss = 3.109319 (0.072 sec/batch), lr: 0.003000
|
| 605 |
-
2025-12-03 15:56:55 INFO: Finished STEP 3460/50000, loss = 5.698625 (0.070 sec/batch), lr: 0.003000
|
| 606 |
-
2025-12-03 15:56:56 INFO: Finished STEP 3480/50000, loss = 4.906111 (0.069 sec/batch), lr: 0.003000
|
| 607 |
-
2025-12-03 15:56:58 INFO: Finished STEP 3500/50000, loss = 2.886077 (0.068 sec/batch), lr: 0.003000
|
| 608 |
-
2025-12-03 15:56:58 INFO: Evaluating on dev set...
|
| 609 |
-
2025-12-03 15:56:58 INFO: LAS MLAS BLEX
|
| 610 |
-
2025-12-03 15:56:58 INFO: 60.64 51.48 55.70
|
| 611 |
-
2025-12-03 15:56:58 INFO: step 3500: train_loss = 3.804499, dev_score = 0.6064
|
| 612 |
-
2025-12-03 15:56:59 INFO: Model saved to saved_models/depparse/sv_diachronic_charlm_parser_checkpoint.pt
|
| 613 |
-
2025-12-03 15:56:59 INFO: new model checkpoint saved.
|
| 614 |
-
2025-12-03 15:57:00 INFO: Finished STEP 3520/50000, loss = 2.301594 (0.065 sec/batch), lr: 0.003000
|
| 615 |
-
2025-12-03 15:57:01 INFO: Finished STEP 3540/50000, loss = 4.423633 (0.075 sec/batch), lr: 0.003000
|
| 616 |
-
2025-12-03 15:57:03 INFO: Finished STEP 3560/50000, loss = 3.613044 (0.063 sec/batch), lr: 0.003000
|
| 617 |
-
2025-12-03 15:57:04 INFO: Finished STEP 3580/50000, loss = 3.496284 (0.067 sec/batch), lr: 0.003000
|
| 618 |
-
2025-12-03 15:57:06 INFO: Finished STEP 3600/50000, loss = 6.527389 (0.066 sec/batch), lr: 0.003000
|
| 619 |
-
2025-12-03 15:57:06 INFO: Evaluating on dev set...
|
| 620 |
-
2025-12-03 15:57:06 INFO: LAS MLAS BLEX
|
| 621 |
-
2025-12-03 15:57:06 INFO: 60.64 51.27 55.08
|
| 622 |
-
2025-12-03 15:57:06 INFO: step 3600: train_loss = 4.020496, dev_score = 0.6064
|
| 623 |
-
2025-12-03 15:57:06 INFO: Training ended with 3600 steps.
|
| 624 |
-
2025-12-03 15:57:06 INFO: Best dev F1 = 62.87, at iteration = 2500
|
| 625 |
-
2025-12-03 15:57:07 INFO: Running dev depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.dev.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
|
| 626 |
-
2025-12-03 15:57:07 INFO: Running parser in predict mode
|
| 627 |
-
2025-12-03 15:57:07 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 628 |
-
2025-12-03 15:57:09 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 629 |
-
2025-12-03 15:57:09 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 630 |
-
2025-12-03 15:57:09 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 631 |
-
2025-12-03 15:57:09 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 632 |
-
2025-12-03 15:57:10 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 633 |
-
2025-12-03 15:57:10 INFO: Loading data with batch size 32...
|
| 634 |
-
2025-12-03 15:57:10 DEBUG: 9 batches created.
|
| 635 |
-
2025-12-03 15:57:10 INFO: F1 scores for each dependency:
|
| 636 |
-
Note that unlabeled attachment errors hurt the labeled attachment scores
|
| 637 |
-
acl: p 0.0000 r 0.0000 f1 0.0000 (3 actual)
|
| 638 |
-
acl:relcl: p 0.2222 r 0.2857 f1 0.2500 (7 actual)
|
| 639 |
-
advcl: p 0.1250 r 0.2000 f1 0.1538 (5 actual)
|
| 640 |
-
advmod: p 0.5357 r 0.6000 f1 0.5660 (25 actual)
|
| 641 |
-
amod: p 0.9231 r 0.7742 f1 0.8421 (31 actual)
|
| 642 |
-
appos: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 643 |
-
aux: p 0.8182 r 0.8182 f1 0.8182 (11 actual)
|
| 644 |
-
case: p 0.9623 r 0.9107 f1 0.9358 (56 actual)
|
| 645 |
-
cc: p 0.6429 r 0.6923 f1 0.6667 (13 actual)
|
| 646 |
-
ccomp: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 647 |
-
conj: p 0.2667 r 0.3333 f1 0.2963 (12 actual)
|
| 648 |
-
cop: p 0.5000 r 0.3333 f1 0.4000 (3 actual)
|
| 649 |
-
csubj: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 650 |
-
det: p 0.8696 r 0.9091 f1 0.8889 (22 actual)
|
| 651 |
-
expl: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 652 |
-
iobj: p 0.5000 r 0.5000 f1 0.5000 (2 actual)
|
| 653 |
-
mark: p 0.4167 r 0.4167 f1 0.4167 (12 actual)
|
| 654 |
-
nmod: p 0.4000 r 0.5333 f1 0.4571 (15 actual)
|
| 655 |
-
nmod:poss: p 1.0000 r 0.8947 f1 0.9444 (19 actual)
|
| 656 |
-
nsubj: p 0.5417 r 0.7647 f1 0.6341 (17 actual)
|
| 657 |
-
nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
|
| 658 |
-
obj: p 0.6071 r 0.7727 f1 0.6800 (22 actual)
|
| 659 |
-
obl: p 0.6154 r 0.5854 f1 0.6000 (41 actual)
|
| 660 |
-
obl:agent: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 661 |
-
orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 662 |
-
parataxis: p 0.1111 r 0.3333 f1 0.1667 (3 actual)
|
| 663 |
-
punct: p 0.4231 r 0.4231 f1 0.4231 (52 actual)
|
| 664 |
-
root: p 0.4444 r 0.4444 f1 0.4444 (9 actual)
|
| 665 |
-
xcomp: p 0.5000 r 0.1250 f1 0.2000 (8 actual)
|
| 666 |
-
2025-12-03 15:57:10 INFO: LAS MLAS BLEX
|
| 667 |
-
2025-12-03 15:57:10 INFO: 62.87 54.66 58.05
|
| 668 |
-
2025-12-03 15:57:10 INFO: Parser score:
|
| 669 |
-
2025-12-03 15:57:10 INFO: sv_diachronic 62.87
|
| 670 |
-
2025-12-03 15:57:10 INFO: Finished running dev set on
|
| 671 |
-
UD_Swedish-diachronic
|
| 672 |
-
UAS LAS CLAS MLAS BLEX
|
| 673 |
-
68.56 62.87 58.05 54.66 58.05
|
| 674 |
-
2025-12-03 15:57:10 INFO: Running test depparse for UD_Swedish-diachronic with args ['--wordvec_dir', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain', '--eval_file', '/mimer/NOBACKUP/groups/dionysus/cleland/stanza-digphil/data/depparse/sv_diachronic.test.in.conllu', '--lang', 'sv', '--shorthand', 'sv_diachronic', '--mode', 'predict', '--charlm', '--charlm_shorthand', 'sv_conll17', '--charlm_forward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt', '--charlm_backward_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt', '--wordvec_pretrain_file', '/cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt', '--batch_size', '32', '--dropout', '0.33']
|
| 675 |
-
2025-12-03 15:57:10 INFO: Running parser in predict mode
|
| 676 |
-
2025-12-03 15:57:10 INFO: Loading model from: saved_models/depparse/sv_diachronic_charlm_parser.pt
|
| 677 |
-
2025-12-03 15:57:12 DEBUG: Loaded pretrain from /cephyr/users/cleland/Alvis/stanza_resources/sv/pretrain/conll17.pt
|
| 678 |
-
2025-12-03 15:57:12 DEBUG: Depparse model loading charmodels: /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt and /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 679 |
-
2025-12-03 15:57:12 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/forward_charlm/conll17.pt
|
| 680 |
-
2025-12-03 15:57:12 DEBUG: Loading charlm from /cephyr/users/cleland/Alvis/stanza_resources/sv/backward_charlm/conll17.pt
|
| 681 |
-
2025-12-03 15:57:12 DEBUG: Building Adam with lr=0.003000, betas=(0.9, 0.95), eps=0.000001
|
| 682 |
-
2025-12-03 15:57:12 INFO: Loading data with batch size 32...
|
| 683 |
-
2025-12-03 15:57:12 DEBUG: 93 batches created.
|
| 684 |
-
2025-12-03 15:57:17 INFO: F1 scores for each dependency:
|
| 685 |
-
Note that unlabeled attachment errors hurt the labeled attachment scores
|
| 686 |
-
acl: p 0.2778 r 0.1562 f1 0.2000 (32 actual)
|
| 687 |
-
acl:cleft: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 688 |
-
acl:relcl: p 0.2075 r 0.1467 f1 0.1719 (75 actual)
|
| 689 |
-
advcl: p 0.0842 r 0.1333 f1 0.1032 (60 actual)
|
| 690 |
-
advcl:relcl: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 691 |
-
advmod: p 0.5520 r 0.5746 f1 0.5631 (268 actual)
|
| 692 |
-
amod: p 0.8596 r 0.8522 f1 0.8559 (230 actual)
|
| 693 |
-
appos: p 0.0000 r 0.0000 f1 0.0000 (13 actual)
|
| 694 |
-
aux: p 0.8734 r 0.8214 f1 0.8466 (84 actual)
|
| 695 |
-
aux:pass: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 696 |
-
case: p 0.8481 r 0.8231 f1 0.8354 (373 actual)
|
| 697 |
-
cc: p 0.5833 r 0.5871 f1 0.5852 (155 actual)
|
| 698 |
-
ccomp: p 0.0000 r 0.0000 f1 0.0000 (35 actual)
|
| 699 |
-
compound:prt: p 0.7727 r 0.8095 f1 0.7907 (21 actual)
|
| 700 |
-
conj: p 0.2397 r 0.2215 f1 0.2303 (158 actual)
|
| 701 |
-
cop: p 0.7714 r 0.5870 f1 0.6667 (46 actual)
|
| 702 |
-
csubj: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 703 |
-
dep: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 704 |
-
det: p 0.8960 r 0.8702 f1 0.8829 (208 actual)
|
| 705 |
-
discourse: p 0.0000 r 0.0000 f1 0.0000 (7 actual)
|
| 706 |
-
dislocated: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 707 |
-
expl: p 0.0000 r 0.0000 f1 0.0000 (11 actual)
|
| 708 |
-
expl:pv: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 709 |
-
fixed: p 0.0000 r 0.0000 f1 0.0000 (8 actual)
|
| 710 |
-
flat: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 711 |
-
flat:name: p 0.0000 r 0.0000 f1 0.0000 (12 actual)
|
| 712 |
-
goeswith: p 0.0000 r 0.0000 f1 0.0000 (2 actual)
|
| 713 |
-
iobj: p 0.3333 r 0.0714 f1 0.1176 (14 actual)
|
| 714 |
-
mark: p 0.6757 r 0.6536 f1 0.6645 (153 actual)
|
| 715 |
-
nmod: p 0.3798 r 0.4804 f1 0.4242 (102 actual)
|
| 716 |
-
nmod:poss: p 0.8707 r 0.9014 f1 0.8858 (142 actual)
|
| 717 |
-
nsubj: p 0.5673 r 0.6321 f1 0.5980 (280 actual)
|
| 718 |
-
nsubj:pass: p 0.0000 r 0.0000 f1 0.0000 (25 actual)
|
| 719 |
-
nummod: p 0.7500 r 0.9000 f1 0.8182 (10 actual)
|
| 720 |
-
obj: p 0.5628 r 0.7104 f1 0.6280 (183 actual)
|
| 721 |
-
obl: p 0.4857 r 0.5504 f1 0.5160 (278 actual)
|
| 722 |
-
obl:agent: p 0.0000 r 0.0000 f1 0.0000 (4 actual)
|
| 723 |
-
orphan: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 724 |
-
parataxis: p 0.0333 r 0.1667 f1 0.0556 (18 actual)
|
| 725 |
-
punct: p 0.4262 r 0.4282 f1 0.4272 (425 actual)
|
| 726 |
-
reparandum: p 0.0000 r 0.0000 f1 0.0000 (1 actual)
|
| 727 |
-
root: p 0.5354 r 0.5354 f1 0.5354 (99 actual)
|
| 728 |
-
vocative: p 0.0000 r 0.0000 f1 0.0000 (5 actual)
|
| 729 |
-
xcomp: p 0.5750 r 0.3067 f1 0.4000 (75 actual)
|
| 730 |
-
2025-12-03 15:57:17 INFO: LAS MLAS BLEX
|
| 731 |
-
2025-12-03 15:57:17 INFO: 58.54 50.12 53.03
|
| 732 |
-
2025-12-03 15:57:17 INFO: Parser score:
|
| 733 |
-
2025-12-03 15:57:17 INFO: sv_diachronic 58.54
|
| 734 |
-
2025-12-03 15:57:17 INFO: Finished running test set on
|
| 735 |
-
UD_Swedish-diachronic
|
| 736 |
-
UAS LAS CLAS MLAS BLEX
|
| 737 |
-
67.16 58.54 53.03 50.12 53.03
|
| 738 |
-
DONE.
|
| 739 |
-
Full log saved to: logs/log_conll17.pt_sv_diachron_de_lit_20251203_154509.txt
|
| 740 |
-
Symlink updated: logs/latest.txt → log_conll17.pt_sv_diachron_de_lit_20251203_154509.txt
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|