anonymous-11 commited on
Commit
2d03f38
·
verified ·
1 Parent(s): 4b0ff82

Upload 12 files

Browse files
Files changed (12) hide show
  1. config.json +54 -0
  2. dict.txt +0 -0
  3. merges.txt +0 -0
  4. model.safetensors +3 -0
  5. optimizer.pt +3 -0
  6. readme.md +3 -0
  7. rng_state.pth +3 -0
  8. scheduler.pt +3 -0
  9. tokenizer.json +0 -0
  10. trainer_state.json +41 -0
  11. training_args.bin +3 -0
  12. vocab.json +0 -0
config.json ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "../model/iter2/roberta-base/",
3
+ "architectures": [
4
+ "RobertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "id2label": {
14
+ "0": "LABEL_0",
15
+ "1": "LABEL_1",
16
+ "2": "LABEL_2",
17
+ "3": "LABEL_3",
18
+ "4": "LABEL_4",
19
+ "5": "LABEL_5",
20
+ "6": "LABEL_6",
21
+ "7": "LABEL_7",
22
+ "8": "LABEL_8",
23
+ "9": "LABEL_9",
24
+ "10": "LABEL_10"
25
+ },
26
+ "initializer_range": 0.02,
27
+ "intermediate_size": 3072,
28
+ "label2id": {
29
+ "LABEL_0": 0,
30
+ "LABEL_1": 1,
31
+ "LABEL_10": 10,
32
+ "LABEL_2": 2,
33
+ "LABEL_3": 3,
34
+ "LABEL_4": 4,
35
+ "LABEL_5": 5,
36
+ "LABEL_6": 6,
37
+ "LABEL_7": 7,
38
+ "LABEL_8": 8,
39
+ "LABEL_9": 9
40
+ },
41
+ "layer_norm_eps": 1e-05,
42
+ "max_position_embeddings": 514,
43
+ "model_type": "roberta",
44
+ "num_attention_heads": 12,
45
+ "num_hidden_layers": 12,
46
+ "pad_token_id": 1,
47
+ "position_embedding_type": "absolute",
48
+ "problem_type": "multi_label_classification",
49
+ "torch_dtype": "float32",
50
+ "transformers_version": "4.40.1",
51
+ "type_vocab_size": 1,
52
+ "use_cache": true,
53
+ "vocab_size": 50265
54
+ }
dict.txt ADDED
The diff for this file is too large to render. See raw diff
 
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b550c872366848a71104d4ffcb9ce38c4b320178a1b07a695d59da7f9adc1978
3
+ size 498640508
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e561880c5a3770db4668b5180db4562be7595aa6290f0c2d5c284c8978412ba
3
+ size 997401227
readme.md ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ Base model for error identification.
2
+
3
+ Please cite our paper when using it.
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc8fc2dac8c511a2d1d138defc5c36988a11a1d9ae89d102f39752f91ea9cd25
3
+ size 14645
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccd870779a2d1f19d526db5eefef955e95050292e6e80cc53c5bd74c19f50dd6
3
+ size 1465
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
trainer_state.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.7021276595744681,
3
+ "best_model_checkpoint": "../model/iter3/roberta-base/checkpoint-216",
4
+ "epoch": 1.0,
5
+ "eval_steps": 500,
6
+ "global_step": 216,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "grad_norm": 4.311070919036865,
14
+ "learning_rate": 1.9333333333333333e-05,
15
+ "loss": 0.1212,
16
+ "step": 216
17
+ },
18
+ {
19
+ "epoch": 1.0,
20
+ "eval_accuracy": 0.6692307692307692,
21
+ "eval_loss": 0.1722601354122162,
22
+ "eval_macro_f1": 0.5433354082106209,
23
+ "eval_micro_f1": 0.7021276595744681,
24
+ "eval_precision": 0.7279411764705882,
25
+ "eval_recall": 0.678082191780822,
26
+ "eval_runtime": 0.9143,
27
+ "eval_samples_per_second": 142.181,
28
+ "eval_steps_per_second": 9.843,
29
+ "step": 216
30
+ }
31
+ ],
32
+ "logging_steps": 500,
33
+ "max_steps": 6480,
34
+ "num_input_tokens_seen": 0,
35
+ "num_train_epochs": 30,
36
+ "save_steps": 500,
37
+ "total_flos": 907806492518400.0,
38
+ "train_batch_size": 16,
39
+ "trial_name": null,
40
+ "trial_params": null
41
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:239272dff4db0695404f8dfa5a4260d488ece61a2c2be8f5469fd7de5953cfc2
3
+ size 5393
vocab.json ADDED
The diff for this file is too large to render. See raw diff