pszemraj commited on
Commit
b970478
·
verified ·
0 Parent(s):

Super-squash branch 'main' using huggingface_hub

Browse files
.gitattributes ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: deepmind/language-perceiver
4
+ tags:
5
+ - book
6
+ - genre
7
+ - book title
8
+ metrics:
9
+ - f1
10
+ widget:
11
+ - text: The Quantum Chip
12
+ example_title: Science Fiction & Fantasy
13
+ - text: One Dollar's Journey
14
+ example_title: Business & Finance
15
+ - text: Timmy The Talking Tree
16
+ example_title: idk fiction
17
+ - text: The Cursed Canvas
18
+ example_title: Arts & Design
19
+ - text: Hoops and Hegel
20
+ example_title: Philosophy & Religion
21
+ - text: Overview of Streams in North Dakota
22
+ example_title: Nature
23
+ - text: Advanced Topology
24
+ example_title: Non-fiction/Math
25
+ - text: Cooking Up Love
26
+ example_title: Food & Cooking
27
+ - text: Dr. Doolittle's Extraplanatary Commute
28
+ example_title: Science & Technology
29
+ pipeline_tag: text-classification
30
+ ---
31
+
32
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
33
+ should probably proofread and complete it, then remove this comment. -->
34
+
35
+ # language-perceiver for title-genre classification
36
+
37
+ This model is a fine-tuned version of [deepmind/language-perceiver](https://huggingface.co/deepmind/language-perceiver) on an unknown dataset.
38
+ It achieves the following results on the evaluation set:
39
+ - Loss: 0.2832
40
+ - F1: 0.5108
41
+
42
+ ## Model description
43
+
44
+ This classifies one or more **genre** labels in a **multi-label** setting for a given book **title**.
45
+
46
+ The 'standard' way of interpreting the predictions is that the predicted labels for a given example are **only the ones with a greater than 50% probability.**
47
+
48
+ ## Training procedure
49
+
50
+ ### Training hyperparameters
51
+
52
+ The following hyperparameters were used during training:
53
+ - learning_rate: 2e-05
54
+ - train_batch_size: 32
55
+ - eval_batch_size: 32
56
+ - seed: 42
57
+ - gradient_accumulation_steps: 4
58
+ - total_train_batch_size: 128
59
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
+ - lr_scheduler_type: linear
61
+ - num_epochs: 8.0
62
+
63
+ ### Training results
64
+
65
+ | Training Loss | Epoch | Step | Validation Loss | F1 |
66
+ |:-------------:|:-----:|:----:|:---------------:|:------:|
67
+ | 0.3059 | 1.0 | 62 | 0.2893 | 0.3263 |
68
+ | 0.2879 | 2.0 | 124 | 0.2795 | 0.4290 |
69
+ | 0.2729 | 3.0 | 186 | 0.2730 | 0.4356 |
70
+ | 0.2606 | 4.0 | 248 | 0.2722 | 0.4590 |
71
+ | 0.2433 | 5.0 | 310 | 0.2747 | 0.4775 |
72
+ | 0.227 | 6.0 | 372 | 0.2777 | 0.4976 |
73
+ | 0.207 | 7.0 | 434 | 0.2814 | 0.5088 |
74
+ | 0.1969 | 8.0 | 496 | 0.2832 | 0.5108 |
75
+
76
+
77
+ ### Framework versions
78
+
79
+ - Transformers 4.33.3
80
+ - Pytorch 2.2.0.dev20231001+cu121
81
+ - Datasets 2.14.5
82
+ - Tokenizers 0.13.3
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
all_results.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.0,
3
+ "eval_f1": 0.5107802874743327,
4
+ "eval_loss": 0.283179372549057,
5
+ "eval_runtime": 3.5813,
6
+ "eval_samples": 989,
7
+ "eval_samples_per_second": 276.158,
8
+ "eval_steps_per_second": 8.656,
9
+ "train_loss": 0.2570930659290283,
10
+ "train_runtime": 822.741,
11
+ "train_samples": 7914,
12
+ "train_samples_per_second": 76.953,
13
+ "train_steps_per_second": 0.603
14
+ }
config.json ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_label_trainable_num_channels": 1024,
3
+ "_name_or_path": "deepmind/language-perceiver",
4
+ "architectures": [
5
+ "PerceiverForSequenceClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.1,
8
+ "audio_samples_per_frame": 1920,
9
+ "cross_attention_shape_for_attention": "kv",
10
+ "cross_attention_widening_factor": 1,
11
+ "d_latents": 1280,
12
+ "d_model": 768,
13
+ "finetuning_task": "text-classification",
14
+ "hidden_act": "gelu",
15
+ "id2label": {
16
+ "0": "History & Politics",
17
+ "1": "Health & Medicine",
18
+ "2": "Mystery & Thriller",
19
+ "3": "Arts & Design",
20
+ "4": "Self-Help & Wellness",
21
+ "5": "Sports & Recreation",
22
+ "6": "Non-Fiction",
23
+ "7": "Science Fiction & Fantasy",
24
+ "8": "Countries & Geography",
25
+ "9": "Other",
26
+ "10": "Nature & Environment",
27
+ "11": "Business & Finance",
28
+ "12": "Romance",
29
+ "13": "Philosophy & Religion",
30
+ "14": "Literature & Fiction",
31
+ "15": "Science & Technology",
32
+ "16": "Children & Young Adult",
33
+ "17": "Food & Cooking"
34
+ },
35
+ "image_size": 56,
36
+ "initializer_range": 0.02,
37
+ "label2id": {
38
+ "Arts & Design": 3,
39
+ "Business & Finance": 11,
40
+ "Children & Young Adult": 16,
41
+ "Countries & Geography": 8,
42
+ "Food & Cooking": 17,
43
+ "Health & Medicine": 1,
44
+ "History & Politics": 0,
45
+ "Literature & Fiction": 14,
46
+ "Mystery & Thriller": 2,
47
+ "Nature & Environment": 10,
48
+ "Non-Fiction": 6,
49
+ "Other": 9,
50
+ "Philosophy & Religion": 13,
51
+ "Romance": 12,
52
+ "Science & Technology": 15,
53
+ "Science Fiction & Fantasy": 7,
54
+ "Self-Help & Wellness": 4,
55
+ "Sports & Recreation": 5
56
+ },
57
+ "layer_norm_eps": 1e-12,
58
+ "max_position_embeddings": 2048,
59
+ "model_type": "perceiver",
60
+ "num_blocks": 1,
61
+ "num_cross_attention_heads": 8,
62
+ "num_frames": 16,
63
+ "num_latents": 256,
64
+ "num_self_attends_per_block": 26,
65
+ "num_self_attention_heads": 8,
66
+ "output_num_channels": 512,
67
+ "output_shape": [
68
+ 1,
69
+ 16,
70
+ 224,
71
+ 224
72
+ ],
73
+ "problem_type": "multi_label_classification",
74
+ "qk_channels": 256,
75
+ "samples_per_patch": 16,
76
+ "self_attention_widening_factor": 1,
77
+ "torch_dtype": "float32",
78
+ "train_size": [
79
+ 368,
80
+ 496
81
+ ],
82
+ "transformers_version": "4.33.3",
83
+ "use_query_residual": true,
84
+ "v_channels": 1280,
85
+ "vocab_size": 262
86
+ }
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.0,
3
+ "eval_f1": 0.5107802874743327,
4
+ "eval_loss": 0.283179372549057,
5
+ "eval_runtime": 3.5813,
6
+ "eval_samples": 989,
7
+ "eval_samples_per_second": 276.158,
8
+ "eval_steps_per_second": 8.656
9
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9312728a4bb10a39a19e2e83c6166b9dc3ec22deee8d403c62ad974aaa094294
3
+ size 824536032
special_tokens_map.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "[BOS]",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "[CLS]",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "[EOS]",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "[MASK]",
25
+ "lstrip": false,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "[PAD]",
32
+ "lstrip": false,
33
+ "normalized": true,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "[SEP]",
39
+ "lstrip": false,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ }
44
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "__type": "AddedToken",
4
+ "content": "[BOS]",
5
+ "lstrip": false,
6
+ "normalized": true,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ "clean_up_tokenization_spaces": true,
11
+ "cls_token": {
12
+ "__type": "AddedToken",
13
+ "content": "[CLS]",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false
18
+ },
19
+ "eos_token": {
20
+ "__type": "AddedToken",
21
+ "content": "[EOS]",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false
26
+ },
27
+ "mask_token": {
28
+ "__type": "AddedToken",
29
+ "content": "[MASK]",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false
34
+ },
35
+ "model_max_length": 2048,
36
+ "pad_token": {
37
+ "__type": "AddedToken",
38
+ "content": "[PAD]",
39
+ "lstrip": false,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "sep_token": {
45
+ "__type": "AddedToken",
46
+ "content": "[SEP]",
47
+ "lstrip": false,
48
+ "normalized": true,
49
+ "rstrip": false,
50
+ "single_word": false
51
+ },
52
+ "tokenizer_class": "PerceiverTokenizer"
53
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.0,
3
+ "train_loss": 0.2570930659290283,
4
+ "train_runtime": 822.741,
5
+ "train_samples": 7914,
6
+ "train_samples_per_second": 76.953,
7
+ "train_steps_per_second": 0.603
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,394 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 8.0,
5
+ "eval_steps": 500,
6
+ "global_step": 496,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.16,
13
+ "learning_rate": 1.9596774193548388e-05,
14
+ "loss": 0.4886,
15
+ "step": 10
16
+ },
17
+ {
18
+ "epoch": 0.32,
19
+ "learning_rate": 1.9193548387096777e-05,
20
+ "loss": 0.3236,
21
+ "step": 20
22
+ },
23
+ {
24
+ "epoch": 0.48,
25
+ "learning_rate": 1.8790322580645163e-05,
26
+ "loss": 0.3137,
27
+ "step": 30
28
+ },
29
+ {
30
+ "epoch": 0.65,
31
+ "learning_rate": 1.838709677419355e-05,
32
+ "loss": 0.3073,
33
+ "step": 40
34
+ },
35
+ {
36
+ "epoch": 0.81,
37
+ "learning_rate": 1.7983870967741936e-05,
38
+ "loss": 0.3068,
39
+ "step": 50
40
+ },
41
+ {
42
+ "epoch": 0.97,
43
+ "learning_rate": 1.7580645161290325e-05,
44
+ "loss": 0.3059,
45
+ "step": 60
46
+ },
47
+ {
48
+ "epoch": 1.0,
49
+ "eval_f1": 0.32627646326276466,
50
+ "eval_loss": 0.2893124222755432,
51
+ "eval_runtime": 3.6709,
52
+ "eval_samples_per_second": 269.418,
53
+ "eval_steps_per_second": 8.445,
54
+ "step": 62
55
+ },
56
+ {
57
+ "epoch": 1.13,
58
+ "learning_rate": 1.717741935483871e-05,
59
+ "loss": 0.3038,
60
+ "step": 70
61
+ },
62
+ {
63
+ "epoch": 1.29,
64
+ "learning_rate": 1.6774193548387098e-05,
65
+ "loss": 0.2959,
66
+ "step": 80
67
+ },
68
+ {
69
+ "epoch": 1.45,
70
+ "learning_rate": 1.6370967741935487e-05,
71
+ "loss": 0.2953,
72
+ "step": 90
73
+ },
74
+ {
75
+ "epoch": 1.61,
76
+ "learning_rate": 1.596774193548387e-05,
77
+ "loss": 0.2908,
78
+ "step": 100
79
+ },
80
+ {
81
+ "epoch": 1.77,
82
+ "learning_rate": 1.556451612903226e-05,
83
+ "loss": 0.2902,
84
+ "step": 110
85
+ },
86
+ {
87
+ "epoch": 1.94,
88
+ "learning_rate": 1.5161290322580646e-05,
89
+ "loss": 0.2879,
90
+ "step": 120
91
+ },
92
+ {
93
+ "epoch": 2.0,
94
+ "eval_f1": 0.429018492176387,
95
+ "eval_loss": 0.2794504165649414,
96
+ "eval_runtime": 3.6769,
97
+ "eval_samples_per_second": 268.977,
98
+ "eval_steps_per_second": 8.431,
99
+ "step": 124
100
+ },
101
+ {
102
+ "epoch": 2.1,
103
+ "learning_rate": 1.4758064516129033e-05,
104
+ "loss": 0.2774,
105
+ "step": 130
106
+ },
107
+ {
108
+ "epoch": 2.26,
109
+ "learning_rate": 1.4354838709677421e-05,
110
+ "loss": 0.2762,
111
+ "step": 140
112
+ },
113
+ {
114
+ "epoch": 2.42,
115
+ "learning_rate": 1.3951612903225809e-05,
116
+ "loss": 0.2811,
117
+ "step": 150
118
+ },
119
+ {
120
+ "epoch": 2.58,
121
+ "learning_rate": 1.3548387096774194e-05,
122
+ "loss": 0.2734,
123
+ "step": 160
124
+ },
125
+ {
126
+ "epoch": 2.74,
127
+ "learning_rate": 1.3145161290322581e-05,
128
+ "loss": 0.279,
129
+ "step": 170
130
+ },
131
+ {
132
+ "epoch": 2.9,
133
+ "learning_rate": 1.274193548387097e-05,
134
+ "loss": 0.2729,
135
+ "step": 180
136
+ },
137
+ {
138
+ "epoch": 3.0,
139
+ "eval_f1": 0.4356266057664859,
140
+ "eval_loss": 0.27300506830215454,
141
+ "eval_runtime": 3.6734,
142
+ "eval_samples_per_second": 269.233,
143
+ "eval_steps_per_second": 8.439,
144
+ "step": 186
145
+ },
146
+ {
147
+ "epoch": 3.06,
148
+ "learning_rate": 1.2338709677419355e-05,
149
+ "loss": 0.2722,
150
+ "step": 190
151
+ },
152
+ {
153
+ "epoch": 3.23,
154
+ "learning_rate": 1.1935483870967743e-05,
155
+ "loss": 0.2605,
156
+ "step": 200
157
+ },
158
+ {
159
+ "epoch": 3.39,
160
+ "learning_rate": 1.1532258064516131e-05,
161
+ "loss": 0.2564,
162
+ "step": 210
163
+ },
164
+ {
165
+ "epoch": 3.55,
166
+ "learning_rate": 1.1129032258064516e-05,
167
+ "loss": 0.264,
168
+ "step": 220
169
+ },
170
+ {
171
+ "epoch": 3.71,
172
+ "learning_rate": 1.0725806451612903e-05,
173
+ "loss": 0.2627,
174
+ "step": 230
175
+ },
176
+ {
177
+ "epoch": 3.87,
178
+ "learning_rate": 1.0322580645161291e-05,
179
+ "loss": 0.2606,
180
+ "step": 240
181
+ },
182
+ {
183
+ "epoch": 4.0,
184
+ "eval_f1": 0.458980044345898,
185
+ "eval_loss": 0.272247850894928,
186
+ "eval_runtime": 3.6724,
187
+ "eval_samples_per_second": 269.307,
188
+ "eval_steps_per_second": 8.441,
189
+ "step": 248
190
+ },
191
+ {
192
+ "epoch": 4.03,
193
+ "learning_rate": 9.919354838709679e-06,
194
+ "loss": 0.2523,
195
+ "step": 250
196
+ },
197
+ {
198
+ "epoch": 4.19,
199
+ "learning_rate": 9.516129032258065e-06,
200
+ "loss": 0.2437,
201
+ "step": 260
202
+ },
203
+ {
204
+ "epoch": 4.35,
205
+ "learning_rate": 9.112903225806451e-06,
206
+ "loss": 0.2451,
207
+ "step": 270
208
+ },
209
+ {
210
+ "epoch": 4.52,
211
+ "learning_rate": 8.70967741935484e-06,
212
+ "loss": 0.2514,
213
+ "step": 280
214
+ },
215
+ {
216
+ "epoch": 4.68,
217
+ "learning_rate": 8.306451612903227e-06,
218
+ "loss": 0.2439,
219
+ "step": 290
220
+ },
221
+ {
222
+ "epoch": 4.84,
223
+ "learning_rate": 7.903225806451613e-06,
224
+ "loss": 0.2378,
225
+ "step": 300
226
+ },
227
+ {
228
+ "epoch": 5.0,
229
+ "learning_rate": 7.500000000000001e-06,
230
+ "loss": 0.2433,
231
+ "step": 310
232
+ },
233
+ {
234
+ "epoch": 5.0,
235
+ "eval_f1": 0.4775401069518716,
236
+ "eval_loss": 0.2747083902359009,
237
+ "eval_runtime": 3.674,
238
+ "eval_samples_per_second": 269.186,
239
+ "eval_steps_per_second": 8.438,
240
+ "step": 310
241
+ },
242
+ {
243
+ "epoch": 5.16,
244
+ "learning_rate": 7.096774193548388e-06,
245
+ "loss": 0.2302,
246
+ "step": 320
247
+ },
248
+ {
249
+ "epoch": 5.32,
250
+ "learning_rate": 6.693548387096774e-06,
251
+ "loss": 0.2292,
252
+ "step": 330
253
+ },
254
+ {
255
+ "epoch": 5.48,
256
+ "learning_rate": 6.290322580645162e-06,
257
+ "loss": 0.223,
258
+ "step": 340
259
+ },
260
+ {
261
+ "epoch": 5.65,
262
+ "learning_rate": 5.887096774193549e-06,
263
+ "loss": 0.2281,
264
+ "step": 350
265
+ },
266
+ {
267
+ "epoch": 5.81,
268
+ "learning_rate": 5.483870967741935e-06,
269
+ "loss": 0.2301,
270
+ "step": 360
271
+ },
272
+ {
273
+ "epoch": 5.97,
274
+ "learning_rate": 5.080645161290323e-06,
275
+ "loss": 0.227,
276
+ "step": 370
277
+ },
278
+ {
279
+ "epoch": 6.0,
280
+ "eval_f1": 0.49764027267960154,
281
+ "eval_loss": 0.2776886522769928,
282
+ "eval_runtime": 3.6732,
283
+ "eval_samples_per_second": 269.25,
284
+ "eval_steps_per_second": 8.44,
285
+ "step": 372
286
+ },
287
+ {
288
+ "epoch": 6.13,
289
+ "learning_rate": 4.67741935483871e-06,
290
+ "loss": 0.2188,
291
+ "step": 380
292
+ },
293
+ {
294
+ "epoch": 6.29,
295
+ "learning_rate": 4.274193548387097e-06,
296
+ "loss": 0.2195,
297
+ "step": 390
298
+ },
299
+ {
300
+ "epoch": 6.45,
301
+ "learning_rate": 3.870967741935484e-06,
302
+ "loss": 0.2123,
303
+ "step": 400
304
+ },
305
+ {
306
+ "epoch": 6.61,
307
+ "learning_rate": 3.4677419354838714e-06,
308
+ "loss": 0.2121,
309
+ "step": 410
310
+ },
311
+ {
312
+ "epoch": 6.77,
313
+ "learning_rate": 3.0645161290322584e-06,
314
+ "loss": 0.2136,
315
+ "step": 420
316
+ },
317
+ {
318
+ "epoch": 6.94,
319
+ "learning_rate": 2.6612903225806454e-06,
320
+ "loss": 0.207,
321
+ "step": 430
322
+ },
323
+ {
324
+ "epoch": 7.0,
325
+ "eval_f1": 0.5087719298245615,
326
+ "eval_loss": 0.28140273690223694,
327
+ "eval_runtime": 3.6742,
328
+ "eval_samples_per_second": 269.173,
329
+ "eval_steps_per_second": 8.437,
330
+ "step": 434
331
+ },
332
+ {
333
+ "epoch": 7.1,
334
+ "learning_rate": 2.2580645161290324e-06,
335
+ "loss": 0.2085,
336
+ "step": 440
337
+ },
338
+ {
339
+ "epoch": 7.26,
340
+ "learning_rate": 1.8548387096774196e-06,
341
+ "loss": 0.2071,
342
+ "step": 450
343
+ },
344
+ {
345
+ "epoch": 7.42,
346
+ "learning_rate": 1.4516129032258066e-06,
347
+ "loss": 0.2027,
348
+ "step": 460
349
+ },
350
+ {
351
+ "epoch": 7.58,
352
+ "learning_rate": 1.0483870967741936e-06,
353
+ "loss": 0.2017,
354
+ "step": 470
355
+ },
356
+ {
357
+ "epoch": 7.74,
358
+ "learning_rate": 6.451612903225807e-07,
359
+ "loss": 0.2017,
360
+ "step": 480
361
+ },
362
+ {
363
+ "epoch": 7.9,
364
+ "learning_rate": 2.4193548387096775e-07,
365
+ "loss": 0.1969,
366
+ "step": 490
367
+ },
368
+ {
369
+ "epoch": 8.0,
370
+ "eval_f1": 0.5107802874743327,
371
+ "eval_loss": 0.283179372549057,
372
+ "eval_runtime": 3.6742,
373
+ "eval_samples_per_second": 269.174,
374
+ "eval_steps_per_second": 8.437,
375
+ "step": 496
376
+ },
377
+ {
378
+ "epoch": 8.0,
379
+ "step": 496,
380
+ "total_flos": 0.0,
381
+ "train_loss": 0.2570930659290283,
382
+ "train_runtime": 822.741,
383
+ "train_samples_per_second": 76.953,
384
+ "train_steps_per_second": 0.603
385
+ }
386
+ ],
387
+ "logging_steps": 10,
388
+ "max_steps": 496,
389
+ "num_train_epochs": 8,
390
+ "save_steps": 500,
391
+ "total_flos": 0.0,
392
+ "trial_name": null,
393
+ "trial_params": null
394
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05febb53989d3609718ff89cfe07adccc4fe13af016c2a02e54484aa6e188756
3
+ size 4600
vocab.txt ADDED
The diff for this file is too large to render. See raw diff