loris3 commited on
Commit
23f7abd
·
verified ·
1 Parent(s): 2b2125c

Upload folder using huggingface_hub

Browse files
checkpoints/checkpoint-147/config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "RobertaForMaskedLM"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "classifier_dropout": null,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 514,
16
+ "model_type": "roberta",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 1,
20
+ "position_embedding_type": "absolute",
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.47.0",
23
+ "type_vocab_size": 1,
24
+ "use_cache": true,
25
+ "vocab_size": 52000
26
+ }
checkpoints/checkpoint-147/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42b44bba0dd9d98dc35e95836cec6a823774d4743286970307e7e3f532e5d9aa
3
+ size 504150808
checkpoints/checkpoint-147/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a17e7ec0ff0fde3b5e60a201243672b6588686100536f8184bdfcea23e120cda
3
+ size 1008422138
checkpoints/checkpoint-147/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d95eee3380b64cac3a27079066a6af6b9eaa4209c7360f42c6ea6a7cb1475cf7
3
+ size 14244
checkpoints/checkpoint-147/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0f6c34a89a716ab5cbf514de19dc73c14b24dd3c036169150147ee856073460
3
+ size 1064
checkpoints/checkpoint-147/trainer_state.json ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 147,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_accuracy": 0.0,
14
+ "eval_loss": 10.322837829589844,
15
+ "eval_normalizer": 95717.0,
16
+ "eval_runtime": 87.7846,
17
+ "eval_samples_per_second": 608.956,
18
+ "eval_steps_per_second": 1.196,
19
+ "step": 49
20
+ },
21
+ {
22
+ "epoch": 1.020460358056266,
23
+ "grad_norm": 46.59355545043945,
24
+ "learning_rate": 2.5e-06,
25
+ "loss": 10.6797,
26
+ "step": 50
27
+ },
28
+ {
29
+ "epoch": 2.0,
30
+ "eval_accuracy": 0.0,
31
+ "eval_loss": 9.83644962310791,
32
+ "eval_normalizer": 95497.0,
33
+ "eval_runtime": 86.1922,
34
+ "eval_samples_per_second": 620.207,
35
+ "eval_steps_per_second": 1.218,
36
+ "step": 98
37
+ },
38
+ {
39
+ "epoch": 2.040920716112532,
40
+ "grad_norm": 37.12765121459961,
41
+ "learning_rate": 5e-06,
42
+ "loss": 9.9107,
43
+ "step": 100
44
+ },
45
+ {
46
+ "epoch": 3.0,
47
+ "eval_accuracy": 0.0,
48
+ "eval_loss": 9.46417236328125,
49
+ "eval_normalizer": 95100.0,
50
+ "eval_runtime": 85.1382,
51
+ "eval_samples_per_second": 627.885,
52
+ "eval_steps_per_second": 1.233,
53
+ "step": 147
54
+ }
55
+ ],
56
+ "logging_steps": 50,
57
+ "max_steps": 480,
58
+ "num_input_tokens_seen": 0,
59
+ "num_train_epochs": 10,
60
+ "save_steps": 500,
61
+ "stateful_callbacks": {
62
+ "TrainerControl": {
63
+ "args": {
64
+ "should_epoch_stop": false,
65
+ "should_evaluate": false,
66
+ "should_log": false,
67
+ "should_save": true,
68
+ "should_training_stop": false
69
+ },
70
+ "attributes": {}
71
+ }
72
+ },
73
+ "total_flos": 3.749219620879565e+16,
74
+ "train_batch_size": 128,
75
+ "trial_name": null,
76
+ "trial_params": null
77
+ }
checkpoints/checkpoint-147/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9aafeddeb70594fd32db941f95ee1d7a2daf58027ecc4c0fb9e57ba222b6f1e
3
+ size 5432