Silicon23 commited on
Commit
cc3c2f4
·
verified ·
1 Parent(s): a54308a

Training in progress, epoch 1

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. model.safetensors +1 -1
  2. run-7/checkpoint-184/model.safetensors +1 -1
  3. run-7/checkpoint-184/optimizer.pt +1 -1
  4. run-7/checkpoint-184/rng_state.pth +2 -2
  5. run-7/checkpoint-184/scheduler.pt +1 -1
  6. run-7/checkpoint-184/trainer_state.json +10 -47
  7. run-7/checkpoint-184/training_args.bin +1 -1
  8. run-8/checkpoint-184/config.json +27 -0
  9. run-8/checkpoint-184/model.safetensors +3 -0
  10. run-8/checkpoint-184/optimizer.pt +3 -0
  11. run-8/checkpoint-184/rng_state.pth +3 -0
  12. run-8/checkpoint-184/scheduler.pt +3 -0
  13. run-8/checkpoint-184/special_tokens_map.json +7 -0
  14. run-8/checkpoint-184/tokenizer.json +0 -0
  15. run-8/checkpoint-184/tokenizer_config.json +55 -0
  16. run-8/checkpoint-184/trainer_state.json +47 -0
  17. run-8/checkpoint-184/training_args.bin +3 -0
  18. run-8/checkpoint-184/vocab.txt +0 -0
  19. run-8/checkpoint-368/config.json +27 -0
  20. run-8/checkpoint-368/model.safetensors +3 -0
  21. run-8/checkpoint-368/optimizer.pt +3 -0
  22. run-8/checkpoint-368/rng_state.pth +3 -0
  23. run-8/checkpoint-368/scheduler.pt +3 -0
  24. run-8/checkpoint-368/special_tokens_map.json +7 -0
  25. run-8/checkpoint-368/tokenizer.json +0 -0
  26. run-8/checkpoint-368/tokenizer_config.json +55 -0
  27. run-8/checkpoint-368/trainer_state.json +56 -0
  28. run-8/checkpoint-368/training_args.bin +3 -0
  29. run-8/checkpoint-368/vocab.txt +0 -0
  30. run-8/checkpoint-552/config.json +27 -0
  31. run-8/checkpoint-552/model.safetensors +3 -0
  32. run-8/checkpoint-552/optimizer.pt +3 -0
  33. run-8/checkpoint-552/rng_state.pth +3 -0
  34. run-8/checkpoint-552/scheduler.pt +3 -0
  35. run-8/checkpoint-552/special_tokens_map.json +7 -0
  36. run-8/checkpoint-552/tokenizer.json +0 -0
  37. run-8/checkpoint-552/tokenizer_config.json +55 -0
  38. run-8/checkpoint-552/trainer_state.json +72 -0
  39. run-8/checkpoint-552/training_args.bin +3 -0
  40. run-8/checkpoint-552/vocab.txt +0 -0
  41. run-8/checkpoint-736/config.json +27 -0
  42. run-8/checkpoint-736/model.safetensors +3 -0
  43. run-8/checkpoint-736/optimizer.pt +3 -0
  44. run-8/checkpoint-736/rng_state.pth +3 -0
  45. run-8/checkpoint-736/scheduler.pt +3 -0
  46. run-8/checkpoint-736/special_tokens_map.json +7 -0
  47. run-8/checkpoint-736/tokenizer.json +0 -0
  48. run-8/checkpoint-736/tokenizer_config.json +55 -0
  49. run-8/checkpoint-736/trainer_state.json +81 -0
  50. run-8/checkpoint-736/training_args.bin +3 -0
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a10f8004a7cf08aaf06d4a2531659d5220ecc605c4286e7c0c9fd8662977ca4
3
  size 437958648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e18b3b55391290fa102826a1e71aba3750b61eafc8d5a853f755688f29389482
3
  size 437958648
run-7/checkpoint-184/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:770348fd8b1253de5a479d14ddd4bb1ce3bc917fb881aeb0c9e189cd214d6466
3
  size 437958648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45bb6c49c2bedd3504a185f0dce951195cf6d38bac23597772f8431210847ba9
3
  size 437958648
run-7/checkpoint-184/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7063058cd8d58576dd8cbe5348e8de13dfc195c4b9eacfc23d9a916dd1c4dc0b
3
  size 876038394
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:868db70dc279c5d8098672ca8d47ee7eb62bdee7cc99a04d925d482941ac4993
3
  size 876038394
run-7/checkpoint-184/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca8bb9bde246d02b70bda4a225a6ccdb5bed7324dacfc3ed50db070926cf9c87
3
- size 14308
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00bf828b1919b530596e9a1734505951eab0a227f4a347b876f1c643c9b2b956
3
+ size 14244
run-7/checkpoint-184/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:83f7bc4fa9d630ccd56f888d0b282720d75d1e8d643c566818be703b8e6de47b
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fb14ba6f788273c29f871dd1124460e8da363a4500daff1fd2b44e559e3ace1
3
  size 1064
run-7/checkpoint-184/trainer_state.json CHANGED
@@ -1,54 +1,17 @@
1
  {
2
- "best_metric": 0.8367096378146102,
3
- "best_model_checkpoint": "BERTForDetectingDepression-Twitter2015/run-7/checkpoint-184",
4
- "epoch": 4.0,
5
  "eval_steps": 500,
6
  "global_step": 184,
7
  "is_hyper_param_search": true,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
- "log_history": [
11
- {
12
- "epoch": 1.0,
13
- "eval_accuracy": 0.7839165131982812,
14
- "eval_loss": 0.44573765993118286,
15
- "eval_runtime": 1.5248,
16
- "eval_samples_per_second": 1068.356,
17
- "eval_steps_per_second": 66.895,
18
- "step": 46
19
- },
20
- {
21
- "epoch": 2.0,
22
- "eval_accuracy": 0.8029465930018416,
23
- "eval_loss": 0.4049816429615021,
24
- "eval_runtime": 1.5307,
25
- "eval_samples_per_second": 1064.206,
26
- "eval_steps_per_second": 66.635,
27
- "step": 92
28
- },
29
- {
30
- "epoch": 3.0,
31
- "eval_accuracy": 0.8305709023941068,
32
- "eval_loss": 0.36977845430374146,
33
- "eval_runtime": 1.5281,
34
- "eval_samples_per_second": 1066.032,
35
- "eval_steps_per_second": 66.75,
36
- "step": 138
37
- },
38
- {
39
- "epoch": 4.0,
40
- "eval_accuracy": 0.8367096378146102,
41
- "eval_loss": 0.38024526834487915,
42
- "eval_runtime": 1.5933,
43
- "eval_samples_per_second": 1022.426,
44
- "eval_steps_per_second": 64.019,
45
- "step": 184
46
- }
47
- ],
48
  "logging_steps": 500,
49
  "max_steps": 184,
50
  "num_input_tokens_seen": 0,
51
- "num_train_epochs": 4,
52
  "save_steps": 500,
53
  "stateful_callbacks": {
54
  "TrainerControl": {
@@ -63,12 +26,12 @@
63
  }
64
  },
65
  "total_flos": 0,
66
- "train_batch_size": 32,
67
  "trial_name": null,
68
  "trial_params": {
69
- "learning_rate": 2.448709525953929e-05,
70
- "num_train_epochs": 4,
71
- "per_device_train_batch_size": 32,
72
- "seed": 21
73
  }
74
  }
 
1
  {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
  "eval_steps": 500,
6
  "global_step": 184,
7
  "is_hyper_param_search": true,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
+ "log_history": [],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  "logging_steps": 500,
12
  "max_steps": 184,
13
  "num_input_tokens_seen": 0,
14
+ "num_train_epochs": 1,
15
  "save_steps": 500,
16
  "stateful_callbacks": {
17
  "TrainerControl": {
 
26
  }
27
  },
28
  "total_flos": 0,
29
+ "train_batch_size": 8,
30
  "trial_name": null,
31
  "trial_params": {
32
+ "learning_rate": 7.059488330688458e-06,
33
+ "num_train_epochs": 1,
34
+ "per_device_train_batch_size": 8,
35
+ "seed": 36
36
  }
37
  }
run-7/checkpoint-184/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7783b64377581c270d4bf6ae57189eb09386ad289a71203380cd72e257c9a682
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:575bb016637d16461e5cbb2d3476fe7a1937baddcde0db5fb1c149eb0a1cf074
3
  size 5176
run-8/checkpoint-184/config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google-bert/bert-base-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "problem_type": "single_label_classification",
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.42.4",
24
+ "type_vocab_size": 2,
25
+ "use_cache": true,
26
+ "vocab_size": 30522
27
+ }
run-8/checkpoint-184/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:871d64bbcfe7e3b7796e49809c2c034ee718a85d48065ba01fb33c8b128ea7b7
3
+ size 437958648
run-8/checkpoint-184/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8bc2aa37c641062526fd14c9e1965c4a0cc15c4e045dff2ff279113327355038
3
+ size 876038394
run-8/checkpoint-184/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e9dbde989aabd7113299b007c70b149b3d27968de69c8cddf8b5f43b150d373
3
+ size 14308
run-8/checkpoint-184/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c0e9beb3c68ca5959fb2ff6789e58f2e43b008984b3dba67ca73c97d2362a84
3
+ size 1064
run-8/checkpoint-184/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
run-8/checkpoint-184/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-8/checkpoint-184/tokenizer_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 512,
49
+ "pad_token": "[PAD]",
50
+ "sep_token": "[SEP]",
51
+ "strip_accents": null,
52
+ "tokenize_chinese_chars": true,
53
+ "tokenizer_class": "BertTokenizer",
54
+ "unk_token": "[UNK]"
55
+ }
run-8/checkpoint-184/trainer_state.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.8182934315531001,
3
+ "best_model_checkpoint": "BERTForDetectingDepression-Twitter2015/run-8/checkpoint-184",
4
+ "epoch": 1.0,
5
+ "eval_steps": 500,
6
+ "global_step": 184,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_accuracy": 0.8182934315531001,
14
+ "eval_loss": 0.39697179198265076,
15
+ "eval_runtime": 1.4648,
16
+ "eval_samples_per_second": 1112.118,
17
+ "eval_steps_per_second": 69.635,
18
+ "step": 184
19
+ }
20
+ ],
21
+ "logging_steps": 500,
22
+ "max_steps": 920,
23
+ "num_input_tokens_seen": 0,
24
+ "num_train_epochs": 5,
25
+ "save_steps": 500,
26
+ "stateful_callbacks": {
27
+ "TrainerControl": {
28
+ "args": {
29
+ "should_epoch_stop": false,
30
+ "should_evaluate": false,
31
+ "should_log": false,
32
+ "should_save": true,
33
+ "should_training_stop": false
34
+ },
35
+ "attributes": {}
36
+ }
37
+ },
38
+ "total_flos": 0,
39
+ "train_batch_size": 8,
40
+ "trial_name": null,
41
+ "trial_params": {
42
+ "learning_rate": 8.186801881082434e-05,
43
+ "num_train_epochs": 5,
44
+ "per_device_train_batch_size": 8,
45
+ "seed": 29
46
+ }
47
+ }
run-8/checkpoint-184/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fa6fdef93917d7fe7ad43bf5df2576ba9b6fddeca1ccee15cf1281a921dc8a8
3
+ size 5176
run-8/checkpoint-184/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
run-8/checkpoint-368/config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google-bert/bert-base-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "problem_type": "single_label_classification",
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.42.4",
24
+ "type_vocab_size": 2,
25
+ "use_cache": true,
26
+ "vocab_size": 30522
27
+ }
run-8/checkpoint-368/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20cf0e9c534cfe1ddf1880e1729493ea338f8433f07635eb68f746a277c26903
3
+ size 437958648
run-8/checkpoint-368/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae84e3db447fdbacf502f50da7d22d04aed3ca1139205c4aa8f0fbeafc8d06c6
3
+ size 876038394
run-8/checkpoint-368/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6fd8b234f80eec82bdaf2cf7e7e24a728f9330e69ee3761cf99a20f53659f2f
3
+ size 14308
run-8/checkpoint-368/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d4813f844974ccfd98fe8d0919aab4850c12580960134f1090fc2b6a9f0caeb
3
+ size 1064
run-8/checkpoint-368/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
run-8/checkpoint-368/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-8/checkpoint-368/tokenizer_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 512,
49
+ "pad_token": "[PAD]",
50
+ "sep_token": "[SEP]",
51
+ "strip_accents": null,
52
+ "tokenize_chinese_chars": true,
53
+ "tokenizer_class": "BertTokenizer",
54
+ "unk_token": "[UNK]"
55
+ }
run-8/checkpoint-368/trainer_state.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.8446899938612645,
3
+ "best_model_checkpoint": "BERTForDetectingDepression-Twitter2015/run-8/checkpoint-368",
4
+ "epoch": 2.0,
5
+ "eval_steps": 500,
6
+ "global_step": 368,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_accuracy": 0.8182934315531001,
14
+ "eval_loss": 0.39697179198265076,
15
+ "eval_runtime": 1.4648,
16
+ "eval_samples_per_second": 1112.118,
17
+ "eval_steps_per_second": 69.635,
18
+ "step": 184
19
+ },
20
+ {
21
+ "epoch": 2.0,
22
+ "eval_accuracy": 0.8446899938612645,
23
+ "eval_loss": 0.4469450116157532,
24
+ "eval_runtime": 1.4697,
25
+ "eval_samples_per_second": 1108.387,
26
+ "eval_steps_per_second": 69.402,
27
+ "step": 368
28
+ }
29
+ ],
30
+ "logging_steps": 500,
31
+ "max_steps": 920,
32
+ "num_input_tokens_seen": 0,
33
+ "num_train_epochs": 5,
34
+ "save_steps": 500,
35
+ "stateful_callbacks": {
36
+ "TrainerControl": {
37
+ "args": {
38
+ "should_epoch_stop": false,
39
+ "should_evaluate": false,
40
+ "should_log": false,
41
+ "should_save": true,
42
+ "should_training_stop": false
43
+ },
44
+ "attributes": {}
45
+ }
46
+ },
47
+ "total_flos": 0,
48
+ "train_batch_size": 8,
49
+ "trial_name": null,
50
+ "trial_params": {
51
+ "learning_rate": 8.186801881082434e-05,
52
+ "num_train_epochs": 5,
53
+ "per_device_train_batch_size": 8,
54
+ "seed": 29
55
+ }
56
+ }
run-8/checkpoint-368/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fa6fdef93917d7fe7ad43bf5df2576ba9b6fddeca1ccee15cf1281a921dc8a8
3
+ size 5176
run-8/checkpoint-368/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
run-8/checkpoint-552/config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google-bert/bert-base-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "problem_type": "single_label_classification",
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.42.4",
24
+ "type_vocab_size": 2,
25
+ "use_cache": true,
26
+ "vocab_size": 30522
27
+ }
run-8/checkpoint-552/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f56724db37427c14a83043f0e20fb7e3442bf5fb1d90c2ede868394566d4f57f
3
+ size 437958648
run-8/checkpoint-552/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95b7451a0fed6ecc8883e0797d051d2556bef8ee15e45e4ca20fdd38d1a934d6
3
+ size 876038394
run-8/checkpoint-552/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a75ff17b5cb476de53a1678bd79961ec1cc3cbc187954498c179da1bbc3bd5a9
3
+ size 14308
run-8/checkpoint-552/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbb1b7053ccbd69e3471cd7f4d3d85c30ff8647ab1075c236d407056fefb245c
3
+ size 1064
run-8/checkpoint-552/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
run-8/checkpoint-552/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-8/checkpoint-552/tokenizer_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 512,
49
+ "pad_token": "[PAD]",
50
+ "sep_token": "[SEP]",
51
+ "strip_accents": null,
52
+ "tokenize_chinese_chars": true,
53
+ "tokenizer_class": "BertTokenizer",
54
+ "unk_token": "[UNK]"
55
+ }
run-8/checkpoint-552/trainer_state.json ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.8496009821976673,
3
+ "best_model_checkpoint": "BERTForDetectingDepression-Twitter2015/run-8/checkpoint-552",
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 552,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_accuracy": 0.8182934315531001,
14
+ "eval_loss": 0.39697179198265076,
15
+ "eval_runtime": 1.4648,
16
+ "eval_samples_per_second": 1112.118,
17
+ "eval_steps_per_second": 69.635,
18
+ "step": 184
19
+ },
20
+ {
21
+ "epoch": 2.0,
22
+ "eval_accuracy": 0.8446899938612645,
23
+ "eval_loss": 0.4469450116157532,
24
+ "eval_runtime": 1.4697,
25
+ "eval_samples_per_second": 1108.387,
26
+ "eval_steps_per_second": 69.402,
27
+ "step": 368
28
+ },
29
+ {
30
+ "epoch": 2.717391304347826,
31
+ "grad_norm": 8.623266220092773,
32
+ "learning_rate": 3.7374530326680673e-05,
33
+ "loss": 0.3328,
34
+ "step": 500
35
+ },
36
+ {
37
+ "epoch": 3.0,
38
+ "eval_accuracy": 0.8496009821976673,
39
+ "eval_loss": 0.7693418264389038,
40
+ "eval_runtime": 1.4817,
41
+ "eval_samples_per_second": 1099.419,
42
+ "eval_steps_per_second": 68.84,
43
+ "step": 552
44
+ }
45
+ ],
46
+ "logging_steps": 500,
47
+ "max_steps": 920,
48
+ "num_input_tokens_seen": 0,
49
+ "num_train_epochs": 5,
50
+ "save_steps": 500,
51
+ "stateful_callbacks": {
52
+ "TrainerControl": {
53
+ "args": {
54
+ "should_epoch_stop": false,
55
+ "should_evaluate": false,
56
+ "should_log": false,
57
+ "should_save": true,
58
+ "should_training_stop": false
59
+ },
60
+ "attributes": {}
61
+ }
62
+ },
63
+ "total_flos": 166182381453960.0,
64
+ "train_batch_size": 8,
65
+ "trial_name": null,
66
+ "trial_params": {
67
+ "learning_rate": 8.186801881082434e-05,
68
+ "num_train_epochs": 5,
69
+ "per_device_train_batch_size": 8,
70
+ "seed": 29
71
+ }
72
+ }
run-8/checkpoint-552/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fa6fdef93917d7fe7ad43bf5df2576ba9b6fddeca1ccee15cf1281a921dc8a8
3
+ size 5176
run-8/checkpoint-552/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
run-8/checkpoint-736/config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google-bert/bert-base-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "problem_type": "single_label_classification",
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.42.4",
24
+ "type_vocab_size": 2,
25
+ "use_cache": true,
26
+ "vocab_size": 30522
27
+ }
run-8/checkpoint-736/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55634ac9308f6162f645a237065d0f0ca7c8361699fd2b0dcf72cc05e60ad73e
3
+ size 437958648
run-8/checkpoint-736/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a579a515b4edb7d7c4c097f8e965cd7d7243e6385f6ca8aa2b7c56f70ec79a8
3
+ size 876038394
run-8/checkpoint-736/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d36cc7e5515baf19e79ba7261a0e995224c4f7be1e0bd9deed1e937d143fb84
3
+ size 14308
run-8/checkpoint-736/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8f2ecdf0377af42389d796b25959693d160849b740ca15bc38faae2c99ff870
3
+ size 1064
run-8/checkpoint-736/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
run-8/checkpoint-736/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-8/checkpoint-736/tokenizer_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 512,
49
+ "pad_token": "[PAD]",
50
+ "sep_token": "[SEP]",
51
+ "strip_accents": null,
52
+ "tokenize_chinese_chars": true,
53
+ "tokenizer_class": "BertTokenizer",
54
+ "unk_token": "[UNK]"
55
+ }
run-8/checkpoint-736/trainer_state.json ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.852670349907919,
3
+ "best_model_checkpoint": "BERTForDetectingDepression-Twitter2015/run-8/checkpoint-736",
4
+ "epoch": 4.0,
5
+ "eval_steps": 500,
6
+ "global_step": 736,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_accuracy": 0.8182934315531001,
14
+ "eval_loss": 0.39697179198265076,
15
+ "eval_runtime": 1.4648,
16
+ "eval_samples_per_second": 1112.118,
17
+ "eval_steps_per_second": 69.635,
18
+ "step": 184
19
+ },
20
+ {
21
+ "epoch": 2.0,
22
+ "eval_accuracy": 0.8446899938612645,
23
+ "eval_loss": 0.4469450116157532,
24
+ "eval_runtime": 1.4697,
25
+ "eval_samples_per_second": 1108.387,
26
+ "eval_steps_per_second": 69.402,
27
+ "step": 368
28
+ },
29
+ {
30
+ "epoch": 2.717391304347826,
31
+ "grad_norm": 8.623266220092773,
32
+ "learning_rate": 3.7374530326680673e-05,
33
+ "loss": 0.3328,
34
+ "step": 500
35
+ },
36
+ {
37
+ "epoch": 3.0,
38
+ "eval_accuracy": 0.8496009821976673,
39
+ "eval_loss": 0.7693418264389038,
40
+ "eval_runtime": 1.4817,
41
+ "eval_samples_per_second": 1099.419,
42
+ "eval_steps_per_second": 68.84,
43
+ "step": 552
44
+ },
45
+ {
46
+ "epoch": 4.0,
47
+ "eval_accuracy": 0.852670349907919,
48
+ "eval_loss": 0.9155253767967224,
49
+ "eval_runtime": 1.4898,
50
+ "eval_samples_per_second": 1093.449,
51
+ "eval_steps_per_second": 68.466,
52
+ "step": 736
53
+ }
54
+ ],
55
+ "logging_steps": 500,
56
+ "max_steps": 920,
57
+ "num_input_tokens_seen": 0,
58
+ "num_train_epochs": 5,
59
+ "save_steps": 500,
60
+ "stateful_callbacks": {
61
+ "TrainerControl": {
62
+ "args": {
63
+ "should_epoch_stop": false,
64
+ "should_evaluate": false,
65
+ "should_log": false,
66
+ "should_save": true,
67
+ "should_training_stop": false
68
+ },
69
+ "attributes": {}
70
+ }
71
+ },
72
+ "total_flos": 166182381453960.0,
73
+ "train_batch_size": 8,
74
+ "trial_name": null,
75
+ "trial_params": {
76
+ "learning_rate": 8.186801881082434e-05,
77
+ "num_train_epochs": 5,
78
+ "per_device_train_batch_size": 8,
79
+ "seed": 29
80
+ }
81
+ }
run-8/checkpoint-736/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fa6fdef93917d7fe7ad43bf5df2576ba9b6fddeca1ccee15cf1281a921dc8a8
3
+ size 5176