Training in progress, epoch 1
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- model.safetensors +1 -1
- run-0/checkpoint-184/config.json +2 -2
- run-0/checkpoint-184/model.safetensors +2 -2
- run-0/checkpoint-184/optimizer.pt +2 -2
- run-0/checkpoint-184/rng_state.pth +1 -1
- run-0/checkpoint-184/scheduler.pt +1 -1
- run-0/checkpoint-184/special_tokens_map.json +35 -5
- run-0/checkpoint-184/tokenizer.json +0 -0
- run-0/checkpoint-184/tokenizer_config.json +5 -1
- run-0/checkpoint-184/trainer_state.json +11 -11
- run-0/checkpoint-184/training_args.bin +1 -1
- run-0/checkpoint-184/vocab.txt +0 -0
- run-0/checkpoint-368/config.json +2 -2
- run-0/checkpoint-368/model.safetensors +2 -2
- run-0/checkpoint-368/optimizer.pt +2 -2
- run-0/checkpoint-368/rng_state.pth +1 -1
- run-0/checkpoint-368/scheduler.pt +1 -1
- run-0/checkpoint-368/special_tokens_map.json +35 -5
- run-0/checkpoint-368/tokenizer.json +0 -0
- run-0/checkpoint-368/tokenizer_config.json +5 -1
- run-0/checkpoint-368/trainer_state.json +16 -16
- run-0/checkpoint-368/training_args.bin +1 -1
- run-0/checkpoint-368/vocab.txt +0 -0
- run-0/checkpoint-552/config.json +2 -2
- run-0/checkpoint-552/model.safetensors +2 -2
- run-0/checkpoint-552/optimizer.pt +2 -2
- run-0/checkpoint-552/rng_state.pth +1 -1
- run-0/checkpoint-552/scheduler.pt +1 -1
- run-0/checkpoint-552/special_tokens_map.json +35 -5
- run-0/checkpoint-552/tokenizer.json +0 -0
- run-0/checkpoint-552/tokenizer_config.json +5 -1
- run-0/checkpoint-552/trainer_state.json +26 -26
- run-0/checkpoint-552/training_args.bin +1 -1
- run-0/checkpoint-552/vocab.txt +0 -0
- run-0/checkpoint-736/config.json +2 -2
- run-0/checkpoint-736/model.safetensors +2 -2
- run-0/checkpoint-736/optimizer.pt +2 -2
- run-0/checkpoint-736/rng_state.pth +1 -1
- run-0/checkpoint-736/scheduler.pt +1 -1
- run-0/checkpoint-736/special_tokens_map.json +35 -5
- run-0/checkpoint-736/tokenizer.json +0 -0
- run-0/checkpoint-736/tokenizer_config.json +5 -1
- run-0/checkpoint-736/trainer_state.json +32 -32
- run-0/checkpoint-736/training_args.bin +1 -1
- run-0/checkpoint-736/vocab.txt +0 -0
- run-1/checkpoint-138/config.json +2 -2
- run-1/checkpoint-138/model.safetensors +2 -2
- run-1/checkpoint-138/optimizer.pt +2 -2
- run-1/checkpoint-138/rng_state.pth +2 -2
- run-1/checkpoint-138/scheduler.pt +1 -1
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 433270768
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fadc71a72125e03f5838a56bfadbd4602fa066c074a14f472a32a55cecf283e9
|
| 3 |
size 433270768
|
run-0/checkpoint-184/config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "
|
| 3 |
"architectures": [
|
| 4 |
"BertForSequenceClassification"
|
| 5 |
],
|
|
@@ -23,5 +23,5 @@
|
|
| 23 |
"transformers_version": "4.42.4",
|
| 24 |
"type_vocab_size": 2,
|
| 25 |
"use_cache": true,
|
| 26 |
-
"vocab_size":
|
| 27 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "AIMH/mental-bert-base-cased",
|
| 3 |
"architectures": [
|
| 4 |
"BertForSequenceClassification"
|
| 5 |
],
|
|
|
|
| 23 |
"transformers_version": "4.42.4",
|
| 24 |
"type_vocab_size": 2,
|
| 25 |
"use_cache": true,
|
| 26 |
+
"vocab_size": 28996
|
| 27 |
}
|
run-0/checkpoint-184/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:14904497aeb819b35d389d767bd0c2c6b738e61c74d9ebfe2eb359d91304867c
|
| 3 |
+
size 433270768
|
run-0/checkpoint-184/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1eb42b00c35b39a9288c94317e38da50c7ea7e66ac7186ca8c4fc93dfa9f87fb
|
| 3 |
+
size 866662650
|
run-0/checkpoint-184/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:21786abe7db04770ad5fe11a75ea383e2f20db4c48e1cbb97b29eadf57919021
|
| 3 |
size 14244
|
run-0/checkpoint-184/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:739a5b91b27dc1034e9137a0f663fb583b541201a408f6780cc84acebe277005
|
| 3 |
size 1064
|
run-0/checkpoint-184/special_tokens_map.json
CHANGED
|
@@ -1,7 +1,37 @@
|
|
| 1 |
{
|
| 2 |
-
"cls_token":
|
| 3 |
-
|
| 4 |
-
|
| 5 |
-
|
| 6 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"cls_token": {
|
| 3 |
+
"content": "[CLS]",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"mask_token": {
|
| 10 |
+
"content": "[MASK]",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"pad_token": {
|
| 17 |
+
"content": "[PAD]",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
},
|
| 23 |
+
"sep_token": {
|
| 24 |
+
"content": "[SEP]",
|
| 25 |
+
"lstrip": false,
|
| 26 |
+
"normalized": false,
|
| 27 |
+
"rstrip": false,
|
| 28 |
+
"single_word": false
|
| 29 |
+
},
|
| 30 |
+
"unk_token": {
|
| 31 |
+
"content": "[UNK]",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false
|
| 36 |
+
}
|
| 37 |
}
|
run-0/checkpoint-184/tokenizer.json
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
run-0/checkpoint-184/tokenizer_config.json
CHANGED
|
@@ -43,13 +43,17 @@
|
|
| 43 |
},
|
| 44 |
"clean_up_tokenization_spaces": true,
|
| 45 |
"cls_token": "[CLS]",
|
| 46 |
-
"do_lower_case":
|
| 47 |
"mask_token": "[MASK]",
|
|
|
|
| 48 |
"model_max_length": 512,
|
| 49 |
"pad_token": "[PAD]",
|
| 50 |
"sep_token": "[SEP]",
|
|
|
|
| 51 |
"strip_accents": null,
|
| 52 |
"tokenize_chinese_chars": true,
|
| 53 |
"tokenizer_class": "BertTokenizer",
|
|
|
|
|
|
|
| 54 |
"unk_token": "[UNK]"
|
| 55 |
}
|
|
|
|
| 43 |
},
|
| 44 |
"clean_up_tokenization_spaces": true,
|
| 45 |
"cls_token": "[CLS]",
|
| 46 |
+
"do_lower_case": false,
|
| 47 |
"mask_token": "[MASK]",
|
| 48 |
+
"max_length": 512,
|
| 49 |
"model_max_length": 512,
|
| 50 |
"pad_token": "[PAD]",
|
| 51 |
"sep_token": "[SEP]",
|
| 52 |
+
"stride": 0,
|
| 53 |
"strip_accents": null,
|
| 54 |
"tokenize_chinese_chars": true,
|
| 55 |
"tokenizer_class": "BertTokenizer",
|
| 56 |
+
"truncation_side": "right",
|
| 57 |
+
"truncation_strategy": "longest_first",
|
| 58 |
"unk_token": "[UNK]"
|
| 59 |
}
|
run-0/checkpoint-184/trainer_state.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
"best_model_checkpoint": "BERTForDetectingDepression-Twitter2015/run-0/checkpoint-184",
|
| 4 |
"epoch": 1.0,
|
| 5 |
"eval_steps": 500,
|
|
@@ -10,18 +10,18 @@
|
|
| 10 |
"log_history": [
|
| 11 |
{
|
| 12 |
"epoch": 1.0,
|
| 13 |
-
"eval_accuracy": 0.
|
| 14 |
-
"eval_loss": 0.
|
| 15 |
-
"eval_runtime": 1.
|
| 16 |
-
"eval_samples_per_second":
|
| 17 |
-
"eval_steps_per_second":
|
| 18 |
"step": 184
|
| 19 |
}
|
| 20 |
],
|
| 21 |
"logging_steps": 500,
|
| 22 |
-
"max_steps":
|
| 23 |
"num_input_tokens_seen": 0,
|
| 24 |
-
"num_train_epochs":
|
| 25 |
"save_steps": 500,
|
| 26 |
"stateful_callbacks": {
|
| 27 |
"TrainerControl": {
|
|
@@ -39,9 +39,9 @@
|
|
| 39 |
"train_batch_size": 8,
|
| 40 |
"trial_name": null,
|
| 41 |
"trial_params": {
|
| 42 |
-
"learning_rate": 1.
|
| 43 |
-
"num_train_epochs":
|
| 44 |
"per_device_train_batch_size": 8,
|
| 45 |
-
"seed":
|
| 46 |
}
|
| 47 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.7826887661141805,
|
| 3 |
"best_model_checkpoint": "BERTForDetectingDepression-Twitter2015/run-0/checkpoint-184",
|
| 4 |
"epoch": 1.0,
|
| 5 |
"eval_steps": 500,
|
|
|
|
| 10 |
"log_history": [
|
| 11 |
{
|
| 12 |
"epoch": 1.0,
|
| 13 |
+
"eval_accuracy": 0.7826887661141805,
|
| 14 |
+
"eval_loss": 0.45181456208229065,
|
| 15 |
+
"eval_runtime": 1.6866,
|
| 16 |
+
"eval_samples_per_second": 965.826,
|
| 17 |
+
"eval_steps_per_second": 60.475,
|
| 18 |
"step": 184
|
| 19 |
}
|
| 20 |
],
|
| 21 |
"logging_steps": 500,
|
| 22 |
+
"max_steps": 736,
|
| 23 |
"num_input_tokens_seen": 0,
|
| 24 |
+
"num_train_epochs": 4,
|
| 25 |
"save_steps": 500,
|
| 26 |
"stateful_callbacks": {
|
| 27 |
"TrainerControl": {
|
|
|
|
| 39 |
"train_batch_size": 8,
|
| 40 |
"trial_name": null,
|
| 41 |
"trial_params": {
|
| 42 |
+
"learning_rate": 1.0817235137868972e-05,
|
| 43 |
+
"num_train_epochs": 4,
|
| 44 |
"per_device_train_batch_size": 8,
|
| 45 |
+
"seed": 5
|
| 46 |
}
|
| 47 |
}
|
run-0/checkpoint-184/training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5176
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3dca1282896774eec430907d299b265f87badfd78ce3fae95cf77e64c477252a
|
| 3 |
size 5176
|
run-0/checkpoint-184/vocab.txt
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
run-0/checkpoint-368/config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "
|
| 3 |
"architectures": [
|
| 4 |
"BertForSequenceClassification"
|
| 5 |
],
|
|
@@ -23,5 +23,5 @@
|
|
| 23 |
"transformers_version": "4.42.4",
|
| 24 |
"type_vocab_size": 2,
|
| 25 |
"use_cache": true,
|
| 26 |
-
"vocab_size":
|
| 27 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "AIMH/mental-bert-base-cased",
|
| 3 |
"architectures": [
|
| 4 |
"BertForSequenceClassification"
|
| 5 |
],
|
|
|
|
| 23 |
"transformers_version": "4.42.4",
|
| 24 |
"type_vocab_size": 2,
|
| 25 |
"use_cache": true,
|
| 26 |
+
"vocab_size": 28996
|
| 27 |
}
|
run-0/checkpoint-368/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:67dce7fb2d4f83c465a3416c279e5500cd2ccda1de19b58a50b8e6c17d9b0db6
|
| 3 |
+
size 433270768
|
run-0/checkpoint-368/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fdb1489ea5df01528b5cb3af607ccea72ca7367b151cc3ea8f4d14d34ce27a5a
|
| 3 |
+
size 866662650
|
run-0/checkpoint-368/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5adc7e0b9329c5c2b8f270a5ad77762ef51cec651ab2b69c5e93b6fdcc032b71
|
| 3 |
size 14244
|
run-0/checkpoint-368/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:66623be9c7e826a500123eb24f5d98d30dbadb581c225227dafd683bbb302ab0
|
| 3 |
size 1064
|
run-0/checkpoint-368/special_tokens_map.json
CHANGED
|
@@ -1,7 +1,37 @@
|
|
| 1 |
{
|
| 2 |
-
"cls_token":
|
| 3 |
-
|
| 4 |
-
|
| 5 |
-
|
| 6 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"cls_token": {
|
| 3 |
+
"content": "[CLS]",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"mask_token": {
|
| 10 |
+
"content": "[MASK]",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"pad_token": {
|
| 17 |
+
"content": "[PAD]",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
},
|
| 23 |
+
"sep_token": {
|
| 24 |
+
"content": "[SEP]",
|
| 25 |
+
"lstrip": false,
|
| 26 |
+
"normalized": false,
|
| 27 |
+
"rstrip": false,
|
| 28 |
+
"single_word": false
|
| 29 |
+
},
|
| 30 |
+
"unk_token": {
|
| 31 |
+
"content": "[UNK]",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false
|
| 36 |
+
}
|
| 37 |
}
|
run-0/checkpoint-368/tokenizer.json
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
run-0/checkpoint-368/tokenizer_config.json
CHANGED
|
@@ -43,13 +43,17 @@
|
|
| 43 |
},
|
| 44 |
"clean_up_tokenization_spaces": true,
|
| 45 |
"cls_token": "[CLS]",
|
| 46 |
-
"do_lower_case":
|
| 47 |
"mask_token": "[MASK]",
|
|
|
|
| 48 |
"model_max_length": 512,
|
| 49 |
"pad_token": "[PAD]",
|
| 50 |
"sep_token": "[SEP]",
|
|
|
|
| 51 |
"strip_accents": null,
|
| 52 |
"tokenize_chinese_chars": true,
|
| 53 |
"tokenizer_class": "BertTokenizer",
|
|
|
|
|
|
|
| 54 |
"unk_token": "[UNK]"
|
| 55 |
}
|
|
|
|
| 43 |
},
|
| 44 |
"clean_up_tokenization_spaces": true,
|
| 45 |
"cls_token": "[CLS]",
|
| 46 |
+
"do_lower_case": false,
|
| 47 |
"mask_token": "[MASK]",
|
| 48 |
+
"max_length": 512,
|
| 49 |
"model_max_length": 512,
|
| 50 |
"pad_token": "[PAD]",
|
| 51 |
"sep_token": "[SEP]",
|
| 52 |
+
"stride": 0,
|
| 53 |
"strip_accents": null,
|
| 54 |
"tokenize_chinese_chars": true,
|
| 55 |
"tokenizer_class": "BertTokenizer",
|
| 56 |
+
"truncation_side": "right",
|
| 57 |
+
"truncation_strategy": "longest_first",
|
| 58 |
"unk_token": "[UNK]"
|
| 59 |
}
|
run-0/checkpoint-368/trainer_state.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
"best_model_checkpoint": "BERTForDetectingDepression-Twitter2015/run-0/checkpoint-368",
|
| 4 |
"epoch": 2.0,
|
| 5 |
"eval_steps": 500,
|
|
@@ -10,27 +10,27 @@
|
|
| 10 |
"log_history": [
|
| 11 |
{
|
| 12 |
"epoch": 1.0,
|
| 13 |
-
"eval_accuracy": 0.
|
| 14 |
-
"eval_loss": 0.
|
| 15 |
-
"eval_runtime": 1.
|
| 16 |
-
"eval_samples_per_second":
|
| 17 |
-
"eval_steps_per_second":
|
| 18 |
"step": 184
|
| 19 |
},
|
| 20 |
{
|
| 21 |
"epoch": 2.0,
|
| 22 |
-
"eval_accuracy": 0.
|
| 23 |
-
"eval_loss": 0.
|
| 24 |
-
"eval_runtime": 1.
|
| 25 |
-
"eval_samples_per_second":
|
| 26 |
-
"eval_steps_per_second":
|
| 27 |
"step": 368
|
| 28 |
}
|
| 29 |
],
|
| 30 |
"logging_steps": 500,
|
| 31 |
-
"max_steps":
|
| 32 |
"num_input_tokens_seen": 0,
|
| 33 |
-
"num_train_epochs":
|
| 34 |
"save_steps": 500,
|
| 35 |
"stateful_callbacks": {
|
| 36 |
"TrainerControl": {
|
|
@@ -48,9 +48,9 @@
|
|
| 48 |
"train_batch_size": 8,
|
| 49 |
"trial_name": null,
|
| 50 |
"trial_params": {
|
| 51 |
-
"learning_rate": 1.
|
| 52 |
-
"num_train_epochs":
|
| 53 |
"per_device_train_batch_size": 8,
|
| 54 |
-
"seed":
|
| 55 |
}
|
| 56 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.8238182934315531,
|
| 3 |
"best_model_checkpoint": "BERTForDetectingDepression-Twitter2015/run-0/checkpoint-368",
|
| 4 |
"epoch": 2.0,
|
| 5 |
"eval_steps": 500,
|
|
|
|
| 10 |
"log_history": [
|
| 11 |
{
|
| 12 |
"epoch": 1.0,
|
| 13 |
+
"eval_accuracy": 0.7826887661141805,
|
| 14 |
+
"eval_loss": 0.45181456208229065,
|
| 15 |
+
"eval_runtime": 1.6866,
|
| 16 |
+
"eval_samples_per_second": 965.826,
|
| 17 |
+
"eval_steps_per_second": 60.475,
|
| 18 |
"step": 184
|
| 19 |
},
|
| 20 |
{
|
| 21 |
"epoch": 2.0,
|
| 22 |
+
"eval_accuracy": 0.8238182934315531,
|
| 23 |
+
"eval_loss": 0.38157492876052856,
|
| 24 |
+
"eval_runtime": 1.6313,
|
| 25 |
+
"eval_samples_per_second": 998.618,
|
| 26 |
+
"eval_steps_per_second": 62.529,
|
| 27 |
"step": 368
|
| 28 |
}
|
| 29 |
],
|
| 30 |
"logging_steps": 500,
|
| 31 |
+
"max_steps": 736,
|
| 32 |
"num_input_tokens_seen": 0,
|
| 33 |
+
"num_train_epochs": 4,
|
| 34 |
"save_steps": 500,
|
| 35 |
"stateful_callbacks": {
|
| 36 |
"TrainerControl": {
|
|
|
|
| 48 |
"train_batch_size": 8,
|
| 49 |
"trial_name": null,
|
| 50 |
"trial_params": {
|
| 51 |
+
"learning_rate": 1.0817235137868972e-05,
|
| 52 |
+
"num_train_epochs": 4,
|
| 53 |
"per_device_train_batch_size": 8,
|
| 54 |
+
"seed": 5
|
| 55 |
}
|
| 56 |
}
|
run-0/checkpoint-368/training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5176
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3dca1282896774eec430907d299b265f87badfd78ce3fae95cf77e64c477252a
|
| 3 |
size 5176
|
run-0/checkpoint-368/vocab.txt
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
run-0/checkpoint-552/config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "
|
| 3 |
"architectures": [
|
| 4 |
"BertForSequenceClassification"
|
| 5 |
],
|
|
@@ -23,5 +23,5 @@
|
|
| 23 |
"transformers_version": "4.42.4",
|
| 24 |
"type_vocab_size": 2,
|
| 25 |
"use_cache": true,
|
| 26 |
-
"vocab_size":
|
| 27 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "AIMH/mental-bert-base-cased",
|
| 3 |
"architectures": [
|
| 4 |
"BertForSequenceClassification"
|
| 5 |
],
|
|
|
|
| 23 |
"transformers_version": "4.42.4",
|
| 24 |
"type_vocab_size": 2,
|
| 25 |
"use_cache": true,
|
| 26 |
+
"vocab_size": 28996
|
| 27 |
}
|
run-0/checkpoint-552/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e738066ddf5b3e7a73d4c63cc786c746ebd7f322a2aff63827136cbd74c3e6c2
|
| 3 |
+
size 433270768
|
run-0/checkpoint-552/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:83868c36b9e545a70e142df07b372f02fc3d7531617e33a0bb98a18ba09bc14a
|
| 3 |
+
size 866662650
|
run-0/checkpoint-552/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d4c47e551cb0fb04f7a3b0097cb93ad2b69429e684a5a15bb1645ad0a8680b9f
|
| 3 |
size 14244
|
run-0/checkpoint-552/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ca67c1719e7ecdf02147ca37365edb98f6fef1d7cfba46ebb4d7610ff244fccf
|
| 3 |
size 1064
|
run-0/checkpoint-552/special_tokens_map.json
CHANGED
|
@@ -1,7 +1,37 @@
|
|
| 1 |
{
|
| 2 |
-
"cls_token":
|
| 3 |
-
|
| 4 |
-
|
| 5 |
-
|
| 6 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"cls_token": {
|
| 3 |
+
"content": "[CLS]",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"mask_token": {
|
| 10 |
+
"content": "[MASK]",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"pad_token": {
|
| 17 |
+
"content": "[PAD]",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
},
|
| 23 |
+
"sep_token": {
|
| 24 |
+
"content": "[SEP]",
|
| 25 |
+
"lstrip": false,
|
| 26 |
+
"normalized": false,
|
| 27 |
+
"rstrip": false,
|
| 28 |
+
"single_word": false
|
| 29 |
+
},
|
| 30 |
+
"unk_token": {
|
| 31 |
+
"content": "[UNK]",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false
|
| 36 |
+
}
|
| 37 |
}
|
run-0/checkpoint-552/tokenizer.json
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
run-0/checkpoint-552/tokenizer_config.json
CHANGED
|
@@ -43,13 +43,17 @@
|
|
| 43 |
},
|
| 44 |
"clean_up_tokenization_spaces": true,
|
| 45 |
"cls_token": "[CLS]",
|
| 46 |
-
"do_lower_case":
|
| 47 |
"mask_token": "[MASK]",
|
|
|
|
| 48 |
"model_max_length": 512,
|
| 49 |
"pad_token": "[PAD]",
|
| 50 |
"sep_token": "[SEP]",
|
|
|
|
| 51 |
"strip_accents": null,
|
| 52 |
"tokenize_chinese_chars": true,
|
| 53 |
"tokenizer_class": "BertTokenizer",
|
|
|
|
|
|
|
| 54 |
"unk_token": "[UNK]"
|
| 55 |
}
|
|
|
|
| 43 |
},
|
| 44 |
"clean_up_tokenization_spaces": true,
|
| 45 |
"cls_token": "[CLS]",
|
| 46 |
+
"do_lower_case": false,
|
| 47 |
"mask_token": "[MASK]",
|
| 48 |
+
"max_length": 512,
|
| 49 |
"model_max_length": 512,
|
| 50 |
"pad_token": "[PAD]",
|
| 51 |
"sep_token": "[SEP]",
|
| 52 |
+
"stride": 0,
|
| 53 |
"strip_accents": null,
|
| 54 |
"tokenize_chinese_chars": true,
|
| 55 |
"tokenizer_class": "BertTokenizer",
|
| 56 |
+
"truncation_side": "right",
|
| 57 |
+
"truncation_strategy": "longest_first",
|
| 58 |
"unk_token": "[UNK]"
|
| 59 |
}
|
run-0/checkpoint-552/trainer_state.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
-
"best_model_checkpoint": "BERTForDetectingDepression-Twitter2015/run-0/checkpoint-
|
| 4 |
"epoch": 3.0,
|
| 5 |
"eval_steps": 500,
|
| 6 |
"global_step": 552,
|
|
@@ -10,43 +10,43 @@
|
|
| 10 |
"log_history": [
|
| 11 |
{
|
| 12 |
"epoch": 1.0,
|
| 13 |
-
"eval_accuracy": 0.
|
| 14 |
-
"eval_loss": 0.
|
| 15 |
-
"eval_runtime": 1.
|
| 16 |
-
"eval_samples_per_second":
|
| 17 |
-
"eval_steps_per_second":
|
| 18 |
"step": 184
|
| 19 |
},
|
| 20 |
{
|
| 21 |
"epoch": 2.0,
|
| 22 |
-
"eval_accuracy": 0.
|
| 23 |
-
"eval_loss": 0.
|
| 24 |
-
"eval_runtime": 1.
|
| 25 |
-
"eval_samples_per_second":
|
| 26 |
-
"eval_steps_per_second":
|
| 27 |
"step": 368
|
| 28 |
},
|
| 29 |
{
|
| 30 |
"epoch": 2.717391304347826,
|
| 31 |
-
"grad_norm":
|
| 32 |
-
"learning_rate":
|
| 33 |
-
"loss": 0.
|
| 34 |
"step": 500
|
| 35 |
},
|
| 36 |
{
|
| 37 |
"epoch": 3.0,
|
| 38 |
-
"eval_accuracy": 0.
|
| 39 |
-
"eval_loss": 0.
|
| 40 |
-
"eval_runtime": 1.
|
| 41 |
-
"eval_samples_per_second":
|
| 42 |
-
"eval_steps_per_second":
|
| 43 |
"step": 552
|
| 44 |
}
|
| 45 |
],
|
| 46 |
"logging_steps": 500,
|
| 47 |
-
"max_steps":
|
| 48 |
"num_input_tokens_seen": 0,
|
| 49 |
-
"num_train_epochs":
|
| 50 |
"save_steps": 500,
|
| 51 |
"stateful_callbacks": {
|
| 52 |
"TrainerControl": {
|
|
@@ -60,13 +60,13 @@
|
|
| 60 |
"attributes": {}
|
| 61 |
}
|
| 62 |
},
|
| 63 |
-
"total_flos":
|
| 64 |
"train_batch_size": 8,
|
| 65 |
"trial_name": null,
|
| 66 |
"trial_params": {
|
| 67 |
-
"learning_rate": 1.
|
| 68 |
-
"num_train_epochs":
|
| 69 |
"per_device_train_batch_size": 8,
|
| 70 |
-
"seed":
|
| 71 |
}
|
| 72 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.8520564763658687,
|
| 3 |
+
"best_model_checkpoint": "BERTForDetectingDepression-Twitter2015/run-0/checkpoint-552",
|
| 4 |
"epoch": 3.0,
|
| 5 |
"eval_steps": 500,
|
| 6 |
"global_step": 552,
|
|
|
|
| 10 |
"log_history": [
|
| 11 |
{
|
| 12 |
"epoch": 1.0,
|
| 13 |
+
"eval_accuracy": 0.7826887661141805,
|
| 14 |
+
"eval_loss": 0.45181456208229065,
|
| 15 |
+
"eval_runtime": 1.6866,
|
| 16 |
+
"eval_samples_per_second": 965.826,
|
| 17 |
+
"eval_steps_per_second": 60.475,
|
| 18 |
"step": 184
|
| 19 |
},
|
| 20 |
{
|
| 21 |
"epoch": 2.0,
|
| 22 |
+
"eval_accuracy": 0.8238182934315531,
|
| 23 |
+
"eval_loss": 0.38157492876052856,
|
| 24 |
+
"eval_runtime": 1.6313,
|
| 25 |
+
"eval_samples_per_second": 998.618,
|
| 26 |
+
"eval_steps_per_second": 62.529,
|
| 27 |
"step": 368
|
| 28 |
},
|
| 29 |
{
|
| 30 |
"epoch": 2.717391304347826,
|
| 31 |
+
"grad_norm": 19.13737678527832,
|
| 32 |
+
"learning_rate": 3.4685699626862464e-06,
|
| 33 |
+
"loss": 0.3706,
|
| 34 |
"step": 500
|
| 35 |
},
|
| 36 |
{
|
| 37 |
"epoch": 3.0,
|
| 38 |
+
"eval_accuracy": 0.8520564763658687,
|
| 39 |
+
"eval_loss": 0.3886027932167053,
|
| 40 |
+
"eval_runtime": 1.6798,
|
| 41 |
+
"eval_samples_per_second": 969.762,
|
| 42 |
+
"eval_steps_per_second": 60.722,
|
| 43 |
"step": 552
|
| 44 |
}
|
| 45 |
],
|
| 46 |
"logging_steps": 500,
|
| 47 |
+
"max_steps": 736,
|
| 48 |
"num_input_tokens_seen": 0,
|
| 49 |
+
"num_train_epochs": 4,
|
| 50 |
"save_steps": 500,
|
| 51 |
"stateful_callbacks": {
|
| 52 |
"TrainerControl": {
|
|
|
|
| 60 |
"attributes": {}
|
| 61 |
}
|
| 62 |
},
|
| 63 |
+
"total_flos": 104085089056320.0,
|
| 64 |
"train_batch_size": 8,
|
| 65 |
"trial_name": null,
|
| 66 |
"trial_params": {
|
| 67 |
+
"learning_rate": 1.0817235137868972e-05,
|
| 68 |
+
"num_train_epochs": 4,
|
| 69 |
"per_device_train_batch_size": 8,
|
| 70 |
+
"seed": 5
|
| 71 |
}
|
| 72 |
}
|
run-0/checkpoint-552/training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5176
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3dca1282896774eec430907d299b265f87badfd78ce3fae95cf77e64c477252a
|
| 3 |
size 5176
|
run-0/checkpoint-552/vocab.txt
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
run-0/checkpoint-736/config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "
|
| 3 |
"architectures": [
|
| 4 |
"BertForSequenceClassification"
|
| 5 |
],
|
|
@@ -23,5 +23,5 @@
|
|
| 23 |
"transformers_version": "4.42.4",
|
| 24 |
"type_vocab_size": 2,
|
| 25 |
"use_cache": true,
|
| 26 |
-
"vocab_size":
|
| 27 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "AIMH/mental-bert-base-cased",
|
| 3 |
"architectures": [
|
| 4 |
"BertForSequenceClassification"
|
| 5 |
],
|
|
|
|
| 23 |
"transformers_version": "4.42.4",
|
| 24 |
"type_vocab_size": 2,
|
| 25 |
"use_cache": true,
|
| 26 |
+
"vocab_size": 28996
|
| 27 |
}
|
run-0/checkpoint-736/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e60433460a3a14f30b3b9c6d8745fcaae38ebb6f550d28d5a9f4eebf6a415624
|
| 3 |
+
size 433270768
|
run-0/checkpoint-736/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c1a7a2a6fa9214a6c6b5e6cdec516e37af1f2b0fdef678ac40c185adf1901b39
|
| 3 |
+
size 866662650
|
run-0/checkpoint-736/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:09111e111459102080ed949ecfe5be2a91e2d04f47d7be008649634b2877c619
|
| 3 |
size 14244
|
run-0/checkpoint-736/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3663481edf83a09f6aee2aacb4de5b4a3928c23d15cb3333b5457a97c31f0986
|
| 3 |
size 1064
|
run-0/checkpoint-736/special_tokens_map.json
CHANGED
|
@@ -1,7 +1,37 @@
|
|
| 1 |
{
|
| 2 |
-
"cls_token":
|
| 3 |
-
|
| 4 |
-
|
| 5 |
-
|
| 6 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"cls_token": {
|
| 3 |
+
"content": "[CLS]",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"mask_token": {
|
| 10 |
+
"content": "[MASK]",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"pad_token": {
|
| 17 |
+
"content": "[PAD]",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
},
|
| 23 |
+
"sep_token": {
|
| 24 |
+
"content": "[SEP]",
|
| 25 |
+
"lstrip": false,
|
| 26 |
+
"normalized": false,
|
| 27 |
+
"rstrip": false,
|
| 28 |
+
"single_word": false
|
| 29 |
+
},
|
| 30 |
+
"unk_token": {
|
| 31 |
+
"content": "[UNK]",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false
|
| 36 |
+
}
|
| 37 |
}
|
run-0/checkpoint-736/tokenizer.json
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
run-0/checkpoint-736/tokenizer_config.json
CHANGED
|
@@ -43,13 +43,17 @@
|
|
| 43 |
},
|
| 44 |
"clean_up_tokenization_spaces": true,
|
| 45 |
"cls_token": "[CLS]",
|
| 46 |
-
"do_lower_case":
|
| 47 |
"mask_token": "[MASK]",
|
|
|
|
| 48 |
"model_max_length": 512,
|
| 49 |
"pad_token": "[PAD]",
|
| 50 |
"sep_token": "[SEP]",
|
|
|
|
| 51 |
"strip_accents": null,
|
| 52 |
"tokenize_chinese_chars": true,
|
| 53 |
"tokenizer_class": "BertTokenizer",
|
|
|
|
|
|
|
| 54 |
"unk_token": "[UNK]"
|
| 55 |
}
|
|
|
|
| 43 |
},
|
| 44 |
"clean_up_tokenization_spaces": true,
|
| 45 |
"cls_token": "[CLS]",
|
| 46 |
+
"do_lower_case": false,
|
| 47 |
"mask_token": "[MASK]",
|
| 48 |
+
"max_length": 512,
|
| 49 |
"model_max_length": 512,
|
| 50 |
"pad_token": "[PAD]",
|
| 51 |
"sep_token": "[SEP]",
|
| 52 |
+
"stride": 0,
|
| 53 |
"strip_accents": null,
|
| 54 |
"tokenize_chinese_chars": true,
|
| 55 |
"tokenizer_class": "BertTokenizer",
|
| 56 |
+
"truncation_side": "right",
|
| 57 |
+
"truncation_strategy": "longest_first",
|
| 58 |
"unk_token": "[UNK]"
|
| 59 |
}
|
run-0/checkpoint-736/trainer_state.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
-
"best_model_checkpoint": "BERTForDetectingDepression-Twitter2015/run-0/checkpoint-
|
| 4 |
"epoch": 4.0,
|
| 5 |
"eval_steps": 500,
|
| 6 |
"global_step": 736,
|
|
@@ -10,52 +10,52 @@
|
|
| 10 |
"log_history": [
|
| 11 |
{
|
| 12 |
"epoch": 1.0,
|
| 13 |
-
"eval_accuracy": 0.
|
| 14 |
-
"eval_loss": 0.
|
| 15 |
-
"eval_runtime": 1.
|
| 16 |
-
"eval_samples_per_second":
|
| 17 |
-
"eval_steps_per_second":
|
| 18 |
"step": 184
|
| 19 |
},
|
| 20 |
{
|
| 21 |
"epoch": 2.0,
|
| 22 |
-
"eval_accuracy": 0.
|
| 23 |
-
"eval_loss": 0.
|
| 24 |
-
"eval_runtime": 1.
|
| 25 |
-
"eval_samples_per_second":
|
| 26 |
-
"eval_steps_per_second":
|
| 27 |
"step": 368
|
| 28 |
},
|
| 29 |
{
|
| 30 |
"epoch": 2.717391304347826,
|
| 31 |
-
"grad_norm":
|
| 32 |
-
"learning_rate":
|
| 33 |
-
"loss": 0.
|
| 34 |
"step": 500
|
| 35 |
},
|
| 36 |
{
|
| 37 |
"epoch": 3.0,
|
| 38 |
-
"eval_accuracy": 0.
|
| 39 |
-
"eval_loss": 0.
|
| 40 |
-
"eval_runtime": 1.
|
| 41 |
-
"eval_samples_per_second":
|
| 42 |
-
"eval_steps_per_second":
|
| 43 |
"step": 552
|
| 44 |
},
|
| 45 |
{
|
| 46 |
"epoch": 4.0,
|
| 47 |
-
"eval_accuracy": 0.
|
| 48 |
-
"eval_loss": 0.
|
| 49 |
-
"eval_runtime": 1.
|
| 50 |
-
"eval_samples_per_second":
|
| 51 |
-
"eval_steps_per_second":
|
| 52 |
"step": 736
|
| 53 |
}
|
| 54 |
],
|
| 55 |
"logging_steps": 500,
|
| 56 |
-
"max_steps":
|
| 57 |
"num_input_tokens_seen": 0,
|
| 58 |
-
"num_train_epochs":
|
| 59 |
"save_steps": 500,
|
| 60 |
"stateful_callbacks": {
|
| 61 |
"TrainerControl": {
|
|
@@ -64,18 +64,18 @@
|
|
| 64 |
"should_evaluate": false,
|
| 65 |
"should_log": false,
|
| 66 |
"should_save": true,
|
| 67 |
-
"should_training_stop":
|
| 68 |
},
|
| 69 |
"attributes": {}
|
| 70 |
}
|
| 71 |
},
|
| 72 |
-
"total_flos":
|
| 73 |
"train_batch_size": 8,
|
| 74 |
"trial_name": null,
|
| 75 |
"trial_params": {
|
| 76 |
-
"learning_rate": 1.
|
| 77 |
-
"num_train_epochs":
|
| 78 |
"per_device_train_batch_size": 8,
|
| 79 |
-
"seed":
|
| 80 |
}
|
| 81 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.8520564763658687,
|
| 3 |
+
"best_model_checkpoint": "BERTForDetectingDepression-Twitter2015/run-0/checkpoint-552",
|
| 4 |
"epoch": 4.0,
|
| 5 |
"eval_steps": 500,
|
| 6 |
"global_step": 736,
|
|
|
|
| 10 |
"log_history": [
|
| 11 |
{
|
| 12 |
"epoch": 1.0,
|
| 13 |
+
"eval_accuracy": 0.7826887661141805,
|
| 14 |
+
"eval_loss": 0.45181456208229065,
|
| 15 |
+
"eval_runtime": 1.6866,
|
| 16 |
+
"eval_samples_per_second": 965.826,
|
| 17 |
+
"eval_steps_per_second": 60.475,
|
| 18 |
"step": 184
|
| 19 |
},
|
| 20 |
{
|
| 21 |
"epoch": 2.0,
|
| 22 |
+
"eval_accuracy": 0.8238182934315531,
|
| 23 |
+
"eval_loss": 0.38157492876052856,
|
| 24 |
+
"eval_runtime": 1.6313,
|
| 25 |
+
"eval_samples_per_second": 998.618,
|
| 26 |
+
"eval_steps_per_second": 62.529,
|
| 27 |
"step": 368
|
| 28 |
},
|
| 29 |
{
|
| 30 |
"epoch": 2.717391304347826,
|
| 31 |
+
"grad_norm": 19.13737678527832,
|
| 32 |
+
"learning_rate": 3.4685699626862464e-06,
|
| 33 |
+
"loss": 0.3706,
|
| 34 |
"step": 500
|
| 35 |
},
|
| 36 |
{
|
| 37 |
"epoch": 3.0,
|
| 38 |
+
"eval_accuracy": 0.8520564763658687,
|
| 39 |
+
"eval_loss": 0.3886027932167053,
|
| 40 |
+
"eval_runtime": 1.6798,
|
| 41 |
+
"eval_samples_per_second": 969.762,
|
| 42 |
+
"eval_steps_per_second": 60.722,
|
| 43 |
"step": 552
|
| 44 |
},
|
| 45 |
{
|
| 46 |
"epoch": 4.0,
|
| 47 |
+
"eval_accuracy": 0.8520564763658687,
|
| 48 |
+
"eval_loss": 0.414229154586792,
|
| 49 |
+
"eval_runtime": 1.6727,
|
| 50 |
+
"eval_samples_per_second": 973.864,
|
| 51 |
+
"eval_steps_per_second": 60.979,
|
| 52 |
"step": 736
|
| 53 |
}
|
| 54 |
],
|
| 55 |
"logging_steps": 500,
|
| 56 |
+
"max_steps": 736,
|
| 57 |
"num_input_tokens_seen": 0,
|
| 58 |
+
"num_train_epochs": 4,
|
| 59 |
"save_steps": 500,
|
| 60 |
"stateful_callbacks": {
|
| 61 |
"TrainerControl": {
|
|
|
|
| 64 |
"should_evaluate": false,
|
| 65 |
"should_log": false,
|
| 66 |
"should_save": true,
|
| 67 |
+
"should_training_stop": true
|
| 68 |
},
|
| 69 |
"attributes": {}
|
| 70 |
}
|
| 71 |
},
|
| 72 |
+
"total_flos": 104085089056320.0,
|
| 73 |
"train_batch_size": 8,
|
| 74 |
"trial_name": null,
|
| 75 |
"trial_params": {
|
| 76 |
+
"learning_rate": 1.0817235137868972e-05,
|
| 77 |
+
"num_train_epochs": 4,
|
| 78 |
"per_device_train_batch_size": 8,
|
| 79 |
+
"seed": 5
|
| 80 |
}
|
| 81 |
}
|
run-0/checkpoint-736/training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5176
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3dca1282896774eec430907d299b265f87badfd78ce3fae95cf77e64c477252a
|
| 3 |
size 5176
|
run-0/checkpoint-736/vocab.txt
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
run-1/checkpoint-138/config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "
|
| 3 |
"architectures": [
|
| 4 |
"BertForSequenceClassification"
|
| 5 |
],
|
|
@@ -23,5 +23,5 @@
|
|
| 23 |
"transformers_version": "4.42.4",
|
| 24 |
"type_vocab_size": 2,
|
| 25 |
"use_cache": true,
|
| 26 |
-
"vocab_size":
|
| 27 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "AIMH/mental-bert-base-cased",
|
| 3 |
"architectures": [
|
| 4 |
"BertForSequenceClassification"
|
| 5 |
],
|
|
|
|
| 23 |
"transformers_version": "4.42.4",
|
| 24 |
"type_vocab_size": 2,
|
| 25 |
"use_cache": true,
|
| 26 |
+
"vocab_size": 28996
|
| 27 |
}
|
run-1/checkpoint-138/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:633dbc2c8e981deacd7b99da5f5c4c9bcdb73230ab14686bda24612491410d67
|
| 3 |
+
size 433270768
|
run-1/checkpoint-138/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5a7aeecf2cd0e13d0bd81cb8043bdcf28d265abb9b5897a7f0b24f2a7a30ca10
|
| 3 |
+
size 866662650
|
run-1/checkpoint-138/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c4673f91a41f7ae8a78df03aba2d1037bf40dc9118a2b888b0da54d02fa0a9df
|
| 3 |
+
size 14244
|
run-1/checkpoint-138/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3c752a24e0c2337854c356490b3502f10b0d4695aedf184215315bb4e0ef1e0a
|
| 3 |
size 1064
|