Delete Finetune-species/super_binary_out
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/config.json +0 -27
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/model.safetensors +0 -3
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/optimizer.pt +0 -3
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/rng_state.pth +0 -3
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/scheduler.pt +0 -3
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/special_tokens_map.json +0 -7
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/tokenizer.json +0 -0
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/tokenizer_config.json +0 -56
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/trainer_state.json +0 -32
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/training_args.bin +0 -3
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/config.json +0 -27
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/model.safetensors +0 -3
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/optimizer.pt +0 -3
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/rng_state.pth +0 -3
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/scheduler.pt +0 -3
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/special_tokens_map.json +0 -7
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/tokenizer.json +0 -0
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/tokenizer_config.json +0 -56
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/trainer_state.json +0 -45
- Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/training_args.bin +0 -3
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/config.json +0 -27
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/model.safetensors +0 -3
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/optimizer.pt +0 -3
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/rng_state.pth +0 -3
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/scheduler.pt +0 -3
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/special_tokens_map.json +0 -7
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/tokenizer.json +0 -0
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/tokenizer_config.json +0 -56
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/trainer_state.json +0 -32
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/training_args.bin +0 -3
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/config.json +0 -27
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/model.safetensors +0 -3
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/optimizer.pt +0 -3
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/rng_state.pth +0 -3
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/scheduler.pt +0 -3
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/special_tokens_map.json +0 -7
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/tokenizer.json +0 -0
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/tokenizer_config.json +0 -56
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/trainer_state.json +0 -45
- Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/training_args.bin +0 -3
- Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/config.json +0 -27
- Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/model.safetensors +0 -3
- Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/optimizer.pt +0 -3
- Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/rng_state.pth +0 -3
- Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/scheduler.pt +0 -3
- Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/special_tokens_map.json +0 -7
- Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/tokenizer.json +0 -0
- Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/tokenizer_config.json +0 -15
- Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/trainer_state.json +0 -110
- Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/training_args.bin +0 -3
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/config.json
DELETED
|
@@ -1,27 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"_name_or_path": "/root/NaN/dna-tokenizer/pretrain/models/base_3072/checkpoint-100000",
|
| 3 |
-
"architectures": [
|
| 4 |
-
"BertForSequenceClassification"
|
| 5 |
-
],
|
| 6 |
-
"attention_probs_dropout_prob": 0.1,
|
| 7 |
-
"classifier_dropout": null,
|
| 8 |
-
"hidden_act": "gelu",
|
| 9 |
-
"hidden_dropout_prob": 0.1,
|
| 10 |
-
"hidden_size": 768,
|
| 11 |
-
"initializer_range": 0.02,
|
| 12 |
-
"intermediate_size": 3072,
|
| 13 |
-
"layer_norm_eps": 1e-12,
|
| 14 |
-
"max_length": 512,
|
| 15 |
-
"max_position_embeddings": 512,
|
| 16 |
-
"model_type": "bert",
|
| 17 |
-
"num_attention_heads": 12,
|
| 18 |
-
"num_hidden_layers": 12,
|
| 19 |
-
"pad_token_id": 0,
|
| 20 |
-
"position_embedding_type": "absolute",
|
| 21 |
-
"problem_type": "single_label_classification",
|
| 22 |
-
"torch_dtype": "float32",
|
| 23 |
-
"transformers_version": "4.35.2",
|
| 24 |
-
"type_vocab_size": 2,
|
| 25 |
-
"use_cache": true,
|
| 26 |
-
"vocab_size": 3072
|
| 27 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/model.safetensors
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:f5fa49caa81ea407bc5c8a5e0602b463a935b2a9661fea7f0478865a1609bf44
|
| 3 |
-
size 353632152
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/optimizer.pt
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:0efacfec096f8ac618c7f0376e8100a84f954fc6185819222ce51deeedbe8d60
|
| 3 |
-
size 707385995
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/rng_state.pth
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:4aa69e954458f68c3b809c27ada0da01d92639b6b332b918900199c6023c4dfd
|
| 3 |
-
size 14709
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/scheduler.pt
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:26a279ce49d50e5c80bc88f986b50d6cff25209e4a7ed53a76c5895b94a339dd
|
| 3 |
-
size 1465
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/special_tokens_map.json
DELETED
|
@@ -1,7 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"cls_token": "[CLS]",
|
| 3 |
-
"mask_token": "[MASK]",
|
| 4 |
-
"pad_token": "[PAD]",
|
| 5 |
-
"sep_token": "[SEP]",
|
| 6 |
-
"unk_token": "[UNK]"
|
| 7 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/tokenizer.json
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/tokenizer_config.json
DELETED
|
@@ -1,56 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"added_tokens_decoder": {
|
| 3 |
-
"0": {
|
| 4 |
-
"content": "[PAD]",
|
| 5 |
-
"lstrip": false,
|
| 6 |
-
"normalized": false,
|
| 7 |
-
"rstrip": false,
|
| 8 |
-
"single_word": false,
|
| 9 |
-
"special": true
|
| 10 |
-
},
|
| 11 |
-
"1": {
|
| 12 |
-
"content": "[UNK]",
|
| 13 |
-
"lstrip": false,
|
| 14 |
-
"normalized": false,
|
| 15 |
-
"rstrip": false,
|
| 16 |
-
"single_word": false,
|
| 17 |
-
"special": true
|
| 18 |
-
},
|
| 19 |
-
"2": {
|
| 20 |
-
"content": "[CLS]",
|
| 21 |
-
"lstrip": false,
|
| 22 |
-
"normalized": false,
|
| 23 |
-
"rstrip": false,
|
| 24 |
-
"single_word": false,
|
| 25 |
-
"special": true
|
| 26 |
-
},
|
| 27 |
-
"3": {
|
| 28 |
-
"content": "[SEP]",
|
| 29 |
-
"lstrip": false,
|
| 30 |
-
"normalized": false,
|
| 31 |
-
"rstrip": false,
|
| 32 |
-
"single_word": false,
|
| 33 |
-
"special": true
|
| 34 |
-
},
|
| 35 |
-
"4": {
|
| 36 |
-
"content": "[MASK]",
|
| 37 |
-
"lstrip": false,
|
| 38 |
-
"normalized": false,
|
| 39 |
-
"rstrip": false,
|
| 40 |
-
"single_word": false,
|
| 41 |
-
"special": true
|
| 42 |
-
}
|
| 43 |
-
},
|
| 44 |
-
"cache_dir": null,
|
| 45 |
-
"clean_up_tokenization_spaces": true,
|
| 46 |
-
"cls_token": "[CLS]",
|
| 47 |
-
"mask_token": "[MASK]",
|
| 48 |
-
"model_max_length": 200,
|
| 49 |
-
"pad_token": "[PAD]",
|
| 50 |
-
"padding_side": "right",
|
| 51 |
-
"sep_token": "[SEP]",
|
| 52 |
-
"tokenizer_class": "PreTrainedTokenizerFast",
|
| 53 |
-
"trust_remote_code": true,
|
| 54 |
-
"unk_token": "[UNK]",
|
| 55 |
-
"use_fast": true
|
| 56 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/trainer_state.json
DELETED
|
@@ -1,32 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"best_metric": null,
|
| 3 |
-
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.8322929671244278,
|
| 5 |
-
"eval_steps": 2000,
|
| 6 |
-
"global_step": 2000,
|
| 7 |
-
"is_hyper_param_search": false,
|
| 8 |
-
"is_local_process_zero": true,
|
| 9 |
-
"is_world_process_zero": true,
|
| 10 |
-
"log_history": [
|
| 11 |
-
{
|
| 12 |
-
"epoch": 0.83,
|
| 13 |
-
"eval_accuracy": 0.7232121953122967,
|
| 14 |
-
"eval_f1": 0.680500985958512,
|
| 15 |
-
"eval_loss": 0.5397241115570068,
|
| 16 |
-
"eval_matthews_correlation": 0.3991911550967513,
|
| 17 |
-
"eval_precision": 0.7268894858780544,
|
| 18 |
-
"eval_recall": 0.6755850185066816,
|
| 19 |
-
"eval_runtime": 6.4825,
|
| 20 |
-
"eval_samples_per_second": 5929.926,
|
| 21 |
-
"eval_steps_per_second": 46.432,
|
| 22 |
-
"step": 2000
|
| 23 |
-
}
|
| 24 |
-
],
|
| 25 |
-
"logging_steps": 100000,
|
| 26 |
-
"max_steps": 19224,
|
| 27 |
-
"num_train_epochs": 8,
|
| 28 |
-
"save_steps": 2000,
|
| 29 |
-
"total_flos": 2.6311104946765824e+16,
|
| 30 |
-
"trial_name": null,
|
| 31 |
-
"trial_params": null
|
| 32 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/training_args.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:9846bc17ea30367d7691aef3ec4a8eae9c29225f580075721efd19eaa9d65c75
|
| 3 |
-
size 5265
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/config.json
DELETED
|
@@ -1,27 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"_name_or_path": "/root/NaN/dna-tokenizer/pretrain/models/base_3072/checkpoint-100000",
|
| 3 |
-
"architectures": [
|
| 4 |
-
"BertForSequenceClassification"
|
| 5 |
-
],
|
| 6 |
-
"attention_probs_dropout_prob": 0.1,
|
| 7 |
-
"classifier_dropout": null,
|
| 8 |
-
"hidden_act": "gelu",
|
| 9 |
-
"hidden_dropout_prob": 0.1,
|
| 10 |
-
"hidden_size": 768,
|
| 11 |
-
"initializer_range": 0.02,
|
| 12 |
-
"intermediate_size": 3072,
|
| 13 |
-
"layer_norm_eps": 1e-12,
|
| 14 |
-
"max_length": 512,
|
| 15 |
-
"max_position_embeddings": 512,
|
| 16 |
-
"model_type": "bert",
|
| 17 |
-
"num_attention_heads": 12,
|
| 18 |
-
"num_hidden_layers": 12,
|
| 19 |
-
"pad_token_id": 0,
|
| 20 |
-
"position_embedding_type": "absolute",
|
| 21 |
-
"problem_type": "single_label_classification",
|
| 22 |
-
"torch_dtype": "float32",
|
| 23 |
-
"transformers_version": "4.35.2",
|
| 24 |
-
"type_vocab_size": 2,
|
| 25 |
-
"use_cache": true,
|
| 26 |
-
"vocab_size": 3072
|
| 27 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/model.safetensors
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:d323c9f0c20d5297cc07df751fb4d897c31a811853c190b42918b235b3c4e8f1
|
| 3 |
-
size 353632152
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/optimizer.pt
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:7a8fb249d5d7f21f763ddfacdee4529a0987646ac1cf02b80d3dad4cd155d789
|
| 3 |
-
size 707385995
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/rng_state.pth
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:022888c8d72c2809ac5a3da8c8ed4390df94dd94ad785d29ade33f64674216b7
|
| 3 |
-
size 14709
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/scheduler.pt
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:d0f78c7b09d5edf997a2ef80368655d4d1525769cf0b855b3af5043e564a39d1
|
| 3 |
-
size 1465
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/special_tokens_map.json
DELETED
|
@@ -1,7 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"cls_token": "[CLS]",
|
| 3 |
-
"mask_token": "[MASK]",
|
| 4 |
-
"pad_token": "[PAD]",
|
| 5 |
-
"sep_token": "[SEP]",
|
| 6 |
-
"unk_token": "[UNK]"
|
| 7 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/tokenizer.json
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/tokenizer_config.json
DELETED
|
@@ -1,56 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"added_tokens_decoder": {
|
| 3 |
-
"0": {
|
| 4 |
-
"content": "[PAD]",
|
| 5 |
-
"lstrip": false,
|
| 6 |
-
"normalized": false,
|
| 7 |
-
"rstrip": false,
|
| 8 |
-
"single_word": false,
|
| 9 |
-
"special": true
|
| 10 |
-
},
|
| 11 |
-
"1": {
|
| 12 |
-
"content": "[UNK]",
|
| 13 |
-
"lstrip": false,
|
| 14 |
-
"normalized": false,
|
| 15 |
-
"rstrip": false,
|
| 16 |
-
"single_word": false,
|
| 17 |
-
"special": true
|
| 18 |
-
},
|
| 19 |
-
"2": {
|
| 20 |
-
"content": "[CLS]",
|
| 21 |
-
"lstrip": false,
|
| 22 |
-
"normalized": false,
|
| 23 |
-
"rstrip": false,
|
| 24 |
-
"single_word": false,
|
| 25 |
-
"special": true
|
| 26 |
-
},
|
| 27 |
-
"3": {
|
| 28 |
-
"content": "[SEP]",
|
| 29 |
-
"lstrip": false,
|
| 30 |
-
"normalized": false,
|
| 31 |
-
"rstrip": false,
|
| 32 |
-
"single_word": false,
|
| 33 |
-
"special": true
|
| 34 |
-
},
|
| 35 |
-
"4": {
|
| 36 |
-
"content": "[MASK]",
|
| 37 |
-
"lstrip": false,
|
| 38 |
-
"normalized": false,
|
| 39 |
-
"rstrip": false,
|
| 40 |
-
"single_word": false,
|
| 41 |
-
"special": true
|
| 42 |
-
}
|
| 43 |
-
},
|
| 44 |
-
"cache_dir": null,
|
| 45 |
-
"clean_up_tokenization_spaces": true,
|
| 46 |
-
"cls_token": "[CLS]",
|
| 47 |
-
"mask_token": "[MASK]",
|
| 48 |
-
"model_max_length": 200,
|
| 49 |
-
"pad_token": "[PAD]",
|
| 50 |
-
"padding_side": "right",
|
| 51 |
-
"sep_token": "[SEP]",
|
| 52 |
-
"tokenizer_class": "PreTrainedTokenizerFast",
|
| 53 |
-
"trust_remote_code": true,
|
| 54 |
-
"unk_token": "[UNK]",
|
| 55 |
-
"use_fast": true
|
| 56 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/trainer_state.json
DELETED
|
@@ -1,45 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"best_metric": null,
|
| 3 |
-
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 1.6645859342488556,
|
| 5 |
-
"eval_steps": 2000,
|
| 6 |
-
"global_step": 4000,
|
| 7 |
-
"is_hyper_param_search": false,
|
| 8 |
-
"is_local_process_zero": true,
|
| 9 |
-
"is_world_process_zero": true,
|
| 10 |
-
"log_history": [
|
| 11 |
-
{
|
| 12 |
-
"epoch": 0.83,
|
| 13 |
-
"eval_accuracy": 0.7232121953122967,
|
| 14 |
-
"eval_f1": 0.680500985958512,
|
| 15 |
-
"eval_loss": 0.5397241115570068,
|
| 16 |
-
"eval_matthews_correlation": 0.3991911550967513,
|
| 17 |
-
"eval_precision": 0.7268894858780544,
|
| 18 |
-
"eval_recall": 0.6755850185066816,
|
| 19 |
-
"eval_runtime": 6.4825,
|
| 20 |
-
"eval_samples_per_second": 5929.926,
|
| 21 |
-
"eval_steps_per_second": 46.432,
|
| 22 |
-
"step": 2000
|
| 23 |
-
},
|
| 24 |
-
{
|
| 25 |
-
"epoch": 1.66,
|
| 26 |
-
"eval_accuracy": 0.721963528524232,
|
| 27 |
-
"eval_f1": 0.7135581221078293,
|
| 28 |
-
"eval_loss": 0.5439369082450867,
|
| 29 |
-
"eval_matthews_correlation": 0.43044300831847054,
|
| 30 |
-
"eval_precision": 0.7116658765696064,
|
| 31 |
-
"eval_recall": 0.7188368602594819,
|
| 32 |
-
"eval_runtime": 6.4688,
|
| 33 |
-
"eval_samples_per_second": 5942.505,
|
| 34 |
-
"eval_steps_per_second": 46.531,
|
| 35 |
-
"step": 4000
|
| 36 |
-
}
|
| 37 |
-
],
|
| 38 |
-
"logging_steps": 100000,
|
| 39 |
-
"max_steps": 19224,
|
| 40 |
-
"num_train_epochs": 8,
|
| 41 |
-
"save_steps": 2000,
|
| 42 |
-
"total_flos": 5.261655771657011e+16,
|
| 43 |
-
"trial_name": null,
|
| 44 |
-
"trial_params": null
|
| 45 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/training_args.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:9846bc17ea30367d7691aef3ec4a8eae9c29225f580075721efd19eaa9d65c75
|
| 3 |
-
size 5265
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/config.json
DELETED
|
@@ -1,27 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"_name_or_path": "/root/NaN/dna-tokenizer/pretrain/models/base_4096/checkpoint-100000",
|
| 3 |
-
"architectures": [
|
| 4 |
-
"BertForSequenceClassification"
|
| 5 |
-
],
|
| 6 |
-
"attention_probs_dropout_prob": 0.1,
|
| 7 |
-
"classifier_dropout": null,
|
| 8 |
-
"hidden_act": "gelu",
|
| 9 |
-
"hidden_dropout_prob": 0.1,
|
| 10 |
-
"hidden_size": 768,
|
| 11 |
-
"initializer_range": 0.02,
|
| 12 |
-
"intermediate_size": 3072,
|
| 13 |
-
"layer_norm_eps": 1e-12,
|
| 14 |
-
"max_length": 512,
|
| 15 |
-
"max_position_embeddings": 512,
|
| 16 |
-
"model_type": "bert",
|
| 17 |
-
"num_attention_heads": 12,
|
| 18 |
-
"num_hidden_layers": 12,
|
| 19 |
-
"pad_token_id": 0,
|
| 20 |
-
"position_embedding_type": "absolute",
|
| 21 |
-
"problem_type": "single_label_classification",
|
| 22 |
-
"torch_dtype": "float32",
|
| 23 |
-
"transformers_version": "4.35.2",
|
| 24 |
-
"type_vocab_size": 2,
|
| 25 |
-
"use_cache": true,
|
| 26 |
-
"vocab_size": 4096
|
| 27 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/model.safetensors
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:8b826da22ddb2986b902fceccdb27ab239795e8f329e4c9b05643627359effff
|
| 3 |
-
size 356777880
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/optimizer.pt
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:852c30787e26294b05d65f00119fe817445711e9d69f39981ab5b01f9c65be69
|
| 3 |
-
size 713677451
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/rng_state.pth
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:4aa69e954458f68c3b809c27ada0da01d92639b6b332b918900199c6023c4dfd
|
| 3 |
-
size 14709
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/scheduler.pt
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:26a279ce49d50e5c80bc88f986b50d6cff25209e4a7ed53a76c5895b94a339dd
|
| 3 |
-
size 1465
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/special_tokens_map.json
DELETED
|
@@ -1,7 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"cls_token": "[CLS]",
|
| 3 |
-
"mask_token": "[MASK]",
|
| 4 |
-
"pad_token": "[PAD]",
|
| 5 |
-
"sep_token": "[SEP]",
|
| 6 |
-
"unk_token": "[UNK]"
|
| 7 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/tokenizer.json
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/tokenizer_config.json
DELETED
|
@@ -1,56 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"added_tokens_decoder": {
|
| 3 |
-
"0": {
|
| 4 |
-
"content": "[PAD]",
|
| 5 |
-
"lstrip": false,
|
| 6 |
-
"normalized": false,
|
| 7 |
-
"rstrip": false,
|
| 8 |
-
"single_word": false,
|
| 9 |
-
"special": true
|
| 10 |
-
},
|
| 11 |
-
"1": {
|
| 12 |
-
"content": "[UNK]",
|
| 13 |
-
"lstrip": false,
|
| 14 |
-
"normalized": false,
|
| 15 |
-
"rstrip": false,
|
| 16 |
-
"single_word": false,
|
| 17 |
-
"special": true
|
| 18 |
-
},
|
| 19 |
-
"2": {
|
| 20 |
-
"content": "[CLS]",
|
| 21 |
-
"lstrip": false,
|
| 22 |
-
"normalized": false,
|
| 23 |
-
"rstrip": false,
|
| 24 |
-
"single_word": false,
|
| 25 |
-
"special": true
|
| 26 |
-
},
|
| 27 |
-
"3": {
|
| 28 |
-
"content": "[SEP]",
|
| 29 |
-
"lstrip": false,
|
| 30 |
-
"normalized": false,
|
| 31 |
-
"rstrip": false,
|
| 32 |
-
"single_word": false,
|
| 33 |
-
"special": true
|
| 34 |
-
},
|
| 35 |
-
"4": {
|
| 36 |
-
"content": "[MASK]",
|
| 37 |
-
"lstrip": false,
|
| 38 |
-
"normalized": false,
|
| 39 |
-
"rstrip": false,
|
| 40 |
-
"single_word": false,
|
| 41 |
-
"special": true
|
| 42 |
-
}
|
| 43 |
-
},
|
| 44 |
-
"cache_dir": null,
|
| 45 |
-
"clean_up_tokenization_spaces": true,
|
| 46 |
-
"cls_token": "[CLS]",
|
| 47 |
-
"mask_token": "[MASK]",
|
| 48 |
-
"model_max_length": 200,
|
| 49 |
-
"pad_token": "[PAD]",
|
| 50 |
-
"padding_side": "right",
|
| 51 |
-
"sep_token": "[SEP]",
|
| 52 |
-
"tokenizer_class": "PreTrainedTokenizerFast",
|
| 53 |
-
"trust_remote_code": true,
|
| 54 |
-
"unk_token": "[UNK]",
|
| 55 |
-
"use_fast": true
|
| 56 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/trainer_state.json
DELETED
|
@@ -1,32 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"best_metric": null,
|
| 3 |
-
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.8322929671244278,
|
| 5 |
-
"eval_steps": 2000,
|
| 6 |
-
"global_step": 2000,
|
| 7 |
-
"is_hyper_param_search": false,
|
| 8 |
-
"is_local_process_zero": true,
|
| 9 |
-
"is_world_process_zero": true,
|
| 10 |
-
"log_history": [
|
| 11 |
-
{
|
| 12 |
-
"epoch": 0.83,
|
| 13 |
-
"eval_accuracy": 0.7271923206992534,
|
| 14 |
-
"eval_f1": 0.6916687987201311,
|
| 15 |
-
"eval_loss": 0.5368459820747375,
|
| 16 |
-
"eval_matthews_correlation": 0.40845552866252005,
|
| 17 |
-
"eval_precision": 0.7245172413236349,
|
| 18 |
-
"eval_recall": 0.6857718341711782,
|
| 19 |
-
"eval_runtime": 6.2785,
|
| 20 |
-
"eval_samples_per_second": 6122.616,
|
| 21 |
-
"eval_steps_per_second": 47.941,
|
| 22 |
-
"step": 2000
|
| 23 |
-
}
|
| 24 |
-
],
|
| 25 |
-
"logging_steps": 100000,
|
| 26 |
-
"max_steps": 19224,
|
| 27 |
-
"num_train_epochs": 8,
|
| 28 |
-
"save_steps": 2000,
|
| 29 |
-
"total_flos": 2.6311104946765824e+16,
|
| 30 |
-
"trial_name": null,
|
| 31 |
-
"trial_params": null
|
| 32 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/training_args.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:6556056323c8c121c72de11a49668bf4bd279f2aa3f171eff1ab941b35cd7d73
|
| 3 |
-
size 5265
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/config.json
DELETED
|
@@ -1,27 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"_name_or_path": "/root/NaN/dna-tokenizer/pretrain/models/base_4096/checkpoint-100000",
|
| 3 |
-
"architectures": [
|
| 4 |
-
"BertForSequenceClassification"
|
| 5 |
-
],
|
| 6 |
-
"attention_probs_dropout_prob": 0.1,
|
| 7 |
-
"classifier_dropout": null,
|
| 8 |
-
"hidden_act": "gelu",
|
| 9 |
-
"hidden_dropout_prob": 0.1,
|
| 10 |
-
"hidden_size": 768,
|
| 11 |
-
"initializer_range": 0.02,
|
| 12 |
-
"intermediate_size": 3072,
|
| 13 |
-
"layer_norm_eps": 1e-12,
|
| 14 |
-
"max_length": 512,
|
| 15 |
-
"max_position_embeddings": 512,
|
| 16 |
-
"model_type": "bert",
|
| 17 |
-
"num_attention_heads": 12,
|
| 18 |
-
"num_hidden_layers": 12,
|
| 19 |
-
"pad_token_id": 0,
|
| 20 |
-
"position_embedding_type": "absolute",
|
| 21 |
-
"problem_type": "single_label_classification",
|
| 22 |
-
"torch_dtype": "float32",
|
| 23 |
-
"transformers_version": "4.35.2",
|
| 24 |
-
"type_vocab_size": 2,
|
| 25 |
-
"use_cache": true,
|
| 26 |
-
"vocab_size": 4096
|
| 27 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/model.safetensors
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:e8e9175d75a4cf7657e7003082bca3688d4e87967591ca1c3ce9473a50a666b8
|
| 3 |
-
size 356777880
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/optimizer.pt
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:ffdc80a05cef66ad2e24134b167749be6f2728f5f61c061c9eb0fe69d9f3166f
|
| 3 |
-
size 713677451
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/rng_state.pth
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:022888c8d72c2809ac5a3da8c8ed4390df94dd94ad785d29ade33f64674216b7
|
| 3 |
-
size 14709
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/scheduler.pt
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:d0f78c7b09d5edf997a2ef80368655d4d1525769cf0b855b3af5043e564a39d1
|
| 3 |
-
size 1465
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/special_tokens_map.json
DELETED
|
@@ -1,7 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"cls_token": "[CLS]",
|
| 3 |
-
"mask_token": "[MASK]",
|
| 4 |
-
"pad_token": "[PAD]",
|
| 5 |
-
"sep_token": "[SEP]",
|
| 6 |
-
"unk_token": "[UNK]"
|
| 7 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/tokenizer.json
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/tokenizer_config.json
DELETED
|
@@ -1,56 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"added_tokens_decoder": {
|
| 3 |
-
"0": {
|
| 4 |
-
"content": "[PAD]",
|
| 5 |
-
"lstrip": false,
|
| 6 |
-
"normalized": false,
|
| 7 |
-
"rstrip": false,
|
| 8 |
-
"single_word": false,
|
| 9 |
-
"special": true
|
| 10 |
-
},
|
| 11 |
-
"1": {
|
| 12 |
-
"content": "[UNK]",
|
| 13 |
-
"lstrip": false,
|
| 14 |
-
"normalized": false,
|
| 15 |
-
"rstrip": false,
|
| 16 |
-
"single_word": false,
|
| 17 |
-
"special": true
|
| 18 |
-
},
|
| 19 |
-
"2": {
|
| 20 |
-
"content": "[CLS]",
|
| 21 |
-
"lstrip": false,
|
| 22 |
-
"normalized": false,
|
| 23 |
-
"rstrip": false,
|
| 24 |
-
"single_word": false,
|
| 25 |
-
"special": true
|
| 26 |
-
},
|
| 27 |
-
"3": {
|
| 28 |
-
"content": "[SEP]",
|
| 29 |
-
"lstrip": false,
|
| 30 |
-
"normalized": false,
|
| 31 |
-
"rstrip": false,
|
| 32 |
-
"single_word": false,
|
| 33 |
-
"special": true
|
| 34 |
-
},
|
| 35 |
-
"4": {
|
| 36 |
-
"content": "[MASK]",
|
| 37 |
-
"lstrip": false,
|
| 38 |
-
"normalized": false,
|
| 39 |
-
"rstrip": false,
|
| 40 |
-
"single_word": false,
|
| 41 |
-
"special": true
|
| 42 |
-
}
|
| 43 |
-
},
|
| 44 |
-
"cache_dir": null,
|
| 45 |
-
"clean_up_tokenization_spaces": true,
|
| 46 |
-
"cls_token": "[CLS]",
|
| 47 |
-
"mask_token": "[MASK]",
|
| 48 |
-
"model_max_length": 200,
|
| 49 |
-
"pad_token": "[PAD]",
|
| 50 |
-
"padding_side": "right",
|
| 51 |
-
"sep_token": "[SEP]",
|
| 52 |
-
"tokenizer_class": "PreTrainedTokenizerFast",
|
| 53 |
-
"trust_remote_code": true,
|
| 54 |
-
"unk_token": "[UNK]",
|
| 55 |
-
"use_fast": true
|
| 56 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/trainer_state.json
DELETED
|
@@ -1,45 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"best_metric": null,
|
| 3 |
-
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 1.6645859342488556,
|
| 5 |
-
"eval_steps": 2000,
|
| 6 |
-
"global_step": 4000,
|
| 7 |
-
"is_hyper_param_search": false,
|
| 8 |
-
"is_local_process_zero": true,
|
| 9 |
-
"is_world_process_zero": true,
|
| 10 |
-
"log_history": [
|
| 11 |
-
{
|
| 12 |
-
"epoch": 0.83,
|
| 13 |
-
"eval_accuracy": 0.7271923206992534,
|
| 14 |
-
"eval_f1": 0.6916687987201311,
|
| 15 |
-
"eval_loss": 0.5368459820747375,
|
| 16 |
-
"eval_matthews_correlation": 0.40845552866252005,
|
| 17 |
-
"eval_precision": 0.7245172413236349,
|
| 18 |
-
"eval_recall": 0.6857718341711782,
|
| 19 |
-
"eval_runtime": 6.2785,
|
| 20 |
-
"eval_samples_per_second": 6122.616,
|
| 21 |
-
"eval_steps_per_second": 47.941,
|
| 22 |
-
"step": 2000
|
| 23 |
-
},
|
| 24 |
-
{
|
| 25 |
-
"epoch": 1.66,
|
| 26 |
-
"eval_accuracy": 0.7199604588850446,
|
| 27 |
-
"eval_f1": 0.7113427696435267,
|
| 28 |
-
"eval_loss": 0.5453317165374756,
|
| 29 |
-
"eval_matthews_correlation": 0.42579813599678606,
|
| 30 |
-
"eval_precision": 0.7094611286264896,
|
| 31 |
-
"eval_recall": 0.716393435153353,
|
| 32 |
-
"eval_runtime": 6.2677,
|
| 33 |
-
"eval_samples_per_second": 6133.182,
|
| 34 |
-
"eval_steps_per_second": 48.024,
|
| 35 |
-
"step": 4000
|
| 36 |
-
}
|
| 37 |
-
],
|
| 38 |
-
"logging_steps": 100000,
|
| 39 |
-
"max_steps": 19224,
|
| 40 |
-
"num_train_epochs": 8,
|
| 41 |
-
"save_steps": 2000,
|
| 42 |
-
"total_flos": 5.261655771657011e+16,
|
| 43 |
-
"trial_name": null,
|
| 44 |
-
"trial_params": null
|
| 45 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/training_args.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:6556056323c8c121c72de11a49668bf4bd279f2aa3f171eff1ab941b35cd7d73
|
| 3 |
-
size 5265
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/config.json
DELETED
|
@@ -1,27 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"_name_or_path": "/root/NaN/dna-tokenizer/pretrain/models/model_len2_4096/checkpoint-100000",
|
| 3 |
-
"architectures": [
|
| 4 |
-
"BertForSequenceClassification"
|
| 5 |
-
],
|
| 6 |
-
"attention_probs_dropout_prob": 0.1,
|
| 7 |
-
"classifier_dropout": null,
|
| 8 |
-
"hidden_act": "gelu",
|
| 9 |
-
"hidden_dropout_prob": 0.1,
|
| 10 |
-
"hidden_size": 768,
|
| 11 |
-
"initializer_range": 0.02,
|
| 12 |
-
"intermediate_size": 3072,
|
| 13 |
-
"layer_norm_eps": 1e-12,
|
| 14 |
-
"max_length": 512,
|
| 15 |
-
"max_position_embeddings": 512,
|
| 16 |
-
"model_type": "bert",
|
| 17 |
-
"num_attention_heads": 12,
|
| 18 |
-
"num_hidden_layers": 12,
|
| 19 |
-
"pad_token_id": 0,
|
| 20 |
-
"position_embedding_type": "absolute",
|
| 21 |
-
"problem_type": "single_label_classification",
|
| 22 |
-
"torch_dtype": "float32",
|
| 23 |
-
"transformers_version": "4.35.2",
|
| 24 |
-
"type_vocab_size": 2,
|
| 25 |
-
"use_cache": true,
|
| 26 |
-
"vocab_size": 4096
|
| 27 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/model.safetensors
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:c9820845952f61cb09e84c1cb783cf1897b8d87077036dfe3a2b923b6c1987ee
|
| 3 |
-
size 356777880
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/optimizer.pt
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:9736aa0cc2060da3baaf882df1d9e80995edf78213162734bd83aded19ae42e0
|
| 3 |
-
size 713677451
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/rng_state.pth
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:1b2060874b23673bac4caf21be1d37304c08dcec292418c4a8e6d201f0fc1932
|
| 3 |
-
size 14709
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/scheduler.pt
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:7ba381e07add5e135217829284fed29b78d6b593ab562e0cb3993675f5b72f6e
|
| 3 |
-
size 1465
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/special_tokens_map.json
DELETED
|
@@ -1,7 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"cls_token": "[CLS]",
|
| 3 |
-
"mask_token": "[MASK]",
|
| 4 |
-
"pad_token": "[PAD]",
|
| 5 |
-
"sep_token": "[SEP]",
|
| 6 |
-
"unk_token": "[UNK]"
|
| 7 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/tokenizer.json
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/tokenizer_config.json
DELETED
|
@@ -1,15 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"added_tokens_decoder": {},
|
| 3 |
-
"cache_dir": null,
|
| 4 |
-
"clean_up_tokenization_spaces": true,
|
| 5 |
-
"cls_token": "[CLS]",
|
| 6 |
-
"mask_token": "[MASK]",
|
| 7 |
-
"model_max_length": 200,
|
| 8 |
-
"pad_token": "[PAD]",
|
| 9 |
-
"padding_side": "right",
|
| 10 |
-
"sep_token": "[SEP]",
|
| 11 |
-
"tokenizer_class": "PreTrainedTokenizerFast",
|
| 12 |
-
"trust_remote_code": true,
|
| 13 |
-
"unk_token": "[UNK]",
|
| 14 |
-
"use_fast": true
|
| 15 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/trainer_state.json
DELETED
|
@@ -1,110 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"best_metric": null,
|
| 3 |
-
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 5.826050769870995,
|
| 5 |
-
"eval_steps": 2000,
|
| 6 |
-
"global_step": 14000,
|
| 7 |
-
"is_hyper_param_search": false,
|
| 8 |
-
"is_local_process_zero": true,
|
| 9 |
-
"is_world_process_zero": true,
|
| 10 |
-
"log_history": [
|
| 11 |
-
{
|
| 12 |
-
"epoch": 0.83,
|
| 13 |
-
"eval_accuracy": 0.7252152649514841,
|
| 14 |
-
"eval_f1": 0.6902950682902858,
|
| 15 |
-
"eval_loss": 0.5391069054603577,
|
| 16 |
-
"eval_matthews_correlation": 0.40402463402188366,
|
| 17 |
-
"eval_precision": 0.7211497658648482,
|
| 18 |
-
"eval_recall": 0.6845309492620895,
|
| 19 |
-
"eval_runtime": 6.2755,
|
| 20 |
-
"eval_samples_per_second": 6125.577,
|
| 21 |
-
"eval_steps_per_second": 47.964,
|
| 22 |
-
"step": 2000
|
| 23 |
-
},
|
| 24 |
-
{
|
| 25 |
-
"epoch": 1.66,
|
| 26 |
-
"eval_accuracy": 0.7247470149059598,
|
| 27 |
-
"eval_f1": 0.7099661938281203,
|
| 28 |
-
"eval_loss": 0.5385260581970215,
|
| 29 |
-
"eval_matthews_correlation": 0.42001605877551784,
|
| 30 |
-
"eval_precision": 0.7107448921110509,
|
| 31 |
-
"eval_recall": 0.709273743081231,
|
| 32 |
-
"eval_runtime": 6.2437,
|
| 33 |
-
"eval_samples_per_second": 6156.719,
|
| 34 |
-
"eval_steps_per_second": 48.208,
|
| 35 |
-
"step": 4000
|
| 36 |
-
},
|
| 37 |
-
{
|
| 38 |
-
"epoch": 2.5,
|
| 39 |
-
"eval_accuracy": 0.7316667100231523,
|
| 40 |
-
"eval_f1": 0.7065806776390386,
|
| 41 |
-
"eval_loss": 0.5326813459396362,
|
| 42 |
-
"eval_matthews_correlation": 0.42215521735721945,
|
| 43 |
-
"eval_precision": 0.7213784019572298,
|
| 44 |
-
"eval_recall": 0.7012561139278983,
|
| 45 |
-
"eval_runtime": 6.2526,
|
| 46 |
-
"eval_samples_per_second": 6147.973,
|
| 47 |
-
"eval_steps_per_second": 48.14,
|
| 48 |
-
"step": 6000
|
| 49 |
-
},
|
| 50 |
-
{
|
| 51 |
-
"epoch": 3.33,
|
| 52 |
-
"eval_accuracy": 0.7316406961317343,
|
| 53 |
-
"eval_f1": 0.707569021943413,
|
| 54 |
-
"eval_loss": 0.5344680547714233,
|
| 55 |
-
"eval_matthews_correlation": 0.422807683821826,
|
| 56 |
-
"eval_precision": 0.7207127636727513,
|
| 57 |
-
"eval_recall": 0.7024875391482028,
|
| 58 |
-
"eval_runtime": 6.253,
|
| 59 |
-
"eval_samples_per_second": 6147.608,
|
| 60 |
-
"eval_steps_per_second": 48.137,
|
| 61 |
-
"step": 8000
|
| 62 |
-
},
|
| 63 |
-
{
|
| 64 |
-
"epoch": 4.16,
|
| 65 |
-
"eval_accuracy": 0.731068390520538,
|
| 66 |
-
"eval_f1": 0.708860051313849,
|
| 67 |
-
"eval_loss": 0.5406153202056885,
|
| 68 |
-
"eval_matthews_correlation": 0.4231214190533388,
|
| 69 |
-
"eval_precision": 0.7190318594698323,
|
| 70 |
-
"eval_recall": 0.7043443995944909,
|
| 71 |
-
"eval_runtime": 6.257,
|
| 72 |
-
"eval_samples_per_second": 6143.684,
|
| 73 |
-
"eval_steps_per_second": 48.106,
|
| 74 |
-
"step": 10000
|
| 75 |
-
},
|
| 76 |
-
{
|
| 77 |
-
"epoch": 4.99,
|
| 78 |
-
"eval_accuracy": 0.7305220988007596,
|
| 79 |
-
"eval_f1": 0.7136770886356509,
|
| 80 |
-
"eval_loss": 0.5410209894180298,
|
| 81 |
-
"eval_matthews_correlation": 0.4282737366015945,
|
| 82 |
-
"eval_precision": 0.7168227358137622,
|
| 83 |
-
"eval_recall": 0.7114842716727794,
|
| 84 |
-
"eval_runtime": 6.241,
|
| 85 |
-
"eval_samples_per_second": 6159.471,
|
| 86 |
-
"eval_steps_per_second": 48.23,
|
| 87 |
-
"step": 12000
|
| 88 |
-
},
|
| 89 |
-
{
|
| 90 |
-
"epoch": 5.83,
|
| 91 |
-
"eval_accuracy": 0.7284930152701543,
|
| 92 |
-
"eval_f1": 0.7061587739629024,
|
| 93 |
-
"eval_loss": 0.5537341237068176,
|
| 94 |
-
"eval_matthews_correlation": 0.41760147710952444,
|
| 95 |
-
"eval_precision": 0.7160942515527404,
|
| 96 |
-
"eval_recall": 0.7017533928262485,
|
| 97 |
-
"eval_runtime": 6.2407,
|
| 98 |
-
"eval_samples_per_second": 6159.732,
|
| 99 |
-
"eval_steps_per_second": 48.232,
|
| 100 |
-
"step": 14000
|
| 101 |
-
}
|
| 102 |
-
],
|
| 103 |
-
"logging_steps": 100000,
|
| 104 |
-
"max_steps": 19224,
|
| 105 |
-
"num_train_epochs": 8,
|
| 106 |
-
"save_steps": 2000,
|
| 107 |
-
"total_flos": 1.841494737425531e+17,
|
| 108 |
-
"trial_name": null,
|
| 109 |
-
"trial_params": null
|
| 110 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/training_args.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:0f82b451f8e860b9e6a2b1ea0895aa9da11751687e479b2bf702e23e26f7a518
|
| 3 |
-
size 5265
|
|
|
|
|
|
|
|
|
|
|
|