nancyH commited on
Commit
5d9baec
·
verified ·
1 Parent(s): e574823

Delete Finetune-species/super_binary_out

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/config.json +0 -27
  2. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/model.safetensors +0 -3
  3. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/optimizer.pt +0 -3
  4. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/rng_state.pth +0 -3
  5. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/scheduler.pt +0 -3
  6. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/special_tokens_map.json +0 -7
  7. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/tokenizer.json +0 -0
  8. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/tokenizer_config.json +0 -56
  9. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/trainer_state.json +0 -32
  10. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/training_args.bin +0 -3
  11. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/config.json +0 -27
  12. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/model.safetensors +0 -3
  13. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/optimizer.pt +0 -3
  14. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/rng_state.pth +0 -3
  15. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/scheduler.pt +0 -3
  16. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/special_tokens_map.json +0 -7
  17. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/tokenizer.json +0 -0
  18. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/tokenizer_config.json +0 -56
  19. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/trainer_state.json +0 -45
  20. Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/training_args.bin +0 -3
  21. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/config.json +0 -27
  22. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/model.safetensors +0 -3
  23. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/optimizer.pt +0 -3
  24. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/rng_state.pth +0 -3
  25. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/scheduler.pt +0 -3
  26. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/special_tokens_map.json +0 -7
  27. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/tokenizer.json +0 -0
  28. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/tokenizer_config.json +0 -56
  29. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/trainer_state.json +0 -32
  30. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/training_args.bin +0 -3
  31. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/config.json +0 -27
  32. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/model.safetensors +0 -3
  33. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/optimizer.pt +0 -3
  34. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/rng_state.pth +0 -3
  35. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/scheduler.pt +0 -3
  36. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/special_tokens_map.json +0 -7
  37. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/tokenizer.json +0 -0
  38. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/tokenizer_config.json +0 -56
  39. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/trainer_state.json +0 -45
  40. Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/training_args.bin +0 -3
  41. Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/config.json +0 -27
  42. Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/model.safetensors +0 -3
  43. Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/optimizer.pt +0 -3
  44. Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/rng_state.pth +0 -3
  45. Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/scheduler.pt +0 -3
  46. Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/special_tokens_map.json +0 -7
  47. Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/tokenizer.json +0 -0
  48. Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/tokenizer_config.json +0 -15
  49. Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/trainer_state.json +0 -110
  50. Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/training_args.bin +0 -3
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/config.json DELETED
@@ -1,27 +0,0 @@
1
- {
2
- "_name_or_path": "/root/NaN/dna-tokenizer/pretrain/models/base_3072/checkpoint-100000",
3
- "architectures": [
4
- "BertForSequenceClassification"
5
- ],
6
- "attention_probs_dropout_prob": 0.1,
7
- "classifier_dropout": null,
8
- "hidden_act": "gelu",
9
- "hidden_dropout_prob": 0.1,
10
- "hidden_size": 768,
11
- "initializer_range": 0.02,
12
- "intermediate_size": 3072,
13
- "layer_norm_eps": 1e-12,
14
- "max_length": 512,
15
- "max_position_embeddings": 512,
16
- "model_type": "bert",
17
- "num_attention_heads": 12,
18
- "num_hidden_layers": 12,
19
- "pad_token_id": 0,
20
- "position_embedding_type": "absolute",
21
- "problem_type": "single_label_classification",
22
- "torch_dtype": "float32",
23
- "transformers_version": "4.35.2",
24
- "type_vocab_size": 2,
25
- "use_cache": true,
26
- "vocab_size": 3072
27
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f5fa49caa81ea407bc5c8a5e0602b463a935b2a9661fea7f0478865a1609bf44
3
- size 353632152
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:0efacfec096f8ac618c7f0376e8100a84f954fc6185819222ce51deeedbe8d60
3
- size 707385995
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:4aa69e954458f68c3b809c27ada0da01d92639b6b332b918900199c6023c4dfd
3
- size 14709
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:26a279ce49d50e5c80bc88f986b50d6cff25209e4a7ed53a76c5895b94a339dd
3
- size 1465
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/special_tokens_map.json DELETED
@@ -1,7 +0,0 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
7
- }
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/tokenizer_config.json DELETED
@@ -1,56 +0,0 @@
1
- {
2
- "added_tokens_decoder": {
3
- "0": {
4
- "content": "[PAD]",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false,
9
- "special": true
10
- },
11
- "1": {
12
- "content": "[UNK]",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": true
18
- },
19
- "2": {
20
- "content": "[CLS]",
21
- "lstrip": false,
22
- "normalized": false,
23
- "rstrip": false,
24
- "single_word": false,
25
- "special": true
26
- },
27
- "3": {
28
- "content": "[SEP]",
29
- "lstrip": false,
30
- "normalized": false,
31
- "rstrip": false,
32
- "single_word": false,
33
- "special": true
34
- },
35
- "4": {
36
- "content": "[MASK]",
37
- "lstrip": false,
38
- "normalized": false,
39
- "rstrip": false,
40
- "single_word": false,
41
- "special": true
42
- }
43
- },
44
- "cache_dir": null,
45
- "clean_up_tokenization_spaces": true,
46
- "cls_token": "[CLS]",
47
- "mask_token": "[MASK]",
48
- "model_max_length": 200,
49
- "pad_token": "[PAD]",
50
- "padding_side": "right",
51
- "sep_token": "[SEP]",
52
- "tokenizer_class": "PreTrainedTokenizerFast",
53
- "trust_remote_code": true,
54
- "unk_token": "[UNK]",
55
- "use_fast": true
56
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/trainer_state.json DELETED
@@ -1,32 +0,0 @@
1
- {
2
- "best_metric": null,
3
- "best_model_checkpoint": null,
4
- "epoch": 0.8322929671244278,
5
- "eval_steps": 2000,
6
- "global_step": 2000,
7
- "is_hyper_param_search": false,
8
- "is_local_process_zero": true,
9
- "is_world_process_zero": true,
10
- "log_history": [
11
- {
12
- "epoch": 0.83,
13
- "eval_accuracy": 0.7232121953122967,
14
- "eval_f1": 0.680500985958512,
15
- "eval_loss": 0.5397241115570068,
16
- "eval_matthews_correlation": 0.3991911550967513,
17
- "eval_precision": 0.7268894858780544,
18
- "eval_recall": 0.6755850185066816,
19
- "eval_runtime": 6.4825,
20
- "eval_samples_per_second": 5929.926,
21
- "eval_steps_per_second": 46.432,
22
- "step": 2000
23
- }
24
- ],
25
- "logging_steps": 100000,
26
- "max_steps": 19224,
27
- "num_train_epochs": 8,
28
- "save_steps": 2000,
29
- "total_flos": 2.6311104946765824e+16,
30
- "trial_name": null,
31
- "trial_params": null
32
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-2000/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9846bc17ea30367d7691aef3ec4a8eae9c29225f580075721efd19eaa9d65c75
3
- size 5265
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/config.json DELETED
@@ -1,27 +0,0 @@
1
- {
2
- "_name_or_path": "/root/NaN/dna-tokenizer/pretrain/models/base_3072/checkpoint-100000",
3
- "architectures": [
4
- "BertForSequenceClassification"
5
- ],
6
- "attention_probs_dropout_prob": 0.1,
7
- "classifier_dropout": null,
8
- "hidden_act": "gelu",
9
- "hidden_dropout_prob": 0.1,
10
- "hidden_size": 768,
11
- "initializer_range": 0.02,
12
- "intermediate_size": 3072,
13
- "layer_norm_eps": 1e-12,
14
- "max_length": 512,
15
- "max_position_embeddings": 512,
16
- "model_type": "bert",
17
- "num_attention_heads": 12,
18
- "num_hidden_layers": 12,
19
- "pad_token_id": 0,
20
- "position_embedding_type": "absolute",
21
- "problem_type": "single_label_classification",
22
- "torch_dtype": "float32",
23
- "transformers_version": "4.35.2",
24
- "type_vocab_size": 2,
25
- "use_cache": true,
26
- "vocab_size": 3072
27
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d323c9f0c20d5297cc07df751fb4d897c31a811853c190b42918b235b3c4e8f1
3
- size 353632152
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a8fb249d5d7f21f763ddfacdee4529a0987646ac1cf02b80d3dad4cd155d789
3
- size 707385995
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:022888c8d72c2809ac5a3da8c8ed4390df94dd94ad785d29ade33f64674216b7
3
- size 14709
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d0f78c7b09d5edf997a2ef80368655d4d1525769cf0b855b3af5043e564a39d1
3
- size 1465
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/special_tokens_map.json DELETED
@@ -1,7 +0,0 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
7
- }
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/tokenizer_config.json DELETED
@@ -1,56 +0,0 @@
1
- {
2
- "added_tokens_decoder": {
3
- "0": {
4
- "content": "[PAD]",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false,
9
- "special": true
10
- },
11
- "1": {
12
- "content": "[UNK]",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": true
18
- },
19
- "2": {
20
- "content": "[CLS]",
21
- "lstrip": false,
22
- "normalized": false,
23
- "rstrip": false,
24
- "single_word": false,
25
- "special": true
26
- },
27
- "3": {
28
- "content": "[SEP]",
29
- "lstrip": false,
30
- "normalized": false,
31
- "rstrip": false,
32
- "single_word": false,
33
- "special": true
34
- },
35
- "4": {
36
- "content": "[MASK]",
37
- "lstrip": false,
38
- "normalized": false,
39
- "rstrip": false,
40
- "single_word": false,
41
- "special": true
42
- }
43
- },
44
- "cache_dir": null,
45
- "clean_up_tokenization_spaces": true,
46
- "cls_token": "[CLS]",
47
- "mask_token": "[MASK]",
48
- "model_max_length": 200,
49
- "pad_token": "[PAD]",
50
- "padding_side": "right",
51
- "sep_token": "[SEP]",
52
- "tokenizer_class": "PreTrainedTokenizerFast",
53
- "trust_remote_code": true,
54
- "unk_token": "[UNK]",
55
- "use_fast": true
56
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/trainer_state.json DELETED
@@ -1,45 +0,0 @@
1
- {
2
- "best_metric": null,
3
- "best_model_checkpoint": null,
4
- "epoch": 1.6645859342488556,
5
- "eval_steps": 2000,
6
- "global_step": 4000,
7
- "is_hyper_param_search": false,
8
- "is_local_process_zero": true,
9
- "is_world_process_zero": true,
10
- "log_history": [
11
- {
12
- "epoch": 0.83,
13
- "eval_accuracy": 0.7232121953122967,
14
- "eval_f1": 0.680500985958512,
15
- "eval_loss": 0.5397241115570068,
16
- "eval_matthews_correlation": 0.3991911550967513,
17
- "eval_precision": 0.7268894858780544,
18
- "eval_recall": 0.6755850185066816,
19
- "eval_runtime": 6.4825,
20
- "eval_samples_per_second": 5929.926,
21
- "eval_steps_per_second": 46.432,
22
- "step": 2000
23
- },
24
- {
25
- "epoch": 1.66,
26
- "eval_accuracy": 0.721963528524232,
27
- "eval_f1": 0.7135581221078293,
28
- "eval_loss": 0.5439369082450867,
29
- "eval_matthews_correlation": 0.43044300831847054,
30
- "eval_precision": 0.7116658765696064,
31
- "eval_recall": 0.7188368602594819,
32
- "eval_runtime": 6.4688,
33
- "eval_samples_per_second": 5942.505,
34
- "eval_steps_per_second": 46.531,
35
- "step": 4000
36
- }
37
- ],
38
- "logging_steps": 100000,
39
- "max_steps": 19224,
40
- "num_train_epochs": 8,
41
- "save_steps": 2000,
42
- "total_flos": 5.261655771657011e+16,
43
- "trial_name": null,
44
- "trial_params": null
45
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_3072/1e-5/checkpoint-4000/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9846bc17ea30367d7691aef3ec4a8eae9c29225f580075721efd19eaa9d65c75
3
- size 5265
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/config.json DELETED
@@ -1,27 +0,0 @@
1
- {
2
- "_name_or_path": "/root/NaN/dna-tokenizer/pretrain/models/base_4096/checkpoint-100000",
3
- "architectures": [
4
- "BertForSequenceClassification"
5
- ],
6
- "attention_probs_dropout_prob": 0.1,
7
- "classifier_dropout": null,
8
- "hidden_act": "gelu",
9
- "hidden_dropout_prob": 0.1,
10
- "hidden_size": 768,
11
- "initializer_range": 0.02,
12
- "intermediate_size": 3072,
13
- "layer_norm_eps": 1e-12,
14
- "max_length": 512,
15
- "max_position_embeddings": 512,
16
- "model_type": "bert",
17
- "num_attention_heads": 12,
18
- "num_hidden_layers": 12,
19
- "pad_token_id": 0,
20
- "position_embedding_type": "absolute",
21
- "problem_type": "single_label_classification",
22
- "torch_dtype": "float32",
23
- "transformers_version": "4.35.2",
24
- "type_vocab_size": 2,
25
- "use_cache": true,
26
- "vocab_size": 4096
27
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8b826da22ddb2986b902fceccdb27ab239795e8f329e4c9b05643627359effff
3
- size 356777880
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:852c30787e26294b05d65f00119fe817445711e9d69f39981ab5b01f9c65be69
3
- size 713677451
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:4aa69e954458f68c3b809c27ada0da01d92639b6b332b918900199c6023c4dfd
3
- size 14709
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:26a279ce49d50e5c80bc88f986b50d6cff25209e4a7ed53a76c5895b94a339dd
3
- size 1465
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/special_tokens_map.json DELETED
@@ -1,7 +0,0 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
7
- }
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/tokenizer_config.json DELETED
@@ -1,56 +0,0 @@
1
- {
2
- "added_tokens_decoder": {
3
- "0": {
4
- "content": "[PAD]",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false,
9
- "special": true
10
- },
11
- "1": {
12
- "content": "[UNK]",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": true
18
- },
19
- "2": {
20
- "content": "[CLS]",
21
- "lstrip": false,
22
- "normalized": false,
23
- "rstrip": false,
24
- "single_word": false,
25
- "special": true
26
- },
27
- "3": {
28
- "content": "[SEP]",
29
- "lstrip": false,
30
- "normalized": false,
31
- "rstrip": false,
32
- "single_word": false,
33
- "special": true
34
- },
35
- "4": {
36
- "content": "[MASK]",
37
- "lstrip": false,
38
- "normalized": false,
39
- "rstrip": false,
40
- "single_word": false,
41
- "special": true
42
- }
43
- },
44
- "cache_dir": null,
45
- "clean_up_tokenization_spaces": true,
46
- "cls_token": "[CLS]",
47
- "mask_token": "[MASK]",
48
- "model_max_length": 200,
49
- "pad_token": "[PAD]",
50
- "padding_side": "right",
51
- "sep_token": "[SEP]",
52
- "tokenizer_class": "PreTrainedTokenizerFast",
53
- "trust_remote_code": true,
54
- "unk_token": "[UNK]",
55
- "use_fast": true
56
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/trainer_state.json DELETED
@@ -1,32 +0,0 @@
1
- {
2
- "best_metric": null,
3
- "best_model_checkpoint": null,
4
- "epoch": 0.8322929671244278,
5
- "eval_steps": 2000,
6
- "global_step": 2000,
7
- "is_hyper_param_search": false,
8
- "is_local_process_zero": true,
9
- "is_world_process_zero": true,
10
- "log_history": [
11
- {
12
- "epoch": 0.83,
13
- "eval_accuracy": 0.7271923206992534,
14
- "eval_f1": 0.6916687987201311,
15
- "eval_loss": 0.5368459820747375,
16
- "eval_matthews_correlation": 0.40845552866252005,
17
- "eval_precision": 0.7245172413236349,
18
- "eval_recall": 0.6857718341711782,
19
- "eval_runtime": 6.2785,
20
- "eval_samples_per_second": 6122.616,
21
- "eval_steps_per_second": 47.941,
22
- "step": 2000
23
- }
24
- ],
25
- "logging_steps": 100000,
26
- "max_steps": 19224,
27
- "num_train_epochs": 8,
28
- "save_steps": 2000,
29
- "total_flos": 2.6311104946765824e+16,
30
- "trial_name": null,
31
- "trial_params": null
32
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-2000/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6556056323c8c121c72de11a49668bf4bd279f2aa3f171eff1ab941b35cd7d73
3
- size 5265
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/config.json DELETED
@@ -1,27 +0,0 @@
1
- {
2
- "_name_or_path": "/root/NaN/dna-tokenizer/pretrain/models/base_4096/checkpoint-100000",
3
- "architectures": [
4
- "BertForSequenceClassification"
5
- ],
6
- "attention_probs_dropout_prob": 0.1,
7
- "classifier_dropout": null,
8
- "hidden_act": "gelu",
9
- "hidden_dropout_prob": 0.1,
10
- "hidden_size": 768,
11
- "initializer_range": 0.02,
12
- "intermediate_size": 3072,
13
- "layer_norm_eps": 1e-12,
14
- "max_length": 512,
15
- "max_position_embeddings": 512,
16
- "model_type": "bert",
17
- "num_attention_heads": 12,
18
- "num_hidden_layers": 12,
19
- "pad_token_id": 0,
20
- "position_embedding_type": "absolute",
21
- "problem_type": "single_label_classification",
22
- "torch_dtype": "float32",
23
- "transformers_version": "4.35.2",
24
- "type_vocab_size": 2,
25
- "use_cache": true,
26
- "vocab_size": 4096
27
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8e9175d75a4cf7657e7003082bca3688d4e87967591ca1c3ce9473a50a666b8
3
- size 356777880
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:ffdc80a05cef66ad2e24134b167749be6f2728f5f61c061c9eb0fe69d9f3166f
3
- size 713677451
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:022888c8d72c2809ac5a3da8c8ed4390df94dd94ad785d29ade33f64674216b7
3
- size 14709
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d0f78c7b09d5edf997a2ef80368655d4d1525769cf0b855b3af5043e564a39d1
3
- size 1465
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/special_tokens_map.json DELETED
@@ -1,7 +0,0 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
7
- }
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/tokenizer_config.json DELETED
@@ -1,56 +0,0 @@
1
- {
2
- "added_tokens_decoder": {
3
- "0": {
4
- "content": "[PAD]",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false,
9
- "special": true
10
- },
11
- "1": {
12
- "content": "[UNK]",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": true
18
- },
19
- "2": {
20
- "content": "[CLS]",
21
- "lstrip": false,
22
- "normalized": false,
23
- "rstrip": false,
24
- "single_word": false,
25
- "special": true
26
- },
27
- "3": {
28
- "content": "[SEP]",
29
- "lstrip": false,
30
- "normalized": false,
31
- "rstrip": false,
32
- "single_word": false,
33
- "special": true
34
- },
35
- "4": {
36
- "content": "[MASK]",
37
- "lstrip": false,
38
- "normalized": false,
39
- "rstrip": false,
40
- "single_word": false,
41
- "special": true
42
- }
43
- },
44
- "cache_dir": null,
45
- "clean_up_tokenization_spaces": true,
46
- "cls_token": "[CLS]",
47
- "mask_token": "[MASK]",
48
- "model_max_length": 200,
49
- "pad_token": "[PAD]",
50
- "padding_side": "right",
51
- "sep_token": "[SEP]",
52
- "tokenizer_class": "PreTrainedTokenizerFast",
53
- "trust_remote_code": true,
54
- "unk_token": "[UNK]",
55
- "use_fast": true
56
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/trainer_state.json DELETED
@@ -1,45 +0,0 @@
1
- {
2
- "best_metric": null,
3
- "best_model_checkpoint": null,
4
- "epoch": 1.6645859342488556,
5
- "eval_steps": 2000,
6
- "global_step": 4000,
7
- "is_hyper_param_search": false,
8
- "is_local_process_zero": true,
9
- "is_world_process_zero": true,
10
- "log_history": [
11
- {
12
- "epoch": 0.83,
13
- "eval_accuracy": 0.7271923206992534,
14
- "eval_f1": 0.6916687987201311,
15
- "eval_loss": 0.5368459820747375,
16
- "eval_matthews_correlation": 0.40845552866252005,
17
- "eval_precision": 0.7245172413236349,
18
- "eval_recall": 0.6857718341711782,
19
- "eval_runtime": 6.2785,
20
- "eval_samples_per_second": 6122.616,
21
- "eval_steps_per_second": 47.941,
22
- "step": 2000
23
- },
24
- {
25
- "epoch": 1.66,
26
- "eval_accuracy": 0.7199604588850446,
27
- "eval_f1": 0.7113427696435267,
28
- "eval_loss": 0.5453317165374756,
29
- "eval_matthews_correlation": 0.42579813599678606,
30
- "eval_precision": 0.7094611286264896,
31
- "eval_recall": 0.716393435153353,
32
- "eval_runtime": 6.2677,
33
- "eval_samples_per_second": 6133.182,
34
- "eval_steps_per_second": 48.024,
35
- "step": 4000
36
- }
37
- ],
38
- "logging_steps": 100000,
39
- "max_steps": 19224,
40
- "num_train_epochs": 8,
41
- "save_steps": 2000,
42
- "total_flos": 5.261655771657011e+16,
43
- "trial_name": null,
44
- "trial_params": null
45
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/base_4096/1e-5/checkpoint-4000/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6556056323c8c121c72de11a49668bf4bd279f2aa3f171eff1ab941b35cd7d73
3
- size 5265
 
 
 
 
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/config.json DELETED
@@ -1,27 +0,0 @@
1
- {
2
- "_name_or_path": "/root/NaN/dna-tokenizer/pretrain/models/model_len2_4096/checkpoint-100000",
3
- "architectures": [
4
- "BertForSequenceClassification"
5
- ],
6
- "attention_probs_dropout_prob": 0.1,
7
- "classifier_dropout": null,
8
- "hidden_act": "gelu",
9
- "hidden_dropout_prob": 0.1,
10
- "hidden_size": 768,
11
- "initializer_range": 0.02,
12
- "intermediate_size": 3072,
13
- "layer_norm_eps": 1e-12,
14
- "max_length": 512,
15
- "max_position_embeddings": 512,
16
- "model_type": "bert",
17
- "num_attention_heads": 12,
18
- "num_hidden_layers": 12,
19
- "pad_token_id": 0,
20
- "position_embedding_type": "absolute",
21
- "problem_type": "single_label_classification",
22
- "torch_dtype": "float32",
23
- "transformers_version": "4.35.2",
24
- "type_vocab_size": 2,
25
- "use_cache": true,
26
- "vocab_size": 4096
27
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9820845952f61cb09e84c1cb783cf1897b8d87077036dfe3a2b923b6c1987ee
3
- size 356777880
 
 
 
 
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9736aa0cc2060da3baaf882df1d9e80995edf78213162734bd83aded19ae42e0
3
- size 713677451
 
 
 
 
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b2060874b23673bac4caf21be1d37304c08dcec292418c4a8e6d201f0fc1932
3
- size 14709
 
 
 
 
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ba381e07add5e135217829284fed29b78d6b593ab562e0cb3993675f5b72f6e
3
- size 1465
 
 
 
 
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/special_tokens_map.json DELETED
@@ -1,7 +0,0 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
7
- }
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/tokenizer_config.json DELETED
@@ -1,15 +0,0 @@
1
- {
2
- "added_tokens_decoder": {},
3
- "cache_dir": null,
4
- "clean_up_tokenization_spaces": true,
5
- "cls_token": "[CLS]",
6
- "mask_token": "[MASK]",
7
- "model_max_length": 200,
8
- "pad_token": "[PAD]",
9
- "padding_side": "right",
10
- "sep_token": "[SEP]",
11
- "tokenizer_class": "PreTrainedTokenizerFast",
12
- "trust_remote_code": true,
13
- "unk_token": "[UNK]",
14
- "use_fast": true
15
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/trainer_state.json DELETED
@@ -1,110 +0,0 @@
1
- {
2
- "best_metric": null,
3
- "best_model_checkpoint": null,
4
- "epoch": 5.826050769870995,
5
- "eval_steps": 2000,
6
- "global_step": 14000,
7
- "is_hyper_param_search": false,
8
- "is_local_process_zero": true,
9
- "is_world_process_zero": true,
10
- "log_history": [
11
- {
12
- "epoch": 0.83,
13
- "eval_accuracy": 0.7252152649514841,
14
- "eval_f1": 0.6902950682902858,
15
- "eval_loss": 0.5391069054603577,
16
- "eval_matthews_correlation": 0.40402463402188366,
17
- "eval_precision": 0.7211497658648482,
18
- "eval_recall": 0.6845309492620895,
19
- "eval_runtime": 6.2755,
20
- "eval_samples_per_second": 6125.577,
21
- "eval_steps_per_second": 47.964,
22
- "step": 2000
23
- },
24
- {
25
- "epoch": 1.66,
26
- "eval_accuracy": 0.7247470149059598,
27
- "eval_f1": 0.7099661938281203,
28
- "eval_loss": 0.5385260581970215,
29
- "eval_matthews_correlation": 0.42001605877551784,
30
- "eval_precision": 0.7107448921110509,
31
- "eval_recall": 0.709273743081231,
32
- "eval_runtime": 6.2437,
33
- "eval_samples_per_second": 6156.719,
34
- "eval_steps_per_second": 48.208,
35
- "step": 4000
36
- },
37
- {
38
- "epoch": 2.5,
39
- "eval_accuracy": 0.7316667100231523,
40
- "eval_f1": 0.7065806776390386,
41
- "eval_loss": 0.5326813459396362,
42
- "eval_matthews_correlation": 0.42215521735721945,
43
- "eval_precision": 0.7213784019572298,
44
- "eval_recall": 0.7012561139278983,
45
- "eval_runtime": 6.2526,
46
- "eval_samples_per_second": 6147.973,
47
- "eval_steps_per_second": 48.14,
48
- "step": 6000
49
- },
50
- {
51
- "epoch": 3.33,
52
- "eval_accuracy": 0.7316406961317343,
53
- "eval_f1": 0.707569021943413,
54
- "eval_loss": 0.5344680547714233,
55
- "eval_matthews_correlation": 0.422807683821826,
56
- "eval_precision": 0.7207127636727513,
57
- "eval_recall": 0.7024875391482028,
58
- "eval_runtime": 6.253,
59
- "eval_samples_per_second": 6147.608,
60
- "eval_steps_per_second": 48.137,
61
- "step": 8000
62
- },
63
- {
64
- "epoch": 4.16,
65
- "eval_accuracy": 0.731068390520538,
66
- "eval_f1": 0.708860051313849,
67
- "eval_loss": 0.5406153202056885,
68
- "eval_matthews_correlation": 0.4231214190533388,
69
- "eval_precision": 0.7190318594698323,
70
- "eval_recall": 0.7043443995944909,
71
- "eval_runtime": 6.257,
72
- "eval_samples_per_second": 6143.684,
73
- "eval_steps_per_second": 48.106,
74
- "step": 10000
75
- },
76
- {
77
- "epoch": 4.99,
78
- "eval_accuracy": 0.7305220988007596,
79
- "eval_f1": 0.7136770886356509,
80
- "eval_loss": 0.5410209894180298,
81
- "eval_matthews_correlation": 0.4282737366015945,
82
- "eval_precision": 0.7168227358137622,
83
- "eval_recall": 0.7114842716727794,
84
- "eval_runtime": 6.241,
85
- "eval_samples_per_second": 6159.471,
86
- "eval_steps_per_second": 48.23,
87
- "step": 12000
88
- },
89
- {
90
- "epoch": 5.83,
91
- "eval_accuracy": 0.7284930152701543,
92
- "eval_f1": 0.7061587739629024,
93
- "eval_loss": 0.5537341237068176,
94
- "eval_matthews_correlation": 0.41760147710952444,
95
- "eval_precision": 0.7160942515527404,
96
- "eval_recall": 0.7017533928262485,
97
- "eval_runtime": 6.2407,
98
- "eval_samples_per_second": 6159.732,
99
- "eval_steps_per_second": 48.232,
100
- "step": 14000
101
- }
102
- ],
103
- "logging_steps": 100000,
104
- "max_steps": 19224,
105
- "num_train_epochs": 8,
106
- "save_steps": 2000,
107
- "total_flos": 1.841494737425531e+17,
108
- "trial_name": null,
109
- "trial_params": null
110
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Finetune-species/super_binary_out/GLIA/len2_4096/1e-5/checkpoint-14000/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f82b451f8e860b9e6a2b1ea0895aa9da11751687e479b2bf702e23e26f7a518
3
- size 5265