CyrilMa commited on
Commit
9a67f6b
·
verified ·
1 Parent(s): 8fb6c1b

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation": "gelu",
3
+ "architectures": [
4
+ "ProteomeLMForMaskedLM"
5
+ ],
6
+ "attention_dropout": 0.1,
7
+ "batch_size": 32,
8
+ "beta1": 0.9,
9
+ "beta2": 0.999,
10
+ "dataloader_num_workers": 8,
11
+ "db_path": "/home/malbrank/training/",
12
+ "dim": 512,
13
+ "dropout": 0.1,
14
+ "dtype": "bfloat16",
15
+ "eval_accumulation_steps": 1,
16
+ "eval_epochs": 1,
17
+ "gradient_accumulation_steps": 1,
18
+ "gradient_checkpointing": false,
19
+ "hidden_dim": 512,
20
+ "initializer_range": 0.02,
21
+ "input_size": 1152,
22
+ "learning_rate": 0.0003,
23
+ "logging_steps": 10,
24
+ "loss_type": "mse",
25
+ "mask_fraction": 0.5,
26
+ "max_grad_norm": 1.0,
27
+ "max_position_embeddings": 512,
28
+ "max_steps": 1000000,
29
+ "min_taxid_size": 200,
30
+ "model_type": "distilbert",
31
+ "n_heads": 8,
32
+ "n_inner": null,
33
+ "n_layers": 6,
34
+ "namedir": "ProteomeLM-Mini-Kuma",
35
+ "num_epochs": 2000,
36
+ "output_dir": "/home/malbrank/ProteomeLM/output/",
37
+ "pad_token_id": 0,
38
+ "qa_dropout": 0.1,
39
+ "save_epochs": 15,
40
+ "scheduler": "cosine",
41
+ "seq_classif_dropout": 0.2,
42
+ "sinusoidal_pos_embds": false,
43
+ "torch_dtype": "bfloat16",
44
+ "transformers_version": "4.51.1",
45
+ "use_cache": true,
46
+ "use_one_gpu": "0",
47
+ "vocab_size": 30522,
48
+ "wandb_project": "proteomelm",
49
+ "warmup_steps": 500,
50
+ "weight_decay": 0.01
51
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1323cde73870e23d47147830eeee65d59827d8da450e90813ced84c6f1081b3
3
+ size 73807700
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23124d384845c83fe8b75c48b289c2a09a903a6702b020ac54ec672328fa2f83
3
+ size 45041601
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06906c303caef61ee809eaaa2ce7a2e300a6d92b018457ab134be7115dc5f699
3
+ size 14443
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b5e96354ec4cf1fcbb3c99de5b7222d2b606c2add9ab73ea93ce9dec5d34bf1
3
+ size 1263
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d44282220cefb4c48efd8d86b5200e5b09034998a3e57f24baec7fac34b2252
3
+ size 5571