yqsu95 commited on
Commit
87b78d1
·
verified ·
1 Parent(s): fd7cc73

Push model using huggingface_hub.

Browse files
Files changed (3) hide show
  1. README.md +6 -11
  2. config.json +6 -6
  3. training_config.json +32 -1
README.md CHANGED
@@ -1,15 +1,10 @@
1
  ---
2
- language: en
3
- license: mit
4
  tags:
5
- - pytorch
6
- - pertTF
7
  ---
8
 
9
- # pertTF-large
10
-
11
- This is model files for pertTF-large
12
-
13
- ## Usage
14
- ```python
15
- model = HFPerturbationTFModel.from_pretrained('weililab/pertTF-large')
 
1
  ---
 
 
2
  tags:
3
+ - model_hub_mixin
4
+ - pytorch_model_hub_mixin
5
  ---
6
 
7
+ This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
+ - Code: [More Information Needed]
9
+ - Paper: [More Information Needed]
10
+ - Docs: [More Information Needed]
 
 
 
config.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "cell_emb_style": "cls",
3
- "d_hid": null,
4
- "d_model": 32,
5
  "do_dab": false,
6
- "do_mvc": false,
7
  "domain_spec_batchnorm": false,
8
  "dropout": 0.1,
9
  "ecs_threshold": 0.7,
@@ -12,8 +12,8 @@
12
  "input_emb_style": "continuous",
13
  "mvc_decoder_style": "inner product",
14
  "n_bins": 51,
15
- "n_cls": 1,
16
- "n_pert": 1,
17
  "n_ps": 1,
18
  "nhead": 8,
19
  "nlayers": 12,
@@ -29,6 +29,6 @@
29
  "pred_lochness_next": false,
30
  "ps_decoder2_nlayer": 3,
31
  "use_batch_labels": false,
32
- "use_fast_transformer": false,
33
  "vocab": null
34
  }
 
1
  {
2
  "cell_emb_style": "cls",
3
+ "d_hid": 512,
4
+ "d_model": 512,
5
  "do_dab": false,
6
+ "do_mvc": true,
7
  "domain_spec_batchnorm": false,
8
  "dropout": 0.1,
9
  "ecs_threshold": 0.7,
 
12
  "input_emb_style": "continuous",
13
  "mvc_decoder_style": "inner product",
14
  "n_bins": 51,
15
+ "n_cls": 20,
16
+ "n_pert": 40,
17
  "n_ps": 1,
18
  "nhead": 8,
19
  "nlayers": 12,
 
29
  "pred_lochness_next": false,
30
  "ps_decoder2_nlayer": 3,
31
  "use_batch_labels": false,
32
+ "use_fast_transformer": true,
33
  "vocab": null
34
  }
training_config.json CHANGED
@@ -54,5 +54,36 @@
54
  "hvg_col": "highly_variable",
55
  "mvc_masked_train": true,
56
  "max_seq_len": 3001,
57
- "pad_value": -2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
58
  }
 
54
  "hvg_col": "highly_variable",
55
  "mvc_masked_train": true,
56
  "max_seq_len": 3001,
57
+ "n_pert": 40,
58
+ "nlayers_pert": 4,
59
+ "n_ps": 1,
60
+ "ntoken": 36604,
61
+ "d_model": 512,
62
+ "nhead": 8,
63
+ "d_hid": 512,
64
+ "nlayers": 12,
65
+ "nlayers_cls": 3,
66
+ "n_cls": 20,
67
+ "vocab": null,
68
+ "dropout": 0.1,
69
+ "pad_token": "<pad>",
70
+ "pad_value": -2,
71
+ "do_mvc": true,
72
+ "do_dab": false,
73
+ "use_batch_labels": false,
74
+ "num_batch_labels": null,
75
+ "domain_spec_batchnorm": false,
76
+ "input_emb_style": "continuous",
77
+ "n_bins": 51,
78
+ "cell_emb_style": "cls",
79
+ "mvc_decoder_style": "inner product",
80
+ "ecs_threshold": 0.7,
81
+ "explicit_zero_prob": true,
82
+ "use_fast_transformer": true,
83
+ "fast_transformer_backend": "flash",
84
+ "pre_norm": false,
85
+ "pred_lochness_next": false,
86
+ "ps_decoder2_nlayer": 3,
87
+ "pert_pad_id": null,
88
+ "pert_dim": null
89
  }