YvanCarre commited on
Commit
328d64e
·
verified ·
1 Parent(s): 3b6c623

Upload InkubaLM multi-task with adapters

Browse files
config.json CHANGED
@@ -2,11 +2,46 @@
2
  "_name_or_path": "lelapa/InkubaLM-0.4B",
3
  "adapters": {
4
  "adapters": {
5
- "machine-translation": "seq_bn",
6
  "sentiment-analysis": "seq_bn",
7
  "xnli": "seq_bn"
8
  },
9
- "config_map": {},
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  "fusion_config_map": {},
11
  "fusion_name_map": {},
12
  "fusions": {}
 
2
  "_name_or_path": "lelapa/InkubaLM-0.4B",
3
  "adapters": {
4
  "adapters": {
5
+ "machine-translation": "26cd1b10db746518",
6
  "sentiment-analysis": "seq_bn",
7
  "xnli": "seq_bn"
8
  },
9
+ "config_map": {
10
+ "26cd1b10db746518": {
11
+ "adapter_residual_before_ln": false,
12
+ "cross_adapter": false,
13
+ "dropout": 0.0,
14
+ "factorized_phm_W": true,
15
+ "factorized_phm_rule": false,
16
+ "hypercomplex_nonlinearity": "glorot-uniform",
17
+ "init_weights": "bert",
18
+ "inv_adapter": null,
19
+ "inv_adapter_reduction_factor": null,
20
+ "is_parallel": false,
21
+ "learn_phm": true,
22
+ "leave_out": [],
23
+ "ln_after": false,
24
+ "ln_before": false,
25
+ "mh_adapter": false,
26
+ "non_linearity": "relu",
27
+ "original_ln_after": true,
28
+ "original_ln_before": true,
29
+ "output_adapter": true,
30
+ "phm_bias": true,
31
+ "phm_c_init": "normal",
32
+ "phm_dim": 4,
33
+ "phm_init_range": 0.0001,
34
+ "phm_layer": false,
35
+ "phm_rank": 1,
36
+ "reduction_factor": 8,
37
+ "residual_before_ln": true,
38
+ "scaling": 1.0,
39
+ "shared_W_phm": false,
40
+ "shared_phm_rule": true,
41
+ "stochastic_depth": 0.0,
42
+ "use_gating": false
43
+ }
44
+ },
45
  "fusion_config_map": {},
46
  "fusion_name_map": {},
47
  "fusions": {}
machine-translation/adapter_config.json CHANGED
@@ -25,7 +25,7 @@
25
  "phm_init_range": 0.0001,
26
  "phm_layer": false,
27
  "phm_rank": 1,
28
- "reduction_factor": 16,
29
  "residual_before_ln": true,
30
  "scaling": 1.0,
31
  "shared_W_phm": false,
 
25
  "phm_init_range": 0.0001,
26
  "phm_layer": false,
27
  "phm_rank": 1,
28
+ "reduction_factor": 8,
29
  "residual_before_ln": true,
30
  "scaling": 1.0,
31
  "shared_W_phm": false,
machine-translation/pytorch_adapter.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc47aca23a01a521202b599519696e10a8b03a4c10b86d7e3a310cbd7dfbdc6d
3
- size 16859030
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05181a303bdf4633323da7cd7cc42c0dc42de0047093fd154bd3d3cf56475ccb
3
+ size 33640406
machine-translation/pytorch_model_head.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a95f248b1ba49ae1ae7000ec31584083fefccc4d66397773748d39d0f3c1e82
3
  size 523218998
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11fd55df1c791afae000989604f2fefa0e12abd715657be35bf2a00a393874cd
3
  size 523218998
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c6afbf71ace664bd28406e25b8ce5afca0eccfc3011e7f72cca267d017e53e9
3
- size 3264041952
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e02a8f87608cbc40d6a46096618ad785ab8dfff3f75d0ef496ede17c994c4901
3
+ size 3280823264
sentiment-analysis/pytorch_adapter.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64cc397b31b51d58f07ce860a91ce71fd759b9ed8084931ca76a951c4146fa4a
3
  size 16859030
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b67fe84a865c161706578ea8e295c8f158b246896b98e8ee767e455f0fe5fded
3
  size 16859030
sentiment-analysis/pytorch_model_head.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e3b4a4fecbb3da2184e373f5f5f826c098bc440b03fee12b528e5e5e4554c3fc
3
  size 16812200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf7eac6525f75894ce9e9adbed2e71322205cbd82d0619042b473d006ab9edbc
3
  size 16812200
tokenizer.json CHANGED
@@ -1,7 +1,21 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 128,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": {
11
+ "Fixed": 128
12
+ },
13
+ "direction": "Left",
14
+ "pad_to_multiple_of": null,
15
+ "pad_id": 2,
16
+ "pad_type_id": 0,
17
+ "pad_token": "</s>"
18
+ },
19
  "added_tokens": [
20
  {
21
  "id": 0,
xnli/pytorch_adapter.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3d91c96805d108be1de37eff8546ba63dad24e08f9366104bcb3a2c402725cf1
3
  size 16858582
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a584e75646d868757e70744f3fed1c83c25d66ff6cdf9994d6775d70874a806
3
  size 16858582
xnli/pytorch_model_head.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:190ceda766cc08d3d537b7ac71c3980ca14c5edccbbffcd6e975928faac12477
3
  size 16812136
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1366e1b83a7b97a418f3f3053659d15a96ac13a53b4114249979ca6fce64fbd
3
  size 16812136