mazesmazes commited on
Commit
dbdcadd
·
verified ·
1 Parent(s): 1297df1

Training in progress, step 500

Browse files
adapter_config.json CHANGED
@@ -29,8 +29,8 @@
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
- "v_proj",
33
- "q_proj"
34
  ],
35
  "target_parameters": null,
36
  "task_type": "CAUSAL_LM",
 
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
+ "q_proj",
33
+ "v_proj"
34
  ],
35
  "target_parameters": null,
36
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ecdec63cb8329cc0dc6fd1aae2e8b137e92332a315a0959f06a9ca2273136cb
3
  size 51395296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dc0bc199fe5d8b06e71a090b2ff0f8fcba898ccf47c4033337276fb3486892b
3
  size 51395296
config.json CHANGED
@@ -181,7 +181,7 @@
181
  "projector_input_noise": 0.0,
182
  "projector_num_layers": 2,
183
  "projector_pool_stride": 4,
184
- "projector_type": "mlp",
185
  "qformer_hidden_size": null,
186
  "qformer_intermediate_size": null,
187
  "qformer_num_heads": 16,
 
181
  "projector_input_noise": 0.0,
182
  "projector_num_layers": 2,
183
  "projector_pool_stride": 4,
184
+ "projector_type": "mosa",
185
  "qformer_hidden_size": null,
186
  "qformer_intermediate_size": null,
187
  "qformer_num_heads": 16,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:72aee860c43af06d76c2722c35d23141fabd615117090895d2fa37ec4b736b3e
3
- size 58732960
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ea3de9ef3b74588fe5d0a87427a4133cab4ee0ec7c826dc05e15bbac1384820
3
+ size 236248080