cmwalker00 commited on
Commit
0dbf5ed
·
1 Parent(s): 603ab67

Upload MixFormerSequentialForCausalLM

Browse files
Files changed (2) hide show
  1. config.json +1 -0
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -14,6 +14,7 @@
14
  "model_type": "mixformer-sequential",
15
  "n_embd": 2048,
16
  "n_head": 32,
 
17
  "n_inner": null,
18
  "n_layer": 24,
19
  "n_positions": 2048,
 
14
  "model_type": "mixformer-sequential",
15
  "n_embd": 2048,
16
  "n_head": 32,
17
+ "n_head_kv": null,
18
  "n_inner": null,
19
  "n_layer": 24,
20
  "n_positions": 2048,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2dce1fe3f2853078363dd692ea28dfa45d49cf82366181dc4de8b7792b54f489
3
- size 5673168010
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3da7ae1d696e9a8afee130ad096affba4b3e56975254e31bcbfc7e2bf7854c2
3
+ size 5673158870