jonathanli commited on
Commit
9b1577c
·
verified ·
1 Parent(s): 21e1fe2

Upload InductionVLForCausalLM

Browse files
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "architectures": [
3
- "InductionVLModel"
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
@@ -52,7 +52,7 @@
52
  "rope_scaling": null,
53
  "rope_theta": 1000000,
54
  "sliding_window": null,
55
- "tie_word_embeddings": true,
56
  "transformers_version": "4.56.2",
57
  "use_cache": true,
58
  "use_mla": false,
 
1
  {
2
  "architectures": [
3
+ "InductionVLForCausalLM"
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
 
52
  "rope_scaling": null,
53
  "rope_theta": 1000000,
54
  "sliding_window": null,
55
+ "tie_word_embeddings": false,
56
  "transformers_version": "4.56.2",
57
  "use_cache": true,
58
  "use_mla": false,
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "eos_token_id": 1,
5
+ "transformers_version": "4.56.2"
6
+ }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf6e3ddf31b8c80415f380fe42cca2f2adf17d76da08aa47254d7aa7c0a86e83
3
- size 4993789648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52c575e26f7e5fd3644837595919dd649287a4bd2c137c28fb68dec5f010cf9e
3
+ size 4993791224
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4842ee12c17907e85fc24a85349da5411b024d0b40c852a22ed6645d7cc4535e
3
- size 1558880888
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35924689a0ac38b57fe46fe8fe68a83e86f1dbec0411056f74bd0453e14f1198
3
+ size 1821053040
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff