echarlaix HF Staff commited on
Commit
ac237d9
·
1 Parent(s): 33e9661

set attention_head_dim to 2

Browse files
Files changed (3) hide show
  1. config.json +5 -1
  2. generation_config.json +4 -0
  3. model.safetensors +2 -2
config.json CHANGED
@@ -5,7 +5,7 @@
5
  "Zamba2ForCausalLM"
6
  ],
7
  "attention_dropout": 0.0,
8
- "attention_head_dim": 1,
9
  "attention_hidden_size": 32,
10
  "bos_token_id": 1,
11
  "chunk_size": 256,
@@ -54,7 +54,11 @@
54
  "time_step_limit": null,
55
  "time_step_max": 0.1,
56
  "time_step_min": 0.001,
 
57
  "transformers_version": "5.0.0",
 
 
 
58
  "use_cache": true,
59
  "use_conv_bias": true,
60
  "use_long_context": false,
 
5
  "Zamba2ForCausalLM"
6
  ],
7
  "attention_dropout": 0.0,
8
+ "attention_head_dim": 2,
9
  "attention_hidden_size": 32,
10
  "bos_token_id": 1,
11
  "chunk_size": 256,
 
54
  "time_step_limit": null,
55
  "time_step_max": 0.1,
56
  "time_step_min": 0.001,
57
+ <<<<<<< Updated upstream
58
  "transformers_version": "5.0.0",
59
+ =======
60
+ "transformers_version": "4.57.6",
61
+ >>>>>>> Stashed changes
62
  "use_cache": true,
63
  "use_conv_bias": true,
64
  "use_long_context": false,
generation_config.json CHANGED
@@ -3,5 +3,9 @@
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
  "pad_token_id": 0,
 
6
  "transformers_version": "5.0.0"
 
 
 
7
  }
 
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
  "pad_token_id": 0,
6
+ <<<<<<< Updated upstream
7
  "transformers_version": "5.0.0"
8
+ =======
9
+ "transformers_version": "4.57.6"
10
+ >>>>>>> Stashed changes
11
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:257244fe95a815ccdfc00b230826eb464c8b6bfc791d6e8c35ec7a0e470ae4fb
3
- size 3398168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a59661b1a44705c94679f15edadb04c60a80eb324d27fd0f81db8ee513779c78
3
+ size 3412504