inferencerlabs commited on
Commit
c06374f
·
verified ·
1 Parent(s): 79ecbbb

Upload complete model

Browse files
Files changed (1) hide show
  1. config.json +47 -0
config.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "IQuestLoopCoderForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "auto_map": {
8
+ "AutoConfig": "configuration_iquestloopcoder.IQuestLoopCoderConfig",
9
+ "AutoModel": "modeling_iquestloopcoder.IQuestLoopCoderModel",
10
+ "AutoModelForCausalLM": "modeling_iquestloopcoder.IQuestLoopCoderForCausalLM"
11
+ },
12
+ "eos_token_id": [
13
+ 2,
14
+ 75864,
15
+ 75869
16
+ ],
17
+ "head_dim": 128,
18
+ "hidden_act": "silu",
19
+ "hidden_size": 5120,
20
+ "initializer_range": 0.02,
21
+ "intermediate_size": 27648,
22
+ "loop_num": 2,
23
+ "loop_window_size": 64,
24
+ "max_position_embeddings": 131072,
25
+ "mlp_bias": false,
26
+ "model_type": "iquestloopcoder",
27
+ "num_attention_heads": 40,
28
+ "num_hidden_layers": 80,
29
+ "num_key_value_heads": 8,
30
+ "quantization": {
31
+ "group_size": 64,
32
+ "bits": 6,
33
+ "mode": "affine"
34
+ },
35
+ "quantization_config": {
36
+ "group_size": 64,
37
+ "bits": 6,
38
+ "mode": "affine"
39
+ },
40
+ "rms_norm_eps": 1e-05,
41
+ "rope_theta": 500000,
42
+ "tie_word_embeddings": false,
43
+ "torch_dtype": "bfloat16",
44
+ "transformers_version": "4.40.0",
45
+ "use_cache": true,
46
+ "vocab_size": 76800
47
+ }