guanwenyu1995 commited on
Commit
2c2fada
·
1 Parent(s): 1520dab

add gptq eagle

Browse files
MiniCPM4_1-8B-Eagle3-GPTQ/config.json ADDED
@@ -0,0 +1,176 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLMEagle3"
4
+ ],
5
+ "bos_token_id": 1,
6
+ "eos_token_id": [2,73440],
7
+ "hidden_act": "silu",
8
+ "hidden_size": 4096,
9
+ "initializer_range": 0.02,
10
+ "intermediate_size": 16384,
11
+ "max_position_embeddings": 65536,
12
+ "model_type": "llama",
13
+ "num_attention_heads": 32,
14
+ "num_key_value_heads": 2,
15
+ "num_hidden_layers": 1,
16
+ "pad_token_id": 2,
17
+ "quantization_config": {
18
+ "bits": 4,
19
+ "checkpoint_format": "gptq",
20
+ "damp_percent": 0.01,
21
+ "desc_act": false,
22
+ "group_size": 128,
23
+ "lm_head": false,
24
+ "model_file_base_name": null,
25
+ "model_name_or_path": null,
26
+ "quant_method": "gptq",
27
+ "static_groups": false,
28
+ "sym": true,
29
+ "true_sequential": true
30
+ },
31
+ "rms_norm_eps": 1e-06,
32
+ "tie_word_embeddings": false,
33
+ "torch_dtype": "bfloat16",
34
+ "transformers_version": "4.28.1",
35
+ "use_cache": true,
36
+ "vocab_size": 73448,
37
+ "draft_vocab_size": 32000,
38
+ "rope_scaling": {
39
+ "factor": 1.0,
40
+ "long_factor": [
41
+ 0.9982316082870437,
42
+ 1.033048153422584,
43
+ 1.0749920956484724,
44
+ 1.1255096879436193,
45
+ 1.1863348602111476,
46
+ 1.259543828902579,
47
+ 1.3476188888731149,
48
+ 1.4535223827776373,
49
+ 1.5807816745852985,
50
+ 1.7335856049489526,
51
+ 1.9168922912975785,
52
+ 2.1365471404135326,
53
+ 2.3994084200118646,
54
+ 2.713475511863602,
55
+ 3.0880118452194134,
56
+ 3.533650295140154,
57
+ 4.062463396503134,
58
+ 4.687974098908333,
59
+ 5.425075306704039,
60
+ 6.289818967956352,
61
+ 7.29902962722721,
62
+ 8.469695779093664,
63
+ 9.81809877306655,
64
+ 11.358657902065282,
65
+ 13.102505860712087,
66
+ 15.055862949967128,
67
+ 17.218348131364184,
68
+ 19.581439255386453,
69
+ 22.127353314656723,
70
+ 24.828633849376587,
71
+ 27.6486820771775,
72
+ 30.54334096108829,
73
+ 33.46345345363812,
74
+ 36.358112337548896,
75
+ 39.17816056534983,
76
+ 41.879441100069684,
77
+ 44.425355159339965,
78
+ 46.78844628336223,
79
+ 48.95093146475928,
80
+ 50.90428855401433,
81
+ 52.648136512661125,
82
+ 54.18869564165987,
83
+ 55.537098635632745,
84
+ 56.7077647874992,
85
+ 57.71697544677006,
86
+ 58.58171910802236,
87
+ 59.31882031581807,
88
+ 59.94433101822328,
89
+ 60.47314411958625,
90
+ 60.918782569507,
91
+ 61.29331890286281,
92
+ 61.60738599471455,
93
+ 61.87024727431288,
94
+ 62.089902123428836,
95
+ 62.27320880977746,
96
+ 62.42601274014111,
97
+ 62.55327203194878,
98
+ 62.65917552585329,
99
+ 62.74725058582382,
100
+ 62.82045955451526,
101
+ 62.88128472678279,
102
+ 62.931802319077946,
103
+ 62.97374626130382,
104
+ 63.008562806439365
105
+ ],
106
+ "original_max_position_embeddings": 65536,
107
+ "rope_type": "longrope",
108
+ "short_factor": [
109
+ 0.9982316082870437,
110
+ 1.033048153422584,
111
+ 1.0749920956484724,
112
+ 1.1255096879436193,
113
+ 1.1863348602111476,
114
+ 1.259543828902579,
115
+ 1.3476188888731149,
116
+ 1.4535223827776373,
117
+ 1.5807816745852985,
118
+ 1.7335856049489526,
119
+ 1.9168922912975785,
120
+ 2.1365471404135326,
121
+ 2.3994084200118646,
122
+ 2.713475511863602,
123
+ 3.0880118452194134,
124
+ 3.533650295140154,
125
+ 4.062463396503134,
126
+ 4.687974098908333,
127
+ 5.425075306704039,
128
+ 6.289818967956352,
129
+ 7.29902962722721,
130
+ 8.469695779093664,
131
+ 9.81809877306655,
132
+ 11.358657902065282,
133
+ 13.102505860712087,
134
+ 15.055862949967128,
135
+ 17.218348131364184,
136
+ 19.581439255386453,
137
+ 22.127353314656723,
138
+ 24.828633849376587,
139
+ 27.6486820771775,
140
+ 30.54334096108829,
141
+ 33.46345345363812,
142
+ 36.358112337548896,
143
+ 39.17816056534983,
144
+ 41.879441100069684,
145
+ 44.425355159339965,
146
+ 46.78844628336223,
147
+ 48.95093146475928,
148
+ 50.90428855401433,
149
+ 52.648136512661125,
150
+ 54.18869564165987,
151
+ 55.537098635632745,
152
+ 56.7077647874992,
153
+ 57.71697544677006,
154
+ 58.58171910802236,
155
+ 59.31882031581807,
156
+ 59.94433101822328,
157
+ 60.47314411958625,
158
+ 60.918782569507,
159
+ 61.29331890286281,
160
+ 61.60738599471455,
161
+ 61.87024727431288,
162
+ 62.089902123428836,
163
+ 62.27320880977746,
164
+ 62.42601274014111,
165
+ 62.55327203194878,
166
+ 62.65917552585329,
167
+ 62.74725058582382,
168
+ 62.82045955451526,
169
+ 62.88128472678279,
170
+ 62.931802319077946,
171
+ 62.97374626130382,
172
+ 63.008562806439365
173
+ ],
174
+ "type": "longrope"
175
+ }
176
+ }
MiniCPM4_1-8B-Eagle3-GPTQ/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3214778ef8aa6add55a037a92439ca91de02db2ccfcdbc19f9a67b800056e21b
3
+ size 421844472