Abdohaaland commited on
Commit
54d983f
·
verified ·
1 Parent(s): 631e001

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -34,3 +34,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  Models/70/tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  Models/70/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ Models/50/tokenizer.json filter=lfs diff=lfs merge=lfs -text
Models/50/chat_template.jinja ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {{- bos_token }}
2
+ {%- if custom_tools is defined %}
3
+ {%- set tools = custom_tools %}
4
+ {%- endif %}
5
+ {%- if not tools_in_user_message is defined %}
6
+ {%- set tools_in_user_message = true %}
7
+ {%- endif %}
8
+ {%- if not date_string is defined %}
9
+ {%- if strftime_now is defined %}
10
+ {%- set date_string = strftime_now("%d %b %Y") %}
11
+ {%- else %}
12
+ {%- set date_string = "26 Jul 2024" %}
13
+ {%- endif %}
14
+ {%- endif %}
15
+ {%- if not tools is defined %}
16
+ {%- set tools = none %}
17
+ {%- endif %}
18
+
19
+ {#- This block extracts the system message, so we can slot it into the right place. #}
20
+ {%- if messages[0]['role'] == 'system' %}
21
+ {%- set system_message = messages[0]['content']|trim %}
22
+ {%- set messages = messages[1:] %}
23
+ {%- else %}
24
+ {%- set system_message = "" %}
25
+ {%- endif %}
26
+
27
+ {#- System message #}
28
+ {{- "<|start_header_id|>system<|end_header_id|>\n\n" }}
29
+ {%- if tools is not none %}
30
+ {{- "Environment: ipython\n" }}
31
+ {%- endif %}
32
+ {{- "Cutting Knowledge Date: December 2023\n" }}
33
+ {{- "Today Date: " + date_string + "\n\n" }}
34
+ {%- if tools is not none and not tools_in_user_message %}
35
+ {{- "You have access to the following functions. To call a function, please respond with JSON for a function call." }}
36
+ {{- 'Respond in the format {"name": function name, "parameters": dictionary of argument name and its value}.' }}
37
+ {{- "Do not use variables.\n\n" }}
38
+ {%- for t in tools %}
39
+ {{- t | tojson(indent=4) }}
40
+ {{- "\n\n" }}
41
+ {%- endfor %}
42
+ {%- endif %}
43
+ {{- system_message }}
44
+ {{- "<|eot_id|>" }}
45
+
46
+ {#- Custom tools are passed in a user message with some extra guidance #}
47
+ {%- if tools_in_user_message and not tools is none %}
48
+ {#- Extract the first user message so we can plug it in here #}
49
+ {%- if messages | length != 0 %}
50
+ {%- set first_user_message = messages[0]['content']|trim %}
51
+ {%- set messages = messages[1:] %}
52
+ {%- else %}
53
+ {{- raise_exception("Cannot put tools in the first user message when there's no first user message!") }}
54
+ {%- endif %}
55
+ {{- '<|start_header_id|>user<|end_header_id|>\n\n' -}}
56
+ {{- "Given the following functions, please respond with a JSON for a function call " }}
57
+ {{- "with its proper arguments that best answers the given prompt.\n\n" }}
58
+ {{- 'Respond in the format {"name": function name, "parameters": dictionary of argument name and its value}.' }}
59
+ {{- "Do not use variables.\n\n" }}
60
+ {%- for t in tools %}
61
+ {{- t | tojson(indent=4) }}
62
+ {{- "\n\n" }}
63
+ {%- endfor %}
64
+ {{- first_user_message + "<|eot_id|>"}}
65
+ {%- endif %}
66
+
67
+ {%- for message in messages %}
68
+ {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}
69
+ {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' }}
70
+ {%- elif 'tool_calls' in message %}
71
+ {%- if not message.tool_calls|length == 1 %}
72
+ {{- raise_exception("This model only supports single tool-calls at once!") }}
73
+ {%- endif %}
74
+ {%- set tool_call = message.tool_calls[0].function %}
75
+ {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}}
76
+ {{- '{"name": "' + tool_call.name + '", ' }}
77
+ {{- '"parameters": ' }}
78
+ {{- tool_call.arguments | tojson }}
79
+ {{- "}" }}
80
+ {{- "<|eot_id|>" }}
81
+ {%- elif message.role == "tool" or message.role == "ipython" %}
82
+ {{- "<|start_header_id|>ipython<|end_header_id|>\n\n" }}
83
+ {%- if message.content is mapping or message.content is iterable %}
84
+ {{- message.content | tojson }}
85
+ {%- else %}
86
+ {{- message.content }}
87
+ {%- endif %}
88
+ {{- "<|eot_id|>" }}
89
+ {%- endif %}
90
+ {%- endfor %}
91
+ {%- if add_generation_prompt %}
92
+ {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' }}
93
+ {%- endif %}
Models/50/config.json ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 128000,
8
+ "dtype": "bfloat16",
9
+ "eos_token_id": [
10
+ 128001,
11
+ 128008,
12
+ 128009
13
+ ],
14
+ "head_dim": 128,
15
+ "hidden_act": "silu",
16
+ "hidden_size": 3072,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 8192,
19
+ "max_position_embeddings": 131072,
20
+ "mlp_bias": false,
21
+ "model_type": "llama",
22
+ "num_attention_heads": 24,
23
+ "num_hidden_layers": 28,
24
+ "num_key_value_heads": 8,
25
+ "pad_token_id": 128009,
26
+ "pretraining_tp": 1,
27
+ "quantization_config": {
28
+ "bits": 4,
29
+ "checkpoint_format": "gptq",
30
+ "desc_act": false,
31
+ "format": "gptq",
32
+ "group_size": 128,
33
+ "lm_head": false,
34
+ "meta": {
35
+ "act_group_aware": true,
36
+ "auto_forward_data_parallel": true,
37
+ "damp_auto_increment": 0.01,
38
+ "damp_percent": 0.05,
39
+ "failsafe": {
40
+ "smooth": null,
41
+ "strategy": "rtn",
42
+ "threshold": "0.5%"
43
+ },
44
+ "gc_mode": "interval",
45
+ "gptaq": null,
46
+ "hessian": {
47
+ "chunk_bytes": null,
48
+ "chunk_size": null,
49
+ "staging_dtype": "float32"
50
+ },
51
+ "mock_quantization": false,
52
+ "mse": 0.0,
53
+ "offload_to_disk": true,
54
+ "offload_to_disk_path": "./gptqmodel_offload/pthyuejg-tgujncci/",
55
+ "pack_impl": "cpu",
56
+ "quantizer": [
57
+ "gptqmodel:5.8.0"
58
+ ],
59
+ "static_groups": false,
60
+ "true_sequential": true,
61
+ "uri": "https://github.com/modelcloud/gptqmodel",
62
+ "vram_strategy": "exclusive",
63
+ "wait_for_submodule_finalizers": false
64
+ },
65
+ "pack_dtype": "int32",
66
+ "quant_method": "gptq",
67
+ "sym": true
68
+ },
69
+ "rms_norm_eps": 1e-05,
70
+ "rope_parameters": {
71
+ "factor": 32.0,
72
+ "high_freq_factor": 4.0,
73
+ "low_freq_factor": 1.0,
74
+ "original_max_position_embeddings": 8192,
75
+ "rope_theta": 500000.0,
76
+ "rope_type": "llama3"
77
+ },
78
+ "tie_word_embeddings": true,
79
+ "transformers_version": "5.2.0",
80
+ "use_cache": true,
81
+ "vocab_size": 128256
82
+ }
Models/50/generation_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 128000,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 128001,
6
+ 128008,
7
+ 128009
8
+ ],
9
+ "temperature": 0.6,
10
+ "top_p": 0.9,
11
+ "transformers_version": "5.2.0"
12
+ }
Models/50/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d756ce8fb90e545e5b47ef1566bc3cbdca3a6ded02e0e31ba7fed2b04dec3d6
3
+ size 2255767880
Models/50/quant_log.csv ADDED
@@ -0,0 +1,197 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ layer,module,loss,samples,damp,time
2
+ 0,self_attn.v_proj,0.0000003655,0.05000,6.135
3
+ 0,self_attn.q_proj,0.0000139163,0.05000,6.251
4
+ 0,self_attn.k_proj,0.0000070522,0.05000,6.287
5
+ 0,self_attn.o_proj,0.0000000161,0.05000,1.438
6
+ 0,mlp.up_proj,0.0000059081,0.05000,3.042
7
+ 0,mlp.gate_proj,0.0000067841,0.05000,3.079
8
+ 0,mlp.down_proj,0.0000000707,0.05000,3.574
9
+ 1,self_attn.q_proj,0.0000158107,0.05000,6.250
10
+ 1,self_attn.k_proj,0.0000093271,0.05000,6.342
11
+ 1,self_attn.v_proj,0.0000009569,0.05000,6.403
12
+ 1,self_attn.o_proj,0.0000000717,0.05000,1.268
13
+ 1,mlp.gate_proj,0.0000085050,0.05000,2.909
14
+ 1,mlp.up_proj,0.0000074407,0.05000,2.938
15
+ 1,mlp.down_proj,0.0000020780,0.05000,3.624
16
+ 2,self_attn.v_proj,0.0000035664,0.05000,6.128
17
+ 2,self_attn.k_proj,0.0000384627,0.05000,6.183
18
+ 2,self_attn.q_proj,0.0000636528,0.05000,6.217
19
+ 2,self_attn.o_proj,0.0000000483,0.05000,1.279
20
+ 2,mlp.gate_proj,0.0000133714,0.05000,2.891
21
+ 2,mlp.up_proj,0.0000112456,0.05000,2.918
22
+ 2,mlp.down_proj,0.0000001836,0.05000,3.592
23
+ 3,self_attn.k_proj,0.0000209325,0.05000,6.104
24
+ 3,self_attn.q_proj,0.0000374590,0.05000,6.210
25
+ 3,self_attn.v_proj,0.0000032397,0.05000,6.243
26
+ 3,self_attn.o_proj,0.0000000519,0.05000,1.242
27
+ 3,mlp.gate_proj,0.0000199032,0.05000,2.985
28
+ 3,mlp.up_proj,0.0000147765,0.05000,3.037
29
+ 3,mlp.down_proj,0.0000002623,0.05000,3.622
30
+ 4,self_attn.k_proj,0.0000209495,0.05000,6.247
31
+ 4,self_attn.v_proj,0.0000035178,0.05000,6.307
32
+ 4,self_attn.q_proj,0.0000402571,0.05000,6.356
33
+ 4,self_attn.o_proj,0.0000001683,0.05000,1.296
34
+ 4,mlp.gate_proj,0.0000267916,0.05000,3.001
35
+ 4,mlp.up_proj,0.0000176318,0.05000,3.032
36
+ 4,mlp.down_proj,0.0000004458,0.05000,3.538
37
+ 5,self_attn.q_proj,0.0000566906,0.05000,6.143
38
+ 5,self_attn.k_proj,0.0000350725,0.05000,6.192
39
+ 5,self_attn.v_proj,0.0000035867,0.05000,6.259
40
+ 5,self_attn.o_proj,0.0000002498,0.05000,1.254
41
+ 5,mlp.up_proj,0.0000200849,0.05000,2.861
42
+ 5,mlp.gate_proj,0.0000287105,0.05000,2.907
43
+ 5,mlp.down_proj,0.0000006270,0.05000,3.494
44
+ 6,self_attn.q_proj,0.0000456653,0.05000,6.164
45
+ 6,self_attn.k_proj,0.0000243565,0.05000,6.213
46
+ 6,self_attn.v_proj,0.0000032720,0.05000,6.258
47
+ 6,self_attn.o_proj,0.0000002966,0.05000,1.271
48
+ 6,mlp.gate_proj,0.0000316156,0.05000,2.851
49
+ 6,mlp.up_proj,0.0000221315,0.05000,2.875
50
+ 6,mlp.down_proj,0.0000007019,0.05000,3.596
51
+ 7,self_attn.k_proj,0.0000232553,0.05000,6.241
52
+ 7,self_attn.q_proj,0.0000390834,0.05000,6.349
53
+ 7,self_attn.v_proj,0.0000030140,0.05000,6.364
54
+ 7,self_attn.o_proj,0.0000005020,0.05000,1.230
55
+ 7,mlp.gate_proj,0.0000300552,0.05000,2.820
56
+ 7,mlp.up_proj,0.0000227827,0.05000,2.857
57
+ 7,mlp.down_proj,0.0000007818,0.05000,3.548
58
+ 8,self_attn.k_proj,0.0000296915,0.05000,6.160
59
+ 8,self_attn.q_proj,0.0000489327,0.05000,6.220
60
+ 8,self_attn.v_proj,0.0000035765,0.05000,6.250
61
+ 8,self_attn.o_proj,0.0000005180,0.05000,1.287
62
+ 8,mlp.gate_proj,0.0000325682,0.05000,3.116
63
+ 8,mlp.up_proj,0.0000241525,0.05000,3.147
64
+ 8,mlp.down_proj,0.0000007777,0.05000,3.697
65
+ 9,self_attn.k_proj,0.0000279815,0.05000,6.365
66
+ 9,self_attn.q_proj,0.0000470294,0.05000,6.415
67
+ 9,self_attn.v_proj,0.0000045795,0.05000,6.444
68
+ 9,self_attn.o_proj,0.0000005254,0.05000,1.274
69
+ 9,mlp.gate_proj,0.0000318904,0.05000,3.054
70
+ 9,mlp.up_proj,0.0000243400,0.05000,3.097
71
+ 9,mlp.down_proj,0.0000007917,0.05000,3.662
72
+ 10,self_attn.v_proj,0.0000035862,0.05000,6.390
73
+ 10,self_attn.k_proj,0.0000299603,0.05000,6.439
74
+ 10,self_attn.q_proj,0.0000476768,0.05000,6.472
75
+ 10,self_attn.o_proj,0.0000006423,0.05000,1.260
76
+ 10,mlp.gate_proj,0.0000323488,0.05000,3.004
77
+ 10,mlp.up_proj,0.0000264711,0.05000,3.029
78
+ 10,mlp.down_proj,0.0000009133,0.05000,3.569
79
+ 11,self_attn.v_proj,0.0000042222,0.05000,6.385
80
+ 11,self_attn.q_proj,0.0000389616,0.05000,6.432
81
+ 11,self_attn.k_proj,0.0000221590,0.05000,6.478
82
+ 11,self_attn.o_proj,0.0000007108,0.05000,1.303
83
+ 11,mlp.gate_proj,0.0000346920,0.05000,3.147
84
+ 11,mlp.up_proj,0.0000293260,0.05000,3.189
85
+ 11,mlp.down_proj,0.0000010680,0.05000,3.578
86
+ 12,self_attn.v_proj,0.0000049074,0.05000,6.397
87
+ 12,self_attn.q_proj,0.0000599314,0.05000,6.425
88
+ 12,self_attn.k_proj,0.0000360839,0.05000,6.448
89
+ 12,self_attn.o_proj,0.0000008251,0.05000,1.306
90
+ 12,mlp.up_proj,0.0000314120,0.05000,3.121
91
+ 12,mlp.gate_proj,0.0000372577,0.05000,3.155
92
+ 12,mlp.down_proj,0.0000012126,0.05000,3.640
93
+ 13,self_attn.k_proj,0.0000389678,0.05000,6.532
94
+ 13,self_attn.q_proj,0.0000606843,0.05000,6.581
95
+ 13,self_attn.v_proj,0.0000056146,0.05000,6.603
96
+ 13,self_attn.o_proj,0.0000009978,0.05000,1.232
97
+ 13,mlp.gate_proj,0.0000445419,0.05000,3.054
98
+ 13,mlp.up_proj,0.0000351055,0.05000,3.085
99
+ 13,mlp.down_proj,0.0000015760,0.05000,3.608
100
+ 14,self_attn.q_proj,0.0000686343,0.05000,6.346
101
+ 14,self_attn.k_proj,0.0000314004,0.05000,6.429
102
+ 14,self_attn.v_proj,0.0000066387,0.05000,6.455
103
+ 14,self_attn.o_proj,0.0000012561,0.05000,1.325
104
+ 14,mlp.up_proj,0.0000375102,0.05000,2.993
105
+ 14,mlp.gate_proj,0.0000478461,0.05000,3.019
106
+ 14,mlp.down_proj,0.0000019180,0.05000,3.639
107
+ 15,self_attn.v_proj,0.0000065177,0.05000,6.382
108
+ 15,self_attn.q_proj,0.0000684232,0.05000,6.478
109
+ 15,self_attn.k_proj,0.0000357806,0.05000,6.490
110
+ 15,self_attn.o_proj,0.0000009062,0.05000,1.253
111
+ 15,mlp.up_proj,0.0000374816,0.05000,3.019
112
+ 15,mlp.gate_proj,0.0000510690,0.05000,3.046
113
+ 15,mlp.down_proj,0.0000019322,0.05000,3.550
114
+ 16,self_attn.q_proj,0.0000697884,0.05000,6.246
115
+ 16,self_attn.v_proj,0.0000073443,0.05000,6.299
116
+ 16,self_attn.k_proj,0.0000398958,0.05000,6.347
117
+ 16,self_attn.o_proj,0.0000005650,0.05000,1.274
118
+ 16,mlp.up_proj,0.0000372290,0.05000,3.057
119
+ 16,mlp.gate_proj,0.0000518827,0.05000,3.110
120
+ 16,mlp.down_proj,0.0000018123,0.05000,3.574
121
+ 17,self_attn.v_proj,0.0000071509,0.05000,6.345
122
+ 17,self_attn.k_proj,0.0000360828,0.05000,6.390
123
+ 17,self_attn.q_proj,0.0000662234,0.05000,6.422
124
+ 17,self_attn.o_proj,0.0000005764,0.05000,1.277
125
+ 17,mlp.gate_proj,0.0000540872,0.05000,3.071
126
+ 17,mlp.up_proj,0.0000382789,0.05000,3.121
127
+ 17,mlp.down_proj,0.0000019723,0.05000,3.628
128
+ 18,self_attn.q_proj,0.0000768960,0.05000,6.232
129
+ 18,self_attn.v_proj,0.0000090087,0.05000,6.312
130
+ 18,self_attn.k_proj,0.0000426919,0.05000,6.348
131
+ 18,self_attn.o_proj,0.0000005129,0.05000,1.306
132
+ 18,mlp.up_proj,0.0000403784,0.05000,3.084
133
+ 18,mlp.gate_proj,0.0000557208,0.05000,3.102
134
+ 18,mlp.down_proj,0.0000020153,0.05000,3.576
135
+ 19,self_attn.k_proj,0.0000391191,0.05000,6.414
136
+ 19,self_attn.q_proj,0.0000672697,0.05000,6.462
137
+ 19,self_attn.v_proj,0.0000086852,0.05000,6.490
138
+ 19,self_attn.o_proj,0.0000007129,0.05000,1.348
139
+ 19,mlp.up_proj,0.0000428687,0.05000,3.091
140
+ 19,mlp.gate_proj,0.0000584434,0.05000,3.129
141
+ 19,mlp.down_proj,0.0000024692,0.05000,3.558
142
+ 20,self_attn.v_proj,0.0000102000,0.05000,6.572
143
+ 20,self_attn.k_proj,0.0000419316,0.05000,6.646
144
+ 20,self_attn.q_proj,0.0000693657,0.05000,6.679
145
+ 20,self_attn.o_proj,0.0000005730,0.05000,1.440
146
+ 20,mlp.up_proj,0.0000435856,0.05000,3.067
147
+ 20,mlp.gate_proj,0.0000564082,0.05000,3.126
148
+ 20,mlp.down_proj,0.0000023884,0.05000,3.670
149
+ 21,self_attn.k_proj,0.0000410597,0.05000,6.395
150
+ 21,self_attn.v_proj,0.0000134331,0.05000,6.524
151
+ 21,self_attn.q_proj,0.0000694150,0.05000,6.547
152
+ 21,self_attn.o_proj,0.0000006532,0.05000,1.324
153
+ 21,mlp.gate_proj,0.0000595703,0.05000,3.093
154
+ 21,mlp.up_proj,0.0000455969,0.05000,3.126
155
+ 21,mlp.down_proj,0.0000025119,0.05000,3.593
156
+ 22,self_attn.k_proj,0.0000379314,0.05000,6.376
157
+ 22,self_attn.q_proj,0.0000656641,0.05000,6.415
158
+ 22,self_attn.v_proj,0.0000129559,0.05000,6.437
159
+ 22,self_attn.o_proj,0.0000006026,0.05000,1.297
160
+ 22,mlp.up_proj,0.0000487634,0.05000,3.087
161
+ 22,mlp.gate_proj,0.0000640544,0.05000,3.115
162
+ 22,mlp.down_proj,0.0000028967,0.05000,3.652
163
+ 23,self_attn.q_proj,0.0000680379,0.05000,6.571
164
+ 23,self_attn.v_proj,0.0000125475,0.05000,6.652
165
+ 23,self_attn.k_proj,0.0000421066,0.05000,6.674
166
+ 23,self_attn.o_proj,0.0000008360,0.05000,1.300
167
+ 23,mlp.gate_proj,0.0000771958,0.05000,3.051
168
+ 23,mlp.up_proj,0.0000553964,0.05000,3.084
169
+ 23,mlp.down_proj,0.0000037663,0.05000,3.648
170
+ 24,self_attn.q_proj,0.0000738199,0.05000,6.379
171
+ 24,self_attn.v_proj,0.0000189734,0.05000,6.423
172
+ 24,self_attn.k_proj,0.0000465050,0.05000,6.477
173
+ 24,self_attn.o_proj,0.0000014266,0.05000,1.281
174
+ 24,mlp.up_proj,0.0000596369,0.05000,3.043
175
+ 24,mlp.gate_proj,0.0000848039,0.05000,3.081
176
+ 24,mlp.down_proj,0.0000043019,0.05000,3.599
177
+ 25,self_attn.k_proj,0.0000343078,0.05000,6.722
178
+ 25,self_attn.q_proj,0.0000664509,0.05000,6.772
179
+ 25,self_attn.v_proj,0.0000153221,0.05000,6.806
180
+ 25,self_attn.o_proj,0.0000016806,0.05000,1.262
181
+ 25,mlp.gate_proj,0.0000898007,0.05000,3.004
182
+ 25,mlp.up_proj,0.0000626860,0.05000,3.051
183
+ 25,mlp.down_proj,0.0000053409,0.05000,3.695
184
+ 26,self_attn.v_proj,0.0000210691,0.05000,6.411
185
+ 26,self_attn.q_proj,0.0000636958,0.05000,6.472
186
+ 26,self_attn.k_proj,0.0000403479,0.05000,6.524
187
+ 26,self_attn.o_proj,0.0000014542,0.05000,1.290
188
+ 26,mlp.up_proj,0.0000650641,0.05000,3.043
189
+ 26,mlp.gate_proj,0.0000945761,0.05000,3.087
190
+ 26,mlp.down_proj,0.0000081644,0.05000,3.669
191
+ 27,self_attn.v_proj,0.0000141159,0.05000,6.830
192
+ 27,self_attn.k_proj,0.0000273797,0.05000,6.913
193
+ 27,self_attn.q_proj,0.0000486121,0.05000,6.940
194
+ 27,self_attn.o_proj,0.0000070236,0.05000,1.345
195
+ 27,mlp.gate_proj,0.0000950958,0.05000,3.230
196
+ 27,mlp.up_proj,0.0000753671,0.05000,3.255
197
+ 27,mlp.down_proj,0.0000416984,0.05000,3.798
Models/50/quantize_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 4,
3
+ "group_size": 128,
4
+ "desc_act": false,
5
+ "lm_head": false,
6
+ "quant_method": "gptq",
7
+ "checkpoint_format": "gptq",
8
+ "pack_dtype": "int32",
9
+ "meta": {
10
+ "quantizer": [
11
+ "gptqmodel:5.8.0"
12
+ ],
13
+ "uri": "https://github.com/modelcloud/gptqmodel",
14
+ "damp_percent": 0.05,
15
+ "damp_auto_increment": 0.01,
16
+ "static_groups": false,
17
+ "true_sequential": true,
18
+ "mse": 0.0,
19
+ "gptaq": null,
20
+ "act_group_aware": true,
21
+ "failsafe": {
22
+ "strategy": "rtn",
23
+ "threshold": "0.5%",
24
+ "smooth": null
25
+ },
26
+ "offload_to_disk": true,
27
+ "offload_to_disk_path": "./gptqmodel_offload/pthyuejg-tgujncci/",
28
+ "pack_impl": "cpu",
29
+ "mock_quantization": false,
30
+ "gc_mode": "interval",
31
+ "wait_for_submodule_finalizers": false,
32
+ "auto_forward_data_parallel": true,
33
+ "hessian": {
34
+ "chunk_size": null,
35
+ "chunk_bytes": null,
36
+ "staging_dtype": "float32"
37
+ },
38
+ "vram_strategy": "exclusive"
39
+ },
40
+ "sym": true,
41
+ "format": "gptq"
42
+ }
Models/50/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c70650b4236027dc8db4abca6b918783a8ed2ee38cd69142f6dbbeb5945f876f
3
+ size 17210195
Models/50/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "bos_token": "<|begin_of_text|>",
4
+ "clean_up_tokenization_spaces": true,
5
+ "eos_token": "<|eot_id|>",
6
+ "is_local": false,
7
+ "model_input_names": [
8
+ "input_ids",
9
+ "attention_mask"
10
+ ],
11
+ "model_max_length": 131072,
12
+ "pad_token": "<|eot_id|>",
13
+ "tokenizer_class": "TokenizersBackend"
14
+ }