Abdohaaland commited on
Commit
f18dd20
·
verified ·
1 Parent(s): f235135

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -34,3 +34,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  Models/25/tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  Models/25/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ Models/50/tokenizer.json filter=lfs diff=lfs merge=lfs -text
Models/50/chat_template.jinja ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {{- bos_token }}
2
+ {%- if custom_tools is defined %}
3
+ {%- set tools = custom_tools %}
4
+ {%- endif %}
5
+ {%- if not tools_in_user_message is defined %}
6
+ {%- set tools_in_user_message = true %}
7
+ {%- endif %}
8
+ {%- if not date_string is defined %}
9
+ {%- if strftime_now is defined %}
10
+ {%- set date_string = strftime_now("%d %b %Y") %}
11
+ {%- else %}
12
+ {%- set date_string = "26 Jul 2024" %}
13
+ {%- endif %}
14
+ {%- endif %}
15
+ {%- if not tools is defined %}
16
+ {%- set tools = none %}
17
+ {%- endif %}
18
+
19
+ {#- This block extracts the system message, so we can slot it into the right place. #}
20
+ {%- if messages[0]['role'] == 'system' %}
21
+ {%- set system_message = messages[0]['content']|trim %}
22
+ {%- set messages = messages[1:] %}
23
+ {%- else %}
24
+ {%- set system_message = "" %}
25
+ {%- endif %}
26
+
27
+ {#- System message #}
28
+ {{- "<|start_header_id|>system<|end_header_id|>\n\n" }}
29
+ {%- if tools is not none %}
30
+ {{- "Environment: ipython\n" }}
31
+ {%- endif %}
32
+ {{- "Cutting Knowledge Date: December 2023\n" }}
33
+ {{- "Today Date: " + date_string + "\n\n" }}
34
+ {%- if tools is not none and not tools_in_user_message %}
35
+ {{- "You have access to the following functions. To call a function, please respond with JSON for a function call." }}
36
+ {{- 'Respond in the format {"name": function name, "parameters": dictionary of argument name and its value}.' }}
37
+ {{- "Do not use variables.\n\n" }}
38
+ {%- for t in tools %}
39
+ {{- t | tojson(indent=4) }}
40
+ {{- "\n\n" }}
41
+ {%- endfor %}
42
+ {%- endif %}
43
+ {{- system_message }}
44
+ {{- "<|eot_id|>" }}
45
+
46
+ {#- Custom tools are passed in a user message with some extra guidance #}
47
+ {%- if tools_in_user_message and not tools is none %}
48
+ {#- Extract the first user message so we can plug it in here #}
49
+ {%- if messages | length != 0 %}
50
+ {%- set first_user_message = messages[0]['content']|trim %}
51
+ {%- set messages = messages[1:] %}
52
+ {%- else %}
53
+ {{- raise_exception("Cannot put tools in the first user message when there's no first user message!") }}
54
+ {%- endif %}
55
+ {{- '<|start_header_id|>user<|end_header_id|>\n\n' -}}
56
+ {{- "Given the following functions, please respond with a JSON for a function call " }}
57
+ {{- "with its proper arguments that best answers the given prompt.\n\n" }}
58
+ {{- 'Respond in the format {"name": function name, "parameters": dictionary of argument name and its value}.' }}
59
+ {{- "Do not use variables.\n\n" }}
60
+ {%- for t in tools %}
61
+ {{- t | tojson(indent=4) }}
62
+ {{- "\n\n" }}
63
+ {%- endfor %}
64
+ {{- first_user_message + "<|eot_id|>"}}
65
+ {%- endif %}
66
+
67
+ {%- for message in messages %}
68
+ {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}
69
+ {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' }}
70
+ {%- elif 'tool_calls' in message %}
71
+ {%- if not message.tool_calls|length == 1 %}
72
+ {{- raise_exception("This model only supports single tool-calls at once!") }}
73
+ {%- endif %}
74
+ {%- set tool_call = message.tool_calls[0].function %}
75
+ {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}}
76
+ {{- '{"name": "' + tool_call.name + '", ' }}
77
+ {{- '"parameters": ' }}
78
+ {{- tool_call.arguments | tojson }}
79
+ {{- "}" }}
80
+ {{- "<|eot_id|>" }}
81
+ {%- elif message.role == "tool" or message.role == "ipython" %}
82
+ {{- "<|start_header_id|>ipython<|end_header_id|>\n\n" }}
83
+ {%- if message.content is mapping or message.content is iterable %}
84
+ {{- message.content | tojson }}
85
+ {%- else %}
86
+ {{- message.content }}
87
+ {%- endif %}
88
+ {{- "<|eot_id|>" }}
89
+ {%- endif %}
90
+ {%- endfor %}
91
+ {%- if add_generation_prompt %}
92
+ {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' }}
93
+ {%- endif %}
Models/50/config.json ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 128000,
8
+ "dtype": "bfloat16",
9
+ "eos_token_id": [
10
+ 128001,
11
+ 128008,
12
+ 128009
13
+ ],
14
+ "head_dim": 128,
15
+ "hidden_act": "silu",
16
+ "hidden_size": 3072,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 8192,
19
+ "max_position_embeddings": 131072,
20
+ "mlp_bias": false,
21
+ "model_type": "llama",
22
+ "num_attention_heads": 24,
23
+ "num_hidden_layers": 28,
24
+ "num_key_value_heads": 8,
25
+ "pad_token_id": 128009,
26
+ "pretraining_tp": 1,
27
+ "quantization_config": {
28
+ "bits": 4,
29
+ "checkpoint_format": "gptq",
30
+ "desc_act": false,
31
+ "format": "gptq",
32
+ "group_size": 128,
33
+ "lm_head": false,
34
+ "meta": {
35
+ "act_group_aware": true,
36
+ "auto_forward_data_parallel": true,
37
+ "damp_auto_increment": 0.01,
38
+ "damp_percent": 0.05,
39
+ "failsafe": {
40
+ "smooth": null,
41
+ "strategy": "rtn",
42
+ "threshold": "0.5%"
43
+ },
44
+ "gc_mode": "interval",
45
+ "gptaq": null,
46
+ "hessian": {
47
+ "chunk_bytes": null,
48
+ "chunk_size": null,
49
+ "staging_dtype": "float32"
50
+ },
51
+ "mock_quantization": false,
52
+ "mse": 0.0,
53
+ "offload_to_disk": true,
54
+ "offload_to_disk_path": "./gptqmodel_offload/hibgcqvy-hpwwspav/",
55
+ "pack_impl": "cpu",
56
+ "quantizer": [
57
+ "gptqmodel:5.8.0"
58
+ ],
59
+ "static_groups": false,
60
+ "true_sequential": true,
61
+ "uri": "https://github.com/modelcloud/gptqmodel",
62
+ "vram_strategy": "exclusive",
63
+ "wait_for_submodule_finalizers": false
64
+ },
65
+ "pack_dtype": "int32",
66
+ "quant_method": "gptq",
67
+ "sym": true
68
+ },
69
+ "rms_norm_eps": 1e-05,
70
+ "rope_parameters": {
71
+ "factor": 32.0,
72
+ "high_freq_factor": 4.0,
73
+ "low_freq_factor": 1.0,
74
+ "original_max_position_embeddings": 8192,
75
+ "rope_theta": 500000.0,
76
+ "rope_type": "llama3"
77
+ },
78
+ "tie_word_embeddings": true,
79
+ "transformers_version": "5.2.0",
80
+ "use_cache": true,
81
+ "vocab_size": 128256
82
+ }
Models/50/generation_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 128000,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 128001,
6
+ 128008,
7
+ 128009
8
+ ],
9
+ "temperature": 0.6,
10
+ "top_p": 0.9,
11
+ "transformers_version": "5.2.0"
12
+ }
Models/50/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4d04048ad86686c28387ddc010af3d4fe33a653e6a99352597d465ee7a6b201
3
+ size 2255767880
Models/50/quant_log.csv ADDED
@@ -0,0 +1,197 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ layer,module,loss,samples,damp,time
2
+ 0,self_attn.q_proj,0.0000139680,0.05000,5.305
3
+ 0,self_attn.v_proj,0.0000003759,0.05000,5.383
4
+ 0,self_attn.k_proj,0.0000070846,0.05000,5.402
5
+ 0,self_attn.o_proj,0.0000000157,0.05000,1.004
6
+ 0,mlp.gate_proj,0.0000068226,0.05000,2.417
7
+ 0,mlp.up_proj,0.0000059404,0.05000,2.418
8
+ 0,mlp.down_proj,0.0000000725,0.05000,2.874
9
+ 1,self_attn.k_proj,0.0000095077,0.05000,5.638
10
+ 1,self_attn.v_proj,0.0000010649,0.05000,5.697
11
+ 1,self_attn.q_proj,0.0000162416,0.05000,5.717
12
+ 1,self_attn.o_proj,0.0000000565,0.05000,1.076
13
+ 1,mlp.gate_proj,0.0000087613,0.05000,2.989
14
+ 1,mlp.up_proj,0.0000076847,0.05000,3.024
15
+ 1,mlp.down_proj,0.0000018727,0.05000,3.125
16
+ 2,self_attn.q_proj,0.0000670399,0.05000,5.696
17
+ 2,self_attn.v_proj,0.0000039832,0.05000,5.768
18
+ 2,self_attn.k_proj,0.0000399288,0.05000,5.786
19
+ 2,self_attn.o_proj,0.0000000645,0.05000,1.049
20
+ 2,mlp.gate_proj,0.0000140579,0.05000,2.393
21
+ 2,mlp.up_proj,0.0000118356,0.05000,2.407
22
+ 2,mlp.down_proj,0.0000002062,0.05000,3.033
23
+ 3,self_attn.q_proj,0.0000403938,0.05000,4.883
24
+ 3,self_attn.v_proj,0.0000036359,0.05000,4.970
25
+ 3,self_attn.k_proj,0.0000225691,0.05000,4.991
26
+ 3,self_attn.o_proj,0.0000000964,0.05000,1.053
27
+ 3,mlp.up_proj,0.0000156097,0.05000,2.391
28
+ 3,mlp.gate_proj,0.0000210012,0.05000,2.412
29
+ 3,mlp.down_proj,0.0000003030,0.05000,3.005
30
+ 4,self_attn.q_proj,0.0000433604,0.05000,4.959
31
+ 4,self_attn.v_proj,0.0000040748,0.05000,5.001
32
+ 4,self_attn.k_proj,0.0000225520,0.05000,5.032
33
+ 4,self_attn.o_proj,0.0000001913,0.05000,1.017
34
+ 4,mlp.gate_proj,0.0000279136,0.05000,2.466
35
+ 4,mlp.up_proj,0.0000184692,0.05000,2.476
36
+ 4,mlp.down_proj,0.0000004594,0.05000,2.948
37
+ 5,self_attn.v_proj,0.0000041336,0.05000,4.881
38
+ 5,self_attn.k_proj,0.0000375154,0.05000,4.976
39
+ 5,self_attn.q_proj,0.0000616927,0.05000,4.982
40
+ 5,self_attn.o_proj,0.0000002548,0.05000,1.100
41
+ 5,mlp.up_proj,0.0000208768,0.05000,2.398
42
+ 5,mlp.gate_proj,0.0000297058,0.05000,2.419
43
+ 5,mlp.down_proj,0.0000006227,0.05000,2.969
44
+ 6,self_attn.k_proj,0.0000261515,0.05000,5.002
45
+ 6,self_attn.q_proj,0.0000491912,0.05000,5.060
46
+ 6,self_attn.v_proj,0.0000037164,0.05000,5.079
47
+ 6,self_attn.o_proj,0.0000003613,0.05000,1.029
48
+ 6,mlp.gate_proj,0.0000324102,0.05000,2.340
49
+ 6,mlp.up_proj,0.0000227860,0.05000,2.352
50
+ 6,mlp.down_proj,0.0000007253,0.05000,2.979
51
+ 7,self_attn.v_proj,0.0000032542,0.05000,4.959
52
+ 7,self_attn.q_proj,0.0000409642,0.05000,4.990
53
+ 7,self_attn.k_proj,0.0000242736,0.05000,5.024
54
+ 7,self_attn.o_proj,0.0000005799,0.05000,1.028
55
+ 7,mlp.up_proj,0.0000234077,0.05000,2.298
56
+ 7,mlp.gate_proj,0.0000306611,0.05000,2.321
57
+ 7,mlp.down_proj,0.0000008050,0.05000,2.968
58
+ 8,self_attn.v_proj,0.0000039600,0.05000,4.906
59
+ 8,self_attn.k_proj,0.0000308434,0.05000,4.953
60
+ 8,self_attn.q_proj,0.0000512178,0.05000,4.986
61
+ 8,self_attn.o_proj,0.0000006049,0.05000,1.060
62
+ 8,mlp.up_proj,0.0000246837,0.05000,2.312
63
+ 8,mlp.gate_proj,0.0000330912,0.05000,2.334
64
+ 8,mlp.down_proj,0.0000008281,0.05000,3.030
65
+ 9,self_attn.q_proj,0.0000488215,0.05000,4.915
66
+ 9,self_attn.k_proj,0.0000289286,0.05000,4.963
67
+ 9,self_attn.v_proj,0.0000049874,0.05000,4.988
68
+ 9,self_attn.o_proj,0.0000006253,0.05000,1.045
69
+ 9,mlp.up_proj,0.0000248506,0.05000,2.336
70
+ 9,mlp.gate_proj,0.0000323409,0.05000,2.342
71
+ 9,mlp.down_proj,0.0000008285,0.05000,3.060
72
+ 10,self_attn.v_proj,0.0000038885,0.05000,4.946
73
+ 10,self_attn.k_proj,0.0000307636,0.05000,4.983
74
+ 10,self_attn.q_proj,0.0000493704,0.05000,5.019
75
+ 10,self_attn.o_proj,0.0000007088,0.05000,1.017
76
+ 10,mlp.gate_proj,0.0000328339,0.05000,2.391
77
+ 10,mlp.up_proj,0.0000270201,0.05000,2.409
78
+ 10,mlp.down_proj,0.0000009400,0.05000,2.972
79
+ 11,self_attn.q_proj,0.0000401640,0.05000,4.869
80
+ 11,self_attn.v_proj,0.0000045547,0.05000,4.928
81
+ 11,self_attn.k_proj,0.0000226958,0.05000,4.982
82
+ 11,self_attn.o_proj,0.0000008010,0.05000,1.034
83
+ 11,mlp.gate_proj,0.0000345544,0.05000,2.392
84
+ 11,mlp.up_proj,0.0000293561,0.05000,2.400
85
+ 11,mlp.down_proj,0.0000010772,0.05000,3.015
86
+ 12,self_attn.k_proj,0.0000361648,0.05000,4.865
87
+ 12,self_attn.q_proj,0.0000607128,0.05000,4.983
88
+ 12,self_attn.v_proj,0.0000052258,0.05000,5.000
89
+ 12,self_attn.o_proj,0.0000010185,0.05000,1.020
90
+ 12,mlp.gate_proj,0.0000371353,0.05000,2.335
91
+ 12,mlp.up_proj,0.0000316075,0.05000,2.347
92
+ 12,mlp.down_proj,0.0000012233,0.05000,2.968
93
+ 13,self_attn.v_proj,0.0000059725,0.05000,4.970
94
+ 13,self_attn.q_proj,0.0000619890,0.05000,5.033
95
+ 13,self_attn.k_proj,0.0000395530,0.05000,5.056
96
+ 13,self_attn.o_proj,0.0000011513,0.05000,1.011
97
+ 13,mlp.gate_proj,0.0000444610,0.05000,2.320
98
+ 13,mlp.up_proj,0.0000352364,0.05000,2.340
99
+ 13,mlp.down_proj,0.0000016179,0.05000,3.016
100
+ 14,self_attn.v_proj,0.0000069888,0.05000,5.068
101
+ 14,self_attn.q_proj,0.0000696170,0.05000,5.098
102
+ 14,self_attn.k_proj,0.0000319154,0.05000,5.126
103
+ 14,self_attn.o_proj,0.0000014018,0.05000,1.034
104
+ 14,mlp.gate_proj,0.0000475400,0.05000,2.334
105
+ 14,mlp.up_proj,0.0000374746,0.05000,2.356
106
+ 14,mlp.down_proj,0.0000019131,0.05000,2.958
107
+ 15,self_attn.v_proj,0.0000069070,0.05000,4.931
108
+ 15,self_attn.q_proj,0.0000694283,0.05000,4.980
109
+ 15,self_attn.k_proj,0.0000361245,0.05000,5.006
110
+ 15,self_attn.o_proj,0.0000009409,0.05000,1.067
111
+ 15,mlp.up_proj,0.0000371522,0.05000,2.324
112
+ 15,mlp.gate_proj,0.0000503156,0.05000,2.343
113
+ 15,mlp.down_proj,0.0000019074,0.05000,3.047
114
+ 16,self_attn.k_proj,0.0000399651,0.05000,4.850
115
+ 16,self_attn.v_proj,0.0000076146,0.05000,4.928
116
+ 16,self_attn.q_proj,0.0000704941,0.05000,4.971
117
+ 16,self_attn.o_proj,0.0000006443,0.05000,1.035
118
+ 16,mlp.up_proj,0.0000365405,0.05000,2.331
119
+ 16,mlp.gate_proj,0.0000506454,0.05000,2.351
120
+ 16,mlp.down_proj,0.0000017757,0.05000,2.968
121
+ 17,self_attn.q_proj,0.0000665387,0.05000,4.831
122
+ 17,self_attn.v_proj,0.0000073697,0.05000,4.925
123
+ 17,self_attn.k_proj,0.0000360675,0.05000,4.959
124
+ 17,self_attn.o_proj,0.0000005999,0.05000,1.135
125
+ 17,mlp.gate_proj,0.0000523846,0.05000,2.339
126
+ 17,mlp.up_proj,0.0000372537,0.05000,2.362
127
+ 17,mlp.down_proj,0.0000019583,0.05000,2.957
128
+ 18,self_attn.v_proj,0.0000092865,0.05000,4.932
129
+ 18,self_attn.q_proj,0.0000767019,0.05000,4.971
130
+ 18,self_attn.k_proj,0.0000423681,0.05000,5.010
131
+ 18,self_attn.o_proj,0.0000005972,0.05000,1.011
132
+ 18,mlp.gate_proj,0.0000545215,0.05000,2.314
133
+ 18,mlp.up_proj,0.0000397141,0.05000,2.335
134
+ 18,mlp.down_proj,0.0000020258,0.05000,2.967
135
+ 19,self_attn.k_proj,0.0000392652,0.05000,4.891
136
+ 19,self_attn.q_proj,0.0000678425,0.05000,4.933
137
+ 19,self_attn.v_proj,0.0000090537,0.05000,4.961
138
+ 19,self_attn.o_proj,0.0000008387,0.05000,1.025
139
+ 19,mlp.gate_proj,0.0000570506,0.05000,2.278
140
+ 19,mlp.up_proj,0.0000419815,0.05000,2.298
141
+ 19,mlp.down_proj,0.0000025156,0.05000,2.970
142
+ 20,self_attn.k_proj,0.0000415792,0.05000,4.948
143
+ 20,self_attn.q_proj,0.0000694669,0.05000,4.995
144
+ 20,self_attn.v_proj,0.0000106951,0.05000,5.019
145
+ 20,self_attn.o_proj,0.0000006934,0.05000,1.078
146
+ 20,mlp.up_proj,0.0000426496,0.05000,2.360
147
+ 20,mlp.gate_proj,0.0000549296,0.05000,2.369
148
+ 20,mlp.down_proj,0.0000024157,0.05000,2.975
149
+ 21,self_attn.q_proj,0.0000692564,0.05000,4.916
150
+ 21,self_attn.k_proj,0.0000408783,0.05000,4.951
151
+ 21,self_attn.v_proj,0.0000139237,0.05000,4.966
152
+ 21,self_attn.o_proj,0.0000008302,0.05000,1.034
153
+ 21,mlp.gate_proj,0.0000580645,0.05000,2.358
154
+ 21,mlp.up_proj,0.0000446525,0.05000,2.373
155
+ 21,mlp.down_proj,0.0000025897,0.05000,2.969
156
+ 22,self_attn.q_proj,0.0000659685,0.05000,4.862
157
+ 22,self_attn.k_proj,0.0000378584,0.05000,4.965
158
+ 22,self_attn.v_proj,0.0000134829,0.05000,4.988
159
+ 22,self_attn.o_proj,0.0000007891,0.05000,1.037
160
+ 22,mlp.up_proj,0.0000472200,0.05000,2.298
161
+ 22,mlp.gate_proj,0.0000617282,0.05000,2.326
162
+ 22,mlp.down_proj,0.0000030136,0.05000,3.058
163
+ 23,self_attn.k_proj,0.0000415418,0.05000,4.939
164
+ 23,self_attn.q_proj,0.0000675351,0.05000,4.964
165
+ 23,self_attn.v_proj,0.0000132672,0.05000,5.006
166
+ 23,self_attn.o_proj,0.0000010529,0.05000,1.018
167
+ 23,mlp.up_proj,0.0000528725,0.05000,2.424
168
+ 23,mlp.gate_proj,0.0000733665,0.05000,2.454
169
+ 23,mlp.down_proj,0.0000039680,0.05000,2.920
170
+ 24,self_attn.k_proj,0.0000455178,0.05000,4.938
171
+ 24,self_attn.v_proj,0.0000195001,0.05000,4.999
172
+ 24,self_attn.q_proj,0.0000731985,0.05000,5.013
173
+ 24,self_attn.o_proj,0.0000015971,0.05000,1.046
174
+ 24,mlp.gate_proj,0.0000816732,0.05000,2.348
175
+ 24,mlp.up_proj,0.0000576196,0.05000,2.363
176
+ 24,mlp.down_proj,0.0000048145,0.05000,2.924
177
+ 25,self_attn.v_proj,0.0000157814,0.05000,4.796
178
+ 25,self_attn.q_proj,0.0000655517,0.05000,4.887
179
+ 25,self_attn.k_proj,0.0000338076,0.05000,4.911
180
+ 25,self_attn.o_proj,0.0000018210,0.05000,1.030
181
+ 25,mlp.up_proj,0.0000605838,0.05000,2.382
182
+ 25,mlp.gate_proj,0.0000865645,0.05000,2.403
183
+ 25,mlp.down_proj,0.0000061517,0.05000,2.949
184
+ 26,self_attn.k_proj,0.0000397186,0.05000,5.017
185
+ 26,self_attn.q_proj,0.0000629629,0.05000,5.040
186
+ 26,self_attn.v_proj,0.0000218369,0.05000,5.064
187
+ 26,self_attn.o_proj,0.0000021068,0.05000,1.017
188
+ 26,mlp.gate_proj,0.0000916392,0.05000,2.277
189
+ 26,mlp.up_proj,0.0000632485,0.05000,2.296
190
+ 26,mlp.down_proj,0.0000102032,0.05000,2.917
191
+ 27,self_attn.v_proj,0.0000139737,0.05000,4.881
192
+ 27,self_attn.k_proj,0.0000263273,0.05000,4.936
193
+ 27,self_attn.q_proj,0.0000469980,0.05000,4.958
194
+ 27,self_attn.o_proj,0.0000071803,0.05000,1.008
195
+ 27,mlp.up_proj,0.0000702180,0.05000,2.307
196
+ 27,mlp.gate_proj,0.0000891004,0.05000,2.325
197
+ 27,mlp.down_proj,0.0000469706,0.05000,2.988
Models/50/quantize_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 4,
3
+ "group_size": 128,
4
+ "desc_act": false,
5
+ "lm_head": false,
6
+ "quant_method": "gptq",
7
+ "checkpoint_format": "gptq",
8
+ "pack_dtype": "int32",
9
+ "meta": {
10
+ "quantizer": [
11
+ "gptqmodel:5.8.0"
12
+ ],
13
+ "uri": "https://github.com/modelcloud/gptqmodel",
14
+ "damp_percent": 0.05,
15
+ "damp_auto_increment": 0.01,
16
+ "static_groups": false,
17
+ "true_sequential": true,
18
+ "mse": 0.0,
19
+ "gptaq": null,
20
+ "act_group_aware": true,
21
+ "failsafe": {
22
+ "strategy": "rtn",
23
+ "threshold": "0.5%",
24
+ "smooth": null
25
+ },
26
+ "offload_to_disk": true,
27
+ "offload_to_disk_path": "./gptqmodel_offload/hibgcqvy-hpwwspav/",
28
+ "pack_impl": "cpu",
29
+ "mock_quantization": false,
30
+ "gc_mode": "interval",
31
+ "wait_for_submodule_finalizers": false,
32
+ "auto_forward_data_parallel": true,
33
+ "hessian": {
34
+ "chunk_size": null,
35
+ "chunk_bytes": null,
36
+ "staging_dtype": "float32"
37
+ },
38
+ "vram_strategy": "exclusive"
39
+ },
40
+ "sym": true,
41
+ "format": "gptq"
42
+ }
Models/50/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c70650b4236027dc8db4abca6b918783a8ed2ee38cd69142f6dbbeb5945f876f
3
+ size 17210195
Models/50/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "bos_token": "<|begin_of_text|>",
4
+ "clean_up_tokenization_spaces": true,
5
+ "eos_token": "<|eot_id|>",
6
+ "is_local": false,
7
+ "model_input_names": [
8
+ "input_ids",
9
+ "attention_mask"
10
+ ],
11
+ "model_max_length": 131072,
12
+ "pad_token": "<|eot_id|>",
13
+ "tokenizer_class": "TokenizersBackend"
14
+ }