witflag armand0e commited on
Commit
7384d3c
·
0 Parent(s):

Duplicate from TeichAI/Qwen3-14B-Gemini-3-Pro-Preview-High-Reasoning-Distill-GGUF

Browse files
.gitattributes ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ qwen3-14b.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
37
+ qwen3-14b.F16.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Qwen3-14B-Gemini-3-Pro-Preview-Distill.bf16.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Qwen3-14B-Gemini-3-Pro-Preview-Distill.q8_0.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Qwen3-14B-Gemini-3-Pro-Preview-Distill.q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Qwen3-14B-Gemini-3-Pro-Preview-Distill.q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Qwen3-14B-Gemini-3-Pro-Preview-Distill.iq4_nl.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Qwen3-14B-Gemini-3-Pro-Preview-Distill.q3_k_s.gguf filter=lfs diff=lfs merge=lfs -text
Modelfile ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ FROM qwen3-14b.Q8_0.gguf
3
+ TEMPLATE """{{- if .Messages }}
4
+ {{- if or .System .Tools }}<|im_start|>system
5
+ {{- if .System }}
6
+ {{ .System }}
7
+ {{- end }}
8
+ {{- if .Tools }}
9
+
10
+ # Tools
11
+
12
+ You may call one or more functions to assist with the user query.
13
+
14
+ You are provided with function signatures within <tools></tools> XML tags:
15
+ <tools>
16
+ {{- range .Tools }}
17
+ {"type": "function", "function": {{ .Function }}}
18
+ {{- end }}
19
+ </tools>
20
+
21
+ For each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:
22
+ <tool_call>
23
+ {"name": <function-name>, "arguments": <args-json-object>}
24
+ </tool_call>
25
+ {{- end }}<|im_end|>
26
+ {{ end }}
27
+ {{- range $i, $_ := .Messages }}
28
+ {{- $last := eq (len (slice $.Messages $i)) 1 -}}
29
+ {{- if eq .Role "user" }}<|im_start|>user
30
+ {{ .Content }}<|im_end|>
31
+ {{ else if eq .Role "assistant" }}<|im_start|>assistant
32
+ {{ if .Content }}{{ .Content }}
33
+ {{- else if .ToolCalls }}<tool_call>
34
+ {{ range .ToolCalls }}{"name": "{{ .Function.Name }}", "arguments": {{ .Function.Arguments }}}
35
+ {{ end }}</tool_call>
36
+ {{- end }}{{ if not $last }}<|im_end|>
37
+ {{ end }}
38
+ {{- else if eq .Role "tool" }}<|im_start|>user
39
+ <tool_response>
40
+ {{ .Content }}
41
+ </tool_response><|im_end|>
42
+ {{ end }}
43
+ {{- if and (ne .Role "assistant") $last }}<|im_start|>assistant
44
+ {{ end }}
45
+ {{- end }}
46
+ {{- else }}
47
+ {{- if .System }}<|im_start|>system
48
+ {{ .System }}<|im_end|>
49
+ {{ end }}{{ if .Prompt }}<|im_start|>user
50
+ {{ .Prompt }}<|im_end|>
51
+ {{ end }}<|im_start|>assistant
52
+ {{ end }}{{ .Response }}{{ if .Response }}<|im_end|>{{ end }}"""
53
+ PARAMETER stop "<|im_end|>"
54
+ PARAMETER stop "<|im_start|>"
55
+ PARAMETER temperature 0.6
56
+ PARAMETER min_p 0.0
57
+ PARAMETER top_k 20
58
+ PARAMETER top_p 0.95
59
+ PARAMETER repeat_penalty 1
Qwen3-14B-Gemini-3-Pro-Preview-Distill.bf16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df7ac3d37cdeaad0176ccd0b340965e1c118c78f82e5cc399bfdb13c1ca09eaf
3
+ size 29543424480
Qwen3-14B-Gemini-3-Pro-Preview-Distill.iq4_nl.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df02594508a577ea01e99317231b1a844f455fbac8dc4e317ebd9122771d7a92
3
+ size 8597069280
Qwen3-14B-Gemini-3-Pro-Preview-Distill.q3_k_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a38fb51d9c2e052b13eab474c4524934b5390ae5201fa17d1aaad039a8fea85
3
+ size 7321313760
Qwen3-14B-Gemini-3-Pro-Preview-Distill.q3_k_s.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:026c90bab5018375e520235716fa19563ea20e04425a17f0a534be00b6da8a4b
3
+ size 6657106400
Qwen3-14B-Gemini-3-Pro-Preview-Distill.q4_k_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc76f75c6e83b5dcfe1eaffca087297d5a5150e990da56d949de9cdc08b216dd
3
+ size 9001754080
Qwen3-14B-Gemini-3-Pro-Preview-Distill.q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc96d145e88a1b4dd50dd01f7e6cc5a383b6e7d5233a93751de27aeffdfe5f03
3
+ size 15698534880
README.md ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - gguf
4
+ - llama.cpp
5
+ - unsloth
6
+ datasets:
7
+ - TeichAI/gemini-3-pro-preview-high-reasoning-1000x
8
+ base_model:
9
+ - TeichAI/Qwen3-14B-Gemini-3-Pro-Preview-High-Reasoning-Distill
10
+ ---
11
+
12
+ # Qwen3 14B Gemini 3 Pro Preview Reasoning Distill
13
+
14
+ This model was trained on a **Gemini 3 Pro Preview** dataset with a high reasoning effort.
15
+
16
+ - &#129302; Related Models:
17
+ | Model | Effective parameters | Active parameters |
18
+ | ------------- | ------------- | ------------- |
19
+ | [`TeichAI/Qwen3-8B-Gemini-3-Pro-Preview-Distill-GGUF`](https://huggingface.co/TeichAI/Qwen3-8B-Gemini-3-Pro-Preview-Distill-GGUF) | 8 B | 8 B |
20
+ | [`TeichAI/Qwen3-4B-Thinking-2507-Gemini-3-Pro-Preview-High-Reasoning-Distill-GGUF`](https://huggingface.co/TeichAI/Qwen3-4B-Thinking-2507-Gemini-3-Pro-Preview-High-Reasoning-Distill-GGUF) | 4 B | 4 B |
21
+
22
+
23
+ - 🧬 Datasets:
24
+ - `TeichAI/gemini-3-pro-preview-high-reasoning-1000x`
25
+
26
+ - 🏗 Base Model:
27
+ - `unsloth/Qwen3-14B`
28
+
29
+ - &#9889; Use cases:
30
+ - Coding
31
+ - Science
32
+
33
+ - &#8721; Stats (Dataset)
34
+ - Costs: $ 32.7 (USD)
35
+ - Total tokens (input + output): 2.73 M
36
+
37
+ ---
38
+
39
+ This qwen3 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
config.json ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "torch_dtype": "bfloat16",
8
+ "eos_token_id": 151645,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 5120,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 17408,
14
+ "layer_types": [
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention",
53
+ "full_attention",
54
+ "full_attention"
55
+ ],
56
+ "max_length": 8192,
57
+ "max_position_embeddings": 40960,
58
+ "max_window_layers": 40,
59
+ "model_type": "qwen3",
60
+ "num_attention_heads": 40,
61
+ "num_hidden_layers": 40,
62
+ "num_key_value_heads": 8,
63
+ "pad_token_id": 151654,
64
+ "rms_norm_eps": 1e-06,
65
+ "rope_scaling": null,
66
+ "rope_theta": 1000000,
67
+ "sliding_window": null,
68
+ "tie_word_embeddings": false,
69
+ "transformers_version": "4.56.2",
70
+ "unsloth_fixed": true,
71
+ "unsloth_version": "2025.11.3",
72
+ "use_cache": true,
73
+ "use_sliding_window": false,
74
+ "vocab_size": 151936
75
+ }
qwen3-14b.F16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1cb62014d7afd9c327ddf51dffae77971523beaa1f4316a27eac00eb7563d7e
3
+ size 29543424000
qwen3-14b.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a84b5ad340f6fe357f4118b5e86444d82d7ced7fb27a7144793eff576ae3c74b
3
+ size 15698534400