mattator juliendenize commited on
Commit
fb82033
·
0 Parent(s):

Duplicate from mistralai/Ministral-3-8B-Instruct-2512-GGUF

Browse files

Co-authored-by: Julien Denize <juliendenize@users.noreply.huggingface.co>

.gitattributes ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Ministral-3-8B-Instruct-2512-BF16-mmproj.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Ministral-3-8B-Instruct-2512-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Ministral-3-8B-Instruct-2512-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Ministral-3-8B-Instruct-2512-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Ministral-3-8B-Instruct-2512-BF16.gguf filter=lfs diff=lfs merge=lfs -text
Ministral-3-8B-Instruct-2512-BF16-mmproj.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e799380f596d152ff4026a72d409ecc89c96cd437676804bc3f2ab3bd6b486ec
3
+ size 858283168
Ministral-3-8B-Instruct-2512-BF16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa785e6a56f240dacbc451ea4aaab501d11f5903005903cec6e0ba11c30e3357
3
+ size 16988084640
Ministral-3-8B-Instruct-2512-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33e7a72cf5e6e2cfc2f2847075acc013d68bba023e35310cef86b5cf8fdca761
3
+ size 5198911904
Ministral-3-8B-Instruct-2512-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a5454127ec772e2389f0e71a77fedb88b83d4366d8a69facd0cfd0898f04d35
3
+ size 6059268512
Ministral-3-8B-Instruct-2512-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd0ca58473b64618df4d08c1410cc25c910ca84e647833ebaf099e0fc0523b45
3
+ size 9029392800
README.md ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: vllm
3
+ language:
4
+ - en
5
+ - fr
6
+ - es
7
+ - de
8
+ - it
9
+ - pt
10
+ - nl
11
+ - zh
12
+ - ja
13
+ - ko
14
+ - ar
15
+ license: apache-2.0
16
+ inference: false
17
+ base_model:
18
+ - mistralai/Ministral-3-8B-Instruct-2512
19
+ extra_gated_description: >-
20
+ If you want to learn more about how we process your personal data, please read
21
+ our <a href="https://mistral.ai/terms/">Privacy Policy</a>.
22
+ tags:
23
+ - mistral-common
24
+ ---
25
+
26
+ # Ministral 3 8B Instruct 2512 GGUF
27
+
28
+ A balanced model in the Ministral 3 family, **Ministral 3 8B** is a powerful, efficient tiny language model with vision capabilities.
29
+
30
+ This model includes different quantization levels of the instruct post-trained version in **GGUF**, fine-tuned for instruction tasks, making it ideal for chat and instruction based use cases.
31
+
32
+ The Ministral 3 family is designed for edge deployment, capable of running on a wide range of hardware. Ministral 3 8B can even be deployed locally, capable of fitting in 12GB of VRAM in FP8, and less if further quantized.
33
+
34
+ Learn more in our [blog post](https://mistral.ai/news/mistral-3) and [paper](https://arxiv.org/abs/2601.08584).
35
+
36
+ ## Key Features
37
+ Ministral 3 8B consists of two main architectural components:
38
+ - **8.4B Language Model**
39
+ - **0.4B Vision Encoder**
40
+
41
+ The Ministral 3 8B Instruct model offers the following capabilities:
42
+ - **Vision**: Enables the model to analyze images and provide insights based on visual content, in addition to text.
43
+ - **Multilingual**: Supports dozens of languages, including English, French, Spanish, German, Italian, Portuguese, Dutch, Chinese, Japanese, Korean, Arabic.
44
+ - **System Prompt**: Maintains strong adherence and support for system prompts.
45
+ - **Agentic**: Offers best-in-class agentic capabilities with native function calling and JSON outputting.
46
+ - **Edge-Optimized**: Delivers best-in-class performance at a small scale, deployable anywhere.
47
+ - **Apache 2.0 License**: Open-source license allowing usage and modification for both commercial and non-commercial purposes.
48
+ - **Large Context Window**: Supports a 256k context window.
49
+
50
+ ### Recommended Settings
51
+
52
+ We recommend deploying with the following best practices:
53
+ - System Prompt: Define a clear environment and use case, including guidance on how to effectively leverage tools in agentic systems.
54
+ - Sampling Parameters: Use a **temperature below 0.1** for daily-driver and production environments ; Higher temperatures may be explored for creative use cases - developers are encouraged to experiment with alternative settings.
55
+ - Tools: Keep the set of tools well-defined and limit their number to the minimum required for the use case - Avoiding overloading the model with an excessive number of tools.
56
+ - Vision: When deploying with vision capabilities, we recommend maintaining an aspect ratio close to 1:1 (width-to-height) for images. Avoiding the use of overly thin or wide images - crop them as needed to ensure optimal performance.
57
+
58
+ ## License
59
+
60
+ This model is licensed under the [Apache 2.0 License](https://www.apache.org/licenses/LICENSE-2.0.txt).
61
+
62
+ *You must not use this model in a manner that infringes, misappropriates, or otherwise violates any third party’s rights, including intellectual property rights.*