|
|
--- |
|
|
base_model: |
|
|
- N-Bot-Int/MistThena7B |
|
|
tags: |
|
|
- text-generation-inference |
|
|
- transformers |
|
|
- mistral |
|
|
- rp |
|
|
- gguf |
|
|
language: |
|
|
- en |
|
|
license: apache-2.0 |
|
|
datasets: |
|
|
- N-Bot-Int/Iris-Uncensored-R1 |
|
|
- N-Bot-Int/Moshpit-Combined-R2-Uncensored |
|
|
- N-Bot-Int/Mushed-Dataset-Uncensored |
|
|
- N-Bot-Int/Muncher-R1-Uncensored |
|
|
- unalignment/toxic-dpo-v0.1 |
|
|
library_name: transformers |
|
|
new_version: N-Bot-Int/MistThena7BV2-GGUF |
|
|
--- |
|
|
# Support Us Through |
|
|
- [](https://ko-fi.com/J3J61D8NHV) |
|
|
- [https://ko-fi.com/nexusnetworkint](Official Ko-FI link!) |
|
|
|
|
|
 |
|
|
# GGUF Version |
|
|
**GGUF** with Quants! Allowing you to run models using KoboldCPP and other AI Environments! |
|
|
|
|
|
|
|
|
# Quantizations: |
|
|
| Quant Type | Benefits | Cons | |
|
|
|---------------|---------------------------------------------------|---------------------------------------------------| |
|
|
| **Q4_K_M** | β
Smallest size (fastest inference) | β Lowest accuracy compared to other quants | |
|
|
| | β
Requires the least VRAM/RAM | β May struggle with complex reasoning | |
|
|
| | β
Ideal for edge devices & low-resource setups | β Can produce slightly degraded text quality | |
|
|
| **Q5_K_M** | β
Better accuracy than Q4, while still compact | β Slightly larger model size than Q4 | |
|
|
| | β
Good balance between speed and precision | β Needs a bit more VRAM than Q4 | |
|
|
| | β
Works well on mid-range GPUs | β Still not as accurate as higher-bit models | |
|
|
| **Q8_0** | β
Highest accuracy (closest to full model) | β Requires significantly more VRAM/RAM | |
|
|
| | β
Best for complex reasoning & detailed outputs | β Slower inference compared to Q4 & Q5 | |
|
|
| | β
Suitable for high-end GPUs & serious workloads | β Larger file size (takes more storage) | |
|
|
|
|
|
# Model Details: |
|
|
Read the Model details on huggingface |
|
|
[Model Detail Here!](https://huggingface.co/N-Bot-Int/MistThena7B) |