mistral-7b-python / README.md
AntoineChatry's picture
Update README.md
450486b verified
---
tags:
- gguf
- llama.cpp
- unsloth
- mistral
- python
base_model:
- mistralai/Mistral-7B-Instruct-v0.3
---
# mistral-7b-python-gguf
Conversational Python fine-tune of Mistral 7B exported to GGUF format for local inference.
- Base model: Mistral 7B
- Fine-tuning framework: Unsloth
- Format: GGUF
- Author: AntoineChatry
---
# ⚠️ Disclaimer
This is an **early experimental fine-tune**.
It is **not production-ready**, not fully aligned, and not optimized for reliability or long-form reasoning.
This project was created primarily for learning and experimentation.
Please do not expect state-of-the-art coding performance.
---
# Model Overview
This model is a conversational fine-tune of Mistral 7B trained primarily on:
- ShareGPT-style conversations
- Python-focused discussions
- Coding Q&A format
The objective was to:
- Experiment with fine-tuning
- Build a conversational Python model
- Export to GGUF for llama.cpp compatibility
- Test local inference workflows
No RLHF or advanced alignment was applied beyond the base model.
---
# Known Limitations
## Repetition Issues
- Frequently repeats phrases like:
> "Here's the code:"
- Can loop or restate similar sentences
- Overuses patterns learned from dataset formatting
## Weak Long-Form Explanations
- Struggles with multi-paragraph structured reasoning
- May repeat itself when asked for detailed explanations
- Limited depth on conceptual explanations
## Instruction Following
- Not fully aligned
- May ignore strict formatting constraints
- Tends to prioritize generating code over detailed explanations
## Dataset Bias
- Strong ShareGPT conversational tone
- Python-heavy bias
- Some templated response structure
---
# What Works Reasonably Well
- Short Python snippets
- Basic debugging help
- Simple function generation
- Conversational coding prompts
Best performance is observed when:
- Prompts are clear and direct
- Expected output is short
- Tasks are code-focused
---
# Training Details
- Base: Mistral 7B
- Dataset format: ShareGPT-style conversational dataset (Python-oriented)
- Fine-tuned using Unsloth notebooks
- Converted to GGUF for llama.cpp compatibility
- Quantized version included (Q4_K_M)
No additional safety tuning or post-training optimization was applied.
---
# Example Usage
This model was finetuned and converted to GGUF format using Unsloth.
## llama.cpp
For text-only LLMs:
```bash
llama-cli -hf AntoineChatry/mistral-7b-python-gguf --jinja
```
For multimodal models:
```bash
llama-mtmd-cli -hf AntoineChatry/mistral-7b-python-gguf --jinja
```
---
## Available Model files:
- `mistral-7b-instruct-v0.3.Q4_K_M.gguf`
---
# Ollama
An Ollama Modelfile is included for easy deployment.
Example:
```bash
ollama create mistral-python -f Modelfile
ollama run mistral-python
```
---
# Why This Model Is Public
This model represents a learning milestone.
Sharing imperfect models helps:
- Document fine-tuning progress
- Enable experimentation
- Collect feedback
- Iterate toward better versions
This is not a finished product.
---
# Unsloth
This model was trained 2x faster using Unsloth.
https://github.com/unslothai/unsloth
<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>
---
# License
Please refer to the original Mistral 7B license from Mistral AI.