Qwen3-4B-MiniMight / README.md
ZeroXClem's picture
Update README.md
fb0edca verified
---
license: apache-2.0
tags:
- merge
- mergekit
- lazymergekit
- ZeroXClem/Qwen3-4B-MiniMight
- MiniMight
- ZeroXClem
- Code
- Cursor
base_model:
- TeichAI/Qwen3-4B-Thinking-2507-Claude-4.5-Opus-High-Reasoning-Distill
- bunnycore/Qwen3-4B-Mini-Merge
- bunnycore/Qwen3-4B-MegaMerge
- bunnycore/Qwen3-4B-Max-Merge
- ertghiu256/qwen3-4b-claude-sonnet-x-gemini-reasoning
- ZeroXClem/Qwen3-4B-Hermes-Axion-Pro
- Qwen/Qwen3-4B-Thinking-2507
pipeline_tag: text-generation
library_name: transformers
---
## 🧠 ZeroXClem/Qwen3-4B-MiniMight
![MiniMight](https://cdn-uploads.huggingface.co/production/uploads/64408cd43e0374802e19f454/Ahz0EEdsPBjHl3MCDH4tS.png)
> **"Small in scale. Mighty in mind."**
> A beautifully blended 4B model with *262k context length*, fusing *deep reasoning*, *code*, *safety*, and *creativity* β€” distilled through MergeKit’s `model_stock` magic.
---
## πŸ”§ Merge Configuration
```yaml
name: ZeroXClem/Qwen3-4B-MiniMight
base_model: Qwen/Qwen3-4B-Thinking-2507
dtype: bfloat16
merge_method: model_stock
models:
- model: TeichAI/Qwen3-4B-Thinking-2507-Claude-4.5-Opus-High-Reasoning-Distill
- model: bunnycore/Qwen3-4B-Mini-Merge
- model: bunnycore/Qwen3-4B-MegaMerge
- model: bunnycore/Qwen3-4B-Max-Merge
- model: ertghiu256/qwen3-4b-claude-sonnet-x-gemini-reasoning
- model: ZeroXClem/Qwen3-4B-Hermes-Axion-Pro
tokenizer_source: Qwen/Qwen3-4B-Thinking-2507
````
---
## 🌟 Highlights
* 🧠 **Reasoning-Centric Intelligence** β€” Optimized for multi-step thought, explanation, STEM logic, and symbolic problem solving.
* πŸͺ **262,144 Token Context Window** β€” Massive memory span for long documents, complex chains, or multi-turn workflows.
* 🧬 **Deep Merge of Code, RP, and Claude/Gemini Logic** β€” Captures both fluency and fidelity in multi-domain generation.
* πŸ” **Safe Alignment from Hermes & Axion** β€” Includes red-teamed safety-conscious merges to guide outputs.
* ✍️ **Elegant Dialogue & Creative Thought** β€” Engages in immersive roleplay, structured writing, and storytelling.
* ⚑ **Efficient 4B Inference** β€” Small enough for local GPUs, quantized formats, and mobile deployments.
---
## πŸ§ͺ Usage Example
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
model = AutoModelForCausalLM.from_pretrained("ZeroXClem/Qwen3-4B-MiniMight", device_map="auto", torch_dtype="auto")
tokenizer = AutoTokenizer.from_pretrained("ZeroXClem/Qwen3-4B-MiniMight")
messages = [{"role": "user", "content": "Explain quantum tunneling with an example."}]
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=384)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
🧠 *Use `enable_thinking=True` or `/think` with compatible frontends for enhanced reasoning control.*
---
## πŸ” Models Merged
| Source Model | Specialization |
| ----------------------------------------------------------------------- | ----------------------------------------------- |
| `TeichAI/Qwen3-4B-Thinking-2507-Claude-4.5-Opus-High-Reasoning-Distill` | High reasoning distilled from Claude Opus |
| `bunnycore/Qwen3-4B-Mini-Merge` | Mini-coder, Qwaifu, Darwin blend |
| `bunnycore/Qwen3-4B-MegaMerge` | AesCoder + OSS GPT + Qwaifu fusion |
| `bunnycore/Qwen3-4B-Max-Merge` | Maximal blend with strong instruction coherence |
| `ertghiu256/qwen3-4b-claude-sonnet-x-gemini-reasoning` | Gemini 3 Pro & Claude Sonnet enhanced |
| `ZeroXClem/Qwen3-4B-Hermes-Axion-Pro` | Reasoning + alignment + safety-core logic |
---
## πŸ’Ό Ideal For
* πŸ” Chain-of-thought reasoning & symbolic logic
* πŸ§ͺ Long-document summarization & analysis
* πŸ’¬ Roleplay & storytelling with high memory
* πŸ“š Tutoring & educational AI (math, code, science)
* βš–οΈ Safety-aligned assistants for deployment
* πŸ‘¨β€πŸ’» Code generation, refactoring, and walkthroughs
---
## 🚫 Limitations
* Long responses may truncate if `max_new_tokens` is too low
* Mixed training sources may result in style variance across domains
* Dialogue tone defaults to helpful/instructive rather than emotional/creative
---
## πŸ“œ License
**Apache 2.0**
Please verify license alignment when adapting for commercial or public-facing use.
---
## πŸ’Œ Credits
Thank you to bunnycore, Qwen Team, ertghiu256, TeichAI for the wonderful fine tunes and base models that made this merge possible!
Built by **ZeroXClem Team** β€” merging distillations, hugging intellect and ideas into tangible weights.
MergeKit mastery allows full distilled insight in one mighty little package.
> *Mini in name. Infinite in mind.* πŸ›Έ