| | --- |
| | license: apache-2.0 |
| | tags: |
| | - merge |
| | - mergekit |
| | - lazymergekit |
| | - ZeroXClem/Qwen3-4B-MiniMight |
| | - MiniMight |
| | - ZeroXClem |
| | - Code |
| | - Cursor |
| | base_model: |
| | - TeichAI/Qwen3-4B-Thinking-2507-Claude-4.5-Opus-High-Reasoning-Distill |
| | - bunnycore/Qwen3-4B-Mini-Merge |
| | - bunnycore/Qwen3-4B-MegaMerge |
| | - bunnycore/Qwen3-4B-Max-Merge |
| | - ertghiu256/qwen3-4b-claude-sonnet-x-gemini-reasoning |
| | - ZeroXClem/Qwen3-4B-Hermes-Axion-Pro |
| | - Qwen/Qwen3-4B-Thinking-2507 |
| | pipeline_tag: text-generation |
| | library_name: transformers |
| | --- |
| | |
| | ## π§ ZeroXClem/Qwen3-4B-MiniMight |
| |
|
| |  |
| |
|
| | > **"Small in scale. Mighty in mind."** |
| | > A beautifully blended 4B model with *262k context length*, fusing *deep reasoning*, *code*, *safety*, and *creativity* β distilled through MergeKitβs `model_stock` magic. |
| | |
| | --- |
| | |
| | ## π§ Merge Configuration |
| | |
| | ```yaml |
| | name: ZeroXClem/Qwen3-4B-MiniMight |
| | base_model: Qwen/Qwen3-4B-Thinking-2507 |
| | dtype: bfloat16 |
| | merge_method: model_stock |
| | models: |
| | - model: TeichAI/Qwen3-4B-Thinking-2507-Claude-4.5-Opus-High-Reasoning-Distill |
| | - model: bunnycore/Qwen3-4B-Mini-Merge |
| | - model: bunnycore/Qwen3-4B-MegaMerge |
| | - model: bunnycore/Qwen3-4B-Max-Merge |
| | - model: ertghiu256/qwen3-4b-claude-sonnet-x-gemini-reasoning |
| | - model: ZeroXClem/Qwen3-4B-Hermes-Axion-Pro |
| | tokenizer_source: Qwen/Qwen3-4B-Thinking-2507 |
| | ```` |
| | |
| | --- |
| | |
| | ## π Highlights |
| | |
| | * π§ **Reasoning-Centric Intelligence** β Optimized for multi-step thought, explanation, STEM logic, and symbolic problem solving. |
| | * πͺ **262,144 Token Context Window** β Massive memory span for long documents, complex chains, or multi-turn workflows. |
| | * 𧬠**Deep Merge of Code, RP, and Claude/Gemini Logic** β Captures both fluency and fidelity in multi-domain generation. |
| | * π **Safe Alignment from Hermes & Axion** β Includes red-teamed safety-conscious merges to guide outputs. |
| | * βοΈ **Elegant Dialogue & Creative Thought** β Engages in immersive roleplay, structured writing, and storytelling. |
| | * β‘ **Efficient 4B Inference** β Small enough for local GPUs, quantized formats, and mobile deployments. |
| | |
| | --- |
| | |
| | ## π§ͺ Usage Example |
| | |
| | ```python |
| | from transformers import AutoTokenizer, AutoModelForCausalLM |
| | |
| | model = AutoModelForCausalLM.from_pretrained("ZeroXClem/Qwen3-4B-MiniMight", device_map="auto", torch_dtype="auto") |
| | tokenizer = AutoTokenizer.from_pretrained("ZeroXClem/Qwen3-4B-MiniMight") |
| | |
| | messages = [{"role": "user", "content": "Explain quantum tunneling with an example."}] |
| | prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) |
| | |
| | inputs = tokenizer(prompt, return_tensors="pt").to(model.device) |
| | outputs = model.generate(**inputs, max_new_tokens=384) |
| | |
| | print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
| | ``` |
| | |
| | π§ *Use `enable_thinking=True` or `/think` with compatible frontends for enhanced reasoning control.* |
| | |
| | --- |
| | |
| | ## π Models Merged |
| | |
| | | Source Model | Specialization | |
| | | ----------------------------------------------------------------------- | ----------------------------------------------- | |
| | | `TeichAI/Qwen3-4B-Thinking-2507-Claude-4.5-Opus-High-Reasoning-Distill` | High reasoning distilled from Claude Opus | |
| | | `bunnycore/Qwen3-4B-Mini-Merge` | Mini-coder, Qwaifu, Darwin blend | |
| | | `bunnycore/Qwen3-4B-MegaMerge` | AesCoder + OSS GPT + Qwaifu fusion | |
| | | `bunnycore/Qwen3-4B-Max-Merge` | Maximal blend with strong instruction coherence | |
| | | `ertghiu256/qwen3-4b-claude-sonnet-x-gemini-reasoning` | Gemini 3 Pro & Claude Sonnet enhanced | |
| | | `ZeroXClem/Qwen3-4B-Hermes-Axion-Pro` | Reasoning + alignment + safety-core logic | |
| | |
| | --- |
| | |
| | ## πΌ Ideal For |
| | |
| | * π Chain-of-thought reasoning & symbolic logic |
| | * π§ͺ Long-document summarization & analysis |
| | * π¬ Roleplay & storytelling with high memory |
| | * π Tutoring & educational AI (math, code, science) |
| | * βοΈ Safety-aligned assistants for deployment |
| | * π¨βπ» Code generation, refactoring, and walkthroughs |
| | |
| | --- |
| | |
| | ## π« Limitations |
| | |
| | * Long responses may truncate if `max_new_tokens` is too low |
| | * Mixed training sources may result in style variance across domains |
| | * Dialogue tone defaults to helpful/instructive rather than emotional/creative |
| | |
| | --- |
| | |
| | ## π License |
| | |
| | **Apache 2.0** |
| | Please verify license alignment when adapting for commercial or public-facing use. |
| |
|
| | --- |
| |
|
| | ## π Credits |
| |
|
| | Thank you to bunnycore, Qwen Team, ertghiu256, TeichAI for the wonderful fine tunes and base models that made this merge possible! |
| | Built by **ZeroXClem Team** β merging distillations, hugging intellect and ideas into tangible weights. |
| | MergeKit mastery allows full distilled insight in one mighty little package. |
| |
|
| | > *Mini in name. Infinite in mind.* πΈ |