|
|
--- |
|
|
license: apache-2.0 |
|
|
tags: |
|
|
- unsloth |
|
|
- Uncensored |
|
|
- text-generation-inference |
|
|
- transformers |
|
|
- unsloth |
|
|
- llama |
|
|
- trl |
|
|
- roleplay |
|
|
- conversational |
|
|
datasets: |
|
|
- iamketan25/roleplay-instructions-dataset |
|
|
- N-Bot-Int/Iris-Uncensored-R1 |
|
|
- N-Bot-Int/Moshpit-Combined-R2-Uncensored |
|
|
- N-Bot-Int/Mushed-Dataset-Uncensored |
|
|
- N-Bot-Int/Muncher-R1-Uncensored |
|
|
- N-Bot-Int/Millia-R1_DPO |
|
|
language: |
|
|
- en |
|
|
base_model: |
|
|
- N-Bot-Int/MiniMaid-L1 |
|
|
pipeline_tag: text-generation |
|
|
library_name: peft |
|
|
metrics: |
|
|
- character |
|
|
--- |
|
|
 |
|
|
# MiniMaid-L2 |
|
|
- MiniMaid-L2 is a Finetuned Model of MiniMaid-L1 model, with even big and higher quality dataset used to generated roleplaying |
|
|
Capabilities, MiniMaid-L2 also were extracted from Knowledge Distilling A Popular Roleplaying Model named NoroMaid-7B-DPO, |
|
|
Which we've used to enchanced its lacking Ends for coherent And Good Roleplaying Capabilities. |
|
|
|
|
|
- MiniMaid-L2 Outcompete its predecessor as it uses a Clever Knowledge distilling to transfer Knowledge from NoroMaid, |
|
|
And Finetuned it, building on top of MiniMaid-L1 to Produce a better AI model. Sacrificing Some Non-noticable |
|
|
Token-Generation speed, with a near perfect and Competitive Model against **3b Alternatives**! |
|
|
|
|
|
# MiniMaid-L1 Base-Model Card Procedure: |
|
|
- **MiniMaid-L1** achieve a good Performance through process of DPO and Combined Heavy Finetuning, To Prevent Overfitting, |
|
|
We used high LR decays, And Introduced Randomization techniques to prevent the AI from learning and memorizing, |
|
|
However since training this on Google Colab is difficult, the Model might underperform or underfit on specific tasks |
|
|
Or overfit on knowledge it manage to latched on! However please be guided that we did our best, and it will improve as we move onwards! |
|
|
|
|
|
- MiniMaid-L2 is Another Instance of Our Smallest Model Yet! if you find any issue, then please don't hesitate to email us at: |
|
|
[nexus.networkinteractives@gmail.com](mailto:nexus.networkinteractives@gmail.com) |
|
|
about any overfitting, or improvements for the future Model **V3**, |
|
|
Once again feel free to Modify the LORA to your likings, However please consider Adding this Page |
|
|
for credits and if you'll increase its **Dataset**, then please handle it with care and ethical considerations |
|
|
|
|
|
- MiniMaid-L2 is |
|
|
- **Developed by:** N-Bot-Int |
|
|
- **License:** apache-2.0 |
|
|
- **Parent Model from model:** unsloth/llama-3.2-3b-instruct-unsloth-bnb-1bit |
|
|
- **Dataset Combined Using:** Mosher-R1(Propietary Software) |
|
|
|
|
|
- MiniMaid-L1 Official Metric Score |
|
|
 |
|
|
- Metrics Made By **ItsMeDevRoland** |
|
|
Which compares: |
|
|
- **MiniMaid-L1 GGUFF** |
|
|
- **MiniMaid-L2 GGUFF** |
|
|
Which are All Ranked with the Same Prompt, Same Temperature, Same Hardware(Google Colab), |
|
|
To Properly Showcase the differences and strength of the Models |
|
|
|
|
|
- **Visit Below to See details!** |
|
|
|
|
|
--- |
|
|
# 🧵 MiniMaid-L2: Small Size, Big Bite — The Next-Gen Roleplay Assistant |
|
|
> She’s sharper, deeper, and more immersive. And this time? She doesn’t just hold her own — she wins. |
|
|
|
|
|
 |
|
|
|
|
|
# MiniMaid-L2 builds on the scrappy L1 foundation and takes the lead over 3B giants like Hermes, Dolphin, and DeepSeek, with better consistency, longer outputs, and a massive boost to immersion. |
|
|
- 💬 Roleplay Evaluation (v1) |
|
|
- 🧠 Character Consistency: 0.84 |
|
|
- 🌊 Immersion: 0.47 |
|
|
-🧮 Overall RP Score: 0.76 |
|
|
- ✏️ Length Score: 1.00 |
|
|
- L2 scored +0.25 higher overall than L1, while beating top-tier 3B models in every major RP metric. |
|
|
|
|
|
|
|
|
# 📊 Efficient AND Smart |
|
|
- Inference Time: 54.2s — still 3x faster than Hermes |
|
|
- Tokens/sec: 6.88 — near-instant on consumer GPUs |
|
|
- BLEU/ROUGE-L: Stronger n-gram overlap than any 3B rival |
|
|
|
|
|
# MiniMaid-L2 shows that distilled models can outperform much larger ones — when trained right, even 1B can be the boss. |
|
|
- 🛠️ MiniMaid is Built For |
|
|
- High-fidelity RP generation |
|
|
- Lower-latency systems |
|
|
- Custom, character-driven storytelling |
|
|
|
|
|
> 🌱 L2 is the turning point — with upgraded conditioning, tighter personality anchoring, and narrative-aware outputs, she's evolving fast. |
|
|
|
|
|
“MiniMaid-L2 doesn’t just punch above her weight — she’s taking belts. A tighter model, a stronger performer, and still tiny enough to run on a toaster. RP just got smarter.” |
|
|
|
|
|
--- |
|
|
|
|
|
- # Notice |
|
|
- **For a Good Experience, Please use** |
|
|
- Low temperature 1.5, min_p = 0.1 and max_new_tokens = 128 |
|
|
|
|
|
|
|
|
- # Detail card: |
|
|
- Parameter |
|
|
- 1 Billion Parameters |
|
|
- (Please visit your GPU Vendor if you can Run 1B models) |
|
|
|
|
|
- Finetuning tool: |
|
|
- Unsloth AI |
|
|
- This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. |
|
|
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth) |
|
|
- Fine-tuned Using: |
|
|
- Google Colab |