---
base_model:
- LiquidAI/LFM2-1.2B
tags:
- text-generation-inference
- transformers
- unsloth
- lfm2
license: apache-2.0
language:
- en
datasets:
- MegaScience/MegaScience
---
# 🧪 MegaSciMoE‑1.2B
**A lightweight open MoE language model focused on science reasoning and instruction-following tasks.**
Fine-tuned using Unsloth on top of LiquidAI’s LFM2‑1.2B architecture.
---
## 🔬 Model Overview
* **Model Type:** Sparse Mixture-of-Experts (MoE)
* **Base:** LiquidAI LFM2‑1.2B
* **Fine-tuned with:** [Unsloth](https://unsloth.ai) LoRA
* **Focus Areas:** Physics, Chemistry, Biology, Scientific QA
* **Parameter Efficiency:** Only 2 experts active (\~250M active params)
* **Design Goal:** Run efficiently on consumer GPUs and edge devices
---
## 🧠 Capabilities
* Respond to structured science-related instructions
* Generate concise, factual explanations in STEM topics
* Follow multiturn question-answer patterns
* Summarize technical descriptions in plain language
---
## 🚀 Example Usage
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model = AutoModelForCausalLM.from_pretrained("yasserrmd/MegaSciMoE-1.2B", torch_dtype=torch.float16)
tokenizer = AutoTokenizer.from_pretrained("yasserrmd/MegaSciMoE-1.2B")
prompt = "What is the difference between an ionic and covalent bond?"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
---
## 📚 Training Data
MegaSciMoE-1.2B was fine-tuned on a blend of:
* Multiturn scientific dialogues
* Instruction-based question answering
* Simulated domain-specific prompts in physics, chemistry, and biology
* Scientific corpus extracted from open datasets
---
## 📎 Sample Prompts
```text
Q: What is a neutron made of?
A: A neutron is composed of three quarks (two down quarks and one up quark)...
---
Q: Define osmosis in simple terms.
A: Osmosis is the movement of water across a membrane...
---
Q: What is the role of mitochondria?
A: Mitochondria are the powerhouses of the cell...
```
---
## 🧪 Benchmark Intent
Planned evaluations on:
* ARC-Challenge
* PubMedQA
* ScienceQA
(*Note: No math-based datasets included in this build*)
---
## ⚙️ Technical Details
* LoRA fine-tuning with Unsloth on 1.2B MoE model
* Memory-efficient inference (fits on T4, L4, RTX 3060, even CPU)
* GGUF and FP16 variants to be released
---
## 📄 License
Apache 2.0 — free to use and modify with attribution. Ensure upstream license compatibility before commercial use.
---
## 🤝 Acknowledgements
* Thanks to [Unsloth](https://unsloth.ai), [LiquidAI](https://huggingface.co/liquidrock), and [Hugging Face](https://huggingface.co) for enabling this build
* Visuals rendered with scientific prompts and surreal-futuristic style
[
](https://github.com/unslothai/unsloth)