SheikhMiniMoE-Medical
A production-ready Mixture of Experts (MoE) model optimized for bilingual medical Q&A in English and Bengali.
Model Details
- Architecture: SheikhMiniMoE - Sparse Mixture of Experts with top-k routing
- Parameters: N/A
- Vocabulary Size: 5000
- Embedding Dimension: 256
- Hidden Dimension: 512
- Number of Experts: 8
- Top-K Routing: 2
Intended Use
This model is designed for humanitarian medical assistance, providing:
- Bilingual medical Q&A (English/Bengali)
- Safety-first responses with appropriate disclaimers
- Emergency keyword detection
- Knowledge base augmentation
Training Configuration
- Training Date: 2026-01-18
- Device: CPU-optimized for deployment in resource-constrained environments
Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
# Load model and tokenizer
model_name = "OsamaBinLikhon/SheikhMiniMoE-Medical"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
# Medical query
prompt = "What are symptoms of fever?"
inputs = tokenizer(prompt, return_tensors="pt")
# Generate response
outputs = model.generate(**inputs, max_new_tokens=100, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
Safety Notice
⚠️ This model provides medical information for educational purposes only. It is not a substitute for professional medical advice, diagnosis, or treatment. Always consult qualified healthcare providers for medical concerns.
License
Apache 2.0 - See LICENSE file for details.
Citation
@misc{SheikhMiniMoE,
title={SheikhMiniMoE: Bilingual Medical Assistant},
author={MiniMax Agent},
year={2025},
url={https://huggingface.co/OsamaBinLikhon/SheikhMiniMoE-Medical}
}
- Downloads last month
- 9
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support