Danfe AI โ Nepal Language Model
Nepal's first open-source AI model for 17+ languages.
Model Details
- Base Model: meta-llama/Llama-3.1-8B-Instruct
- Method: QLoRA (4-bit quantization + LoRA rank 16)
- Training Data: Nepali Wikipedia + curated Nepal knowledge
- Languages: Nepali, English, Maithili, Doteli, Bajhangi, Achhami, +11 more
Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
base = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-3.1-8B-Instruct")
model = PeftModel.from_pretrained(base, "danfe-ai/danfe-nepali-8b")
tokenizer = AutoTokenizer.from_pretrained("danfe-ai/danfe-nepali-8b")
Part of Danfe AI Ecosystem
- Chat UI, Translation, Speech, OCR
- Crowdsource training platform
- Focus on Sudurpashchim languages (Doteli, Bajhangi, Achhami)
Built with love for Nepal.
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Model tree for danfe-ai/danfe-nepali-8b
Base model
meta-llama/Llama-3.1-8B
Finetuned
meta-llama/Llama-3.1-8B-Instruct