You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Naturecode Dhivehi 270m

The first open-source Dhivehi language model optimized for natural conversations.

Model Description

Naturecode Dhivehi 270m is a fine-tuned language model specifically designed for Dhivehi, the official language of the Maldives. Built on Google's FunctionGemma-270m architecture, this model has been trained through a 6-phase curriculum learning approach on authentic Dhivehi text data.

Key Features

  • Native Dhivehi Support: Trained on authentic Dhivehi text including formal and informal language
  • Conversational: Optimized for natural dialogue in Dhivehi
  • Lightweight: 270M parameters - efficient for deployment
  • Curriculum Trained: Progressive training through 6 phases for robust language understanding

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

base_model = AutoModelForCausalLM.from_pretrained("google/functiongemma-270m-it")
model = PeftModel.from_pretrained(base_model, "hilarl/naturecode-dhivehi-270m")
tokenizer = AutoTokenizer.from_pretrained("google/functiongemma-270m-it")

prompt = "<start_of_turn>user\nކިހިނެއް؟<end_of_turn>\n<start_of_turn>model\n"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0]))

Training Details

  • Base Model: google/functiongemma-270m-it
  • Training Method: LoRA fine-tuning with curriculum learning (6 phases)
  • Training Data: Curated Dhivehi corpus including dictionary entries, formal text, informal conversations, and SFT data

Intended Use

  • Dhivehi language chatbots and assistants
  • Dhivehi text generation
  • Educational applications for Dhivehi language learning
  • Research on low-resource language models

Limitations

  • Primarily trained for conversational use; may not excel at technical domains
  • Performance may vary with highly specialized vocabulary
  • Should not be used for generating harmful or misleading content

License

Apache 2.0

Citation

@misc{naturecode-dhivehi-270m,
  author = {Naturecode},
  title = {Naturecode Dhivehi 270m},
  year = {2025},
  publisher = {HuggingFace},
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for hilarl/naturecode-dhivehi-270m

Finetuned
(282)
this model