๐ท๏ธ MahaBERT v2 โ DAFT Fine-Tuned Model
Authors: Tanvi Somani , Aryan Babare , Samyak Bora
Model Type: Masked Language Model (MLM)
Base Model: MahaBERT v2
Training Method: DAFT (Domain-Adaptive Fine-Tuning)
Language: Marathi ๐ฎ๐ณ
Framework: HuggingFace Transformers
๐ Overview
mahabert-v2-daft is a Domain-Adaptive Fine-Tuned (DAFT) version of the original MahaBERT v2 model.
The model has been further trained on a large collection of unlabeled Marathi text to adapt it to the specific domain used in downstream Marathi NLP tasks such as:
- Sentiment analysis
- Text classification
- Emotion detection
- Topic analysis
- General Marathi language understanding tasks
DAFT helps the model learn domain-specific vocabulary, patterns, and semantics, improving downstream task performance without requiring labeled data.
๐ง What is DAFT?
DAFT (Domain Adaptive Fine-Tuning) is a continued pre-training method where:
- You take a pretrained language model
- Feed it large amounts of unlabeled domain text
- Train it again using the Masked Language Modeling (MLM) objective
This process improves the modelโs understanding of domain-specific words, idioms, and sentence structures.
๐ Training Details
| Item | Details |
|---|---|
| Base Model | MahaBERT v2 |
| Method | DAFT (continued pre-training) |
| Objective | Masked Language Modeling (MLM) |
| Dataset | Unlabeled Marathi text (domain-specific) |
| Batch Size | As per Colab training setup |
| Epochs | Several thousand steps (as seen in training logs) |
| Hardware | Google Colab (T4 GPU) |
| Optimizer | AdamW |
| Precision | FP32 |
๐ Files Included
The repository contains:
model.safetensorsโ model weightsconfig.jsonโ model architecturetokenizer.json,tokenizer_config.jsonโ tokenizer settingsvocab.txtโ BERT vocabularyspecial_tokens_map.jsonโ CLS, SEP, PAD, MASK tokenstraining_args.binโ training configuration
๐งฉ Usage
๐น Load the Model
from transformers import AutoTokenizer, AutoModelForMaskedLM
model_id = "aryanx16/mahabert-v2-daft"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForMaskedLM.from_pretrained(model_id)
- Downloads last month
- 2
Model tree for aryanx16/mahabert-v2-daft
Base model
l3cube-pune/marathi-bert-v2