MuRIL_WR / README.md
Raj411's picture
Create README.md
172c671 verified
---
license: cc-by-4.0
datasets:
- DSL-13-SRMAP/Telugu-Dataset
language:
- te
tags:
- sentiment-analysis
- text-classification
- telugu
- indic-languages
- muril
- rationale-supervision
- explainable-ai
base_model: google/muril-base-cased
pipeline_tag: text-classification
metrics:
- accuracy
- f1
- auroc
---
# MuRIL_WR
## Model Description
**MuRIL_WR** is a Telugu sentiment classification model built on **MuRIL (Multilingual Representations for Indian Languages)**, a Transformer-based BERT model specifically designed for **Indian languages**, including Telugu and English.
MuRIL is pretrained on a **large and diverse corpus of Indian language text**, including web data, religious scriptures, and news content. In contrast to general multilingual models such as mBERT and XLM-R, MuRIL is better suited to capture **Telugu morphology, syntax, and linguistic structure**.
The suffix **WR** denotes **With Rationale supervision**. This model is fine-tuned using both **sentiment labels and human-annotated rationales**, enabling improved alignment between model predictions and human-identified evidence.
---
## Pretraining Details
- **Pretraining corpus:** Indian language text from web sources, religious texts, and news data
- **Training objectives:**
- Masked Language Modeling (MLM)
- Translation Language Modeling (TLM)
- **Language coverage:** 17+ Indian languages, including Telugu and English
---
## Training Data
- **Fine-tuning dataset:** Telugu-Dataset
- **Task:** Sentiment classification
- **Supervision type:** Label + rationale supervision
- **Rationales:** Token-level human-annotated evidence spans
---
## Rationale Supervision
During fine-tuning, **human-provided rationales** guide model learning. Alongside the standard classification loss, an **auxiliary rationale loss** encourages the model’s attention or explanation scores to align with annotated rationale tokens.
This approach improves:
- Interpretability of sentiment predictions
- Alignment between model explanations and human judgment
- Plausibility of generated explanations
---
## Intended Use
This model is intended for:
- Explainable Telugu sentiment classification
- Rationale-supervised learning experiments
- Indian-language explainability research
- Comparative evaluation against label-only (WOR) baselines
MuRIL_WR is particularly effective for **informal, conversational, and social media Telugu text**, where rationale supervision further enhances explanation quality.
---
## Performance Characteristics
Compared to label-only training, rationale supervision typically improves **explanation plausibility** while maintaining competitive sentiment classification performance.
### Strengths
- Strong Telugu-specific linguistic modeling
- Human-aligned explanations via rationale supervision
- Suitable for explainable AI benchmarking in Indian languages
### Limitations
- Requires human-annotated rationales, increasing annotation effort
- Pretraining data bias toward informal text may affect formal Telugu tasks
- Classification gains over WOR may be modest
---
## Use in Explainability Evaluation
**MuRIL_WR** is well-suited for evaluation with explanation frameworks such as FERRET, enabling:
- **Faithfulness evaluation:** How well explanations support model predictions
- **Plausibility evaluation:** How closely explanations align with human rationales
---
## References
- Khanuja et al., 2021
- Joshi, 2022
- Das et al., 2022
- Rajalakshmi et al., 2023