Raj411 commited on
Commit
172c671
·
verified ·
1 Parent(s): 89b29e4

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +111 -0
README.md ADDED
@@ -0,0 +1,111 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-4.0
3
+ datasets:
4
+ - DSL-13-SRMAP/Telugu-Dataset
5
+ language:
6
+ - te
7
+ tags:
8
+ - sentiment-analysis
9
+ - text-classification
10
+ - telugu
11
+ - indic-languages
12
+ - muril
13
+ - rationale-supervision
14
+ - explainable-ai
15
+ base_model: google/muril-base-cased
16
+ pipeline_tag: text-classification
17
+ metrics:
18
+ - accuracy
19
+ - f1
20
+ - auroc
21
+ ---
22
+
23
+ # MuRIL_WR
24
+
25
+ ## Model Description
26
+
27
+ **MuRIL_WR** is a Telugu sentiment classification model built on **MuRIL (Multilingual Representations for Indian Languages)**, a Transformer-based BERT model specifically designed for **Indian languages**, including Telugu and English.
28
+
29
+ MuRIL is pretrained on a **large and diverse corpus of Indian language text**, including web data, religious scriptures, and news content. In contrast to general multilingual models such as mBERT and XLM-R, MuRIL is better suited to capture **Telugu morphology, syntax, and linguistic structure**.
30
+
31
+ The suffix **WR** denotes **With Rationale supervision**. This model is fine-tuned using both **sentiment labels and human-annotated rationales**, enabling improved alignment between model predictions and human-identified evidence.
32
+
33
+ ---
34
+
35
+ ## Pretraining Details
36
+
37
+ - **Pretraining corpus:** Indian language text from web sources, religious texts, and news data
38
+ - **Training objectives:**
39
+ - Masked Language Modeling (MLM)
40
+ - Translation Language Modeling (TLM)
41
+ - **Language coverage:** 17+ Indian languages, including Telugu and English
42
+
43
+ ---
44
+
45
+ ## Training Data
46
+
47
+ - **Fine-tuning dataset:** Telugu-Dataset
48
+ - **Task:** Sentiment classification
49
+ - **Supervision type:** Label + rationale supervision
50
+ - **Rationales:** Token-level human-annotated evidence spans
51
+
52
+ ---
53
+
54
+ ## Rationale Supervision
55
+
56
+ During fine-tuning, **human-provided rationales** guide model learning. Alongside the standard classification loss, an **auxiliary rationale loss** encourages the model’s attention or explanation scores to align with annotated rationale tokens.
57
+
58
+ This approach improves:
59
+
60
+ - Interpretability of sentiment predictions
61
+ - Alignment between model explanations and human judgment
62
+ - Plausibility of generated explanations
63
+
64
+ ---
65
+
66
+ ## Intended Use
67
+
68
+ This model is intended for:
69
+
70
+ - Explainable Telugu sentiment classification
71
+ - Rationale-supervised learning experiments
72
+ - Indian-language explainability research
73
+ - Comparative evaluation against label-only (WOR) baselines
74
+
75
+ MuRIL_WR is particularly effective for **informal, conversational, and social media Telugu text**, where rationale supervision further enhances explanation quality.
76
+
77
+ ---
78
+
79
+ ## Performance Characteristics
80
+
81
+ Compared to label-only training, rationale supervision typically improves **explanation plausibility** while maintaining competitive sentiment classification performance.
82
+
83
+ ### Strengths
84
+
85
+ - Strong Telugu-specific linguistic modeling
86
+ - Human-aligned explanations via rationale supervision
87
+ - Suitable for explainable AI benchmarking in Indian languages
88
+
89
+ ### Limitations
90
+
91
+ - Requires human-annotated rationales, increasing annotation effort
92
+ - Pretraining data bias toward informal text may affect formal Telugu tasks
93
+ - Classification gains over WOR may be modest
94
+
95
+ ---
96
+
97
+ ## Use in Explainability Evaluation
98
+
99
+ **MuRIL_WR** is well-suited for evaluation with explanation frameworks such as FERRET, enabling:
100
+
101
+ - **Faithfulness evaluation:** How well explanations support model predictions
102
+ - **Plausibility evaluation:** How closely explanations align with human rationales
103
+
104
+ ---
105
+
106
+ ## References
107
+
108
+ - Khanuja et al., 2021
109
+ - Joshi, 2022
110
+ - Das et al., 2022
111
+ - Rajalakshmi et al., 2023