Bubbli's picture
Create README.md
25c1f63 verified
---
language:
- en
- tl
tags:
- sarcasm-detection
- mock-politeness
- multi-task-learning
- code-mixed
license: mit
base_model:
- FacebookAI/xlm-roberta-base
---
# XLM-RoBERTa with Multi-Task Learning for Sarcasm and Mock Politeness Detection
## Model Description
This project fine-tunes **XLM-RoBERTa** for detecting **sarcasm** and **mock politeness** in **Filipino (English, Tagalog, or code-mixed (Taglish))** faculty evaluation texts.
Two models are included:
- **MTL model** → sarcasm detection (main task) + mock politeness detection (auxiliary task)
- **STL model** → sarcasm detection only
The models are packaged into a **desktop app (Tkinter + Python)** for easy testing.
---
## Intended Uses & Limitations
### Intended Use
- Demonstrating multi-task learning in NLP
- Exploring sarcasm and politeness detection in Taglish text
- Academic/research purposes only
### Limitations
- Trained on a domain-specific dataset (faculty evaluations)
- May not generalize well outside Taglish or academic settings
- Predictions are not guaranteed to be accurate for all contexts
---
## How to Use
1. Download the **XLM-R folder** from this repository.
2. Inside the folder, locate and open: XLM-R/XLM-R.exe
3. Use the GUI to input text or upload a `.csv` file (see included `INPUT_SAMPLE.csv`).
4. The app will output predictions for sarcasm (and mock politeness if using MTL).
*(No coding required — the `.exe` is standalone on Windows.)*
---
## Training Data
- Collected faculty evaluation texts written in **Filipino** (English, Tagalog, or code-mixed (Taglish))
- Annotated for sarcasm and mock politeness
---
## Evaluation
- Compared **Single-Task (STL)** vs **Multi-Task (MTL)**
- Metrics: accuracy, precision, recall, F1