Bubbli commited on
Commit
25c1f63
·
verified ·
1 Parent(s): 0564338

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +61 -0
README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - tl
5
+ tags:
6
+ - sarcasm-detection
7
+ - mock-politeness
8
+ - multi-task-learning
9
+ - code-mixed
10
+ license: mit
11
+ base_model:
12
+ - FacebookAI/xlm-roberta-base
13
+ ---
14
+
15
+ # XLM-RoBERTa with Multi-Task Learning for Sarcasm and Mock Politeness Detection
16
+
17
+ ## Model Description
18
+ This project fine-tunes **XLM-RoBERTa** for detecting **sarcasm** and **mock politeness** in **Filipino (English, Tagalog, or code-mixed (Taglish))** faculty evaluation texts.
19
+
20
+ Two models are included:
21
+ - **MTL model** → sarcasm detection (main task) + mock politeness detection (auxiliary task)
22
+ - **STL model** → sarcasm detection only
23
+
24
+ The models are packaged into a **desktop app (Tkinter + Python)** for easy testing.
25
+
26
+ ---
27
+
28
+ ## Intended Uses & Limitations
29
+
30
+ ### Intended Use
31
+ - Demonstrating multi-task learning in NLP
32
+ - Exploring sarcasm and politeness detection in Taglish text
33
+ - Academic/research purposes only
34
+
35
+ ### Limitations
36
+ - Trained on a domain-specific dataset (faculty evaluations)
37
+ - May not generalize well outside Taglish or academic settings
38
+ - Predictions are not guaranteed to be accurate for all contexts
39
+
40
+ ---
41
+
42
+ ## How to Use
43
+
44
+ 1. Download the **XLM-R folder** from this repository.
45
+ 2. Inside the folder, locate and open: XLM-R/XLM-R.exe
46
+ 3. Use the GUI to input text or upload a `.csv` file (see included `INPUT_SAMPLE.csv`).
47
+ 4. The app will output predictions for sarcasm (and mock politeness if using MTL).
48
+
49
+ *(No coding required — the `.exe` is standalone on Windows.)*
50
+
51
+ ---
52
+
53
+ ## Training Data
54
+ - Collected faculty evaluation texts written in **Filipino** (English, Tagalog, or code-mixed (Taglish))
55
+ - Annotated for sarcasm and mock politeness
56
+
57
+ ---
58
+
59
+ ## Evaluation
60
+ - Compared **Single-Task (STL)** vs **Multi-Task (MTL)**
61
+ - Metrics: accuracy, precision, recall, F1