Update README.md
Browse files
README.md
CHANGED
|
@@ -70,7 +70,7 @@ from datasets import load_dataset
|
|
| 70 |
|
| 71 |
dataset = load_dataset("your-hf-username/combined-dataset")
|
| 72 |
print(dataset['train'][0])
|
| 73 |
-
|
| 74 |
|
| 75 |
# [Your Model/Dataset Name]
|
| 76 |
|
|
@@ -85,33 +85,13 @@ This resource accompanies our paper accepted in the **Late Breaking Work** track
|
|
| 85 |
|
| 86 |
## ✨ Description
|
| 87 |
|
| 88 |
-
As AI systems become more integrated into daily
|
| 89 |
-
|
| 90 |
-
|
| 91 |
-
|
| 92 |
-
and
|
| 93 |
-
|
| 94 |
-
|
| 95 |
-
advancements, LLMs remain prone to errors, particularly in
|
| 96 |
-
areas requiring nuanced moral reasoning. They struggle with
|
| 97 |
-
detecting implicit hate, offensive language, and gender biases
|
| 98 |
-
due to the subjective and context-dependent nature of these
|
| 99 |
-
issues. Moreover, their reliance on training data can inadvertently
|
| 100 |
-
reinforce societal biases, leading to inconsistencies and ethical
|
| 101 |
-
concerns in their outputs. To explore the limitations of LLMs
|
| 102 |
-
in this role, we developed an experimental framework based
|
| 103 |
-
on state-of-the-art (SOTA) models to assess human emotions
|
| 104 |
-
and offensive behaviors. The framework introduces a unified
|
| 105 |
-
benchmark dataset encompassing 49 distinct categories spanning
|
| 106 |
-
the wide spectrum of human emotions, offensive and hateful
|
| 107 |
-
text, and gender and racial biases. Furthermore, we introduced
|
| 108 |
-
SafePhi, a QLoRA fine-tuned version of Phi-4, adapting diverse
|
| 109 |
-
ethical contexts and outperforming benchmark moderators by
|
| 110 |
-
achieving a Marco F1 score of 0.89, where OpenAI Moderator
|
| 111 |
-
and Llama Guard score 0.77 and 0.74, respectively. This research
|
| 112 |
-
also highlights the critical domains where LLM moderators
|
| 113 |
-
consistently underperformed, pressing the need to incorporate
|
| 114 |
-
more heterogeneous and representative data with human-in-theloop, for better model robustness and explainability.
|
| 115 |
|
| 116 |
## 🚀 Usage
|
| 117 |
|
|
@@ -128,4 +108,3 @@ more heterogeneous and representative data with human-in-theloop, for better mod
|
|
| 128 |
note = {Accepted. Session [XX], Paper ID [XYZ]},
|
| 129 |
url = {https://2025.hci.international/proceedings.html}
|
| 130 |
}
|
| 131 |
-
|
|
|
|
| 70 |
|
| 71 |
dataset = load_dataset("your-hf-username/combined-dataset")
|
| 72 |
print(dataset['train'][0])
|
| 73 |
+
```
|
| 74 |
|
| 75 |
# [Your Model/Dataset Name]
|
| 76 |
|
|
|
|
| 85 |
|
| 86 |
## ✨ Description
|
| 87 |
|
| 88 |
+
As AI systems become more integrated into daily life, the need for safer and more reliable moderation has never been greater. Large Language Models (LLMs) have demonstrated
|
| 89 |
+
remarkable capabilities, surpassing earlier models in complexity and performance. Their evaluation across diverse tasks has consistently showcased their potential, enabling the development of adaptive and personalized agents. However, despite these
|
| 90 |
+
advancements, LLMs remain prone to errors, particularly in areas requiring nuanced moral reasoning. They struggle with detecting implicit hate, offensive language, and gender biases due to the subjective and context-dependent nature of these
|
| 91 |
+
issues. Moreover, their reliance on training data can inadvertently reinforce societal biases, leading to inconsistencies and ethical concerns in their outputs. To explore the limitations of LLMs in this role, we developed an experimental framework based
|
| 92 |
+
on state-of-the-art (SOTA) models to assess human emotions and offensive behaviors. The framework introduces a unified benchmark dataset encompassing 49 distinct categories spanning the wide spectrum of human emotions, offensive and hateful
|
| 93 |
+
text, and gender and racial biases. Furthermore, we introduced SafePhi, a QLoRA fine-tuned version of Phi-4, adapting diverse ethical contexts and outperforming benchmark moderators by achieving a Marco F1 score of 0.89, where OpenAI Moderator
|
| 94 |
+
and Llama Guard score 0.77 and 0.74, respectively. This research also highlights the critical domains where LLM moderators consistently underperformed, pressing the need to incorporate more heterogeneous and representative data with human-in-theloop, for better model robustness and explainability.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 95 |
|
| 96 |
## 🚀 Usage
|
| 97 |
|
|
|
|
| 108 |
note = {Accepted. Session [XX], Paper ID [XYZ]},
|
| 109 |
url = {https://2025.hci.international/proceedings.html}
|
| 110 |
}
|
|
|