Machlovi commited on
Commit
8798353
·
verified ·
1 Parent(s): d5534c5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -29
README.md CHANGED
@@ -70,7 +70,7 @@ from datasets import load_dataset
70
 
71
  dataset = load_dataset("your-hf-username/combined-dataset")
72
  print(dataset['train'][0])
73
-
74
 
75
  # [Your Model/Dataset Name]
76
 
@@ -85,33 +85,13 @@ This resource accompanies our paper accepted in the **Late Breaking Work** track
85
 
86
  ## ✨ Description
87
 
88
- As AI systems become more integrated into daily
89
- life, the need for safer and more reliable moderation has never
90
- been greater. Large Language Models (LLMs) have demonstrated
91
- remarkable capabilities, surpassing earlier models in complexity
92
- and performance. Their evaluation across diverse tasks has
93
- consistently showcased their potential, enabling the development
94
- of adaptive and personalized agents. However, despite these
95
- advancements, LLMs remain prone to errors, particularly in
96
- areas requiring nuanced moral reasoning. They struggle with
97
- detecting implicit hate, offensive language, and gender biases
98
- due to the subjective and context-dependent nature of these
99
- issues. Moreover, their reliance on training data can inadvertently
100
- reinforce societal biases, leading to inconsistencies and ethical
101
- concerns in their outputs. To explore the limitations of LLMs
102
- in this role, we developed an experimental framework based
103
- on state-of-the-art (SOTA) models to assess human emotions
104
- and offensive behaviors. The framework introduces a unified
105
- benchmark dataset encompassing 49 distinct categories spanning
106
- the wide spectrum of human emotions, offensive and hateful
107
- text, and gender and racial biases. Furthermore, we introduced
108
- SafePhi, a QLoRA fine-tuned version of Phi-4, adapting diverse
109
- ethical contexts and outperforming benchmark moderators by
110
- achieving a Marco F1 score of 0.89, where OpenAI Moderator
111
- and Llama Guard score 0.77 and 0.74, respectively. This research
112
- also highlights the critical domains where LLM moderators
113
- consistently underperformed, pressing the need to incorporate
114
- more heterogeneous and representative data with human-in-theloop, for better model robustness and explainability.
115
 
116
  ## 🚀 Usage
117
 
@@ -128,4 +108,3 @@ more heterogeneous and representative data with human-in-theloop, for better mod
128
  note = {Accepted. Session [XX], Paper ID [XYZ]},
129
  url = {https://2025.hci.international/proceedings.html}
130
  }
131
-
 
70
 
71
  dataset = load_dataset("your-hf-username/combined-dataset")
72
  print(dataset['train'][0])
73
+ ```
74
 
75
  # [Your Model/Dataset Name]
76
 
 
85
 
86
  ## ✨ Description
87
 
88
+ As AI systems become more integrated into daily life, the need for safer and more reliable moderation has never been greater. Large Language Models (LLMs) have demonstrated
89
+ remarkable capabilities, surpassing earlier models in complexity and performance. Their evaluation across diverse tasks has consistently showcased their potential, enabling the development of adaptive and personalized agents. However, despite these
90
+ advancements, LLMs remain prone to errors, particularly in areas requiring nuanced moral reasoning. They struggle with detecting implicit hate, offensive language, and gender biases due to the subjective and context-dependent nature of these
91
+ issues. Moreover, their reliance on training data can inadvertently reinforce societal biases, leading to inconsistencies and ethical concerns in their outputs. To explore the limitations of LLMs in this role, we developed an experimental framework based
92
+ on state-of-the-art (SOTA) models to assess human emotions and offensive behaviors. The framework introduces a unified benchmark dataset encompassing 49 distinct categories spanning the wide spectrum of human emotions, offensive and hateful
93
+ text, and gender and racial biases. Furthermore, we introduced SafePhi, a QLoRA fine-tuned version of Phi-4, adapting diverse ethical contexts and outperforming benchmark moderators by achieving a Marco F1 score of 0.89, where OpenAI Moderator
94
+ and Llama Guard score 0.77 and 0.74, respectively. This research also highlights the critical domains where LLM moderators consistently underperformed, pressing the need to incorporate more heterogeneous and representative data with human-in-theloop, for better model robustness and explainability.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
95
 
96
  ## 🚀 Usage
97
 
 
108
  note = {Accepted. Session [XX], Paper ID [XYZ]},
109
  url = {https://2025.hci.international/proceedings.html}
110
  }