Update README.md
Browse files
README.md
CHANGED
|
@@ -12,8 +12,7 @@ The model assigns one of five possible labels:
|
|
| 12 |
3 (**Refusal Capability**): The model refuses to answer due to its own limitations, lack of information, or lack of ability to provide an adequate response. <br />
|
| 13 |
4 (**Disclaimer Capability**): The model signals its limitations but attempts to provide an answer within its capacity <br />
|
| 14 |
|
| 15 |
-
Please cite: <br />
|
| 16 |
-
Pasch, S. (2025). LLM Content Moderation and User Satisfaction: Evidence from Response Refusals in Chatbot Arena. arXiv preprint arXiv:2501.03266.
|
| 17 |
|
| 18 |
References <br />
|
| 19 |
[1] Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; ... & Stoyanov, V. (2019). Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692. <br />
|
|
|
|
| 12 |
3 (**Refusal Capability**): The model refuses to answer due to its own limitations, lack of information, or lack of ability to provide an adequate response. <br />
|
| 13 |
4 (**Disclaimer Capability**): The model signals its limitations but attempts to provide an answer within its capacity <br />
|
| 14 |
|
| 15 |
+
Please cite: <br /> LLM Content Moderation and User Satisfaction: Evidence from Response Refusals in Chatbot Arena (under review).
|
|
|
|
| 16 |
|
| 17 |
References <br />
|
| 18 |
[1] Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; ... & Stoyanov, V. (2019). Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692. <br />
|