Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,12 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
---
|
| 4 |
+
# Better Uncensored
|
| 5 |
+
|
| 6 |
+
"Uncensored" datasets and models based on them (like *-dolphin) have been haphazardly (or maliciously) curated to remove examples of model refusals, and what the authors call "AI moralizing", but above all, to remove any mention of terms they disliked, hated or feared like feminism, lgbt, racism, and a long etc.
|
| 7 |
+
|
| 8 |
+
At first I considered this to be plain laziness but I've come to learn that is a concerted effort of them to remove their percived liberal bias and make the models not only more compliant, but more conservative, like themselves.
|
| 9 |
+
|
| 10 |
+
This project provides a pipeline and datasets that better remove refusals and unsolisited moralizing comments, without censoring anyparticular content, and attempting to recover messages that would otherwise be discarded. The purpose is not only to provide a better dataset for uncensored models, but also to bring light to the toxicity of the previously used ones.
|
| 11 |
+
|
| 12 |
+
See [Better Uncensored github](https://github.com/sudoaza/better-uncensored) for code, for the moment here are only text classifier models for moralizing and refusal detection. Trained on 300 char length strings, probably can take double or triple that and still work decently.
|