Upload README.md
Browse files
README.md
CHANGED
|
@@ -7,7 +7,6 @@ tags:
|
|
| 7 |
- manipulation-detection
|
| 8 |
- pytorch
|
| 9 |
- transformers
|
| 10 |
-
- interpersonal-relationships
|
| 11 |
library_name: transformers
|
| 12 |
pipeline_tag: text-classification
|
| 13 |
metrics:
|
|
@@ -48,9 +47,9 @@ These two classifier models are fine-tuned to flag possible manipulation in mess
|
|
| 48 |
|
| 49 |
The smaller model is based on microsoft/xtremedistil-l6-h256-uncased and has 12.75M total parameters. The larger uses microsoft/deberta-v3-xsmall and is at 70.83M total parameters. Both models achieve +99% F1 score on the held out test split.
|
| 50 |
|
| 51 |
-
The
|
| 52 |
|
| 53 |
-
The training data was augmented to make the models robust to typos and adversarial attacks
|
| 54 |
|
| 55 |
|
| 56 |
Both models are released under the MIT license.
|
|
|
|
| 7 |
- manipulation-detection
|
| 8 |
- pytorch
|
| 9 |
- transformers
|
|
|
|
| 10 |
library_name: transformers
|
| 11 |
pipeline_tag: text-classification
|
| 12 |
metrics:
|
|
|
|
| 47 |
|
| 48 |
The smaller model is based on microsoft/xtremedistil-l6-h256-uncased and has 12.75M total parameters. The larger uses microsoft/deberta-v3-xsmall and is at 70.83M total parameters. Both models achieve +99% F1 score on the held out test split.
|
| 49 |
|
| 50 |
+
The confidence score of the predictions are scaled to reflect the probability of the prediction being true, however there are instances when the models predict a blatantly wrong answer with full confidence. Furthermore, if the message requires additional context to be manipulative, then it is considered bening.
|
| 51 |
|
| 52 |
+
The training data was augmented to make the models robust to typos and adversarial attacks, but highest accuracy is achieved on clean text.
|
| 53 |
|
| 54 |
|
| 55 |
Both models are released under the MIT license.
|