Josephaimodels commited on
Commit
31798eb
·
verified ·
1 Parent(s): ca80854

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -7,7 +7,6 @@ tags:
7
  - manipulation-detection
8
  - pytorch
9
  - transformers
10
- - interpersonal-relationships
11
  library_name: transformers
12
  pipeline_tag: text-classification
13
  metrics:
@@ -48,9 +47,9 @@ These two classifier models are fine-tuned to flag possible manipulation in mess
48
 
49
  The smaller model is based on microsoft/xtremedistil-l6-h256-uncased and has 12.75M total parameters. The larger uses microsoft/deberta-v3-xsmall and is at 70.83M total parameters. Both models achieve +99% F1 score on the held out test split.
50
 
51
- The models are scaled to for the confidence to reflect the probability of the prediction being true, however there are instances when the models predict a blatantly wrong answer with full confidence. Furthermore, if the message requires additional context to be manipulative, then it is not flagged.
52
 
53
- The training data was augmented to make the models robust to typos and adversarial attacks. However, highest accuracy is achieved on clean text.
54
 
55
 
56
  Both models are released under the MIT license.
 
7
  - manipulation-detection
8
  - pytorch
9
  - transformers
 
10
  library_name: transformers
11
  pipeline_tag: text-classification
12
  metrics:
 
47
 
48
  The smaller model is based on microsoft/xtremedistil-l6-h256-uncased and has 12.75M total parameters. The larger uses microsoft/deberta-v3-xsmall and is at 70.83M total parameters. Both models achieve +99% F1 score on the held out test split.
49
 
50
+ The confidence score of the predictions are scaled to reflect the probability of the prediction being true, however there are instances when the models predict a blatantly wrong answer with full confidence. Furthermore, if the message requires additional context to be manipulative, then it is considered bening.
51
 
52
+ The training data was augmented to make the models robust to typos and adversarial attacks, but highest accuracy is achieved on clean text.
53
 
54
 
55
  Both models are released under the MIT license.