Update README.md
Browse files
README.md
CHANGED
|
@@ -58,7 +58,7 @@ model_name = "b4c0n/KAi-toxicity-filter"
|
|
| 58 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 59 |
model = AutoModelForSequenceClassification.from_pretrained(model_name)
|
| 60 |
|
| 61 |
-
text = "
|
| 62 |
inputs = tokenizer(text, return_tensors="pt")
|
| 63 |
outputs = model(**inputs)
|
| 64 |
|
|
@@ -140,7 +140,7 @@ model_name = "your-username/KAi-toxicity-filter"
|
|
| 140 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 141 |
model = AutoModelForSequenceClassification.from_pretrained(model_name)
|
| 142 |
|
| 143 |
-
text = "
|
| 144 |
inputs = tokenizer(text, return_tensors="pt")
|
| 145 |
outputs = model(**inputs)
|
| 146 |
|
|
|
|
| 58 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 59 |
model = AutoModelForSequenceClassification.from_pretrained(model_name)
|
| 60 |
|
| 61 |
+
text = "終わってる暴言"
|
| 62 |
inputs = tokenizer(text, return_tensors="pt")
|
| 63 |
outputs = model(**inputs)
|
| 64 |
|
|
|
|
| 140 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 141 |
model = AutoModelForSequenceClassification.from_pretrained(model_name)
|
| 142 |
|
| 143 |
+
text = "toxic expression"
|
| 144 |
inputs = tokenizer(text, return_tensors="pt")
|
| 145 |
outputs = model(**inputs)
|
| 146 |
|