Update README.md
Browse files
README.md
CHANGED
|
@@ -10,7 +10,7 @@ tags:
|
|
| 10 |
pipeline_tag: text-classification
|
| 11 |
---
|
| 12 |
|
| 13 |
-
# helizac/
|
| 14 |
|
| 15 |
This model is a fine-tuned version of `dbmdz/bert-base-turkish-cased` for classifying the acceptability of a Turkish text output given a Turkish text input.
|
| 16 |
It was developed as part of the "Evaluation of the Acceptability of Model Outputs" (May 2025).
|
|
@@ -40,7 +40,7 @@ import torch
|
|
| 40 |
from transformers import AutoTokenizer, AutoModelForSequenceClassification
|
| 41 |
|
| 42 |
TOKEN_KEY = "YOUR_HF_TOKEN_HERE" # Replace with your Hugging Face token or set to None
|
| 43 |
-
MODEL_NAME = "helizac/
|
| 44 |
|
| 45 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 46 |
MAX_LENGTH = 64
|
|
@@ -113,7 +113,7 @@ print(f"Input: {input_text_2}\nOutput: {output_text_2}\nPrediction: {prediction_
|
|
| 113 |
|
| 114 |
# Example 3: Unacceptable (grammatically poor)
|
| 115 |
input_text_3 = "Hayalindeki meslek ne büyük."
|
| 116 |
-
output_text_3 = "Olmak ben istemek büyük.
|
| 117 |
prediction_3, confidence_3 = predict_pair_acceptability(input_text_3, output_text_3, model, tokenizer, device, MAX_LENGTH)
|
| 118 |
print(f"Input: {input_text_3}\nOutput: {output_text_3}\nPrediction: {prediction_3} (Confidence: {confidence_3:.4f})\n")
|
| 119 |
```
|
|
@@ -125,7 +125,7 @@ The model was fine-tuned on a dataset of approximately 460,000 Turkish input-out
|
|
| 125 |
All pairs were truncated/padded to a maximum sequence length of 64 tokens for the combined input and output.
|
| 126 |
|
| 127 |
## Evaluation Results
|
| 128 |
-
On a manually curated, independent Turkish test set (89 pairs evaluated due to token limits), this model (helizac/
|
| 129 |
|
| 130 |
The stress test for this model showed:
|
| 131 |
|
|
|
|
| 10 |
pipeline_tag: text-classification
|
| 11 |
---
|
| 12 |
|
| 13 |
+
# helizac/pair-acceptability-turkish
|
| 14 |
|
| 15 |
This model is a fine-tuned version of `dbmdz/bert-base-turkish-cased` for classifying the acceptability of a Turkish text output given a Turkish text input.
|
| 16 |
It was developed as part of the "Evaluation of the Acceptability of Model Outputs" (May 2025).
|
|
|
|
| 40 |
from transformers import AutoTokenizer, AutoModelForSequenceClassification
|
| 41 |
|
| 42 |
TOKEN_KEY = "YOUR_HF_TOKEN_HERE" # Replace with your Hugging Face token or set to None
|
| 43 |
+
MODEL_NAME = "helizac/pair-acceptability-turkish"
|
| 44 |
|
| 45 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 46 |
MAX_LENGTH = 64
|
|
|
|
| 113 |
|
| 114 |
# Example 3: Unacceptable (grammatically poor)
|
| 115 |
input_text_3 = "Hayalindeki meslek ne büyük."
|
| 116 |
+
output_text_3 = "Olmak ben istemek büyük."
|
| 117 |
prediction_3, confidence_3 = predict_pair_acceptability(input_text_3, output_text_3, model, tokenizer, device, MAX_LENGTH)
|
| 118 |
print(f"Input: {input_text_3}\nOutput: {output_text_3}\nPrediction: {prediction_3} (Confidence: {confidence_3:.4f})\n")
|
| 119 |
```
|
|
|
|
| 125 |
All pairs were truncated/padded to a maximum sequence length of 64 tokens for the combined input and output.
|
| 126 |
|
| 127 |
## Evaluation Results
|
| 128 |
+
On a manually curated, independent Turkish test set (89 pairs evaluated due to token limits), this model (helizac/pair-acceptability-turkish) achieved an accuracy of 87%.
|
| 129 |
|
| 130 |
The stress test for this model showed:
|
| 131 |
|