helizac commited on
Commit
602615f
·
verified ·
1 Parent(s): 9fc7436

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -10,7 +10,7 @@ tags:
10
  pipeline_tag: text-classification
11
  ---
12
 
13
- # helizac/berturk-pair-acceptability
14
 
15
  This model is a fine-tuned version of `dbmdz/bert-base-turkish-cased` for classifying the acceptability of a Turkish text output given a Turkish text input.
16
  It was developed as part of the "Evaluation of the Acceptability of Model Outputs" (May 2025).
@@ -40,7 +40,7 @@ import torch
40
  from transformers import AutoTokenizer, AutoModelForSequenceClassification
41
 
42
  TOKEN_KEY = "YOUR_HF_TOKEN_HERE" # Replace with your Hugging Face token or set to None
43
- MODEL_NAME = "helizac/berturk-pair-acceptability"
44
 
45
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
46
  MAX_LENGTH = 64
@@ -113,7 +113,7 @@ print(f"Input: {input_text_2}\nOutput: {output_text_2}\nPrediction: {prediction_
113
 
114
  # Example 3: Unacceptable (grammatically poor)
115
  input_text_3 = "Hayalindeki meslek ne büyük."
116
- output_text_3 = "Olmak ben istemek büyük.
117
  prediction_3, confidence_3 = predict_pair_acceptability(input_text_3, output_text_3, model, tokenizer, device, MAX_LENGTH)
118
  print(f"Input: {input_text_3}\nOutput: {output_text_3}\nPrediction: {prediction_3} (Confidence: {confidence_3:.4f})\n")
119
  ```
@@ -125,7 +125,7 @@ The model was fine-tuned on a dataset of approximately 460,000 Turkish input-out
125
  All pairs were truncated/padded to a maximum sequence length of 64 tokens for the combined input and output.
126
 
127
  ## Evaluation Results
128
- On a manually curated, independent Turkish test set (89 pairs evaluated due to token limits), this model (helizac/berturk-pair-acceptability) achieved an accuracy of 87%.
129
 
130
  The stress test for this model showed:
131
 
 
10
  pipeline_tag: text-classification
11
  ---
12
 
13
+ # helizac/pair-acceptability-turkish
14
 
15
  This model is a fine-tuned version of `dbmdz/bert-base-turkish-cased` for classifying the acceptability of a Turkish text output given a Turkish text input.
16
  It was developed as part of the "Evaluation of the Acceptability of Model Outputs" (May 2025).
 
40
  from transformers import AutoTokenizer, AutoModelForSequenceClassification
41
 
42
  TOKEN_KEY = "YOUR_HF_TOKEN_HERE" # Replace with your Hugging Face token or set to None
43
+ MODEL_NAME = "helizac/pair-acceptability-turkish"
44
 
45
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
46
  MAX_LENGTH = 64
 
113
 
114
  # Example 3: Unacceptable (grammatically poor)
115
  input_text_3 = "Hayalindeki meslek ne büyük."
116
+ output_text_3 = "Olmak ben istemek büyük."
117
  prediction_3, confidence_3 = predict_pair_acceptability(input_text_3, output_text_3, model, tokenizer, device, MAX_LENGTH)
118
  print(f"Input: {input_text_3}\nOutput: {output_text_3}\nPrediction: {prediction_3} (Confidence: {confidence_3:.4f})\n")
119
  ```
 
125
  All pairs were truncated/padded to a maximum sequence length of 64 tokens for the combined input and output.
126
 
127
  ## Evaluation Results
128
+ On a manually curated, independent Turkish test set (89 pairs evaluated due to token limits), this model (helizac/pair-acceptability-turkish) achieved an accuracy of 87%.
129
 
130
  The stress test for this model showed:
131