prabhaskenche's picture
Update app.py
1e225c4 verified
raw
history blame contribute delete
983 Bytes
import gradio as gr
from transformers import pipeline, AutoTokenizer
# Load the tokenizer and model
tokenizer = AutoTokenizer.from_pretrained('prabhaskenche/toxic-comment-classification-using-RoBERTa')
classifier = pipeline(
'text-classification',
model='prabhaskenche/toxic-comment-classification-using-RoBERTa',
tokenizer=tokenizer,
top_k=None # Use top_k=None to get all scores
)
def classify(text):
results = classifier(text)
# Assuming LABEL_0 is non-toxic and LABEL_1 is toxic
non_toxic_score = next((item['score'] for item in results[0] if item['label'] == 'LABEL_0'), 0)
toxic_score = next((item['score'] for item in results[0] if item['label'] == 'LABEL_1'), 0)
return f"{non_toxic_score:.3f} non-toxic, {toxic_score:.3f} toxic"
# Create the Gradio interface
interface = gr.Interface(
fn=classify,
inputs=gr.Textbox(lines=2, placeholder="Enter text here..."),
outputs="text"
)
# Launch the interface
interface.launch()