| # DeBERTa-v3-base Fine-Tuned for Hallucination Detection | |
| Model Details | |
| Model Name: DeBERTa-v3-base | |
| Architecture: DeBERTa (Decoding-enhanced BERT with disentangled attention) | |
| Base Model: DeBERTa-v3-base | |
| Fine-tuned Dataset: PAWS (Paraphrase Adversaries from Word Scrambling) | |
| Task: Sentence Pair Classification (Hallucination Detection) | |
| Model Description | |
| This model is a fine-tuned version of the DeBERTa-v3-base model specifically for the task of detecting hallucinations between pairs of sentences. Hallucinations in this context refer to statements or information present in one sentence but not supported or contradicted by the other. | |
| Fine-Tuning Dataset | |
| Dataset Name: PAWS (Paraphrase Adversaries from Word Scrambling) | |
| Dataset Description: The PAWS dataset contains pairs of sentences with high lexical overlap but different meanings, designed to challenge models' understanding of semantic content. | |
| Dataset: https://huggingface.co/datasets/paws | |
| Training Procedure | |
| Number of Epochs: 10 | |
| Hardware Used: NVIDIA -A 100 | |
| Performance: | |
| Accuracy: 94.88% | |
| F1 Score: 92.3% | |
| Precision: 92.82% | |
| Recall: 95.81% | |
| from transformers import AutoTokenizer, AutoModelForSequenceClassification | |
| import torch | |
| # Load model and tokenizer | |
| tokenizer = AutoTokenizer.from_pretrained("Varun-Chowdary/hallucination_detect") | |
| model = AutoModelForSequenceClassification.from_pretrained("Varun-Chowdary/hallucination_detect") | |
| # Define the sentences | |
| sentence1 = "Maradona was born in Argentina, South America." | |
| sentence2 = "Maradona was born in Brazil, South America. " | |
| # Tokenize and prepare input | |
| inputs = tokenizer(sentence1, sentence2, return_tensors='pt', truncation=True, padding=True) | |
| # Perform inference | |
| with torch.no_grad(): | |
| outputs = model(**inputs) | |
| logits = outputs.logits | |
| probabilities = torch.softmax(logits, dim=1) | |
| # Get the predicted label | |
| predicted_label = torch.argmax(probabilities, dim=1).item() | |
| labels = ["No Hallucination", "Hallucination"] | |
| print(f"Predicted label: {labels[predicted_label]}") |