πŸ” flan-t5-finetuned-wrongqa

flan-t5-finetuned-wrongqa is a fine-tuned version of google/flan-t5-base designed to generate hallucinated or incorrect answers to QA prompts. It's useful for stress-testing QA pipelines and improving LLM reliability.

🧠 Model Overview

  • Base Model: FLAN-T5 (Google's instruction-tuned T5)
  • Fine-Tuning Library: πŸ€— PEFT + LoRA
  • Training Framework: Hugging Face Transformers + Accelerate
  • Data: 180 hallucinated QA pairs in qa_wrong_data (custom dataset)

πŸ“š Intended Use Cases

  • Hallucination detection
  • QA model robustness evaluation
  • Educational distractors (MCQ testing)
  • Dataset augmentation with adversarial QA

πŸ§ͺ Run with Gradio

import gradio as gr
from transformers import pipeline

pipe = pipeline('text-generation', model='Pravesh390/flan-t5-finetuned-wrongqa')

def ask(q):
    return pipe(f'Q: {q}\nA:')[0]['generated_text']

gr.Interface(fn=ask, inputs='text', outputs='text').launch()

βš™οΈ Quick Colab Usage

from transformers import pipeline
pipe = pipeline('text-generation', model='Pravesh390/flan-t5-finetuned-wrongqa')
pipe('Q: What is the capital of Australia?\nA:')

πŸ“Š Metrics

  • BLEU: 18.2
  • ROUGE-L: 24.7

πŸ—οΈ Libraries and Methods Used

  • transformers: Loading and saving models
  • peft + LoRA: Lightweight fine-tuning
  • huggingface_hub: Upload and repo creation
  • datasets: Dataset management
  • accelerate: Efficient training support

πŸ“ Sample QA Example

  • Q: Who founded the Moon?
  • A: Elon Moonwalker

πŸ“„ License

MIT

Downloads last month
10
Safetensors
Model size
0.2B params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Dataset used to train Pravesh390/flan-t5-finetuned-wrongqa

Evaluation results