SQuAD_BERT / README.md
tmt3103's picture
Update README.md
e295d47 verified
---
license: apache-2.0
tags:
- question-answering
- squad
- transformers
- pytorch
- evaluation
- hf-course
- fine-tuned
datasets:
- squad
metrics:
- exact_match
- f1
model-index:
- name: QA-SQuAD-BERT
results:
- task:
type: question-answering
name: Question Answering
dataset:
name: SQuAD v1.1
type: squad
metrics:
- name: Exact Match
type: exact_match
value: 82.7
- name: F1
type: f1
value: 87.0039
---
# QA-SQuAD-BERT
A BERT-based model fine-tuned on SQuAD v1.1 for extractive QA
## Model Description
This model is based on bert-base-uncased and was fine-tuned on the **SQuAD v1.1** dataset for extractive question answering. It takes a question and a context passage as input and predicts the span of text in the passage that most likely answers the question.
The model was trained using the Hugging Face 馃 Transformers library.
## Intended Uses & Limitations
### Intended Uses
- Extractive question answering on Wikipedia-style passages.
- As a downstream component in information retrieval pipelines.
- Educational purposes or experimentation with fine-tuning on QA tasks.
### Limitations
- The model may not generalize well to out-of-domain datasets.
- It does not handle unanswerable questions (not trained on SQuAD v2.0).
- It may produce incorrect or misleading answers if context is ambiguous.
## Training Details
- **Base model**: bert-base-uncased
- **Dataset**: [SQuAD v1.1](https://huggingface.co/datasets/squad)
- **Epochs**: 3
- **Batch size**: 8
- **Learning rate**: 2e-5
- **Optimizer**: AdamW
- **Max length**: 384
- **Hardware used**: Colab/GPU T4
## Evaluation Results
The model was evaluated on the SQuAD v1.1 development set using the standard metrics: Exact Match (EM) and F1.
| Metric | Score |
|--------------|-------|
| Exact Match | 82.7 |
| F1 | 87.0039 |
## How to Use
You can load this model using the `pipeline` API:
```python
from transformers import pipeline
qa_pipeline = pipeline("question-answering", model="tmt3103/SQuAD_BERT")
result = qa_pipeline({
"context": "Hugging Face is creating a tool that democratizes AI.",
"question": "What is Hugging Face creating?"
})
print(result)