|
|
--- |
|
|
license: apache-2.0 |
|
|
tags: |
|
|
- question-answering |
|
|
- squad |
|
|
- transformers |
|
|
- pytorch |
|
|
- evaluation |
|
|
- hf-course |
|
|
- fine-tuned |
|
|
datasets: |
|
|
- squad |
|
|
metrics: |
|
|
- exact_match |
|
|
- f1 |
|
|
model-index: |
|
|
- name: QA-SQuAD-BERT |
|
|
results: |
|
|
- task: |
|
|
type: question-answering |
|
|
name: Question Answering |
|
|
dataset: |
|
|
name: SQuAD v1.1 |
|
|
type: squad |
|
|
metrics: |
|
|
- name: Exact Match |
|
|
type: exact_match |
|
|
value: 82.7 |
|
|
- name: F1 |
|
|
type: f1 |
|
|
value: 87.0039 |
|
|
--- |
|
|
|
|
|
# QA-SQuAD-BERT |
|
|
|
|
|
A BERT-based model fine-tuned on SQuAD v1.1 for extractive QA |
|
|
|
|
|
## Model Description |
|
|
|
|
|
This model is based on bert-base-uncased and was fine-tuned on the **SQuAD v1.1** dataset for extractive question answering. It takes a question and a context passage as input and predicts the span of text in the passage that most likely answers the question. |
|
|
|
|
|
The model was trained using the Hugging Face 馃 Transformers library. |
|
|
|
|
|
## Intended Uses & Limitations |
|
|
|
|
|
### Intended Uses |
|
|
|
|
|
- Extractive question answering on Wikipedia-style passages. |
|
|
- As a downstream component in information retrieval pipelines. |
|
|
- Educational purposes or experimentation with fine-tuning on QA tasks. |
|
|
|
|
|
### Limitations |
|
|
|
|
|
- The model may not generalize well to out-of-domain datasets. |
|
|
- It does not handle unanswerable questions (not trained on SQuAD v2.0). |
|
|
- It may produce incorrect or misleading answers if context is ambiguous. |
|
|
|
|
|
## Training Details |
|
|
|
|
|
- **Base model**: bert-base-uncased |
|
|
- **Dataset**: [SQuAD v1.1](https://huggingface.co/datasets/squad) |
|
|
- **Epochs**: 3 |
|
|
- **Batch size**: 8 |
|
|
- **Learning rate**: 2e-5 |
|
|
- **Optimizer**: AdamW |
|
|
- **Max length**: 384 |
|
|
- **Hardware used**: Colab/GPU T4 |
|
|
|
|
|
## Evaluation Results |
|
|
|
|
|
The model was evaluated on the SQuAD v1.1 development set using the standard metrics: Exact Match (EM) and F1. |
|
|
|
|
|
| Metric | Score | |
|
|
|--------------|-------| |
|
|
| Exact Match | 82.7 | |
|
|
| F1 | 87.0039 | |
|
|
|
|
|
## How to Use |
|
|
|
|
|
You can load this model using the `pipeline` API: |
|
|
|
|
|
```python |
|
|
from transformers import pipeline |
|
|
|
|
|
qa_pipeline = pipeline("question-answering", model="tmt3103/SQuAD_BERT") |
|
|
result = qa_pipeline({ |
|
|
"context": "Hugging Face is creating a tool that democratizes AI.", |
|
|
"question": "What is Hugging Face creating?" |
|
|
}) |
|
|
print(result) |
|
|
|