| ---
|
| language:
|
| - en
|
| tags:
|
| - question-answering
|
| - transformers
|
| - bert
|
| - squad
|
| license: apache-2.0
|
| datasets:
|
| - squad
|
| model-name: bert-base-uncased-finetuned-squad
|
| library_name: transformers
|
| ---
|
|
|
| # BERT-Base Uncased Fine-Tuned on SQuAD
|
|
|
| ## Overview
|
|
|
| This repository contains a **BERT-Base Uncased** model fine-tuned on the **SQuAD (Stanford Question Answering Dataset)** for **Question Answering (QA) tasks**. The model has been fine-tuned for **2 epochs**, making it suitable for extracting answers from given contexts by predicting start and end token positions.
|
|
|
| ## The Model predicts 2 probabilities among all the tokens in the vocab , One indicating the start token and the other indicating the end token, Then the answer between both these tokens are extracted.
|
|
|
| ## Model Details
|
|
|
| - **Model Type**: BERT-Base Uncased
|
| - **Fine-Tuning Dataset**: SQuAD (Stanford Question Answering Dataset)
|
| - **Number of Epochs**: 2
|
| - **Task**: Question Answering
|
| - **Base Model**: [BERT-Base Uncased](https://huggingface.co/bert-base-uncased)
|
|
|
| ---
|
|
|
| ## Usage
|
|
|
| ### How to Load the Model
|
|
|
| You can load the model using the `transformers` library from Hugging Face:
|
|
|
| ```python
|
| from transformers import BertForQuestionAnswering, BertTokenizer
|
|
|
| # Load the tokenizer and model
|
| tokenizer = BertTokenizer.from_pretrained("Abdo36/Bert-SquAD-QA")
|
| model = BertForQuestionAnswering.from_pretrained("Abdo36/Bert-SquAD-QA")
|
|
|
| context = "BERT is a method of pre-training language representations."
|
| question = "What is BERT?"
|
|
|
| inputs = tokenizer.encode_plus(question, context, return_tensors="pt")
|
|
|
| # Perform inference
|
| outputs = model(**inputs)
|
| start_scores = outputs.start_logits
|
| end_scores = outputs.end_logits
|
|
|
| # Extract answer
|
| start_index = start_scores.argmax()
|
| end_index = end_scores.argmax()
|
| answer = tokenizer.decode(inputs["input_ids"][0][start_index:end_index + 1])
|
|
|
| print("Answer:", answer)
|
| ```
|
|
|
| ## Citation
|
|
|
| If you use this model in your research, please cite the original BERT paper:
|
|
|
| ```bibtex
|
| @article{devlin2018bert,
|
| title={BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding},
|
| author={Devlin, Jacob and Chang, Ming-Wei and Lee, Kenton and Toutanova, Kristina},
|
| journal={arXiv preprint arXiv:1810.04805},
|
| year={2018}
|
| }
|
| ```
|
|
|