|
---
|
|
language:
|
|
- en
|
|
tags:
|
|
- question-answering
|
|
- transformers
|
|
- bert
|
|
- squad
|
|
license: apache-2.0
|
|
datasets:
|
|
- squad
|
|
model-name: bert-base-uncased-finetuned-squad
|
|
library_name: transformers
|
|
---
|
|
|
|
# BERT-Base Uncased Fine-Tuned on SQuAD
|
|
|
|
## Overview
|
|
|
|
This repository contains a **BERT-Base Uncased** model fine-tuned on the **SQuAD (Stanford Question Answering Dataset)** for **Question Answering (QA) tasks**. The model has been fine-tuned for **2 epochs**, making it suitable for extracting answers from given contexts by predicting start and end token positions.
|
|
|
|
## The Model predicts 2 probabilities among all the tokens in the vocab , One indicating the start token and the other indicating the end token, Then the answer between both these tokens are extracted.
|
|
|
|
## Model Details
|
|
|
|
- **Model Type**: BERT-Base Uncased
|
|
- **Fine-Tuning Dataset**: SQuAD (Stanford Question Answering Dataset)
|
|
- **Number of Epochs**: 2
|
|
- **Task**: Question Answering
|
|
- **Base Model**: [BERT-Base Uncased](https://huggingface.co/bert-base-uncased)
|
|
|
|
---
|
|
|
|
## Usage
|
|
|
|
### How to Load the Model
|
|
|
|
You can load the model using the `transformers` library from Hugging Face:
|
|
|
|
```python
|
|
from transformers import BertForQuestionAnswering, BertTokenizer
|
|
|
|
# Load the tokenizer and model
|
|
tokenizer = BertTokenizer.from_pretrained("Abdo36/Bert-SquAD-QA")
|
|
model = BertForQuestionAnswering.from_pretrained("Abdo36/Bert-SquAD-QA")
|
|
|
|
context = "BERT is a method of pre-training language representations."
|
|
question = "What is BERT?"
|
|
|
|
inputs = tokenizer.encode_plus(question, context, return_tensors="pt")
|
|
|
|
# Perform inference
|
|
outputs = model(**inputs)
|
|
start_scores = outputs.start_logits
|
|
end_scores = outputs.end_logits
|
|
|
|
# Extract answer
|
|
start_index = start_scores.argmax()
|
|
end_index = end_scores.argmax()
|
|
answer = tokenizer.decode(inputs["input_ids"][0][start_index:end_index + 1])
|
|
|
|
print("Answer:", answer)
|
|
```
|
|
|
|
## Citation
|
|
|
|
If you use this model in your research, please cite the original BERT paper:
|
|
|
|
```bibtex
|
|
@article{devlin2018bert,
|
|
title={BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding},
|
|
author={Devlin, Jacob and Chang, Ming-Wei and Lee, Kenton and Toutanova, Kristina},
|
|
journal={arXiv preprint arXiv:1810.04805},
|
|
year={2018}
|
|
}
|
|
```
|
|
|