Bert-SquAD-QA / README.md
Abdo36's picture
Updated model card
fe0a17e verified
---
language:
- en
tags:
- question-answering
- transformers
- bert
- squad
license: apache-2.0
datasets:
- squad
model-name: bert-base-uncased-finetuned-squad
library_name: transformers
---
# BERT-Base Uncased Fine-Tuned on SQuAD
## Overview
This repository contains a **BERT-Base Uncased** model fine-tuned on the **SQuAD (Stanford Question Answering Dataset)** for **Question Answering (QA) tasks**. The model has been fine-tuned for **2 epochs**, making it suitable for extracting answers from given contexts by predicting start and end token positions.
## The Model predicts 2 probabilities among all the tokens in the vocab , One indicating the start token and the other indicating the end token, Then the answer between both these tokens are extracted.
## Model Details
- **Model Type**: BERT-Base Uncased
- **Fine-Tuning Dataset**: SQuAD (Stanford Question Answering Dataset)
- **Number of Epochs**: 2
- **Task**: Question Answering
- **Base Model**: [BERT-Base Uncased](https://huggingface.co/bert-base-uncased)
---
## Usage
### How to Load the Model
You can load the model using the `transformers` library from Hugging Face:
```python
from transformers import BertForQuestionAnswering, BertTokenizer
# Load the tokenizer and model
tokenizer = BertTokenizer.from_pretrained("Abdo36/Bert-SquAD-QA")
model = BertForQuestionAnswering.from_pretrained("Abdo36/Bert-SquAD-QA")
context = "BERT is a method of pre-training language representations."
question = "What is BERT?"
inputs = tokenizer.encode_plus(question, context, return_tensors="pt")
# Perform inference
outputs = model(**inputs)
start_scores = outputs.start_logits
end_scores = outputs.end_logits
# Extract answer
start_index = start_scores.argmax()
end_index = end_scores.argmax()
answer = tokenizer.decode(inputs["input_ids"][0][start_index:end_index + 1])
print("Answer:", answer)
```
## Citation
If you use this model in your research, please cite the original BERT paper:
```bibtex
@article{devlin2018bert,
title={BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding},
author={Devlin, Jacob and Chang, Ming-Wei and Lee, Kenton and Toutanova, Kristina},
journal={arXiv preprint arXiv:1810.04805},
year={2018}
}
```