summarization-model / README.md
ChijoTheDatascientist's picture
Update README.md
2f1750d verified
---
library_name: transformers
license: mit
language:
- en
pipeline_tag: summarization
---
# BART Base Text Summarization Modeli
<!-- Provide a quick summary of what the model is/does. -->
This model is based on the Facebook BART (Bidirectional and Auto-Regressive Transformers) architecture. BART is particularly effective when fine-tuned for text generation tasks like summarization but also works well for comprehension tasks.
BART is a transformer encoder-decoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder. BART is pre-trained by (1) corrupting text with an arbitrary noising function,
and (2) learning a model to reconstruct the original text.
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Architecture:** [BART Base]
- **Pre-trained model:** [facebook/bart-base]
- **Fine-tuned for:** [Summarization]
- **License:** [MIT]
- **Finetuned from model:** [facebook/bart-base]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
- **Installation:**
pip install transformers
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
Here is a simple snippet oon how to use the model directly.
# Load model directly
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
tokenizer = AutoTokenizer.from_pretrained("ChijoTheDatascientist/summarization-model")
model = AutoModelForSeq2SeqLM.from_pretrained("ChijoTheDatascientist/summarization-model")