Update README.md
Browse files
README.md
CHANGED
@@ -39,18 +39,23 @@ parameters:
|
|
39 |
|
40 |
# Grammar-Synthesis-Enhanced: FLAN-t5
|
41 |
|
42 |
-
|
43 |
<img src="https://colab.research.google.com/assets/colab-badge.svg" alt="Open In Colab"/>
|
44 |
</a>
|
45 |
|
46 |
-
|
47 |
|
|
|
48 |
|
49 |
-
|
50 |
|
51 |
-
|
52 |
|
53 |
-
|
|
|
|
|
|
|
|
|
54 |
|
55 |
```python
|
56 |
from transformers import pipeline
|
|
|
39 |
|
40 |
# Grammar-Synthesis-Enhanced: FLAN-t5
|
41 |
|
42 |
+
<a href="https://colab.research.google.com/gist/pszemraj/5dc89199a631a9c6cfd7e386011452a0/demo-flan-t5-large-grammar-synthesis.ipynb">
|
43 |
<img src="https://colab.research.google.com/assets/colab-badge.svg" alt="Open In Colab"/>
|
44 |
</a>
|
45 |
|
46 |
+
This model is a fine-tuned version of [pszemraj/flan-t5-large-grammar-synthesis](https://huggingface.co/pszemraj/flan-t5-large-grammar-synthesis) using the C4 200M dataset for the NaraSpeak Bangkit 2024 ENTR-H130 application.
|
47 |
|
48 |
+
## T5 Model Overview
|
49 |
|
50 |
+
The T5 (Text-To-Text Transfer Transformer) model, introduced by Google Research, is a transformer-based model that treats every NLP task as a text-to-text problem. This unified approach allows T5 to excel at a variety of tasks, such as translation, summarization, and question answering, by converting inputs and outputs into text format.
|
51 |
|
52 |
+
### Transformer Architecture
|
53 |
|
54 |
+
Transformers are a type of deep learning model designed for sequence-to-sequence tasks. They utilize a mechanism called "attention" to weigh the influence of different words in a sequence, allowing the model to focus on relevant parts of the input when generating each word in the output. This architecture is highly parallelizable and has proven effective in NLP tasks.
|
55 |
+
|
56 |
+
## Usage in Python
|
57 |
+
|
58 |
+
After `pip install transformers`, run the following code:
|
59 |
|
60 |
```python
|
61 |
from transformers import pipeline
|