Update README.md
Browse files
README.md
CHANGED
@@ -8,7 +8,7 @@ datasets:
|
|
8 |
This is a fine-tuned version of the GPT2 model. It's best suited for text-generation.
|
9 |
|
10 |
## Model Description
|
11 |
-
gpt2-finetuned-
|
12 |
|
13 |
## Intended Uses and Limitations
|
14 |
Given the magnitude of the [Microsoft DialoGPT-large](https://huggingface.co/microsoft/DialoGPT-large) model, the author resorted to fine-tuning the gpt2 model for the creation of a chatbot. The intent was for the chatbot to emulate a banking customer agent, hence the use of the banking77 dataset. However, when the fine-tuned model was deployed in the chatbot, the results were undesirable. Its responses were inappropriate and unnecessarily long. The last word of its response is repeated numerously, a major glitch in it. The model performs better in text-generation but is prone to generating banking-related text because of the corpus it was trained on.
|
@@ -20,7 +20,7 @@ You can use this model directly with a pipeline for text generation:
|
|
20 |
```python
|
21 |
>>>from transformers import pipeline
|
22 |
|
23 |
-
>>> model_name = "Kwaku/gpt2-finetuned-
|
24 |
>>> generator = pipeline("text-generation", model=model_name)
|
25 |
>>> result = generator("My money is", max_length=15, num_return_sequences=2)
|
26 |
>>> print(result)
|
|
|
8 |
This is a fine-tuned version of the GPT2 model. It's best suited for text-generation.
|
9 |
|
10 |
## Model Description
|
11 |
+
Kwaku/gpt2-finetuned-banking77 was fine tuned on the [banking77](https://huggingface.co/datasets/banking77) dataset, which is "composed of online banking queries annotated with their corresponding intents."
|
12 |
|
13 |
## Intended Uses and Limitations
|
14 |
Given the magnitude of the [Microsoft DialoGPT-large](https://huggingface.co/microsoft/DialoGPT-large) model, the author resorted to fine-tuning the gpt2 model for the creation of a chatbot. The intent was for the chatbot to emulate a banking customer agent, hence the use of the banking77 dataset. However, when the fine-tuned model was deployed in the chatbot, the results were undesirable. Its responses were inappropriate and unnecessarily long. The last word of its response is repeated numerously, a major glitch in it. The model performs better in text-generation but is prone to generating banking-related text because of the corpus it was trained on.
|
|
|
20 |
```python
|
21 |
>>>from transformers import pipeline
|
22 |
|
23 |
+
>>> model_name = "Kwaku/gpt2-finetuned-banking77"
|
24 |
>>> generator = pipeline("text-generation", model=model_name)
|
25 |
>>> result = generator("My money is", max_length=15, num_return_sequences=2)
|
26 |
>>> print(result)
|