Update README.md
Browse files
README.md
CHANGED
@@ -7,7 +7,7 @@ license: mit
|
|
7 |
---
|
8 |
|
9 |
|
10 |
-
#
|
11 |
|
12 |
Test the whole generation capabilities here: https://transformer.huggingface.co/doc/gpt2-large
|
13 |
|
@@ -104,7 +104,7 @@ Here's an example of how the model can have biased predictions:
|
|
104 |
>>> from transformers import pipeline, set_seed
|
105 |
>>> generator = pipeline('text-generation', model='gpt2')
|
106 |
>>> set_seed(42)
|
107 |
-
>>> generator("The
|
108 |
|
109 |
[{'generated_text': 'The White man worked as a mannequin for'},
|
110 |
{'generated_text': 'The White man worked as a maniser of the'},
|
@@ -134,6 +134,7 @@ this dataset, so the model was not trained on any part of Wikipedia. The resulti
|
|
134 |
|
135 |
## Training procedure
|
136 |
|
|
|
137 |
### Preprocessing
|
138 |
|
139 |
The texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a
|
|
|
7 |
---
|
8 |
|
9 |
|
10 |
+
# CHATBOT
|
11 |
|
12 |
Test the whole generation capabilities here: https://transformer.huggingface.co/doc/gpt2-large
|
13 |
|
|
|
104 |
>>> from transformers import pipeline, set_seed
|
105 |
>>> generator = pipeline('text-generation', model='gpt2')
|
106 |
>>> set_seed(42)
|
107 |
+
>>> generator("The black man worked as a", max_length=10, num_return_sequences=5)
|
108 |
|
109 |
[{'generated_text': 'The White man worked as a mannequin for'},
|
110 |
{'generated_text': 'The White man worked as a maniser of the'},
|
|
|
134 |
|
135 |
## Training procedure
|
136 |
|
137 |
+
|
138 |
### Preprocessing
|
139 |
|
140 |
The texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a
|