mkthoma commited on
Commit
5bb6892
·
1 Parent(s): 939dd1c

readme update

Browse files
Files changed (1) hide show
  1. README.md +4 -0
README.md CHANGED
@@ -42,3 +42,7 @@ The Bigram Language Model is based on the Transformer architecture, which has be
42
 
43
  This architecture allows the Bigram Language Model to capture complex dependencies in the input data and generate text that follows the patterns and style of the training data. It is a fundamental building block for a wide range of natural language processing tasks and can be further extended and fine-tuned for specific applications.
44
 
 
 
 
 
 
42
 
43
  This architecture allows the Bigram Language Model to capture complex dependencies in the input data and generate text that follows the patterns and style of the training data. It is a fundamental building block for a wide range of natural language processing tasks and can be further extended and fine-tuned for specific applications.
44
 
45
+ ## Training Data
46
+ The data used for training the GPT models can be found here:
47
+ - [Shakespeare dataset](https://github.com/karpathy/char-rnn/blob/6f9487a6fe5b420b7ca9afb0d7c078e37c1d1b4e/data/tinyshakespeare/input.txt)
48
+ - [Wikiedpia dataset](https://www.kaggle.com/datasets/mikeortman/wikipedia-sentences)