--- task_categories: - text-generation - summarization language: - en - hi - ja - fr tags: - textdataset - text - youtube - webscrapped data - youtube transcripts - llm training - transformer models size_categories: - 1B This dataset contains transcripts of around 167K youtube videos that include coding lectures, podcasts, interviews, news videos, commentary and song lyrics. Also there are multiple files that have been generated using webscrapping. - **Curated by:** [Shivendra Singh](https://linktr.ee/shivendrra_) - **License:** [none] ### Dataset Sources - **Repository:** [SmallLanguageModel](https://github.com/shivendrra/SmallLanguageModel-project) - **Demo [optional]:** [More Information Needed] ## Uses - Can be used to train Transformer model/BPE tokenizers - Also for learning and research purposes - whatever you can think of, do whatever the fuck you want. ### Direct Use Used to train a 76million parameter transformer model. [Github repo](https://github.com/shivendrra/SmallLanguageModel-project) ### Out-of-Scope Use Not suitable for finetuning any base model or pre-trained models. Only NLP and base model training from scratch. ## Dataset Structure I'll add some finetuning data and then will update this section ## Dataset Creation ### Curation Rationale I wanted to create an app that would help me write script for my youtube videos. I fucked around a little with gpt-3.5 finetuning and langchain, and Youtube/Google APIs and got an idea to make a model and train it from scratch, all by myself. [Youtube video](https://youtu.be/PVpyN_2z5II?si=Q1yl-sVp8kxaGyre) ### Source Data Youtube Videos: -podcasts like Lex Fridman's, Waveform, Joe Rogan, vergecast, bill gates, etc. -videos from candaian lad, aevy tv, SNL, lemmino, mrwhosetheboss, johnny harris, and many more. -news videos from vox, wallstreetjournal, newyorktimes, the guardian, etc. -interviews from variety, wired, y-combinator, eo, etc. -lectures from mit opencourseware, cs50, freecodecamp, crashcourse, etc. -tech and science from kurzgesagt, real engineering, arvin ash, vsause, veritasium, etc. Britannica.com: -articles on various topics like Covid, Nuclear reactions, Antarctica, Nobel prize, Great leaders, countries, etc. #### Data Collection and Processing Used [Youtube V3 API](https://console.cloud.google.com/apis/api/youtube.googleapis.com/) to fetch video ids from a particular Youtube channel and generated a traget url. Then used [Youtube Transcript API](https://pypi.org/project/youtube-transcript-api/) to fetch transcripts from the videos and write it in a .txt file. Made a json file containing channel ids of around 45channels and fetched transcipts from around 167K videos Webscrapping data was generated using webscrapper that scrapped data from britannica.com and some sites that were fetched by GoogleCustomSearch API. [More Information Needed](https://medium.com/@shivendrra_/build-your-own-llm-using-youtube-transcript-data-87c04469c5e2)