--- license: cc-by-4.0 datasets: - clarin-knext/msmarco-pl - clarin-knext/nq-pl - clarin-knext/hotpotqa-pl - clarin-knext/scidocs-pl - clarin-knext/nfcorpus-pl - clarin-knext/dbpedia-pl - clarin-knext/trec-covid-pl - clarin-knext/quora-pl - clarin-knext/arguana-pl - clarin-knext/fiqa-pl - radlab/wikipedia-pl - radlab/legal-mc4-pl language: - pl library_name: transformers tags: - gpt2 - from-scratch - polish-gpt2 --- ## Description This is the polish gpt2 model in medium architecture. This model was released on 30.11.2023. ## Datasets Data which are used to train this model: - clarin-knext/msmarco-pl - clarin-knext/nq-pl - clarin-knext/hotpotqa-pl - clarin-knext/scidocs-pl - clarin-knext/nfcorpus-pl - clarin-knext/dbpedia-pl - clarin-knext/trec-covid-pl - clarin-knext/quora-pl - clarin-knext/arguana-pl - clarin-knext/fiqa-pl - radlab/wikipedia-pl - radlab/legal-mc4-pl - own corpora not published yet It is about 30,5 GB of data which is 3 times more than the prevoius version. ## Metrics from W&B ![image/png](https://cdn-uploads.huggingface.co/production/uploads/644addfe9279988e0cbc296b/if9o_UUUEevA9pG4wYraQ.png) ![image/png](https://cdn-uploads.huggingface.co/production/uploads/644addfe9279988e0cbc296b/EuAz5vK-pwyKpXqVhYo8Q.png) ![image/png](https://cdn-uploads.huggingface.co/production/uploads/644addfe9279988e0cbc296b/b4EXRzvYoZJi-i7I_jbop.png) ## Changelog - _2023.11.30_ - publishing the first version of the model