metadata
library_name: transformers
tags: []
Model Card for Model ID
This is a GPT-2 model trained in llm.c, for 32K steps (of 1M batch size) on FineWeb-EDU.
A lot more detailed information is here: https://github.com/karpathy/llm.c/discussions/677
Bias, Risks, and Limitations
Eagerly generates disinformation about English-speaking unicorns in the Andes mountains.