benlipkin's picture
args and setup.
30b356b
raw
history blame contribute delete
410 Bytes
{
"model": "gpt2",
"dataset": "wikitext",
"subset": "wikitext-103-v1",
"output_dir": "output",
"num_epochs": 20,
"num_tokens": 100000000,
"window_size": 32,
"batch_size": 512,
"learning_rate": 1e-05,
"warmup_steps": 3000,
"scheduler": "cosine",
"weight_decay": 0.1,
"random_seed": 42,
"eval_steps": 1000,
"patience": 5,
"id": "4271d55d34c8c387"
}