ttt-mc-pythia-2.8b / all_results.json
ricdomolm's picture
Add files from e3
e5e5663
[{"train_loss": 0.09130859375, "epoch": 2, "step": 3072, "lr": 4.750621824247943e-06}, {"train_loss": 0.0087890625, "epoch": 2, "step": 3200, "lr": 4.0073511784396335e-06}, {"train_loss": 0.032470703125, "epoch": 2, "step": 3328, "lr": 3.312524130702509e-06}, {"train_loss": 0.0289306640625, "epoch": 2, "step": 3456, "lr": 2.6717575498422943e-06}, {"train_loss": 0.10791015625, "epoch": 2, "step": 3584, "lr": 2.0902312886461217e-06}, {"train_loss": 0.03369140625, "epoch": 2, "step": 3712, "lr": 1.572646310795336e-06}, {"train_loss": 0.17578125, "epoch": 2, "step": 3840, "lr": 1.12318668903995e-06}, {"train_loss": 0.2333984375, "epoch": 2, "step": 3968, "lr": 7.45485781835279e-07}, {"train_loss": 0.003631591796875, "epoch": 2, "step": 4096, "lr": 4.425968618633292e-07}, {"train_loss": 0.8515625, "epoch": 2, "step": 4224, "lr": 2.1696843387321142e-07}, {"train_loss": 0.01953125, "epoch": 2, "step": 4352, "lr": 7.04244413671129e-08}, {"train_loss": 0.004241943359375, "epoch": 2, "step": 4480, "lr": 4.149522137787409e-09}]