{ "Huggy": { "checkpoints": [ { "steps": 499883, "file_path": "results/Huggy/Huggy/Huggy-499883.onnx", "reward": 3.995732213060061, "creation_time": 1671041010.5275624, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-499883.pt" ] }, { "steps": 999685, "file_path": "results/Huggy/Huggy/Huggy-999685.onnx", "reward": 3.6068526417948306, "creation_time": 1671041540.508205, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-999685.pt" ] }, { "steps": 1499883, "file_path": "results/Huggy/Huggy/Huggy-1499883.onnx", "reward": 3.4947310465833414, "creation_time": 1671042063.871647, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-1499883.pt" ] }, { "steps": 1999883, "file_path": "results/Huggy/Huggy/Huggy-1999883.onnx", "reward": 3.3967123214155435, "creation_time": 1671042589.2499971, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-1999883.pt" ] }, { "steps": 2000002, "file_path": "results/Huggy/Huggy/Huggy-2000002.onnx", "reward": 3.3955315688382024, "creation_time": 1671042589.363981, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-2000002.pt" ] } ], "final_checkpoint": { "steps": 2000002, "file_path": "results/Huggy/Huggy.onnx", "reward": 3.3955315688382024, "creation_time": 1671042589.363981, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-2000002.pt" ] } }, "metadata": { "stats_format_version": "0.3.0", "mlagents_version": "0.29.0.dev0", "torch_version": "1.8.1+cu102" } }