{ "Huggy": { "checkpoints": [ { "steps": 199912, "file_path": "results/Huggy/Huggy/Huggy-199912.onnx", "reward": 3.6211049343858446, "creation_time": 1673382013.424851, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-199912.pt" ] }, { "steps": 254283, "file_path": "results/Huggy/Huggy/Huggy-254283.onnx", "reward": 3.355522334575653, "creation_time": 1673382076.152711, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-254283.pt" ] } ], "final_checkpoint": { "steps": 254283, "file_path": "results/Huggy/Huggy.onnx", "reward": 3.355522334575653, "creation_time": 1673382076.152711, "auxillary_file_paths": [ "results/Huggy/Huggy/Huggy-254283.pt" ] } }, "metadata": { "stats_format_version": "0.3.0", "mlagents_version": "0.29.0.dev0", "torch_version": "1.8.1+cu102" } }