ppo-Huggy / run_logs /training_status.json
D3MI4N's picture
Huggy
f705db5 verified
{
"Huggy": {
"checkpoints": [
{
"steps": 199908,
"file_path": "results/Huggy2/Huggy/Huggy-199908.onnx",
"reward": 3.409526271606559,
"creation_time": 1718950708.8969884,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-199908.pt"
]
},
{
"steps": 399983,
"file_path": "results/Huggy2/Huggy/Huggy-399983.onnx",
"reward": 3.8304943469437687,
"creation_time": 1718950943.5999355,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-399983.pt"
]
},
{
"steps": 599861,
"file_path": "results/Huggy2/Huggy/Huggy-599861.onnx",
"reward": 3.8438420626852245,
"creation_time": 1718951184.717459,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-599861.pt"
]
},
{
"steps": 799952,
"file_path": "results/Huggy2/Huggy/Huggy-799952.onnx",
"reward": 3.944607486378961,
"creation_time": 1718951419.7593048,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-799952.pt"
]
},
{
"steps": 999923,
"file_path": "results/Huggy2/Huggy/Huggy-999923.onnx",
"reward": 3.803531540494387,
"creation_time": 1718951658.3907535,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-999923.pt"
]
},
{
"steps": 1199957,
"file_path": "results/Huggy2/Huggy/Huggy-1199957.onnx",
"reward": 3.5539825840128794,
"creation_time": 1718951896.5814931,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-1199957.pt"
]
},
{
"steps": 1399941,
"file_path": "results/Huggy2/Huggy/Huggy-1399941.onnx",
"reward": 3.839244754132578,
"creation_time": 1718952132.814045,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-1399941.pt"
]
},
{
"steps": 1599662,
"file_path": "results/Huggy2/Huggy/Huggy-1599662.onnx",
"reward": 3.740521822397302,
"creation_time": 1718952372.9722133,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-1599662.pt"
]
},
{
"steps": 1799535,
"file_path": "results/Huggy2/Huggy/Huggy-1799535.onnx",
"reward": 3.7319080554522,
"creation_time": 1718952608.6633627,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-1799535.pt"
]
},
{
"steps": 1999944,
"file_path": "results/Huggy2/Huggy/Huggy-1999944.onnx",
"reward": 4.094827184677124,
"creation_time": 1718952845.8949244,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-1999944.pt"
]
},
{
"steps": 2000057,
"file_path": "results/Huggy2/Huggy/Huggy-2000057.onnx",
"reward": 4.194906739088205,
"creation_time": 1718952846.0218408,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-2000057.pt"
]
}
],
"final_checkpoint": {
"steps": 2000057,
"file_path": "results/Huggy2/Huggy.onnx",
"reward": 4.194906739088205,
"creation_time": 1718952846.0218408,
"auxillary_file_paths": [
"results/Huggy2/Huggy/Huggy-2000057.pt"
]
}
},
"metadata": {
"stats_format_version": "0.3.0",
"mlagents_version": "1.1.0.dev0",
"torch_version": "2.3.0+cu121"
}
}