ppo-Huggy / run_logs /training_status.json
zpattdev's picture
Huggy first
2ecd642
{
"Huggy": {
"checkpoints": [
{
"steps": 199933,
"file_path": "results/Huggy/Huggy/Huggy-199933.onnx",
"reward": 3.2989487666350144,
"creation_time": 1685893372.9858158,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-199933.pt"
]
},
{
"steps": 399928,
"file_path": "results/Huggy/Huggy/Huggy-399928.onnx",
"reward": 3.643261293233451,
"creation_time": 1685893633.5695724,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-399928.pt"
]
},
{
"steps": 599866,
"file_path": "results/Huggy/Huggy/Huggy-599866.onnx",
"reward": 3.459470510482788,
"creation_time": 1685893902.4512067,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-599866.pt"
]
},
{
"steps": 799938,
"file_path": "results/Huggy/Huggy/Huggy-799938.onnx",
"reward": 3.646937182523254,
"creation_time": 1685894169.699844,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-799938.pt"
]
},
{
"steps": 999942,
"file_path": "results/Huggy/Huggy/Huggy-999942.onnx",
"reward": 3.2755108109439712,
"creation_time": 1685894436.1574187,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-999942.pt"
]
},
{
"steps": 1199988,
"file_path": "results/Huggy/Huggy/Huggy-1199988.onnx",
"reward": 2.8745163624936882,
"creation_time": 1685894705.5082548,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1199988.pt"
]
},
{
"steps": 1399613,
"file_path": "results/Huggy/Huggy/Huggy-1399613.onnx",
"reward": 3.591893242735441,
"creation_time": 1685894970.9892201,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1399613.pt"
]
},
{
"steps": 1599968,
"file_path": "results/Huggy/Huggy/Huggy-1599968.onnx",
"reward": 3.9741455993869086,
"creation_time": 1685895236.080857,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1599968.pt"
]
},
{
"steps": 1799907,
"file_path": "results/Huggy/Huggy/Huggy-1799907.onnx",
"reward": 3.0265474021434784,
"creation_time": 1685895501.8614807,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1799907.pt"
]
},
{
"steps": 1999999,
"file_path": "results/Huggy/Huggy/Huggy-1999999.onnx",
"reward": 3.415660858918459,
"creation_time": 1685895764.7821167,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1999999.pt"
]
},
{
"steps": 2000009,
"file_path": "results/Huggy/Huggy/Huggy-2000009.onnx",
"reward": 3.4004826241997397,
"creation_time": 1685895764.9056213,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-2000009.pt"
]
}
],
"final_checkpoint": {
"steps": 2000009,
"file_path": "results/Huggy/Huggy.onnx",
"reward": 3.4004826241997397,
"creation_time": 1685895764.9056213,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-2000009.pt"
]
}
},
"metadata": {
"stats_format_version": "0.3.0",
"mlagents_version": "0.31.0.dev0",
"torch_version": "1.11.0+cu102"
}
}