ppo-Huggy / run_logs /training_status.json
yanka9's picture
Huggy
9af548e
{
"Huggy": {
"checkpoints": [
{
"steps": 199951,
"file_path": "results/Huggy/Huggy/Huggy-199951.onnx",
"reward": 3.2104335543182163,
"creation_time": 1696944845.8444088,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-199951.pt"
]
},
{
"steps": 399949,
"file_path": "results/Huggy/Huggy/Huggy-399949.onnx",
"reward": 4.057730617431494,
"creation_time": 1696945104.2671864,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-399949.pt"
]
},
{
"steps": 599983,
"file_path": "results/Huggy/Huggy/Huggy-599983.onnx",
"reward": 4.3233839472134905,
"creation_time": 1696945363.3151207,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-599983.pt"
]
},
{
"steps": 799998,
"file_path": "results/Huggy/Huggy/Huggy-799998.onnx",
"reward": 3.9448208121395614,
"creation_time": 1696945619.821488,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-799998.pt"
]
},
{
"steps": 999955,
"file_path": "results/Huggy/Huggy/Huggy-999955.onnx",
"reward": 4.296731462923147,
"creation_time": 1696945878.133612,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-999955.pt"
]
},
{
"steps": 1199985,
"file_path": "results/Huggy/Huggy/Huggy-1199985.onnx",
"reward": 3.540226794206179,
"creation_time": 1696946136.3619473,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1199985.pt"
]
},
{
"steps": 1399912,
"file_path": "results/Huggy/Huggy/Huggy-1399912.onnx",
"reward": 4.00359134636228,
"creation_time": 1696946392.5333617,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1399912.pt"
]
},
{
"steps": 1599900,
"file_path": "results/Huggy/Huggy/Huggy-1599900.onnx",
"reward": 4.113592889419822,
"creation_time": 1696946652.009029,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1599900.pt"
]
},
{
"steps": 1799993,
"file_path": "results/Huggy/Huggy/Huggy-1799993.onnx",
"reward": 3.865091286599636,
"creation_time": 1696946914.5095882,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1799993.pt"
]
},
{
"steps": 1999975,
"file_path": "results/Huggy/Huggy/Huggy-1999975.onnx",
"reward": 3.7766625590440706,
"creation_time": 1696947172.2142975,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1999975.pt"
]
},
{
"steps": 2000054,
"file_path": "results/Huggy/Huggy/Huggy-2000054.onnx",
"reward": 3.765142315909976,
"creation_time": 1696947172.374847,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-2000054.pt"
]
}
],
"final_checkpoint": {
"steps": 2000054,
"file_path": "results/Huggy/Huggy.onnx",
"reward": 3.765142315909976,
"creation_time": 1696947172.374847,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-2000054.pt"
]
}
},
"metadata": {
"stats_format_version": "0.3.0",
"mlagents_version": "1.1.0.dev0",
"torch_version": "2.0.1+cu118"
}
}