ppo-Huggy / run_logs /training_status.json
tg825590's picture
Huggy
35aa63b
raw
history blame
4.38 kB
{
"Huggy": {
"checkpoints": [
{
"steps": 199912,
"file_path": "results/Huggy/Huggy/Huggy-199912.onnx",
"reward": 3.4976556259724827,
"creation_time": 1682476294.718137,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-199912.pt"
]
},
{
"steps": 399859,
"file_path": "results/Huggy/Huggy/Huggy-399859.onnx",
"reward": 3.709299368293662,
"creation_time": 1682476539.4453225,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-399859.pt"
]
},
{
"steps": 599991,
"file_path": "results/Huggy/Huggy/Huggy-599991.onnx",
"reward": 3.640692188459284,
"creation_time": 1682476782.4181116,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-599991.pt"
]
},
{
"steps": 799954,
"file_path": "results/Huggy/Huggy/Huggy-799954.onnx",
"reward": 3.6978484108334495,
"creation_time": 1682477018.8469977,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-799954.pt"
]
},
{
"steps": 999939,
"file_path": "results/Huggy/Huggy/Huggy-999939.onnx",
"reward": 4.042499766899989,
"creation_time": 1682477261.7573707,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-999939.pt"
]
},
{
"steps": 1199650,
"file_path": "results/Huggy/Huggy/Huggy-1199650.onnx",
"reward": 4.274921178817749,
"creation_time": 1682477501.7055073,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1199650.pt"
]
},
{
"steps": 1399960,
"file_path": "results/Huggy/Huggy/Huggy-1399960.onnx",
"reward": 3.830636463165283,
"creation_time": 1682477741.149825,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1399960.pt"
]
},
{
"steps": 1599891,
"file_path": "results/Huggy/Huggy/Huggy-1599891.onnx",
"reward": 3.720150049621537,
"creation_time": 1682477974.4985833,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1599891.pt"
]
},
{
"steps": 1799937,
"file_path": "results/Huggy/Huggy/Huggy-1799937.onnx",
"reward": 3.3248569141612965,
"creation_time": 1682478218.9197025,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1799937.pt"
]
},
{
"steps": 1999976,
"file_path": "results/Huggy/Huggy/Huggy-1999976.onnx",
"reward": 2.481512943903605,
"creation_time": 1682478464.649547,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-1999976.pt"
]
},
{
"steps": 2000008,
"file_path": "results/Huggy/Huggy/Huggy-2000008.onnx",
"reward": 2.5112491101026535,
"creation_time": 1682478464.7755363,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-2000008.pt"
]
}
],
"final_checkpoint": {
"steps": 2000008,
"file_path": "results/Huggy/Huggy.onnx",
"reward": 2.5112491101026535,
"creation_time": 1682478464.7755363,
"auxillary_file_paths": [
"results/Huggy/Huggy/Huggy-2000008.pt"
]
}
},
"metadata": {
"stats_format_version": "0.3.0",
"mlagents_version": "0.31.0.dev0",
"torch_version": "1.11.0+cu102"
}
}