| {"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVNwAAAAAAAACMHnN0YWJsZV9iYXNlbGluZXMzLnRkMy5wb2xpY2llc5SMEE11bHRpSW5wdXRQb2xpY3mUk5Qu", "__module__": "stable_baselines3.td3.policies", "__firstlineno__": 314, "__doc__": "\nPolicy class (with both actor and critic) for TD3 to be used with Dict observation spaces.\n\n:param observation_space: Observation space\n:param action_space: Action space\n:param lr_schedule: Learning rate schedule (could be constant)\n:param net_arch: The specification of the policy and value networks.\n:param activation_fn: Activation function\n:param features_extractor_class: Features extractor to use.\n:param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n:param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n:param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n:param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n:param n_critics: Number of critic networks to create.\n:param share_features_extractor: Whether to share or not the features extractor\n between the actor and the critic (this saves computation time)\n", "__init__": "<function MultiInputPolicy.__init__ at 0x11f451080>", "__static_attributes__": [], "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x11f449b80>"}, "verbose": 1, "policy_kwargs": {"n_critics": 1}, "num_timesteps": 10000, "_total_timesteps": 10000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1739907473751650000, "learning_rate": 0.001, "tensorboard_log": "runs/i33fvrb0", "_last_obs": {":type:": "<class 'collections.OrderedDict'>", ":serialized:": "gAWVLAEAAAAAAACMC2NvbGxlY3Rpb25zlIwLT3JkZXJlZERpY3SUk5QpUpQojA1hY2hpZXZlZF9nb2FslIwTbnVtcHkuX2NvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYMAAAAAAAAAAdauj0rW6c9722wPZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJLAUsDhpSMAUOUdJRSlIwMZGVzaXJlZF9nb2FslGgHKJYMAAAAAAAAAB0bCj4faNi8OmWgPZRoDksBSwOGlGgSdJRSlIwLb2JzZXJ2YXRpb26UaAcolhgAAAAAAAAAB1q6PStbpz3vbbA9SZm3vRYIXD/z6fe7lGgOSwFLBoaUaBJ0lFKUdS4=", "achieved_goal": "[[0.09099203 0.08171686 0.08614718]]", "desired_goal": "[[ 0.13486905 -0.02641684 0.07831807]]", "observation": "[[ 0.09099203 0.08171686 0.08614718 -0.08964784 0.8594984 -0.00756573]]"}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQAAAAAAAACME251bXB5Ll9jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWAQAAAAAAAAABlIwFbnVtcHmUjAVkdHlwZZSTlIwCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksBhZSMAUOUdJRSlC4="}, "_last_original_obs": {":type:": "<class 'collections.OrderedDict'>", ":serialized:": "gAWVLAEAAAAAAACMC2NvbGxlY3Rpb25zlIwLT3JkZXJlZERpY3SUk5QpUpQojA1hY2hpZXZlZF9nb2FslIwTbnVtcHkuX2NvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYMAAAAAAAAAMDiwD1yQI09BD7EPZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJLAUsDhpSMAUOUdJRSlIwMZGVzaXJlZF9nb2FslGgHKJYMAAAAAAAAAB0bCj4faNi8OmWgPZRoDksBSwOGlGgSdJRSlIwLb2JzZXJ2YXRpb26UaAcolhgAAAAAAAAAwOLAPXJAjT0EPsQ9nDavuzh1qr5rdvC+lGgOSwFLBoaUaBJ0lFKUdS4=", "achieved_goal": "[[0.09418249 0.06897058 0.09582141]]", "desired_goal": "[[ 0.13486905 -0.02641684 0.07831807]]", "observation": "[[ 0.09418249 0.06897058 0.09582141 -0.00534709 -0.33292556 -0.46965346]]"}, "_episode_num": 987, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": 0.0, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWV4AsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHv+AYfnwG4ZyMAWyUSwaMAXSUR0BCLptix3V1dX2UKGgGR7/cuNxVAAyVaAdLBGgIR0BCMONHYpUhdX2UKGgGR7/ij2Bas6q9aAdLBmgIR0BCNACOmzjWdX2UKGgGR7/ByFPBSDRMaAdLAmgIR0BCNPUrkKeDdX2UKGgGR7/VXKr7wazeaAdLBGgIR0BCNwFC9h7WdX2UKGgGR7/iacAiml67aAdLBmgIR0BCOnVPN3W4dX2UKGgGR7/ihWo3rD64aAdLBmgIR0BCPXai9IwudX2UKGgGR7/fXokiUxEfaAdLBmgIR0BCQH6/IsAedX2UKGgGR7+mRkmQbMouaAdLAWgIR0BCQSjpLVWkdX2UKGgGR7/a2SMcZLqVaAdLBWgIR0BCQ+o1k1/EdX2UKGgGR7/HXuE25xzaaAdLA2gIR0BCRb2L5ylvdX2UKGgGR7/R23KB/ZuiaAdLBGgIR0BCR9WZJCjUdX2UKGgGR7/Kll9Sde6aaAdLA2gIR0BCSUgB91EFdX2UKGgGR7/V1A7gbZOBaAdLBWgIR0BCS+cx0uDjdX2UKGgGR7/QMefZmI0qaAdLBGgIR0BCTe9SMtK7dX2UKGgGR7/U384xUNrkaAdLBGgIR0BCT+pOvdM1dX2UKGgGR7/Nd0q6OHWSaAdLBGgIR0BCUev6j323dX2UKGgGR7/saTGHYYixaAdLCGgIR0BCVjXvphWpdX2UKGgGR7/hbkn1FpfyaAdLBWgIR0BCWKmKqGUOdX2UKGgGR7/yNEkSmIj4aAdLEWgIR0BCYb349HMEdX2UKGgGR7/cTcqOLiuMaAdLBGgIR0BCY6c7QswtdX2UKGgGR7/OYHgP3BYWaAdLA2gIR0BCZVdxAB1cdX2UKGgGR8ATgtz0Yj0MaAdLMmgIR0BCfhTOxB3SdX2UKGgGR7+pIz3yqdYoaAdLAWgIR0BCfpZW7voedX2UKGgGR7/GvIOpbUw0aAdLA2gIR0BCgHWz4UN8dX2UKGgGR7/IcinpB5X2aAdLA2gIR0BCghllK9PDdX2UKGgGR8ANwxk/bCaaaAdLMmgIR0BCnBoVVPvbdX2UKGgGR7/YowEhaC+UaAdLBWgIR0BCnpQUHpr2dX2UKGgGR8AUavUz9CNTaAdLMmgIR0BCt6BI4EOidX2UKGgGR7/lbjtG/etTaAdLB2gIR0BCu42jwhGIdX2UKGgGR7/D1jiGWUr1aAdLA2gIR0BCvTDO1OTJdX2UKGgGR7/gPaL4vexfaAdLB2gIR0BCwLLhaTwEdX2UKGgGR7/fX3QD3dsSaAdLBmgIR0BCw6kRBeHBdX2UKGgGR8APGXiR4hUzaAdLMmgIR0BC3SGahHskdX2UKGgGR7/go/RmbsniaAdLB2gIR0BC4MTnJT2ndX2UKGgGR7/gVL8Jlar4aAdLBmgIR0BC46+evpyIdX2UKGgGR7/hLP+n62v0aAdLB2gIR0BC5yYgJTl1dX2UKGgGR7/RfYBeXzDoaAdLBGgIR0BC6X2ugYgrdX2UKGgGR7/wSPQv6CUYaAdLDGgIR0BC70jcEeQudX2UKGgGR7/V9Sde6ZpjaAdLBWgIR0BC8bhFVktmdX2UKGgGR7/swK0D2alUaAdLCmgIR0BC9pTMqz7edX2UKGgGR7++o1k1/DtPaAdLAmgIR0BC99qDbrTqdX2UKGgGR7/BrjYI0IkaaAdLAmgIR0BC+RFRYRukdX2UKGgGR7/hSflIVdonaAdLB2gIR0BC/Hhjvuw5dX2UKGgGR8AStx82Jiy6aAdLMmgIR0BDFOiWVu76dX2UKGgGR7/x801qFh5PaAdLEGgIR0BDHPK+zt1IdX2UKGgGR7/iebmU4aP0aAdLB2gIR0BDIFm4AjptdX2UKGgGR7/Zgntv4ubraAdLBWgIR0BDIsSkCV8kdX2UKGgGR8AHg1vVEuxsaAdLMmgIR0BDPPIn0CiidX2UKGgGR7/ZYf4h2W6caAdLBWgIR0BDP78FY+0PdX2UKGgGR7/dzIFNcnmaaAdLBWgIR0BDQjPGACnxdX2UKGgGR8ANCVdHDrJKaAdLMmgIR0BDWzrmhdt3dX2UKGgGR7/oFFDv3JxOaAdLCGgIR0BDXz+vQnhLdX2UKGgGR7+4uscQyylfaAdLAmgIR0BDYHtv4ubrdX2UKGgGR7/h5vtMPBi1aAdLBmgIR0BDY4i5d4VzdX2UKGgGR7/GEIPbwjMWaAdLA2gIR0BDZPJJXhfjdX2UKGgGR7/cTVDrqt5laAdLBmgIR0BDZ+Z5Rjz7dX2UKGgGR7/bRdhRZU1iaAdLBWgIR0BDapFCswL3dX2UKGgGR7/f09hZyMkyaAdLBmgIR0BDbX2dupCKdX2UKGgGR7+YjSofjjrBaAdLAWgIR0BDcAqmTC+DdX2UKGgGR7/oRyOq//NraAdLCmgIR0BDd2L5ylvZdX2UKGgGR8AJ9+qioKlYaAdLMmgIR0BDmtM495hSdX2UKGgGR8AQ+VxCIDYAaAdLMmgIR0BDtKbSZ0CBdX2UKGgGR7/GPVd5Y5ktaAdLA2gIR0BDtj9n9NvgdX2UKGgGR7/UyeqaPS2IaAdLBGgIR0BDuKVII4VAdX2UKGgGR7+o9RrJr+HaaAdLAWgIR0BDuUEHMUypdX2UKGgGR7/sQs5GSZBtaAdLCmgIR0BDvpp35eqrdX2UKGgGR7/wCI55qubJaAdLEGgIR0BDxs495hScdX2UKGgGR7/pBE0BOpKjaAdLC2gIR0BDzIkqtozvdX2UKGgGR7/WUBnzxwyZaAdLBWgIR0BDzzkhib2EdX2UKGgGR7+/zJ6po9LYaAdLAmgIR0BD0E4vN/vwdX2UKGgGR7/KZssQNCqqaAdLA2gIR0BD0oEB8x9HdX2UKGgGR7/Lz9S/CZWraAdLBGgIR0BD1RlpXZGsdX2UKGgGR8AS3IvJzT4MaAdLMmgIR0BD8r0jC53DdX2UKGgGR7/P09yLhrFgaAdLA2gIR0BD9FSCOFQEdX2UKGgGR7/SZML4N7SiaAdLBGgIR0BD9lIuoP07dX2UKGgGR7/YiJO32EkCaAdLBGgIR0BD+EAggX/HdX2UKGgGR7+i4OMERraeaAdLAWgIR0BD+OC5EtuldX2UKGgGR7/1+f29L6DXaAdLDWgIR0BD/5DArQPadX2UKGgGR7/UktmL9/BnaAdLBGgIR0BEAaya/h2odX2UKGgGR7/dYb83uNPyaAdLBmgIR0BEBI+W4Vh1dX2UKGgGR7/c3RG+bmU4aAdLBWgIR0BEB1BUrCm/dX2UKGgGR7/pFar3j+72aAdLB2gIR0BECunuRcNZdX2UKGgGR8AQKr2g3974aAdLMmgIR0BEI2fK6nR+dX2UKGgGR7+8+aBqbjLkaAdLAmgIR0BEJLdFfAsTdX2UKGgGR7/MydFvybx3aAdLA2gIR0BEJmN70Fr3dX2UKGgGR7/j5kTYdyT7aAdLB2gIR0BEKfjCHh0hdX2UKGgGR7+gvexfOUt7aAdLAWgIR0BEKn2AXl8xdX2UKGgGR7/NOcDr7fpEaAdLA2gIR0BELATIvJzUdX2UKGgGR8ABQ4bS7Xg+aAdLFGgIR0BEN4Q8OkLydX2UKGgGR7/f2AG0NSZSaAdLBWgIR0BEOz+NtIkJdX2UKGgGR7/a7zCk43m3aAdLBWgIR0BEP0bLlmvodX2UKGgGR7/SwcHWz4UOaAdLBGgIR0BEQjOLR8c/dX2UKGgGR7/XKU3XI2fkaAdLBWgIR0BERlCkXUH6dX2UKGgGR7/r+umrKeTWaAdLCmgIR0BEUlXRw6yTdX2UKGgGR7/OOd5IH1OCaAdLBGgIR0BEVOfVZs9CdX2UKGgGR7/NS8an752yaAdLA2gIR0BEVq//NqxkdX2UKGgGR7/gDwYtQKrraAdLBmgIR0BEWtxlxwQ2dX2UKGgGR7/Qwqy4Wk8BaAdLBGgIR0BEXcMNMGordX2UKGgGR7/q/rB0p3HJaAdLCGgIR0BEYnSOR1YAdWUu"}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVhgAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKIiIiIiIiIiIiIiIiIiIiIiIiIiIiIiJiIiIiYiJiIiIiImIiIiIiIiIiIiIiYiIiImIiImIiIiIiIiIiIiJiYiIiIiIiIiIiIiJiIiIiIiIiIiIiYiIiIiIiIiIiIiIiIiIiIhlLg=="}, "_n_updates": 9900, "observation_space": {":type:": "<class 'gymnasium.spaces.dict.Dict'>", ":serialized:": "gAWVkgMAAAAAAACMFWd5bW5hc2l1bS5zcGFjZXMuZGljdJSMBERpY3SUk5QpgZR9lCiMBnNwYWNlc5R9lCiMDWFjaGlldmVkX2dvYWyUjBRneW1uYXNpdW0uc3BhY2VzLmJveJSMA0JveJSTlCmBlH2UKIwFZHR5cGWUjAVudW1weZSMBWR0eXBllJOUjAJmNJSJiIeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRijAZfc2hhcGWUSwOFlIwDbG93lIwTbnVtcHkuX2NvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYMAAAAAAAAAAAAIMEAACDBAAAgwZRoE0sDhZSMAUOUdJRSlIwNYm91bmRlZF9iZWxvd5RoGyiWAwAAAAAAAAABAQGUaBCMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLA4WUaB50lFKUjARoaWdolGgbKJYMAAAAAAAAAAAAIEEAACBBAAAgQZRoE0sDhZRoHnSUUpSMDWJvdW5kZWRfYWJvdmWUaBsolgMAAAAAAAAAAQEBlGglSwOFlGgedJRSlIwIbG93X3JlcHKUjAUtMTAuMJSMCWhpZ2hfcmVwcpSMBDEwLjCUjApfbnBfcmFuZG9tlE51YowMZGVzaXJlZF9nb2FslGgKKYGUfZQoaA1oE2gWSwOFlGgYaBsolgwAAAAAAAAAAAAgwQAAIMEAACDBlGgTSwOFlGgedJRSlGghaBsolgMAAAAAAAAAAQEBlGglSwOFlGgedJRSlGgraBsolgwAAAAAAAAAAAAgQQAAIEEAACBBlGgTSwOFlGgedJRSlGgwaBsolgMAAAAAAAAAAQEBlGglSwOFlGgedJRSlGg1jAUtMTAuMJRoN4wEMTAuMJRoOU51YowLb2JzZXJ2YXRpb26UaAopgZR9lChoDWgTaBZLBoWUaBhoGyiWGAAAAAAAAAAAACDBAAAgwQAAIMEAACDBAAAgwQAAIMGUaBNLBoWUaB50lFKUaCFoGyiWBgAAAAAAAAABAQEBAQGUaCVLBoWUaB50lFKUaCtoGyiWGAAAAAAAAAAAACBBAAAgQQAAIEEAACBBAAAgQQAAIEGUaBNLBoWUaB50lFKUaDBoGyiWBgAAAAAAAAABAQEBAQGUaCVLBoWUaB50lFKUaDWMBS0xMC4wlGg3jAQxMC4wlGg5TnVidWgWTmgNTmg5TnViLg==", "spaces": "{'achieved_goal': Box(-10.0, 10.0, (3,), float32), 'desired_goal': Box(-10.0, 10.0, (3,), float32), 'observation': Box(-10.0, 10.0, (6,), float32)}", "_shape": null, "dtype": null, "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVZwMAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLB4WUjANsb3eUjBNudW1weS5fY29yZS5udW1lcmljlIwLX2Zyb21idWZmZXKUk5QolhwAAAAAAAAAAACAvwAAgL8AAIC/AACAvwAAgL8AAIC/AACAv5RoC0sHhZSMAUOUdJRSlIwNYm91bmRlZF9iZWxvd5RoEyiWBwAAAAAAAAABAQEBAQEBlGgIjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSweFlGgWdJRSlIwEaGlnaJRoEyiWHAAAAAAAAAAAAIA/AACAPwAAgD8AAIA/AACAPwAAgD8AAIA/lGgLSweFlGgWdJRSlIwNYm91bmRlZF9hYm92ZZRoEyiWBwAAAAAAAAABAQEBAQEBlGgdSweFlGgWdJRSlIwIbG93X3JlcHKUjAQtMS4wlIwJaGlnaF9yZXBylIwDMS4wlIwKX25wX3JhbmRvbZSMFG51bXB5LnJhbmRvbS5fcGlja2xllIwQX19nZW5lcmF0b3JfY3RvcpSTlGgyjBRfX2JpdF9nZW5lcmF0b3JfY3RvcpSTlIwTbnVtcHkucmFuZG9tLl9wY2c2NJSMBVBDRzY0lJOUhZRSlH2UKIwNYml0X2dlbmVyYXRvcpSMBVBDRzY0lIwFc3RhdGWUfZQoaD+KEHS3SoZSClQKIrbmwFXIhkKMA2luY5SKEKsvqyG5X72SNsfzt6c0FTZ1jApoYXNfdWludDMylEsAjAh1aW50ZWdlcpRLAHWMGm51bXB5LnJhbmRvbS5iaXRfZ2VuZXJhdG9ylIwbX19weXhfdW5waWNrbGVfU2VlZFNlcXVlbmNllJOUaESMDFNlZWRTZXF1ZW5jZZSTlEoiouoDToeUUpQoihCz+J7J9jrjfMkEbvqyE9IFSwBoEyiWEAAAAAAAAAA/T8/afKjTa/UwLVj7PBPUlGgIjAJ1NJSJiIeUUpQoSwNoDE5OTkr/////Sv////9LAHSUYksEhZRoFnSUUpRLBCl0lGKGlGKFlFKUdWIu", "dtype": "float32", "_shape": [7], "low": "[-1. -1. -1. -1. -1. -1. -1.]", "bounded_below": "[ True True True True True True True]", "high": "[1. 1. 1. 1. 1. 1. 1.]", "bounded_above": "[ True True True True True True True]", "low_repr": "-1.0", "high_repr": "1.0", "_np_random": "Generator(PCG64)"}, "n_envs": 1, "buffer_size": 1000000, "batch_size": 256, "learning_starts": 100, "tau": 0.005, "gamma": 0.99, "gradient_steps": 1, "optimize_memory_usage": false, "replay_buffer_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVPwAAAAAAAACMJ3N0YWJsZV9iYXNlbGluZXMzLmhlci5oZXJfcmVwbGF5X2J1ZmZlcpSMD0hlclJlcGxheUJ1ZmZlcpSTlC4=", "__module__": "stable_baselines3.her.her_replay_buffer", "__firstlineno__": 15, "__annotations__": "{'env': typing.Optional[stable_baselines3.common.vec_env.base_vec_env.VecEnv]}", "__doc__": "\nHindsight Experience Replay (HER) buffer.\nPaper: https://arxiv.org/abs/1707.01495\n\nReplay buffer for sampling HER (Hindsight Experience Replay) transitions.\n\n.. note::\n\n Compared to other implementations, the ``future`` goal sampling strategy is inclusive:\n the current transition can be used when re-sampling.\n\n:param buffer_size: Max number of element in the buffer\n:param observation_space: Observation space\n:param action_space: Action space\n:param env: The training environment\n:param device: PyTorch device\n:param n_envs: Number of parallel environments\n:param optimize_memory_usage: Enable a memory efficient variant\n Disabled for now (see https://github.com/DLR-RM/stable-baselines3/pull/243#discussion_r531535702)\n:param handle_timeout_termination: Handle timeout termination (due to timelimit)\n separately and treat the task as infinite horizon task.\n https://github.com/DLR-RM/stable-baselines3/issues/284\n:param n_sampled_goal: Number of virtual transitions to create per real transition,\n by sampling new goals.\n:param goal_selection_strategy: Strategy for sampling goals for replay.\n One of ['episode', 'final', 'future']\n:param copy_info_dict: Whether to copy the info dictionary and pass it to\n ``compute_reward()`` method.\n Please note that the copy may cause a slowdown.\n False by default.\n", "__init__": "<function HerReplayBuffer.__init__ at 0x11f451da0>", "__getstate__": "<function HerReplayBuffer.__getstate__ at 0x11f451e40>", "__setstate__": "<function HerReplayBuffer.__setstate__ at 0x11f451ee0>", "set_env": "<function HerReplayBuffer.set_env at 0x11f451f80>", "add": "<function HerReplayBuffer.add at 0x11f4520c0>", "_compute_episode_length": "<function HerReplayBuffer._compute_episode_length at 0x11f452160>", "sample": "<function HerReplayBuffer.sample at 0x11f452200>", "_get_real_samples": "<function HerReplayBuffer._get_real_samples at 0x11f4522a0>", "_get_virtual_samples": "<function HerReplayBuffer._get_virtual_samples at 0x11f452340>", "_sample_goals": "<function HerReplayBuffer._sample_goals at 0x11f4523e0>", "truncate_last_trajectory": "<function HerReplayBuffer.truncate_last_trajectory at 0x11f452480>", "__static_attributes__": ["_current_ep_start", "copy_info_dict", "env", "ep_length", "ep_start", "goal_selection_strategy", "her_ratio", "infos", "n_sampled_goal"], "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x11f464f80>"}, "replay_buffer_kwargs": {}, "train_freq": {":type:": "<class 'stable_baselines3.common.type_aliases.TrainFreq'>", ":serialized:": "gAWVYQAAAAAAAACMJXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi50eXBlX2FsaWFzZXOUjAlUcmFpbkZyZXGUk5RLAWgAjBJUcmFpbkZyZXF1ZW5jeVVuaXSUk5SMBHN0ZXCUhZRSlIaUgZQu"}, "use_sde_at_warmup": false, "policy_delay": 1, "target_noise_clip": 0.0, "target_policy_noise": 0.1, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVigQAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLBUsTQyY+AZUAWwEAAAAAAAAAAFQBIgBVADUBAAAAAAAANQEAAAAAAAAkAJROhZSMBWZsb2F0lIWUjBJwcm9ncmVzc19yZW1haW5pbmeUhZSMii9Vc2Vycy9zdGV2YW4vRG9jdW1lbnRzL0NlbnRyYWxlIEx5b24vM0EvTWFjaGluZSBMZWFybmluZy9oYW5kcy1vbi1ybC8udmVudi9saWIvcHl0aG9uMy4xMy9zaXRlLXBhY2thZ2VzL3N0YWJsZV9iYXNlbGluZXMzL2NvbW1vbi91dGlscy5weZSMCDxsYW1iZGE+lIwhZ2V0X3NjaGVkdWxlX2ZuLjxsb2NhbHM+LjxsYW1iZGE+lEtiQxD4gACkZalO0DtN0yxO1CZPlEMAlIwOdmFsdWVfc2NoZWR1bGWUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flIyKL1VzZXJzL3N0ZXZhbi9Eb2N1bWVudHMvQ2VudHJhbGUgTHlvbi8zQS9NYWNoaW5lIExlYXJuaW5nL2hhbmRzLW9uLXJsLy52ZW52L2xpYi9weXRob24zLjEzL3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUaACMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGgjfZR9lChoGowIPGxhbWJkYT6UjAxfX3F1YWxuYW1lX1+UjCFnZXRfc2NoZWR1bGVfZm4uPGxvY2Fscz4uPGxhbWJkYT6UjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgbjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOUaAIoaAcoSwFLAEsASwFLAUsTQwg+AZUAVAEkAJRoCSmMAV+UhZRoDowEZnVuY5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUS4ZDCPiAANgPEogKlGgSjAN2YWyUhZQpdJRSlGgXTk5oHylSlIWUdJRSlGglaEF9lH2UKGgajARmdW5jlGgpjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlGgrfZRoLU5oLk5oL2gbaDBOaDFoM0c/UGJN0vGp/IWUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwhZRSlIWUaEpdlGhMfZR1hpSGUjAu"}, "actor_batch_norm_stats": [], "critic_batch_norm_stats": [], "actor_batch_norm_stats_target": [], "critic_batch_norm_stats_target": [], "system_info": {"OS": "macOS-15.3-arm64-arm-64bit-Mach-O Darwin Kernel Version 24.3.0: Thu Jan 2 20:24:23 PST 2025; root:xnu-11215.81.4~3/RELEASE_ARM64_T8122", "Python": "3.13.1", "Stable-Baselines3": "2.5.0", "PyTorch": "2.6.0", "GPU Enabled": "False", "Numpy": "2.2.3", "Cloudpickle": "3.1.1", "Gymnasium": "1.0.0", "OpenAI Gym": "0.26.2"}} |