jfelgate's picture
Initial commit
89ccb8c
raw
history blame contribute delete
572 Bytes
!!python/object/apply:collections.OrderedDict
- - - batch_size
- 128
- - buffer_size
- 10000
- - env_wrapper
- - stable_baselines3.common.atari_wrappers.AtariWrapper
- - exploration_final_eps
- 0.1
- - exploration_fraction
- 0.12
- - gamma
- 0.99
- - gradient_steps
- 1
- - learning_rate
- 0.001
- - learning_starts
- 0
- - n_timesteps
- 100000.0
- - policy
- MlpPolicy
- - policy_kwargs
- dict(net_arch=[256, 256])
- - target_update_interval
- 250
- - tau
- 0.001
- - train_freq
- 4