ppo-bin-picking-v2 / config.yml
Quentin Gallouédec
Initial commit
ae1c614
raw
history blame contribute delete
288 Bytes
!!python/object/apply:collections.OrderedDict
- - - batch_size
- 32
- - gamma
- 0.99
- - learning_rate
- 0.0005
- - n_envs
- 4
- - n_steps
- 512
- - n_timesteps
- 1000000
- - normalize
- true
- - policy
- MlpPolicy
- - target_kl
- 0.04