Newer
Older
env_id: "overcooked"
policy_type: "MlpPolicy"
model_type: "PPO"
total_timesteps: 3_000_000 # hendric sagt eher so 300_000_000 schritte
number_envs_parallel: 64
learning_rate: 0.0003
n_steps: 2048
batch_size: 64
n_epochs: 10
gamma: 0.99
gae_lambda: 0.95
clip_range: 0.2
clip_range_vf: None
normalize_advantage: True
ent_coef: 0.0
vf_coef: 0.5
max_grad_norm: 0.5
use_sde: False
sde_sample_freq: -1
rollout_buffer_class: None
rollout_buffer_kwargs: None
target_kl: None
stats_window_size: 100
tensorboard_log: None
policy_kwargs: None
verbose: 0
seed: None
device: 'auto'
_init_setup_model: True