Commit 59d14f5b authored by Dipam Chakraborty's avatar Dipam Chakraborty
Browse files

multi optim

parent fc7b6895
...@@ -45,13 +45,13 @@ procgen-ppo: ...@@ -45,13 +45,13 @@ procgen-ppo:
no_done_at_end: False no_done_at_end: False
# Custom switches # Custom switches
skips: 6 skips: 2
n_pi: 10 n_pi: 16
num_retunes: 16 num_retunes: 14
retune_epochs: 6 retune_epochs: 7
standardize_rewards: True standardize_rewards: True
aux_mbsize: 2 aux_mbsize: 4
aux_num_accumulates: 4 aux_num_accumulates: 2
augment_buffer: True augment_buffer: True
scale_reward: 1.0 scale_reward: 1.0
reset_returns: False reset_returns: False
...@@ -61,7 +61,7 @@ procgen-ppo: ...@@ -61,7 +61,7 @@ procgen-ppo:
value_lr: 1.0e-3 value_lr: 1.0e-3
same_lr_everywhere: False same_lr_everywhere: False
aux_phase_mixed_precision: True aux_phase_mixed_precision: True
single_optimizer: True single_optimizer: False
max_time: 7200 max_time: 7200
pi_phase_mixed_precision: False pi_phase_mixed_precision: False
...@@ -72,7 +72,7 @@ procgen-ppo: ...@@ -72,7 +72,7 @@ procgen-ppo:
entropy_schedule: False entropy_schedule: False
# Memory management, if batch size overflow, batch splitting is done to handle it # Memory management, if batch size overflow, batch splitting is done to handle it
max_minibatch_size: 500 max_minibatch_size: 1500
updates_per_batch: 8 updates_per_batch: 8
normalize_actions: False normalize_actions: False
...@@ -89,10 +89,8 @@ procgen-ppo: ...@@ -89,10 +89,8 @@ procgen-ppo:
model: model:
custom_model: impala_torch_ppg custom_model: impala_torch_ppg
custom_model_config: custom_model_config:
# depths: [32, 64, 64] depths: [32, 64, 64]
# nlatents: 512 nlatents: 512
depths: [64, 128, 128]
nlatents: 1024
init_normed: True init_normed: True
use_layernorm: False use_layernorm: False
diff_framestack: True diff_framestack: True
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment