Commit fc7b6895 authored by Dipam Chakraborty's avatar Dipam Chakraborty
Browse files

ppg 4x

parent 429dfb5d
...@@ -45,13 +45,13 @@ procgen-ppo: ...@@ -45,13 +45,13 @@ procgen-ppo:
no_done_at_end: False no_done_at_end: False
# Custom switches # Custom switches
skips: 0 skips: 6
n_pi: 16 n_pi: 10
num_retunes: 16 num_retunes: 16
retune_epochs: 6 retune_epochs: 6
standardize_rewards: True standardize_rewards: True
aux_mbsize: 4 aux_mbsize: 2
aux_num_accumulates: 2 aux_num_accumulates: 4
augment_buffer: True augment_buffer: True
scale_reward: 1.0 scale_reward: 1.0
reset_returns: False reset_returns: False
...@@ -72,7 +72,7 @@ procgen-ppo: ...@@ -72,7 +72,7 @@ procgen-ppo:
entropy_schedule: False entropy_schedule: False
# Memory management, if batch size overflow, batch splitting is done to handle it # Memory management, if batch size overflow, batch splitting is done to handle it
max_minibatch_size: 1000 max_minibatch_size: 500
updates_per_batch: 8 updates_per_batch: 8
normalize_actions: False normalize_actions: False
...@@ -89,10 +89,10 @@ procgen-ppo: ...@@ -89,10 +89,10 @@ procgen-ppo:
model: model:
custom_model: impala_torch_ppg custom_model: impala_torch_ppg
custom_model_config: custom_model_config:
depths: [32, 64, 64] # depths: [32, 64, 64]
nlatents: 512 # nlatents: 512
# depths: [64, 128, 128] depths: [64, 128, 128]
# nlatents: 1024 nlatents: 1024
init_normed: True init_normed: True
use_layernorm: False use_layernorm: False
diff_framestack: True diff_framestack: True
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment