text stringlengths 0 93.6k |
|---|
os.environ["WANDB_BASE_URL"] = "https://api.fairwandb.ai" |
os.environ["WANDB_API_KEY"] = "092a14187f6f01d8d2df67e8145ed4b16ba8bc9d" |
num_levels = 1 |
level_sampler_args = dict( |
num_actors=args.num_processes, |
strategy=args.level_replay_strategy, |
) |
envs, level_sampler = make_dqn_lr_venv( |
num_envs=args.num_processes, |
env_name=args.env_name, |
seeds=seeds, |
device=args.device, |
num_levels=num_levels, |
start_level=args.start_level, |
no_ret_normalization=args.no_ret_normalization, |
distribution_mode=args.distribution_mode, |
paint_vel_info=args.paint_vel_info, |
use_sequential_levels=args.use_sequential_levels, |
level_sampler_args=level_sampler_args, |
attach_task_id=args.attach_task_id |
) |
if args.atc: |
args.drq = True |
agent = ATCAgent(args, envs) |
else: |
agent = DQNAgent(args, envs) |
replay_buffer = make_buffer(args, envs) |
level_seeds = torch.zeros(args.num_processes) |
if level_sampler: |
state, level_seeds = envs.reset() |
else: |
state = envs.reset() |
level_seeds = level_seeds.unsqueeze(-1) |
if args.autodrq: |
rollouts = RolloutStorage(256, args.num_processes, envs.observation_space.shape, envs.action_space) |
rollouts.obs[0].copy_(state) |
rollouts.to(args.device) |
estimates = [0 for _ in range(args.num_train_seeds)] |
returns = [0 for _ in range(args.num_train_seeds)] |
gaps = [0 for _ in range(args.num_train_seeds)] |
episode_reward = 0 |
state_deque: List[deque] = [deque(maxlen=args.multi_step) for _ in range(args.num_processes)] |
reward_deque: List[deque] = [deque(maxlen=args.multi_step) for _ in range(args.num_processes)] |
action_deque: List[deque] = [deque(maxlen=args.multi_step) for _ in range(args.num_processes)] |
expect_new_seed: List[bool] = [False for _ in range(args.num_processes)] |
reward_stats_deque: List[deque] = [deque(maxlen=500) for _ in range(args.num_processes)] |
num_steps = int(args.T_max // args.num_processes) |
epsilon_start = 1.0 |
epsilon_final = args.end_eps |
epsilon_decay = args.eps_decay_period |
def epsilon(t): |
return epsilon_final + (epsilon_start - epsilon_final) * np.exp( |
-1.0 * (t - args.start_timesteps) / epsilon_decay |
) |
start_time = time.time() |
curr_index = 0 |
#### Log uniform parameters #### |
loguniform_decay = args.ucb_c * args.diff_eps_schedule_base ** ( |
1 + np.arange(args.num_processes)/(args.num_processes-1) * args.diff_eps_schedule_exp) |
loguniform_decay = torch.from_numpy(loguniform_decay).to(args.device).unsqueeze(1) |
#### epsilon-z parameters #### |
n = np.zeros(args.num_processes) |
omega = np.zeros(args.num_processes) |
ez_prob = 1 / np.arange(1, args.eps_z_n+1)**args.eps_z_mu |
ez_prob /= np.sum(ez_prob) |
ez_n = np.arange(1, args.eps_z_n+1) |
for t in range(num_steps): |
if t < args.start_timesteps: |
action = ( |
torch.LongTensor([envs.action_space.sample() for _ in range(args.num_processes)]) |
.reshape(-1, 1) |
.to(args.device) |
) |
value = agent.get_value(state) |
elif args.explore_strat == "qrdqn_ucb": |
_, mean, var, upper_var = agent.get_quantile(state) |
decay = args.ucb_c * np.sqrt(np.log(t+1) / (t+1)) |
value = mean + decay * var |
# print(value.shape) |
action = value.argmax(1).reshape(-1, 1) |
# print(torch.max(mean, 1)) |
if t % 500 == 0: |
stats = { |
"ucb / facotr": decay, |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.