Reactive Control for Discrete-Action VLN Paper
Collection
2 items • Updated • 1
| Parameter | Value |
|---|---|
| n_rollout | 2048 |
| n_epochs | 4 |
| batch_size | 256 |
| lr | 0.0003 |
| gamma | 0.99 |
| gae_lambda | 0.95 |
| clip_eps | 0.2 |
| entropy_coef | 0.01 |
import torch
from ppo_policy import PPOPolicy
policy = PPOPolicy()
ckpt = torch.load("policy_update_XXXXX.pt", map_location="cpu")
policy.load_state_dict(ckpt["policy_state"])
policy.eval()
# obs: dict with keys depth (64,64), command (4,), proprioception (3,)
action, log_prob, entropy, value = policy.get_action_and_value(
depth.unsqueeze(0),
command.unsqueeze(0),
prop.unsqueeze(0),
)
Trained on wenavigatecontroller-long-episodes — HM3D minival scenes 00800–00809, 160 train episodes, 160 eval episodes.