Skip to content

Commit

Permalink
Fixed textual output of example scripts
Browse files Browse the repository at this point in the history
  • Loading branch information
boris-il-forte committed Feb 5, 2025
1 parent de6f703 commit b2fab67
Show file tree
Hide file tree
Showing 7 changed files with 14 additions and 14 deletions.
4 changes: 2 additions & 2 deletions examples/habitat/habitat_rearrange_sac.py
Original file line number Diff line number Diff line change
Expand Up @@ -150,7 +150,7 @@ def experiment(alg, n_epochs, n_steps, n_episodes_test):

J = np.mean(dataset.discounted_return)
R = np.mean(dataset.undiscounted_return)
E = agent.policy.entropy(dataset.state)
E = agent.policy.entropy(dataset.state).item()

logger.epoch_info(0, J=J, R=R, entropy=E)

Expand All @@ -162,7 +162,7 @@ def experiment(alg, n_epochs, n_steps, n_episodes_test):

J = np.mean(dataset.discounted_return)
R = np.mean(dataset.undiscounted_return)
E = agent.policy.entropy(dataset.state)
E = agent.policy.entropy(dataset.state).item()

logger.epoch_info(n+1, J=J, R=R, entropy=E)

Expand Down
4 changes: 2 additions & 2 deletions examples/isaac_example.py
Original file line number Diff line number Diff line change
Expand Up @@ -81,7 +81,7 @@ def experiment(cfg_dict, headless, alg, n_epochs, n_steps, n_steps_per_fit, n_ep

J = torch.mean(dataset.discounted_return)
R = torch.mean(dataset.undiscounted_return)
E = agent.policy.entropy()
E = agent.policy.entropy().item()

logger.epoch_info(0, J=J, R=R, entropy=E)

Expand All @@ -91,7 +91,7 @@ def experiment(cfg_dict, headless, alg, n_epochs, n_steps, n_steps_per_fit, n_ep

J = torch.mean(dataset.discounted_return)
R = torch.mean(dataset.undiscounted_return)
E = agent.policy.entropy()
E = agent.policy.entropy().item()

logger.epoch_info(it+1, J=J, R=R, entropy=E)

Expand Down
4 changes: 2 additions & 2 deletions examples/mujoco_locomotion_ppo.py
Original file line number Diff line number Diff line change
Expand Up @@ -97,7 +97,7 @@ def experiment(env, n_epochs, n_steps, n_episodes_test):

J = np.mean(dataset.discounted_return)
R = np.mean(dataset.undiscounted_return)
E = agent.policy.entropy()
E = agent.policy.entropy().item()

logger.epoch_info(0, J=J, R=R, entropy=E)

Expand All @@ -107,7 +107,7 @@ def experiment(env, n_epochs, n_steps, n_episodes_test):

J = np.mean(dataset.discounted_return)
R = np.mean(dataset.undiscounted_return)
E = agent.policy.entropy()
E = agent.policy.entropy().item()

logger.epoch_info(it + 1, J=J, R=R, entropy=E)

Expand Down
4 changes: 2 additions & 2 deletions examples/pendulum_a2c.py
Original file line number Diff line number Diff line change
Expand Up @@ -73,7 +73,7 @@ def experiment(alg, env_id, horizon, gamma, n_epochs, n_steps, n_steps_per_fit,

J = np.mean(dataset.discounted_return)
R = np.mean(dataset.undiscounted_return)
E = agent.policy.entropy()
E = agent.policy.entropy().item()

logger.epoch_info(0, J=J, R=R, entropy=E)

Expand All @@ -83,7 +83,7 @@ def experiment(alg, env_id, horizon, gamma, n_epochs, n_steps, n_steps_per_fit,

J = np.mean(dataset.discounted_return)
R = np.mean(dataset.undiscounted_return)
E = agent.policy.entropy()
E = agent.policy.entropy().item()

logger.epoch_info(it+1, J=J, R=R, entropy=E)

Expand Down
4 changes: 2 additions & 2 deletions examples/pendulum_sac.py
Original file line number Diff line number Diff line change
Expand Up @@ -127,7 +127,7 @@ def experiment(alg, n_epochs, n_steps, n_steps_test, save, load):

J = np.mean(dataset.discounted_return)
R = np.mean(dataset.undiscounted_return)
E = agent.policy.entropy(dataset.state)
E = agent.policy.entropy(dataset.state).item()

logger.epoch_info(0, J=J, R=R, entropy=E)

Expand All @@ -139,7 +139,7 @@ def experiment(alg, n_epochs, n_steps, n_steps_test, save, load):

J = np.mean(dataset.discounted_return)
R = np.mean(dataset.undiscounted_return)
E = agent.policy.entropy(dataset.state)
E = agent.policy.entropy(dataset.state).item()

logger.epoch_info(n+1, J=J, R=R, entropy=E)

Expand Down
4 changes: 2 additions & 2 deletions examples/pendulum_trust_region.py
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,7 @@ def experiment(alg, env_id, horizon, gamma, n_epochs, n_steps, n_steps_per_fit,

J = np.mean(dataset.discounted_return)
R = np.mean(dataset.undiscounted_return)
E = agent.policy.entropy()
E = agent.policy.entropy().item()

logger.epoch_info(0, J=J, R=R, entropy=E)

Expand All @@ -80,7 +80,7 @@ def experiment(alg, env_id, horizon, gamma, n_epochs, n_steps, n_steps_per_fit,

J = np.mean(dataset.discounted_return)
R = np.mean(dataset.undiscounted_return)
E = agent.policy.entropy()
E = agent.policy.entropy().item()

logger.epoch_info(it+1, J=J, R=R, entropy=E)

Expand Down
4 changes: 2 additions & 2 deletions examples/vectorized_core/pendulum_trust_region.py
Original file line number Diff line number Diff line change
Expand Up @@ -73,7 +73,7 @@ def experiment(alg, env_id, horizon, gamma, n_epochs, n_steps, n_steps_per_fit,

J = np.mean(dataset.discounted_return)
R = np.mean(dataset.undiscounted_return)
E = agent.policy.entropy()
E = agent.policy.entropy().item()

logger.epoch_info(0, J=J, R=R, entropy=E)

Expand All @@ -83,7 +83,7 @@ def experiment(alg, env_id, horizon, gamma, n_epochs, n_steps, n_steps_per_fit,

J = np.mean(dataset.discounted_return)
R = np.mean(dataset.undiscounted_return)
E = agent.policy.entropy()
E = agent.policy.entropy().item()

logger.epoch_info(it+1, J=J, R=R, entropy=E)

Expand Down

0 comments on commit b2fab67

Please sign in to comment.