You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
- Upgraded to new `HerReplayBuffer` implementation that supports multiple envs
50
+
- Removed `TimeFeatureWrapper` for Panda and Fetch envs, as the new replay buffer should handle timeout.
51
+
- Tuned hyperparameters for RecurrentPPO on Swimmer
52
+
- Documentation is now built using Sphinx and hosted on read the doc
53
+
- Removed `use_auth_token` for push to hub util
54
+
- Reverted from v3 to v2 for HumanoidStandup, Reacher, InvertedPendulum and InvertedDoublePendulum since they were not part of the mujoco refactoring (see https://github.com/openai/gym/pull/1304)
55
+
- Fixed `gym-minigrid` policy (from `MlpPolicy` to `MultiInputPolicy`)
56
+
- Replaced deprecated `optuna.suggest_loguniform(...)` by `optuna.suggest_float(..., log=True)`
57
+
- Switched to `ruff` and `pyproject.toml`
58
+
- Removed `online_sampling` and `max_episode_length` argument when using `HerReplayBuffer`
43
59
44
60
Bug Fixes:
45
61
^^^^^^^^^^
46
62
- Fixed Atari wrapper that missed the reset condition (@luizapozzobon)
47
63
- Added the argument ``dtype`` (default to ``float32``) to the noise for consistency with gym action (@sidney-tio)
48
64
- Fixed PPO train/n_updates metric not accounting for early stopping (@adamfrly)
49
65
- Fixed loading of normalized image-based environments
50
-
- Fixed `DictRolloutBuffer.add` with multidimensional action space (@younik)
66
+
- Fixed ``DictRolloutBuffer.add`` with multidimensional action space (@younik)
0 commit comments