You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+15-3Lines changed: 15 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,10 +1,10 @@
1
1
# Adaptive Transformers in RL
2
2
3
-
In this experiment we replicate several results from [Stabilizing Transformers for RL](https://arxiv.org/abs/1910.06764) on both [Pong](https://gym.openai.com/envs/Pong-v0/) and [rooms_select_nonmatching_object](https://github.com/deepmind/lab/tree/master/game_scripts/levels/contributed/dmlab30#select-non-matching-object) from DMLab30.
3
+
Official implementation of [Adaptive Transformers in RL](http://arxiv.org/abs/2004.03761)
4
4
5
-
We also extend the Stable Transformer architecture with [Adaptive Attention Span](https://arxiv.org/abs/1905.07799) on a partially observable (POMDP) setting of Reinforcement Learning. To our knowledge this is one of the first attempts to stabilize and explore Adaptive Attention Span in an RL domain.
5
+
In this work we replicate several results from [Stabilizing Transformers for RL](https://arxiv.org/abs/1910.06764) on both [Pong](https://gym.openai.com/envs/Pong-v0/) and [rooms_select_nonmatching_object](https://github.com/deepmind/lab/tree/master/game_scripts/levels/contributed/dmlab30#select-non-matching-object) from DMLab30.
6
6
7
-
The arxiv preprint for this work can be found here [Adaptive Transformers in RL](http://arxiv.org/abs/2004.03761)
7
+
We also extend the Stable Transformer architecture with [Adaptive Attention Span](https://arxiv.org/abs/1905.07799) on a partially observable (POMDP) setting of Reinforcement Learning. To our knowledge this is one of the first attempts to stabilize and explore Adaptive Attention Span in an RL domain.
8
8
9
9
### Steps to replicate what we did on your own machine
0 commit comments