You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,7 +4,7 @@ In this experiment we replicate several results from [Stabilizing Transformers f
4
4
5
5
We also extend the Stable Transformer architecture with [Adaptive Attention Span](https://arxiv.org/abs/1905.07799) on a partially observable (POMDP) setting of Reinforcement Learning. To our knowledge this is one of the first attempts to stabilize and explore Adaptive Attention Span in an RL domain.
6
6
7
-
The paper for this work will be attached here once it is public on arxiv.
7
+
The arxiv preprint for this work can be found here [Adaptive Transformers in RL](http://arxiv.org/abs/2004.03761)
8
8
9
9
### Steps to replicate what we did on your own machine
0 commit comments