Positional Embedding in DeepSpeed Transformer Kernel #1692
Unanswered
sarvghotra
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
HuggingFace's
BertLayerhas an option to apply positional embedding in itsBertSelfAttention(here), but I couldn't find it in DeepSpeed's Transformer kernel (here). In other words, I was trying to find DeepSpeed's equivalent for Hugginface's position_embedding_type (link) argument in their BertConfig but had no luck. Could you please help with this?Context: I am trying to implement Swin Transformer by reusing DeepSpeed's Kernel code, but Swin has relative positional embegging in it which I couldn't find in the kernel's code.
Beta Was this translation helpful? Give feedback.
All reactions