Questions about Transconv's attention design #7409
HelloWorldLTY
started this conversation in
General
Replies: 2 comments
-
I would suggest running the script with PyTorch profiler to see which part of the execution is taking so long: |
Beta Was this translation helpful? Give feedback.
0 replies
-
Also answered you on Slack. I think @akihironitta is right. From looking at the code, I am not sure your formula is correct. It looks like you will compute attention over all edges rather than local neighborhoods. I am also not sure |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi, I am trying to replace the transformconv's attention manchism with current fast attention design. However, the speed becomes slower, and my output is quite different.
The original code is:
My design is:
Beta Was this translation helpful? Give feedback.
All reactions