[Bug] CUDA error: an illegal memory access was encountered on v.0.0.18
#406
Replies: 2 comments 1 reply
-
I'll try to see if I can get access to an L4 VM and maybe reproduce this, but I am a little swamped. Also a little confused if 0.0.17 works, since there shouldn't really be any changes to inference for 0.0.18. |
Beta Was this translation helpful? Give feedback.
1 reply
-
Unclear why, but reinstalling today worked.. Very weird. Couldn't find why. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hello, I have encountered this new CUDA error on the latest tag. I initially transitioned from
v0.0.12
(which worked) tov0.0.18
(which got this error).v0.0.17
from source worked.My env is a GCP L4 VM, cuda 11.8. I've made sure to use
torch==2.2.0
and clear torch extensions cache. I also tried changing VM gpus, updating dependencies, installing from wheels vs source.Full stack:
I do not see this function
safe_move_tensor
being changed itself. I tried thestreaming.py
example, and it usually runs, but it also sometimes throw this error.The code that generates this issue can be trimmed to:
Beta Was this translation helpful? Give feedback.
All reactions