Replies: 1 comment 1 reply
-
There is a discussion about how to add support for fusing operations: #5413 However, I don't think we have a good answer yet for the specific case of |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
I've asked this question in the GGML discussion board with no reply so decided to try the llama.cpp one
I'm working on a NPU backend and the NPU support 1D and 2D convolutions as a native functionality. AFAIK GGML implements convolution as a combination of im2col and matmul. I looked into the backend data structure and found graph_plan_create but it is documented to be not used right now.
Is there a place I can walk and rewrite the compute graph? So I can transform im2col + matmul into corresponding Conv1D and Conv2D operations.
Thanks
Beta Was this translation helpful? Give feedback.
All reactions