-
Couldn't load subscription status.
- Fork 415
Open
Labels
Description
Hi
As shown in the figure, during the decoding phase, the 2:4 sparsity model is about 12% slower than the dense model, the questions are as follows:
- Is the decode phase dominated by GEMV / small‑N GEMM operations, and does cuSPARSELt not support SpMV operations?
- Are there any sparse kernels or recommended practices for GEMV (matrix‑vector) that can take advantage of 2:4 sparsity?
Thanks!
