🎯
Focusing
AI Researcher Intern at @KaliberAI | PhD Student at Virginia Tech | Ex-MLE at @IQVIA
-
Virginia Tech
- Blacksburg, VA, USA
-
12:36
(UTC -12:00) - https://hasanar1f.github.io/
- @hasanar1f
- http://linkedin.com/in/kazi-hasan-ibn-arif-8b78a61a3
Highlights
- Pro
Pinned Loading
-
-
flash-attention-optimized
flash-attention-optimized PublicForked from tspeterkim/flash-attention-minimal
This repository aims to optimize the forward pass of the Flash Attention implementation in CUDA. This is a part of a graduate course project titled “Emerging Topics in CS: High-Performance Code Gen…
Cuda 1
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.