Skip to content

feat(torch): Support compute capability 12.0 #99

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 8 commits into from
Jun 18, 2025
Merged

feat(torch): Support compute capability 12.0 #99

merged 8 commits into from
Jun 18, 2025

Conversation

Eta0
Copy link
Collaborator

@Eta0 Eta0 commented Jun 18, 2025

Compute Capability 12.0 + CUDA 12.9 Fixes

This change adds compute capability 12.0 (sm_120) to the list of compiled architectures for libraries in the ml-containers/torch and ml-containers/torch-extras images. It additionally fixes an issue with CUDA 12.9 support when compiling PyTorch extension modules by cherry-picking a patch from the main branch of the PyTorch repo into our torch build, and gets torch to use the system NVTX library when available (i.e. on CUDA 12.9).

xformers in ml-containers/torch-extras has also been updated to its v0.0.30 release, offering better compatibility with newer GPU architectures, PyTorch versions, and with vLLM.

Finally, this update drops compute capability 7.0 (e.g. V100), leaving 7.5 (e.g. Quadro RTX 4000/5000) as the lowest supported architecture.

@Eta0 Eta0 requested review from wbrown and sangstar June 18, 2025 19:34
@Eta0 Eta0 self-assigned this Jun 18, 2025
@Eta0 Eta0 added the enhancement New feature or request label Jun 18, 2025
Copy link
Contributor

@sangstar sangstar left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM!

@Eta0 Eta0 merged commit f359507 into main Jun 18, 2025
83 of 92 checks passed
@Eta0 Eta0 deleted the es/compute-12.0 branch June 18, 2025 21:00
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants