-
Notifications
You must be signed in to change notification settings - Fork 743
[NVIDIA] Fix build xformers >= cu129 (torch 2.9.0) #1344
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
@jiawenliu64 @bottler @sgrigory could you run and merge? |
|
just a little message to tell you that i believe that PR is needed hard for people on cuda 13, basically pytorch 2.9 is now the stable one, but the latest available xformers doesn't support cuda 13. so some software ( like comfyui ) using tensor related stuff can't work properly. ps: i'm not a dev at all, but i just tried for an entire day to use comfyui with TensorRT acceleration, and it's basically impossible with Cuda 13 drivers... ( did not managed to compile successfully this PR sadly ) |
You have to export cccl to new ones. For me it is working. |
could you try again? it should be fixed |
|
ping you again |
|
Note: When this branch is build inside of nvcr.io/nvidia/pytorch:25.09-py3 on DGX Spark - it is not working as TORCH_CUDA_ARCH_LIST in it has value "8.0 8.6 9.0 10.0 11.0 12.0+PTX" - +PTX probably breaks recognition of compute capabilities and build is executed for sm_80 and sm_90 to install it on DGX Spark from sources inside of recommended image (Nvidia recommends it instead of installing pytorch manually) unset TORCH_CUDA_ARCH_LIST or export TORCH_CUDA_ARCH_LIST=12.0 is needed |
|
Thanks for your PR! Let's check if the wheels build before merging - tests are running now :) |
|
Also probably we should update python to 3.10 in the CI so that the linter is able to run:
|
true, i miss it |
|
@danthe3rd i upgrade to minimum version 3.10 that is the minimum right now with torch 2.9.0 |
|
Thanks! Can you also look at the windows build? |
seems a bug in jimver action Jimver/cuda-toolkit#395 |
Switched from Jimver to N-Storm fork of cuda-toolkit.
|
Thanks! Looks like we're getting these errors in the CI now (related to PyTorch's CUDAExtension?):
The only supported archs in PyTorch are the following at the moment? |
Yes, i’ve seen that Blackwell Family, it is not compatible yet there. I’m going to change when i arrive at home(im come back from pytorch conference) |
|
Thanks! Let's see how it goes. I'm a bit worried we might hit a timeout on the CI with that many architectures (especially for the build of FA3) |
maybe we can filter for FA3? FA3 is only compatible with 80 and 90, fa4 only still with 100/103 |
|
Hum the FAv3 windows build for cuda 13 seems to be broken. Maybe we could disable FAv3 on windows/cuda13 for now?
https://github.com/facebookresearch/xformers/actions/runs/18913502491/job/53996546199?pr=1344 |
Yes, maybe we can disable it. I don’t have windows right now to test it |
|
I would say we just need to set Lines 282 to 283 in 51aa071
|
i tried to fix it, run CI if it works, if not, we can skip it |
|
@danthe3rd seems that fails again. I reported it internally. We can merge avoiding cu130 at this point |
|
Sure - let me merge if the CI is green :) |
Due action runner in windows is not updates
|
You can cancel the previous one. Now should works @danthe3rd |
Co-authored-by: dan_the_3rd <[email protected]>
|
Everything seems alright, we can fix the linters later on our side. |
What does this PR do?
Fixes #1320 #1308 #1323 #1335 and includes fixes for flash-attention >= CUDA 12.9 and adds cutlass v4.2.1 that fixes some kernels for Blackwell.
Also add support for Spark and Thor.
Added Blackwell family support. https://developer.nvidia.com/blog/nvidia-blackwell-and-nvidia-cuda-12-9-introduce-family-specific-architecture-features/
Thanks to: #1285 #1262 that are included here.
Fixes in flash-attention to support CUDA 13:
Pytorch 2.9.0 https://dev-discuss.pytorch.org/t/pytorch-2-9-rc1-produced-for-pytorch-audio-vision/3234
cc @sgrigory