-
Notifications
You must be signed in to change notification settings - Fork 24.3k
Enable CUDA 12.8.0, Disable CUDA 12.4 #145570
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
this will be amazing for test new blackwell. |
@tinglvv new cudnn and tensorrt is out |
Can we get 12.8 support and 2.7 in nightly? |
They are working on that: |
|
@ofirkris it will be available in the following days |
Breaking #145557 into two parts. Need to have manylinux-cuda12.8 in order to build magma. Issue: #145570 Pull Request resolved: #145567 Approved by: https://github.com/nWEIdia, https://github.com/atalman
Breaking pytorch#145557 into two parts. Need to have manylinux-cuda12.8 in order to build magma. Issue: pytorch#145570 Pull Request resolved: pytorch#145567 Approved by: https://github.com/nWEIdia, https://github.com/atalman
#145570 Builds 12.8 libtorch docker/deprecate 12.1 meanwhile Pull Request resolved: #145789 Approved by: https://github.com/nWEIdia, https://github.com/atalman
#145570 Pull Request resolved: #145765 Approved by: https://github.com/malfet
#145570 Build failure for libtorch wheel `CUDAContext.cpp:(.text+0x157): additional relocation overflows omitted from the output /usr/bin/ld: failed to convert GOTPCREL relocation; relink with --no-relax collect2: error: ld returned 1 exit status` Unsure if this is related, fixing as a start Pull Request resolved: #146019 Approved by: https://github.com/eqy
#145570 Adding cuda 12.8.0 x86 builds first Pull Request resolved: #145792 Approved by: https://github.com/nWEIdia, https://github.com/malfet, https://github.com/atalman
waiting for the windows build as well :) |
Nightly manywheel torch builds are available https://download.pytorch.org/whl/nightly/cu128/torch |
#145570 removing `.ci/pytorch/windows/internal/cuda_install.bat` as it is a duplicate with` .github/scripts/windows/cuda_install.bat`. The later one is the one in use - https://github.com/pytorch/pytorch/pull/146653/files#diff-613791f266f2f7b81148ca8f447b0cd6c6544f824f5f46a78a2794006c78957bR8 Pull Request resolved: #146653 Approved by: https://github.com/atalman
#145570 removing `.ci/pytorch/windows/internal/cuda_install.bat` as it is a duplicate with` .github/scripts/windows/cuda_install.bat`. The later one is the one in use - https://github.com/pytorch/pytorch/pull/146653/files#diff-613791f266f2f7b81148ca8f447b0cd6c6544f824f5f46a78a2794006c78957bR8 Pull Request resolved: #146653 Approved by: https://github.com/atalman Co-authored-by: atalman <atalman@fb.com>
#145570 removing `.ci/pytorch/windows/internal/cuda_install.bat` as it is a duplicate with` .github/scripts/windows/cuda_install.bat`. The later one is the one in use - https://github.com/pytorch/pytorch/pull/146653/files#diff-613791f266f2f7b81148ca8f447b0cd6c6544f824f5f46a78a2794006c78957bR8 Pull Request resolved: #146653 Approved by: https://github.com/atalman Co-authored-by: atalman <atalman@fb.com>
#145570 windows AMI is deployed to prod today, prepping the windows cuda 12.8 build Pull Request resolved: #147037 Approved by: https://github.com/atalman
Try removing sm50 and sm60 to shrink binary size, and resolve the ld --relink error "Architecture support for Maxwell, Pascal, and Volta is considered feature-complete and will be frozen in an upcoming release." from 12.8 release note. Also updating the runner for cuda 12.8 test to g4dn (T4, sm75) due to the drop in sm50/60 support. #145570 Pull Request resolved: #146265 Approved by: https://github.com/atalman
I really think we should try to keep the lower SM arches in the 12.8 build while they are supported. We just need to fix the |
#145570 windows AMI is deployed to prod today, prepping the windows cuda 12.8 build Pull Request resolved: #147037 Approved by: https://github.com/atalman
Try removing sm50 and sm60 to shrink binary size, and resolve the ld --relink error "Architecture support for Maxwell, Pascal, and Volta is considered feature-complete and will be frozen in an upcoming release." from 12.8 release note. Also updating the runner for cuda 12.8 test to g4dn (T4, sm75) due to the drop in sm50/60 support. #145570 Pull Request resolved: #146265 Approved by: https://github.com/atalman
when release the version for cu128? |
pytorch/pytorch#145570 deprecate cuda 12.4 nightly build cc @atalman
Please be patient and kind |
pytorch#145570 removing `.ci/pytorch/windows/internal/cuda_install.bat` as it is a duplicate with` .github/scripts/windows/cuda_install.bat`. The later one is the one in use - https://github.com/pytorch/pytorch/pull/146653/files#diff-613791f266f2f7b81148ca8f447b0cd6c6544f824f5f46a78a2794006c78957bR8 Pull Request resolved: pytorch#146653 Approved by: https://github.com/atalman Co-authored-by: atalman <atalman@fb.com>
pytorch#145570 removing `.ci/pytorch/windows/internal/cuda_install.bat` as it is a duplicate with` .github/scripts/windows/cuda_install.bat`. The later one is the one in use - https://github.com/pytorch/pytorch/pull/146653/files#diff-613791f266f2f7b81148ca8f447b0cd6c6544f824f5f46a78a2794006c78957bR8 Pull Request resolved: pytorch#146653 Approved by: https://github.com/atalman Co-authored-by: atalman <atalman@fb.com>
pytorch#145570 windows AMI is deployed to prod today, prepping the windows cuda 12.8 build Pull Request resolved: pytorch#147037 Approved by: https://github.com/atalman
Try removing sm50 and sm60 to shrink binary size, and resolve the ld --relink error "Architecture support for Maxwell, Pascal, and Volta is considered feature-complete and will be frozen in an upcoming release." from 12.8 release note. Also updating the runner for cuda 12.8 test to g4dn (T4, sm75) due to the drop in sm50/60 support. pytorch#145570 Pull Request resolved: pytorch#146265 Approved by: https://github.com/atalman
follow up for https://github.com/pytorch/pytorch/pull/146265/files, dropping sm_70 as well, since "Architecture support for Maxwell, Pascal, and Volta is considered feature-complete and will be frozen in an upcoming release." #145570 Pull Request resolved: #147607 Approved by: https://github.com/atalman
) #145570 breaking #140793 into eager and inductor benchmarks to unblock Pull Request resolved: #148602 Approved by: https://github.com/atalman, https://github.com/malfet Co-authored-by: atalman <atalman@fb.com>
#145570 breaking #140793 into eager and inductor benchmarks to unblock Seems many inductor yml are added after initial change was prepared. Pull Request resolved: #148612 Approved by: https://github.com/nWEIdia, https://github.com/atalman Co-authored-by: atalman <atalman@fb.com>
#145570 removes cuda 12.4 nightly builds Pull Request resolved: #148625 Approved by: https://github.com/atalman
Cudnn 9.8.0 just dropped: https://pypi.org/project/nvidia-cudnn-cu12/ Might be good to include it in this release given all the Blackwell performance optimizations that were made to it, especially regarding SDPA and given FlashAttention4 is still a ways off. |
#145570 redo #148625 Pull Request resolved: #148895 Approved by: https://github.com/atalman
Uh oh!
There was an error while loading. Please reload this page.
🚀 The feature, motivation and pitch
CUDA 12.8.0 is out, adding to CI/CD.
Docker Images & Windows AMI Update
CD Update
CUDA 12.4 deprecation and CUDA 12.6 CI benchmarks
Alternatives
No response
Additional context
No response
cc @atalman @malfet @ptrblck @msaroufim @eqy @nWEIdia
The text was updated successfully, but these errors were encountered: