-
Notifications
You must be signed in to change notification settings - Fork 24.3k
test_scatter_bf16_cuda
fails on V100
#118581
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Probably related to this #118122. |
@malfet, still working on this ? |
Comments from triage meeting
|
We're doing the weekly check-in on hi-pri issues that haven't been updated in a month. @malfet, any update? |
This sounds like a duplicate of #118122 |
Volta(sm_7x) do not have a HW support for bfloat16 datatype, and while it is is emulated to ted in software, so PyTorch eager can use bfloat16 tensors, but not in Triton. So if graph with either CUDA bf16 input or output tensors is used, raise warnings and skip the frame. Add optional parameter `including_emulation` to `torch.cuda.is_bf16_supported` method and call it from `torch._inductor.compile_fx. _check_triton_bf16_support`. Test plan: Modify `is_bf16_supported` to return False and see that warning is generated Fixes #118122 and #118581 Pull Request resolved: #129288 Approved by: https://github.com/eqy, https://github.com/jansel
Closing, were fixed by skipping compile on Volta's for bf16 dtype |
Volta(sm_7x) do not have a HW support for bfloat16 datatype, and while it is is emulated to ted in software, so PyTorch eager can use bfloat16 tensors, but not in Triton. So if graph with either CUDA bf16 input or output tensors is used, raise warnings and skip the frame. Add optional parameter `including_emulation` to `torch.cuda.is_bf16_supported` method and call it from `torch._inductor.compile_fx. _check_triton_bf16_support`. Test plan: Modify `is_bf16_supported` to return False and see that warning is generated Fixes #118122 and #118581 Pull Request resolved: #129288 Approved by: https://github.com/eqy, https://github.com/jansel (cherry picked from commit 14dc08d)
Inductor to fail gracefully on Voltas for bf16 tensors (#129288) Volta(sm_7x) do not have a HW support for bfloat16 datatype, and while it is is emulated to ted in software, so PyTorch eager can use bfloat16 tensors, but not in Triton. So if graph with either CUDA bf16 input or output tensors is used, raise warnings and skip the frame. Add optional parameter `including_emulation` to `torch.cuda.is_bf16_supported` method and call it from `torch._inductor.compile_fx. _check_triton_bf16_support`. Test plan: Modify `is_bf16_supported` to return False and see that warning is generated Fixes #118122 and #118581 Pull Request resolved: #129288 Approved by: https://github.com/eqy, https://github.com/jansel (cherry picked from commit 14dc08d) Co-authored-by: Nikita Shulga <nshulga@meta.com>
Validated in Colab:
|
same error |
Uh oh!
There was an error while loading. Please reload this page.
🐛 Describe the bug
While running inductor CI on V100, I've found that above-mentioned test fails with unsupported PTX instructions:
Versions
2.2, nightly
cc @ezyang @gchanan @zou3519 @kadeng @msaroufim @bdhirsh @anijain2305 @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @peterbell10 @ipiszy @yf225 @chenyang78 @muchulee8 @aakhundov @ColinPeppler
The text was updated successfully, but these errors were encountered: