8000 [CUDA][avgpool2d] Fix backward launch bounds again for `sm100`, `sm120` by eqy · Pull Request #150640 · pytorch/pytorch · GitHub
[go: up one dir, main page]

Skip to content

[CUDA][avgpool2d] Fix backward launch bounds again for sm100, sm120 #150640

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 1 commit into from

Conversation

eqy
Copy link
Collaborator
@eqy eqy commented Apr 3, 2025

__CUDA_ARCH__ is not visible in host code, which causes incorrect launch bounds and too many resources requested for launch on blackwell

CC @atalman @malfet as we would want this in 2.7 @nWEIdia

cc @ptrblck @msaroufim

@eqy eqy added module: cuda Related to torch.cuda, and CUDA support in general open source topic: bug fixes topic category topic: not user facing topic category labels Apr 3, 2025
@eqy eqy requested a review from syed-ahmed as a code owner April 3, 2025 21:54
Copy link
pytorch-bot bot commented Apr 3, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/150640

Note: Links to docs will display an error until the docs builds have been completed.

✅ You can merge normally! (2 Unrelated Failures)

As of commit 2ede186 with merge base 51da241 (image):

BROKEN TRUNK - The following job failed but were present on the merge base:

👉 Rebase onto the `viable/strict` branch to avoid these failures

UNSTABLE - The following job is marked as unstable, possibly due to flakiness on trunk:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@eqy eqy added the ciflow/trunk Trigger trunk jobs on your pull request label Apr 3, 2025
Copy link
Contributor
@malfet malfet left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@malfet
Copy link
Contributor
malfet commented Apr 3, 2025

@pytorchbot merge

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@pytorchmergebot
Copy link
Collaborator

Merge failed

Reason: 1 mandatory check(s) failed. The first few are:

Dig deeper by viewing the failures on hud

Details for Dev Infra team Raised by workflow job

Failing merge rule: Core Maintainers

Copy link
Contributor
@atalman atalman left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

lgtm

@atalman
Copy link
Contributor
atalman commented Apr 4, 2025

@pytorchmergebot merge -f "failure already existing, lint is green"

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged immediately since you used the force (-f) flag, bypassing any CI checks (ETA: 1-5 minutes). Please use -f as last resort and instead consider -i/--ignore-current to continue the merge ignoring current failures. This will allow currently pending tests to finish and report signal before the merge.

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@atalman
Copy link
Contributor
atalman commented Apr 4, 2025

@pytorchbot cherry-pick --onto release/2.7 -c critical

pytorchbot pushed a commit that referenced this pull request Apr 4, 2025
…0` (#150640)

`__CUDA_ARCH__` is not visible in host code, which causes incorrect launch bounds and `too many resources requested for launch` on blackwell

CC @atalman @malfet as we would want this in 2.7 @nWEIdia

Pull Request resolved: #150640
Approved by: https://github.com/malfet, https://github.com/drisspg, https://github.com/atalman

(cherry picked from commit 09c4da9)
@pytorchbot
Copy link
Collaborator

Cherry picking #150640

The cherry pick PR is at #150676 and it is recommended to link a critical cherry pick PR with an issue. The following tracker issues are updated:

Details for Dev Infra team Raised by workflow job

malfet pushed a commit that referenced this pull request Apr 4, 2025
…0` (#150676)

[CUDA][avgpool2d] Fix backward launch bounds again for `sm100`, `sm120` (#150640)

`__CUDA_ARCH__` is not visible in host code, which causes incorrect launch bounds and `too many resources requested for launch` on blackwell

Pull Request resolved: #150640
Approved by: https://github.com/malfet, https://github.com/drisspg, https://github.com/atalman

(cherry picked from commit 09c4da9)

Co-authored-by: Eddie Yan <eddiey@nvidia.com>
timocafe pushed a commit to timocafe/pytorch that referenced this pull request Apr 16, 2025
…0` (pytorch#150640)

`__CUDA_ARCH__` is not visible in host code, which causes incorrect launch bounds and `too many resources requested for launch` on blackwell

CC @atalman @malfet as we would want this in 2.7 @nWEIdia

Pull Request resolved: pytorch#150640
Approved by: https://github.com/malfet, https://github.com/drisspg, https://github.com/atalman
amathewc pushed a commit to amathewc/pytorch that referenced this pull request Apr 17, 2025
…0` (pytorch#150640)

`__CUDA_ARCH__` is not visible in host code, which causes incorrect launch bounds and `too many resources requested for launch` on blackwell

CC @atalman @malfet as we would want this in 2.7 @nWEIdia

Pull Request resolved: pytorch#150640
Approved by: https://github.com/malfet, https://github.com/drisspg, https://github.com/atalman
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ciflow/trunk Trigger trunk jobs on your pull request Merged module: cuda Related to torch.cuda, and CUDA support in general open source topic: bug fixes topic category topic: not user facing topic category
Projects
None yet
Development

Successfully merging this pull request may close these issues.

6 participants
0