8000 Ensure GPU isolation for kubernetes pod MI300 runners. by saienduri · Pull Request #145829 · pytorch/pytorch · GitHub
[go: up one dir, main page]

Skip to content

Ensure GPU isolation for kubernetes pod MI300 runners. #145829

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 1 commit into from

Conversation

saienduri
Copy link
Contributor
@saienduri saienduri commented Jan 28, 2025

Fixes the reason behind moving the tests to unstable initially. (#145790)
We ensure gpu isolation for each pod within kubernetes by propagating the drivers selected for the pod from the Kubernetes layer up to the docker run in pytorch here.
Now we stick with the GPUs assigned to the pod in the first place and there is no overlap between the test runners.

cc @jeffdaily @sunway513 @jithunnair-amd @pruthvistony @ROCmSupport @dllehr-amd @jataylo @hongxiayang @naromero77amd

@saienduri saienduri requested a review from a team as a code owner January 28, 2025 09:15
Copy link
pytorch-bot bot commented Jan 28, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/145829

Note: Links to docs will display an error until the docs builds have been completed.

⏳ No Failures, 6 Pending

As of commit 363aabf with merge base 5c5306e (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@pytorch-bot pytorch-bot bot added the topic: not user facing topic category label Jan 28, 2025
@facebook-github-bot facebook-github-bot added the module: rocm AMD GPU support for Pytorch label Jan 28, 2025
@jataylo jataylo added the ciflow/unstable Run all experimental or flaky jobs on PyTorch unstable workflow label Jan 28, 2025
@jataylo
Copy link
Collaborator
jataylo commented Jan 28, 2025

Rerunning the failed lint job.

@jeffdaily jeffdaily added the ciflow/rocm Trigger "default" config CI on ROCm label Jan 28, 2025
@pytorch-bot pytorch-bot bot temporarily deployed to upload-benchmark-results January 28, 2025 17:11 Inactive
@pytorch-bot pytorch-bot bot temporarily deployed to upload-benchmark-results January 28, 2025 17:11 Inactive
@pytorch-bot pytorch-bot bot temporarily deployed to upload-benchmark-results January 28, 2025 17:11 Inactive
@pytorch-bot pytorch-bot bot temporarily deployed to upload-benchmark-results January 28, 2025 17:11 Inactive
@pytorch-bot pytorch-bot bot temporarily deployed to upload-benchmark-results January 28, 2025 17:11 Inactive
@pytorch-bot pytorch-bot bot temporarily deployed to upload-benchmark-results January 28, 2025 17:11 Inactive
@jeffdaily
Copy link
Collaborator

@pytorchbot merge -f "workflow-only change. confirmed mi200 runners still work with this change, mi300 can only be tested post-merge"

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged immediately since you used the force (-f) flag, bypassing any CI checks (ETA: 1-5 minutes). Please use -f as last resort and instead consider -i/--ignore-current to continue the merge ignoring current failures. This will allow currently pending tests to finish and report signal before the merge.

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ciflow/rocm Trigger "default" config CI on ROCm ciflow/unstable Run all experimental or flaky jobs on PyTorch unstable workflow Merged module: rocm AMD GPU support for Pytorch open source rocm This tag is for PRs from ROCm team topic: not user facing topic category
Projects
None yet
Development

Successfully merging this pull request may close these issues.

7 participants
0