-
Notifications
You must be signed in to change notification settings - Fork 24.8k
Fix DLPack stream logic. #150217
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix DLPack stream logic. #150217
Changes from 9 commits
1eadb0e
cf810ee
c101dbf
23422b3
56e2c32
3c4f7da
b2de390
d32a334
476b1ea
3338ea1
d121ce9
c2e1184
089f42f
57e26fd
5ffb2ee
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -1717,23 +1717,37 @@ def __dlpack__(self, stream=None, max_version=None): | |
# Stream pointers in CUDA/ROCm are uniquely numbered and can | ||
# be retrieved from their integer value. | ||
raise TypeError("stream must be ``int`` or ``none``") | ||
elif stream is not None and stream != -1: | ||
elif stream != -1: | ||
if self.device.type == "cuda": | ||
# NB: This logic handles the special case values for default | ||
# streams and must be kept in sync with from_dlpack in | ||
# torch/utils/dlpack.py | ||
if stream == 1 and torch.version.hip is None: | ||
stream = torch.cuda.default_stream() | ||
elif stream == 0 and torch.version.hip is not None: | ||
is_rocm = torch.version.hip is not None | ||
is_cuda = not is_rocm | ||
|
||
if ( | ||
stream is None | ||
or (is_rocm and stream == 0) | ||
or (is_cuda and stream == 1) | ||
): | ||
stream = torch.cuda.default_stream() | ||
else: | ||
if is_cuda and stream == 2: | ||
raise BufferError("per-thread default stream is not supported.") | ||
|
||
assert is_cuda or (is_rocm and stream not in (1, 2)), ( | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Shouldn't this be a BufferError like above instead of AssertionError? There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I don't think so. The reason being that this assertion checks something the standard explicitly states as "unsupported" or "disallowed", i.e. something the consumer should know about. Moreover, the standard also says that:
|
||
f"unsupported stream {stream} for ROCm." | ||
) | ||
|
||
stream = torch.cuda.ExternalStream(stream) | ||
|
||
# Only synchronize on different streams | ||
sync_stream = torch.cuda.current_stream() | ||
if stream != sync_stream: | ||
current_stream = torch.cuda.current_stream() | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Do we care if There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Good point. I think we should. I will add a check for that. |
||
if stream != current_stream: | ||
event = torch.cuda.Event() | ||
event.record(sync_stream) | ||
event.record(current_stream) | ||
stream.wait_event(event) | ||
|
||
if self.device.type == "xla": | ||
import torch_xla | ||
import torch_xla.utils.dlpack as xla_dlpack | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
No update to dlpack.py ? :D
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
No need. If
stream
is None, we still need to synchronize, assuming the legacy default stream.