File tree Expand file tree Collapse file tree 1 file changed +1
-3
lines changed
test/distributed/tensor/experimental Expand file tree Collapse file tree 1 file changed +1
-3
lines changed Original file line number Diff line number Diff line change 1414)
1515from torch .distributed .tensor .debug import CommDebugMode
1616from torch .distributed .tensor .experimental import local_map
17+ from torch .testing ._internal .common_distributed import skip_if_lt_x_gpu
1718from torch .testing ._internal .common_utils import run_tests
1819from torch .testing ._internal .distributed ._tensor .common_dtensor import (
1920 DTensorTestBase ,
2021 with_comms ,
2122)
22- from torch .testing ._internal .common_distributed import skip_if_lt_x_gpu
2323
2424
2525funcol_py = torch .ops .c10d_functional
@@ -393,8 +393,6 @@ def test_multi_mesh_inputs(self):
393393 Test the function can be applied to accept DTensors that lives
394394 on different device meshes.
395395 """
396- import os
397- os .environ ["NCCL_DEBUG" ] = "WARN"
398396 mesh_full = init_device_mesh (
399397 device_type = self .device_type , mesh_shape = (self .world_size ,)
400398 )
You can’t perform that action at this time.
0 commit comments