@@ -12622,7 +12622,10 @@ def error_inputs_mean(op_info, device, **kwargs):
12622
12622
decorators=[
12623
12623
skipCUDAIf(_get_torch_cuda_version() < (11, 4), "not available before CUDA 11.3.1"),
12624
12624
skipCUDAIfNoCusolver, skipCUDAIfRocm, skipCPUIfNoLapack],
12625
- ),
12625
+ skips=(
12626
+ DecorateInfo(unittest.skip("memory hog"), 'TestCommon', 'test_out_warning'),
12627
+ DecorateInfo(unittest.skip("memory hog"), 'TestJit', 'test_variant_consistency_jit'),
12628
+ )),
12626
12629
OpInfo('linalg.lstsq',
12627
12630
aten_name='linalg_lstsq',
12628
12631
dtypes=floating_and_complex_types(),
@@ -12952,7 +12955,10 @@ def error_inputs_mean(op_info, device, **kwargs):
12952
12955
supports_forward_ad=True,
12953
12956
supports_fwgrad_bwgrad=True,
12954
12957
sample_inputs_func=sample_inputs_linalg_lu,
12955
- decorators=[skipCUDAIfNoMagmaAndNoCusolver, skipCPUIfNoLapack]),
12958
+ decorators=[skipCUDAIfNoMagmaAndNoCusolver, skipCPUIfNoLapack],
12959
+ skips=(
12960
+ DecorateInfo(unittest.skip("memory hog"), 'TestGradients', 'test_fn_fwgrad_bwgrad'),
12961
+ )),
12956
12962
OpInfo('lu_unpack',
12957
12963
op=torch.lu_unpack,
12958
12964
dtypes=floating_and_complex_types(),
0 commit comments