mirror of
https://github.com/pytorch/pytorch.git
synced 2025-11-16 15:34:57 +08:00
skip failed case
This commit is contained in:
committed by
PyTorch MergeBot
parent
528d3fc4ce
commit
d83d25dee4
@ -39,6 +39,7 @@ from torch.distributed.tensor.parallel import (
|
||||
RowwiseParallel,
|
||||
)
|
||||
from torch.distributed.tensor.placement_types import _StridedShard
|
||||
from torch.testing._internal.common_device_type import skipXPUIf
|
||||
from torch.testing._internal.common_distributed import skip_if_lt_x_gpu
|
||||
from torch.testing._internal.common_fsdp import get_devtype
|
||||
from torch.testing._internal.common_utils import (
|
||||
@ -376,6 +377,7 @@ def forward(self, b_parametrizations_buffer_original0, x):
|
||||
self.assertEqual(res, ref)
|
||||
|
||||
@skipIfHpu
|
||||
@skipXPUIf(True, "https://github.com/intel/torch-xpu-ops/issues/1981")
|
||||
def test_dtensor_dynamic_loss_parallel_log_softmax(self):
|
||||
mesh = DeviceMesh(self.device_type, torch.arange(self.world_size))
|
||||
|
||||
|
||||
Reference in New Issue
Block a user