mirror of
https://github.com/pytorch/pytorch.git
synced 2025-10-21 05:34:18 +08:00
add complex32 to v3_dtypes (#120388)
Fixes [#120290](https://github.com/pytorch/pytorch/issues/120290) Fixes https://github.com/pytorch/pytorch/issues/73502 use `v3_dtypes` and `torch._utils._rebuild_tensor_v3` to handle torch.save(complex32) result:  Pull Request resolved: https://github.com/pytorch/pytorch/pull/120388 Approved by: https://github.com/albanD
This commit is contained in:
@ -3524,8 +3524,6 @@ module_db: List[ModuleInfo] = [
|
||||
# Not implmented for chalf on CPU
|
||||
DecorateInfo(unittest.expectedFailure, 'TestModule', 'test_cpu_gpu_parity',
|
||||
dtypes=(torch.chalf,), device_type='cuda'),
|
||||
# Ref: https://github.com/pytorch/pytorch/issues/73502
|
||||
DecorateInfo(unittest.expectedFailure, 'TestModule', 'test_pickle', dtypes=(torch.chalf,)),
|
||||
# See #119108: MPSNDArrayConvolutionA14.mm:3976: failed assertion `destination datatype must be fp32'
|
||||
# xfail does not work due to Fatal Python error: Aborted
|
||||
DecorateInfo(skipIfMps, "TestModule", "test_memory_format",
|
||||
@ -3559,8 +3557,6 @@ module_db: List[ModuleInfo] = [
|
||||
# Not implemented for chalf on CPU
|
||||
DecorateInfo(unittest.expectedFailure, 'TestModule', 'test_cpu_gpu_parity',
|
||||
dtypes=(torch.chalf,), device_type='cuda'),
|
||||
# Ref: https://github.com/pytorch/pytorch/issues/73502
|
||||
DecorateInfo(unittest.expectedFailure, 'TestModule', 'test_pickle', dtypes=(torch.chalf,)),
|
||||
# See #119108: MPSNDArrayConvolutionA14.mm:3976: failed assertion `destination datatype must be fp32'
|
||||
# xfail does not work due to Fatal Python error: Aborted
|
||||
DecorateInfo(skipIfMps, "TestModule", "test_memory_format",
|
||||
@ -3593,8 +3589,6 @@ module_db: List[ModuleInfo] = [
|
||||
# Not implmented for chalf on CPU
|
||||
DecorateInfo(unittest.expectedFailure, 'TestModule', 'test_cpu_gpu_parity',
|
||||
dtypes=(torch.chalf,), device_type='cuda'),
|
||||
# Ref: https://github.com/pytorch/pytorch/issues/73502
|
||||
DecorateInfo(unittest.expectedFailure, 'TestModule', 'test_pickle', dtypes=(torch.chalf,)),
|
||||
),
|
||||
decorators=(
|
||||
DecorateInfo(precisionOverride({torch.float32: 1e-04}), 'TestModule', 'test_memory_format'),
|
||||
|
Reference in New Issue
Block a user