[Dynamo] Support torch.{cuda/cpu}.amp.autocast (#95416)

For Meta internal use cases.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/95416
Approved by: https://github.com/jansel
This commit is contained in:
Yanbo Liang
2023-03-08 01:40:27 +00:00
committed by PyTorch MergeBot
parent b8f7bd593c
commit c88aa336aa
8 changed files with 66 additions and 8 deletions

View File

@ -7,12 +7,13 @@ from typing import Optional, Tuple
import unittest
from test_jit import JitTestCase
from torch.testing._internal.common_cuda import TEST_CUDA
from torch.testing._internal.common_utils import run_tests
from torch.testing._internal.common_utils import run_tests, skipIfTorchDynamo
from torch.testing import FileCheck
from jit.test_models import MnistNet
TEST_BFLOAT16 = TEST_CUDA and torch.cuda.is_bf16_supported()
@skipIfTorchDynamo("Not a TorchDynamo suitable test")
class TestAutocast(JitTestCase):
def setUp(self):
# common input tensors
@ -757,6 +758,7 @@ class convbn(torch.nn.Module):
def forward(self, x):
return self.bn(self.conv(x))
@skipIfTorchDynamo("Not a TorchDynamo suitable test")
class TestJitTraceAutocast(JitTestCase):
def setUp(self):
super().setUp()