mirror of
https://github.com/pytorch/pytorch.git
synced 2025-10-20 12:54:11 +08:00
This PR follows up on the discussion in #159399 where @Akabbaj and @janeyx99 mentioned renaming HAS_XPU to HAS_XPU_AND_TRITON for consistency. Pull Request resolved: https://github.com/pytorch/pytorch/pull/159908 Approved by: https://github.com/janeyx99, https://github.com/guangyey
60 lines
1.5 KiB
Python
60 lines
1.5 KiB
Python
# Owner(s): ["module: inductor"]
|
|
import importlib
|
|
import os
|
|
import sys
|
|
|
|
import torch
|
|
|
|
|
|
importlib.import_module("filelock")
|
|
|
|
pytorch_test_dir = os.path.dirname(os.path.dirname(os.path.realpath(__file__)))
|
|
sys.path.append(pytorch_test_dir)
|
|
from inductor.test_torchinductor import ( # @manual=fbcode//caffe2/test/inductor:test_inductor-library
|
|
check_model_gpu,
|
|
TestCase,
|
|
)
|
|
|
|
|
|
# TODO: Remove this file.
|
|
# This is a temporary test case to test the base functionality of first Intel GPU Inductor integration.
|
|
# We are working on reuse and pass the test cases in test/inductor/* step by step.
|
|
# Will remove this file when pass full test in test/inductor/*.
|
|
|
|
|
|
class XpuBasicTests(TestCase):
|
|
common = check_model_gpu
|
|
device = "xpu"
|
|
|
|
def test_add(self):
|
|
def fn(a, b):
|
|
return a + b
|
|
|
|
self.common(fn, (torch.rand(2, 3, 16, 16), torch.rand(2, 3, 16, 16)))
|
|
|
|
def test_sub(self):
|
|
def fn(a, b):
|
|
return a - b
|
|
|
|
self.common(fn, (torch.rand(2, 3, 16, 16), torch.rand(2, 3, 16, 16)))
|
|
|
|
def test_mul(self):
|
|
def fn(a, b):
|
|
return a * b
|
|
|
|
self.common(fn, (torch.rand(2, 3, 16, 16), torch.rand(2, 3, 16, 16)))
|
|
|
|
def test_div(self):
|
|
def fn(a, b):
|
|
return a / b
|
|
|
|
self.common(fn, (torch.rand(2, 3, 16, 16), torch.rand(2, 3, 16, 16)))
|
|
|
|
|
|
if __name__ == "__main__":
|
|
from torch._dynamo.test_case import run_tests
|
|
from torch.testing._internal.inductor_utils import HAS_XPU_AND_TRITON
|
|
|
|
if HAS_XPU_AND_TRITON:
|
|
run_tests(needs="filelock")
|