mirror of
https://github.com/pytorch/pytorch.git
synced 2025-10-20 21:14:14 +08:00
[precompile] Fix frame construction for wrapped model. (#165454)
Summary: If a function is wrapped with functools, we should not look at the wrapped function signature but rather the wrapper, since we need to construct the frame for the top level function here. Test Plan: test_decorated_function_with_functools_wrap_aot Differential Revision: D84626752 Pull Request resolved: https://github.com/pytorch/pytorch/pull/165454 Approved by: https://github.com/yiming0416
This commit is contained in:
committed by
PyTorch MergeBot
parent
ca65023b90
commit
839f6facdb
@ -1,5 +1,6 @@
|
||||
# Owner(s): ["module: dynamo"]
|
||||
|
||||
import functools
|
||||
import inspect
|
||||
import os
|
||||
import pickle
|
||||
@ -203,6 +204,39 @@ class TestAOTCompile(torch._inductor.test_case.TestCase):
|
||||
actual = compiled_fn(*example_inputs)
|
||||
self.assertEqual(expected, actual)
|
||||
|
||||
def test_decorated_function_with_functools_wrap_aot(self):
|
||||
def check_inputs(fn):
|
||||
@functools.wraps(fn)
|
||||
def _fn(*args, **kwargs):
|
||||
for arg in args:
|
||||
assert arg.shape[0] > 1
|
||||
|
||||
return fn(*args, **kwargs)
|
||||
|
||||
return _fn
|
||||
|
||||
@check_inputs
|
||||
def foo(x, y):
|
||||
a = x + x
|
||||
b = y + y
|
||||
c = a + b
|
||||
return c
|
||||
|
||||
example_inputs = (torch.ones(3), torch.ones(3))
|
||||
expected = foo(*example_inputs)
|
||||
|
||||
def backend(gm, example_inputs):
|
||||
return CustomCompiledFunction(gm, example_inputs)
|
||||
|
||||
with torch.compiler.set_stance("fail_on_recompile"):
|
||||
compiled_fn = torch.compile(
|
||||
foo,
|
||||
fullgraph=True,
|
||||
backend=backend,
|
||||
).aot_compile((example_inputs, {}))
|
||||
actual = compiled_fn(*example_inputs)
|
||||
self.assertEqual(expected, actual)
|
||||
|
||||
def test_aot_compile_disable_guard_check(self):
|
||||
def fn(x, y):
|
||||
return x + y
|
||||
|
Reference in New Issue
Block a user