mirror of
https://github.com/pytorch/pytorch.git
synced 2025-11-05 00:14:54 +08:00
Using fake tensor with AOTAutograd is now mandatory, simplifying our logic. Unfortunately, this means debug_fake_cross_ref must go, but I don't think anyone has used it recently. Signed-off-by: Edward Z. Yang <ezyang@meta.com> Pull Request resolved: https://github.com/pytorch/pytorch/pull/99314 Approved by: https://github.com/eellison, https://github.com/zou3519
43 lines
1.2 KiB
Python
43 lines
1.2 KiB
Python
# Copyright (c) Facebook, Inc. and its affiliates.
|
|
# All rights reserved.
|
|
#
|
|
# This source code is licensed under the BSD-style license found in the
|
|
# LICENSE file in the root directory of this source tree.
|
|
|
|
"""
|
|
Global flags for aot autograd
|
|
"""
|
|
import os
|
|
import sys
|
|
|
|
use_functionalize = True
|
|
|
|
# Converts torch rng ops to their functional philox rng equivalents. Note that
|
|
# we functionalize only CUDA rng ops today.
|
|
functionalize_rng_ops = False
|
|
|
|
# can be useful for debugging if we are incorrectly creating meta fake tensors
|
|
fake_tensor_allow_meta = os.environ.get("FAKE_ALLOW_META", True)
|
|
|
|
# Enables optional asserts in hotpath code to check for errors. If
|
|
# you are seeing weird accuracy problems, try turning this on.
|
|
# This is currently off by default as it will harm tracing time,
|
|
# but it is on by default for aot_eager.
|
|
debug_assert = False
|
|
|
|
debug_partitioner = os.environ.get("AOT_PARTITIONER_DEBUG", False)
|
|
|
|
static_weight_shapes = True
|
|
|
|
# Applies CSE to the graph before partitioning
|
|
cse = True
|
|
|
|
# Restricts the amount of computation AOTAutograd can do.
|
|
max_dist_from_bw = 3
|
|
|
|
|
|
from .._dynamo.config_utils import install_config_module
|
|
|
|
# adds patch, save_config, invalid config checks, etc
|
|
install_config_module(sys.modules[__name__])
|