mirror of
https://github.com/pytorch/pytorch.git
synced 2025-10-20 12:54:11 +08:00
That tells whether or not PyTorch was compiled with Arm Compute Library Pull Request resolved: https://github.com/pytorch/pytorch/pull/165678 Approved by: https://github.com/Skylion007, https://github.com/atalman, https://github.com/albanD ghstack dependencies: #165583, #165584, #165676
140 lines
4.2 KiB
Python
140 lines
4.2 KiB
Python
# mypy: allow-untyped-defs
|
|
import sys
|
|
from contextlib import contextmanager
|
|
from typing import TYPE_CHECKING
|
|
|
|
import torch
|
|
from torch.backends import (
|
|
__allow_nonbracketed_mutation,
|
|
_FP32Precision,
|
|
_get_fp32_precision_getter,
|
|
_set_fp32_precision_setter,
|
|
ContextProp,
|
|
PropModule,
|
|
)
|
|
|
|
|
|
def is_available():
|
|
r"""Return whether PyTorch is built with MKL-DNN support."""
|
|
return torch._C._has_mkldnn
|
|
|
|
|
|
def is_acl_available():
|
|
r"""Return whether PyTorch is built with MKL-DNN + ACL support."""
|
|
return torch._C._has_mkldnn_acl
|
|
|
|
|
|
VERBOSE_OFF = 0
|
|
VERBOSE_ON = 1
|
|
VERBOSE_ON_CREATION = 2
|
|
|
|
|
|
class verbose:
|
|
"""
|
|
On-demand oneDNN (former MKL-DNN) verbosing functionality.
|
|
|
|
To make it easier to debug performance issues, oneDNN can dump verbose
|
|
messages containing information like kernel size, input data size and
|
|
execution duration while executing the kernel. The verbosing functionality
|
|
can be invoked via an environment variable named `DNNL_VERBOSE`. However,
|
|
this methodology dumps messages in all steps. Those are a large amount of
|
|
verbose messages. Moreover, for investigating the performance issues,
|
|
generally taking verbose messages for one single iteration is enough.
|
|
This on-demand verbosing functionality makes it possible to control scope
|
|
for verbose message dumping. In the following example, verbose messages
|
|
will be dumped out for the second inference only.
|
|
|
|
.. highlight:: python
|
|
.. code-block:: python
|
|
|
|
import torch
|
|
|
|
model(data)
|
|
with torch.backends.mkldnn.verbose(torch.backends.mkldnn.VERBOSE_ON):
|
|
model(data)
|
|
|
|
Args:
|
|
level: Verbose level
|
|
- ``VERBOSE_OFF``: Disable verbosing
|
|
- ``VERBOSE_ON``: Enable verbosing
|
|
- ``VERBOSE_ON_CREATION``: Enable verbosing, including oneDNN kernel creation
|
|
"""
|
|
|
|
def __init__(self, level):
|
|
self.level = level
|
|
|
|
def __enter__(self):
|
|
if self.level == VERBOSE_OFF:
|
|
return
|
|
st = torch._C._verbose.mkldnn_set_verbose(self.level)
|
|
assert st, (
|
|
"Failed to set MKLDNN into verbose mode. Please consider to disable this verbose scope."
|
|
)
|
|
return self
|
|
|
|
def __exit__(self, exc_type, exc_val, exc_tb):
|
|
torch._C._verbose.mkldnn_set_verbose(VERBOSE_OFF)
|
|
return False
|
|
|
|
|
|
def set_flags(
|
|
_enabled=None, _deterministic=None, _allow_tf32=None, _fp32_precision="none"
|
|
):
|
|
orig_flags = (
|
|
torch._C._get_mkldnn_enabled(),
|
|
torch._C._get_mkldnn_deterministic(),
|
|
torch._C._get_onednn_allow_tf32(),
|
|
torch._C._get_fp32_precision_getter("mkldnn", "all"),
|
|
)
|
|
if _enabled is not None:
|
|
torch._C._set_mkldnn_enabled(_enabled)
|
|
if _deterministic is not None:
|
|
torch._C._set_mkldnn_deterministic(_deterministic)
|
|
if _allow_tf32 is not None:
|
|
torch._C._set_onednn_allow_tf32(_allow_tf32)
|
|
if _fp32_precision is not None:
|
|
torch._C._set_fp32_precision_setter("mkldnn", "all", _fp32_precision)
|
|
return orig_flags
|
|
|
|
|
|
@contextmanager
|
|
def flags(enabled=False, deterministic=False, allow_tf32=True, fp32_precision="none"):
|
|
with __allow_nonbracketed_mutation():
|
|
orig_flags = set_flags(enabled, deterministic, allow_tf32, fp32_precision)
|
|
try:
|
|
yield
|
|
finally:
|
|
with __allow_nonbracketed_mutation():
|
|
set_flags(*orig_flags)
|
|
|
|
|
|
class MkldnnModule(PropModule):
|
|
def __init__(self, m, name):
|
|
super().__init__(m, name)
|
|
|
|
def is_available(self):
|
|
return is_available()
|
|
|
|
enabled = ContextProp(torch._C._get_mkldnn_enabled, torch._C._set_mkldnn_enabled)
|
|
deterministic = ContextProp(
|
|
torch._C._get_mkldnn_deterministic, torch._C._set_mkldnn_deterministic
|
|
)
|
|
allow_tf32 = ContextProp(
|
|
torch._C._get_onednn_allow_tf32, torch._C._set_onednn_allow_tf32
|
|
)
|
|
matmul = _FP32Precision("mkldnn", "matmul")
|
|
conv = _FP32Precision("mkldnn", "conv")
|
|
rnn = _FP32Precision("mkldnn", "rnn")
|
|
fp32_precision = ContextProp(
|
|
_get_fp32_precision_getter("mkldnn", "all"),
|
|
_set_fp32_precision_setter("generic", "all"),
|
|
)
|
|
|
|
|
|
if TYPE_CHECKING:
|
|
enabled: ContextProp
|
|
deterministic: ContextProp
|
|
allow_tf32: ContextProp
|
|
|
|
sys.modules[__name__] = MkldnnModule(sys.modules[__name__], __name__)
|