mirror of
https://github.com/pytorch/pytorch.git
synced 2025-10-20 21:14:14 +08:00
[OpenReg] Implement device autoload mechanism (#158555)
# Implement OpenReg device autoload mechanism ## Overview The **Autoload** mechanism in PyTorch simplifies the integration of third-party device backends by enabling automatic discovery and initialization at runtime. Traditionally, integrating a new backend required explicit imports or manual initialization, which could be cumbersome and error-prone. With Autoload, PyTorch dynamically detects and initializes device backends, providing a seamless user experience. This mechanism leverages Python entry points (e.g., `torch.backends`) and dynamic module loading. When PyTorch starts, it scans for registered entry points and invokes their initialization hooks, ensuring that all available backends are ready for use without requiring explicit imports. ## Motivation This PR aims to apply [device autoload mechanism](https://github.com/pytorch/pytorch/issues/122468) to the OpenReg module with some simple changes. ## Change ### Before ```python import torch import torch_openreg x = torch.tensor([1, 2, 3], device="openreg") print(x) ``` ### After ```python import torch # No need to import torch_openreg manually! x = torch.tensor([1, 2, 3], device="openreg") print(x) ``` Pull Request resolved: https://github.com/pytorch/pytorch/pull/158555 Approved by: https://github.com/FFFrog, https://github.com/albanD Co-authored-by: Jiawei Li <ljw1101.vip@gmail.com>
This commit is contained in:
committed by
PyTorch MergeBot
parent
da954f10d6
commit
95191522e0
86
docs/source/accelerator/autoload.md
Normal file
86
docs/source/accelerator/autoload.md
Normal file
@ -0,0 +1,86 @@
|
||||
# Autoload Mechanism
|
||||
|
||||
The **Autoload** mechanism in PyTorch simplifies the integration of a custom backend by enabling automatic discovery and initialization at runtime. This eliminates the need for explicit imports or manual initialization, allowing developers to seamlessly integrate a new accelerator or backend into PyTorch.
|
||||
|
||||
## Background
|
||||
|
||||
The **Autoload Device Extension** proposal in PyTorch is centered on improving support for various hardware backend devices, especially those implemented as out-of-the-tree extensions (not part of PyTorch’s main codebase). Currently, users must manually import or load these device-specific extensions to use them, which complicates the experience and increases cognitive overhead.
|
||||
|
||||
In contrast, in-tree devices (devices officially supported within PyTorch) are seamlessly integrated—users don’t need extra imports or steps. The goal of autoloading is to make out-of-the-tree devices just as easy to use, so users can follow the standard PyTorch device programming model without explicit loading or code changes. This would allow existing PyTorch applications to run on new devices without any modification, making hardware support more user-friendly and reducing barriers to adoption.
|
||||
|
||||
For more information about the background of **Autoload**, please refer to its [RFC](https://github.com/pytorch/pytorch/issues/122468).
|
||||
|
||||
## Design
|
||||
|
||||
The core idea of **Autoload** is to Use Python’s plugin discovery (entry points) so PyTorch automatically loads out-of-tree device extensions when torch is imported—no explicit user import needed.
|
||||
|
||||
For more instructions of the design of **Autoload**, please refer to [**How it works**](https://docs.pytorch.org/tutorials/unstable/python_extension_autoload.html#how-it-works).
|
||||
|
||||
## Implementation
|
||||
|
||||
This tutorial will take **OpenReg** as a new out-of-the-tree device and guide you through the steps to enable and use the **Autoload** mechanism.
|
||||
|
||||
### Entry Point Setup
|
||||
|
||||
To enable **Autoload**, register the `_autoload` function as an entry point in `setup.py` file.
|
||||
|
||||
::::{tab-set}
|
||||
|
||||
:::{tab-item} Python
|
||||
|
||||
```{eval-rst}
|
||||
.. literalinclude:: ../../../test/cpp_extensions/open_registration_extension/torch_openreg/setup.py
|
||||
:language: python
|
||||
:start-after: LITERALINCLUDE START: SETUP
|
||||
:end-before: LITERALINCLUDE END: SETUP
|
||||
:linenos:
|
||||
:emphasize-lines: 9-13
|
||||
```
|
||||
|
||||
:::
|
||||
|
||||
::::
|
||||
|
||||
### Backend Setup
|
||||
|
||||
Define the initialization hook `_autoload` for backend initialization. This hook will be automatically invoked by PyTorch during startup.
|
||||
|
||||
::::{tab-set-code}
|
||||
```{eval-rst}
|
||||
.. literalinclude:: ../../../test/cpp_extensions/open_registration_extension/torch_openreg/torch_openreg/__init__.py
|
||||
:language: python
|
||||
:start-after: LITERALINCLUDE START: AUTOLOAD
|
||||
:end-before: LITERALINCLUDE END: AUTOLOAD
|
||||
:linenos:
|
||||
:emphasize-lines: 10-12
|
||||
```
|
||||
|
||||
|
||||
::::
|
||||
|
||||
## Result
|
||||
|
||||
After setting up the entry point and backend, build and install your backend. Now, we can use the new accelerator without explicitly importing it.
|
||||
|
||||
```{eval-rst}
|
||||
.. grid:: 2
|
||||
|
||||
.. grid-item-card:: :octicon:`terminal;1em;` Without Autoload
|
||||
:class-card: card-prerequisites
|
||||
|
||||
::
|
||||
|
||||
>>> import torch
|
||||
>>> import torch_openreg
|
||||
>>> torch.tensor(1, device="openreg")
|
||||
tensor(1, device='openreg:0')
|
||||
|
||||
.. grid-item-card:: :octicon:`terminal;1em;` With Autoload
|
||||
:class-card: card-prerequisites
|
||||
|
||||
::
|
||||
|
||||
>>> import torch # Automatically import torch_openreg
|
||||
>>> torch.tensor(1, device="openreg")
|
||||
tensor(1, device='openreg:0')
|
||||
```
|
@ -42,6 +42,7 @@ Next, we will delve into each chapter of this guide. Each chapter focuses on a k
|
||||
:glob:
|
||||
:maxdepth: 1
|
||||
|
||||
autoload
|
||||
operators
|
||||
```
|
||||
|
||||
|
@ -124,6 +124,16 @@ There are 4 DSOs in torch_openreg, and the dependencies between them are as foll
|
||||
- Per-operator Fallback: See `sub.Tensor`
|
||||
- Global Fallback: See `wrapper_cpu_fallback`
|
||||
|
||||
### Autoload
|
||||
|
||||
- Autoload Machanism
|
||||
|
||||
When `import torch`, installed accelerators (such as `torch_openreg`) will be automatically loaded, achieving the same experience as the built-in backends.
|
||||
|
||||
- Registering the backend with Python `entry points`: See `setup` in `setup.py`
|
||||
- Adding a callable function for backend initialization: See `_autoload` in `torch_openreg/__init__.py`
|
||||
- Dynamically loading the backend without explicit imports: See [Usage Example](#usage-example)
|
||||
|
||||
## Installation and Usage
|
||||
|
||||
### Installation
|
||||
@ -139,7 +149,6 @@ After installation, you can use the `openreg` device in Python just like any oth
|
||||
|
||||
```python
|
||||
import torch
|
||||
import torch_openreg
|
||||
|
||||
if not torch.openreg.is_available():
|
||||
print("OpenReg backend is not available in this build.")
|
||||
|
@ -28,6 +28,12 @@ def make_relative_rpath_args(path):
|
||||
|
||||
|
||||
def get_pytorch_dir():
|
||||
# Disable autoload of the accelerator
|
||||
|
||||
# We must do this for two reasons:
|
||||
# We only need to get the PyTorch installation directory, so whether the accelerator is loaded or not is irrelevant
|
||||
# If the accelerator has been previously built and not uninstalled, importing torch will cause a circular import error
|
||||
os.environ["TORCH_DEVICE_BACKEND_AUTOLOAD"] = "0"
|
||||
import torch
|
||||
|
||||
return os.path.dirname(os.path.realpath(torch.__file__))
|
||||
@ -127,6 +133,7 @@ def main():
|
||||
]
|
||||
}
|
||||
|
||||
# LITERALINCLUDE START: SETUP
|
||||
setup(
|
||||
packages=find_packages(),
|
||||
package_data=package_data,
|
||||
@ -135,7 +142,13 @@ def main():
|
||||
"clean": BuildClean, # type: ignore[misc]
|
||||
},
|
||||
include_package_data=False,
|
||||
entry_points={
|
||||
"torch.backends": [
|
||||
"torch_openreg = torch_openreg:_autoload",
|
||||
],
|
||||
},
|
||||
)
|
||||
# LITERALINCLUDE END: SETUP
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
|
@ -9,7 +9,7 @@ if sys.platform == "win32":
|
||||
_load_dll_libraries()
|
||||
del _load_dll_libraries
|
||||
|
||||
|
||||
# LITERALINCLUDE START: AUTOLOAD
|
||||
import torch_openreg._C # type: ignore[misc]
|
||||
import torch_openreg.openreg
|
||||
|
||||
@ -17,3 +17,11 @@ import torch_openreg.openreg
|
||||
torch.utils.rename_privateuse1_backend("openreg")
|
||||
torch._register_device_module("openreg", torch_openreg.openreg)
|
||||
torch.utils.generate_methods_for_privateuse1_backend(for_storage=True)
|
||||
|
||||
|
||||
def _autoload():
|
||||
# It is a placeholder function here to be registered as an entry point.
|
||||
pass
|
||||
|
||||
|
||||
# LITERALINCLUDE END: AUTOLOAD
|
||||
|
@ -10,7 +10,6 @@ from unittest.mock import patch
|
||||
|
||||
import numpy as np
|
||||
import psutil
|
||||
import torch_openreg # noqa: F401
|
||||
|
||||
import torch
|
||||
from torch.serialization import safe_globals
|
||||
|
@ -4,8 +4,6 @@ import unittest
|
||||
from collections import namedtuple
|
||||
from functools import partial
|
||||
|
||||
import torch_openreg # noqa: F401
|
||||
|
||||
import torch
|
||||
from torch.nn.attention import SDPBackend
|
||||
from torch.testing._internal.common_nn import NNTestCase
|
||||
|
Reference in New Issue
Block a user