[OpenReg] Implement device autoload mechanism (#158555)

# Implement OpenReg device autoload mechanism

## Overview
The **Autoload** mechanism in PyTorch simplifies the integration of third-party device backends by enabling automatic discovery and initialization at runtime. Traditionally, integrating a new backend required explicit imports or manual initialization, which could be cumbersome and error-prone. With Autoload, PyTorch dynamically detects and initializes device backends, providing a seamless user experience.

This mechanism leverages Python entry points (e.g., `torch.backends`) and dynamic module loading. When PyTorch starts, it scans for registered entry points and invokes their initialization hooks, ensuring that all available backends are ready for use without requiring explicit imports.

## Motivation

This PR aims to apply [device autoload mechanism](https://github.com/pytorch/pytorch/issues/122468) to the OpenReg module with some simple changes.

## Change
### Before
```python
import torch
import torch_openreg

x = torch.tensor([1, 2, 3], device="openreg")
print(x)
```
### After
```python
import torch

# No need to import torch_openreg manually!
x = torch.tensor([1, 2, 3], device="openreg")
print(x)
```

Pull Request resolved: https://github.com/pytorch/pytorch/pull/158555
Approved by: https://github.com/FFFrog, https://github.com/albanD

Co-authored-by: Jiawei Li <ljw1101.vip@gmail.com>
This commit is contained in:
can-gaa-hou
2025-09-12 04:24:08 +00:00
committed by PyTorch MergeBot
parent da954f10d6
commit 95191522e0
7 changed files with 119 additions and 5 deletions

View File

@ -0,0 +1,86 @@
# Autoload Mechanism
The **Autoload** mechanism in PyTorch simplifies the integration of a custom backend by enabling automatic discovery and initialization at runtime. This eliminates the need for explicit imports or manual initialization, allowing developers to seamlessly integrate a new accelerator or backend into PyTorch.
## Background
The **Autoload Device Extension** proposal in PyTorch is centered on improving support for various hardware backend devices, especially those implemented as out-of-the-tree extensions (not part of PyTorchs main codebase). Currently, users must manually import or load these device-specific extensions to use them, which complicates the experience and increases cognitive overhead.
In contrast, in-tree devices (devices officially supported within PyTorch) are seamlessly integrated—users dont need extra imports or steps. The goal of autoloading is to make out-of-the-tree devices just as easy to use, so users can follow the standard PyTorch device programming model without explicit loading or code changes. This would allow existing PyTorch applications to run on new devices without any modification, making hardware support more user-friendly and reducing barriers to adoption.
For more information about the background of **Autoload**, please refer to its [RFC](https://github.com/pytorch/pytorch/issues/122468).
## Design
The core idea of **Autoload** is to Use Pythons plugin discovery (entry points) so PyTorch automatically loads out-of-tree device extensions when torch is imported—no explicit user import needed.
For more instructions of the design of **Autoload**, please refer to [**How it works**](https://docs.pytorch.org/tutorials/unstable/python_extension_autoload.html#how-it-works).
## Implementation
This tutorial will take **OpenReg** as a new out-of-the-tree device and guide you through the steps to enable and use the **Autoload** mechanism.
### Entry Point Setup
To enable **Autoload**, register the `_autoload` function as an entry point in `setup.py` file.
::::{tab-set}
:::{tab-item} Python
```{eval-rst}
.. literalinclude:: ../../../test/cpp_extensions/open_registration_extension/torch_openreg/setup.py
:language: python
:start-after: LITERALINCLUDE START: SETUP
:end-before: LITERALINCLUDE END: SETUP
:linenos:
:emphasize-lines: 9-13
```
:::
::::
### Backend Setup
Define the initialization hook `_autoload` for backend initialization. This hook will be automatically invoked by PyTorch during startup.
::::{tab-set-code}
```{eval-rst}
.. literalinclude:: ../../../test/cpp_extensions/open_registration_extension/torch_openreg/torch_openreg/__init__.py
:language: python
:start-after: LITERALINCLUDE START: AUTOLOAD
:end-before: LITERALINCLUDE END: AUTOLOAD
:linenos:
:emphasize-lines: 10-12
```
::::
## Result
After setting up the entry point and backend, build and install your backend. Now, we can use the new accelerator without explicitly importing it.
```{eval-rst}
.. grid:: 2
.. grid-item-card:: :octicon:`terminal;1em;` Without Autoload
:class-card: card-prerequisites
::
>>> import torch
>>> import torch_openreg
>>> torch.tensor(1, device="openreg")
tensor(1, device='openreg:0')
.. grid-item-card:: :octicon:`terminal;1em;` With Autoload
:class-card: card-prerequisites
::
>>> import torch # Automatically import torch_openreg
>>> torch.tensor(1, device="openreg")
tensor(1, device='openreg:0')
```

View File

@ -42,6 +42,7 @@ Next, we will delve into each chapter of this guide. Each chapter focuses on a k
:glob:
:maxdepth: 1
autoload
operators
```

View File

@ -124,6 +124,16 @@ There are 4 DSOs in torch_openreg, and the dependencies between them are as foll
- Per-operator Fallback: See `sub.Tensor`
- Global Fallback: See `wrapper_cpu_fallback`
### Autoload
- Autoload Machanism
When `import torch`, installed accelerators (such as `torch_openreg`) will be automatically loaded, achieving the same experience as the built-in backends.
- Registering the backend with Python `entry points`: See `setup` in `setup.py`
- Adding a callable function for backend initialization: See `_autoload` in `torch_openreg/__init__.py`
- Dynamically loading the backend without explicit imports: See [Usage Example](#usage-example)
## Installation and Usage
### Installation
@ -139,7 +149,6 @@ After installation, you can use the `openreg` device in Python just like any oth
```python
import torch
import torch_openreg
if not torch.openreg.is_available():
print("OpenReg backend is not available in this build.")

View File

@ -28,6 +28,12 @@ def make_relative_rpath_args(path):
def get_pytorch_dir():
# Disable autoload of the accelerator
# We must do this for two reasons:
# We only need to get the PyTorch installation directory, so whether the accelerator is loaded or not is irrelevant
# If the accelerator has been previously built and not uninstalled, importing torch will cause a circular import error
os.environ["TORCH_DEVICE_BACKEND_AUTOLOAD"] = "0"
import torch
return os.path.dirname(os.path.realpath(torch.__file__))
@ -127,6 +133,7 @@ def main():
]
}
# LITERALINCLUDE START: SETUP
setup(
packages=find_packages(),
package_data=package_data,
@ -135,7 +142,13 @@ def main():
"clean": BuildClean, # type: ignore[misc]
},
include_package_data=False,
entry_points={
"torch.backends": [
"torch_openreg = torch_openreg:_autoload",
],
},
)
# LITERALINCLUDE END: SETUP
if __name__ == "__main__":

View File

@ -9,7 +9,7 @@ if sys.platform == "win32":
_load_dll_libraries()
del _load_dll_libraries
# LITERALINCLUDE START: AUTOLOAD
import torch_openreg._C # type: ignore[misc]
import torch_openreg.openreg
@ -17,3 +17,11 @@ import torch_openreg.openreg
torch.utils.rename_privateuse1_backend("openreg")
torch._register_device_module("openreg", torch_openreg.openreg)
torch.utils.generate_methods_for_privateuse1_backend(for_storage=True)
def _autoload():
# It is a placeholder function here to be registered as an entry point.
pass
# LITERALINCLUDE END: AUTOLOAD

View File

@ -10,7 +10,6 @@ from unittest.mock import patch
import numpy as np
import psutil
import torch_openreg # noqa: F401
import torch
from torch.serialization import safe_globals

View File

@ -4,8 +4,6 @@ import unittest
from collections import namedtuple
from functools import partial
import torch_openreg # noqa: F401
import torch
from torch.nn.attention import SDPBackend
from torch.testing._internal.common_nn import NNTestCase