# Installation vLLM supports the following hardware platforms: - [GPU](gpu.md) - [NVIDIA CUDA](gpu.md#nvidia-cuda) - [AMD ROCm](gpu.md#amd-rocm) - [Intel XPU](gpu.md#intel-xpu) - [CPU](cpu.md) - [Intel/AMD x86](cpu.md#intelamd-x86) - [ARM AArch64](cpu.md#arm-aarch64) - [Apple silicon](cpu.md#apple-silicon) - [IBM Z (S390X)](cpu.md#ibm-z-s390x) - [Google TPU](google_tpu.md) - [AWS Neuron](aws_neuron.md) ## Hardware Plugins The backends below live **outside** the main `vllm` repository and follow the [Hardware-Pluggable RFC](../../design/plugin_system.md). | Accelerator | PyPI / package | Repository | |-------------|----------------|------------| | Ascend NPU | `vllm-ascend` | | | Intel Gaudi (HPU) | N/A, install from source | | | MetaX MACA GPU | N/A, install from source | | | Rebellions ATOM / REBEL NPU | `vllm-rbln` | | | IBM Spyre AIU | `vllm-spyre` | |