This repo is designed as a vLLM plugin which provides custom kernels for Intel GPU (known as XPU in PyTorch).
Currently we use PyTorch 2.8, oneapi 2025.1.
python3 setup.py build - will build a _C.abi3.so
under build directory
python3 setup.py install - will copy above .so to vllm_xpu_kernels
folder
python3 setup.py develop - will be local install if we use develop build or system/virtual env lib path if we use install.
On vllm side, we will import vllm_xpu_kernels._C
at start time which should register all custom ops so we can directly use.
Install oneapi 2025.1 deep learning essential dependency.
Create a new virtual env, install build dependency and torch dependency
pip install -r requirements.txt
Build development installation to current directory:
VLLM_TARGET_DEVICE=xpu python3 setup.py develop
or installation to system directory:
VLLM_TARGET_DEVICE=xpu python3 setup.py install
or build wheel (generated .whl in dist folder)
VLLM_TARGET_DEVICE=xpu python3 setup.py bdist_wheel
Please refer to temporary branch https://github.com/jikunshang/vllm/tree/xpu_kernel to install & test vllm which replaces rms_norm
kernel from IPEX to vllm-xpu-kernels.