vLLM wheel
vLLM for arm64
- Create vllm library wheel
sudo apt-get install build-essential libnuma-dev -y sudo apt remove cmake -y wget https://github.com/Kitware/CMake/releases/download/v3.31.8/cmake-3.31.8-linux-aarch64.sh chmod +x cmake-3.31.8-linux-aarch64.sh sudo ./cmake-3.31.8-linux-aarch64.sh --prefix=/usr/local --exclude-subdir pip uninstall torch torchvision torchaudio pip install torch==2.7.1 torchaudio==2.7.1 torchvision --index-url https://download.pytorch.org/whl/cu128 git clone https://github.com/vllm-project/vllm.git cd vllm python use_existing_torch.py pip install --upgrade setuptools twine setuptools-scm pip install -r requirements/cuda.txt export MAX_JOBS=16 export NVCC_THREADS=4 export TORCH_CUDA_ARCH_LIST="" export VLLM_TARGET_DEVICE=cuda python setup.py bdist_wheel pip install dist/*.whl