Blame
435d27 | lhl | 2025-09-13 03:01:48 | 1 | # vLLM |
2 | ||||
9f4fd9 | lhl | 2025-09-13 03:02:21 | 3 | [@lhl](https://github.com/lhl) got implemented the [first public vLLM build recipe](https://github.com/lhl/strix-halo-testing/tree/main/vllm) and shortly after Discord member @ssweens created Arch-based dockerfiles. [@kyuz0](https://github.com/kyuz0) adapted these into a [amd-strix-halo-vllm-toolboxes](https://github.com/kyuz0/amd-strix-halo-vllm-toolboxes) and that is probably the easiest way currently (2025-09-13) to bring up vLLM on Strix Halo. |
435d27 | lhl | 2025-09-13 03:01:48 | 4 | |
5 | ## Current Status |
|||
6 | ||||
7 | While vLLM can be built, and runs for basic models like llama2, newer models (gpt-oss for example) or those with different kernel/code dependencies may not work. |
|||
8 | ||||
9 | ## Build Steps |
|||
10 | ||||
11 | A general description if you want to build vLLM. You should work in a clean env if possible: |
|||
12 | ||||
13 | - First have ROCm setup (latest prod or TheRock build should be ok) |
|||
14 | - Build a [PyTorch w/ AOTriton support](https://github.com/lhl/strix-halo-testing/tree/main/torch-therock) - currently, TheRock PyTorch builds do not have this, so you need to build your own |
|||
15 | - [Build vLLM](https://github.com/lhl/strix-halo-testing/blob/main/vllm/01-build-vllm.sh) - there are many patches you might need to apply |
|||
16 | - use_existing_torch.py |
|||
17 | - Add gfx1151 to CMakeLists.txt |
|||
18 | - build isolation fixes |
|||
19 | - patch out amdsmi (not gfx1151 compatible) |
|||
20 | - `pip install -e . --no-build-isolation` |
|||
21 | - either set constraints or reinstall torch/triton after as some intermediate steps may try to overwrite; alternatively you can `--no-deps` or manual installs of some packages but life is short |