Use vllm cu129 wheel on dash0
This commit is contained in:
@@ -14,7 +14,7 @@
|
|||||||
"engine": {
|
"engine": {
|
||||||
"engine_name": "vllm",
|
"engine_name": "vllm",
|
||||||
"engine_version": "0.20.0",
|
"engine_version": "0.20.0",
|
||||||
"exec_path": "/tmp/wjh/venvs/vllm-0.20.0-auto/bin/vllm",
|
"exec_path": "/tmp/wjh/venvs/vllm-0.20.0-cu129/bin/vllm",
|
||||||
"cwd": "/home/admin/cpfs/wjh/aituner/aituner",
|
"cwd": "/home/admin/cpfs/wjh/aituner/aituner",
|
||||||
"host": "127.0.0.1",
|
"host": "127.0.0.1",
|
||||||
"port": 18230,
|
"port": 18230,
|
||||||
@@ -57,7 +57,7 @@
|
|||||||
"allowed_data_parallel_sizes": [1, 2, 4, 8],
|
"allowed_data_parallel_sizes": [1, 2, 4, 8],
|
||||||
"allowed_expert_parallel_sizes": [1, 2, 4, 8]
|
"allowed_expert_parallel_sizes": [1, 2, 4, 8]
|
||||||
},
|
},
|
||||||
"python_executable": "/tmp/wjh/venvs/vllm-0.20.0-auto/bin/python"
|
"python_executable": "/tmp/wjh/venvs/vllm-0.20.0-cu129/bin/python"
|
||||||
},
|
},
|
||||||
"trace": {
|
"trace": {
|
||||||
"windows_path": "/home/admin/cpfs/wjh/aituner/aituner/trace_windows/windows.json",
|
"windows_path": "/home/admin/cpfs/wjh/aituner/aituner/trace_windows/windows.json",
|
||||||
|
|||||||
@@ -14,7 +14,7 @@
|
|||||||
"engine": {
|
"engine": {
|
||||||
"engine_name": "vllm",
|
"engine_name": "vllm",
|
||||||
"engine_version": "0.20.0",
|
"engine_version": "0.20.0",
|
||||||
"exec_path": "/tmp/wjh/venvs/vllm-0.20.0-auto/bin/vllm",
|
"exec_path": "/tmp/wjh/venvs/vllm-0.20.0-cu129/bin/vllm",
|
||||||
"cwd": "/home/admin/cpfs/wjh/aituner/aituner",
|
"cwd": "/home/admin/cpfs/wjh/aituner/aituner",
|
||||||
"host": "127.0.0.1",
|
"host": "127.0.0.1",
|
||||||
"port": 18231,
|
"port": 18231,
|
||||||
@@ -57,7 +57,7 @@
|
|||||||
"allowed_data_parallel_sizes": [1, 2, 4, 8],
|
"allowed_data_parallel_sizes": [1, 2, 4, 8],
|
||||||
"allowed_expert_parallel_sizes": [1, 2, 4, 8]
|
"allowed_expert_parallel_sizes": [1, 2, 4, 8]
|
||||||
},
|
},
|
||||||
"python_executable": "/tmp/wjh/venvs/vllm-0.20.0-auto/bin/python"
|
"python_executable": "/tmp/wjh/venvs/vllm-0.20.0-cu129/bin/python"
|
||||||
},
|
},
|
||||||
"trace": {
|
"trace": {
|
||||||
"windows_path": "/home/admin/cpfs/wjh/aituner/aituner/trace_windows/windows.json",
|
"windows_path": "/home/admin/cpfs/wjh/aituner/aituner/trace_windows/windows.json",
|
||||||
|
|||||||
@@ -19,9 +19,9 @@ Both specs start from the same base vLLM configuration. The base contains only s
|
|||||||
|
|
||||||
PyPI reports `vllm==0.20.0` as the current community release checked on 2026-05-02. The dash0 runtime venv is on local rootfs rather than CPFS, because installing torch/CUDA wheels into CPFS was I/O-bound:
|
PyPI reports `vllm==0.20.0` as the current community release checked on 2026-05-02. The dash0 runtime venv is on local rootfs rather than CPFS, because installing torch/CUDA wheels into CPFS was I/O-bound:
|
||||||
|
|
||||||
`/tmp/wjh/venvs/vllm-0.20.0-auto`
|
`/tmp/wjh/venvs/vllm-0.20.0-cu129`
|
||||||
|
|
||||||
The first plain `pip install vllm==0.20.0` smoke pulled `torch 2.11.0+cu130` and failed on dash0's driver (`570.133.20`, CUDA 12.9). The active install uses the vLLM-documented `uv pip install vllm==0.20.0 --torch-backend=auto` path so uv selects a CUDA backend compatible with the installed driver.
|
The first plain `pip install vllm==0.20.0` smoke pulled `torch 2.11.0+cu130` and failed on dash0's driver (`570.133.20`, CUDA 12.9). The active install uses the vLLM 0.20.0 GitHub release `+cu129` wheel and the PyTorch CUDA 12.9 index, matching the vLLM documented CUDA 12.9 install path for this driver.
|
||||||
|
|
||||||
Install log:
|
Install log:
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user