Align qwen27b baseline proposal with topology study

This commit is contained in:
2026-04-10 17:43:02 +08:00
parent 8d0777e5e2
commit 06d4c380b3

View File

@@ -3,20 +3,18 @@
"diagnosis": "This model uses a long-context hybrid stack and fp8 quantization. The safest first measurement is to preserve the existing warmup, hybrid-model, chunked-prefill, and prefix-caching behavior from run_qwen27b.sh, while keeping a conservative sequence cap.", "diagnosis": "This model uses a long-context hybrid stack and fp8 quantization. The safest first measurement is to preserve the existing warmup, hybrid-model, chunked-prefill, and prefix-caching behavior from run_qwen27b.sh, while keeping a conservative sequence cap.",
"config_patch": { "config_patch": {
"env_patch": { "env_patch": {
"VLLM_ATTENTION_BACKEND": "FLASH_ATTN", "VLLM_ENABLE_TORCH_COMPILE": "1"
"VLLM_ENABLE_TORCH_COMPILE": "1",
"VLLM_USE_FLASHINFER_SAMPLER": "0",
"VLLM_ENABLE_MODEL_RUNNER_WARMUP": "1"
}, },
"flag_patch": { "flag_patch": {
"tensor-parallel-size": 4, "tensor-parallel-size": 4,
"data-parallel-size": 1,
"expert-parallel-size": 1,
"gpu-memory-utilization": 0.9, "gpu-memory-utilization": 0.9,
"block-size": 64, "block-size": 64,
"max-num-batched-tokens": 8192, "max-num-batched-tokens": 8192,
"max-num-seqs": 16, "max-num-seqs": 16,
"enable-prefix-caching": true, "enable-prefix-caching": true,
"enable-chunked-prefill": true, "enable-chunked-prefill": true
"disable-cascade-attn": true
} }
}, },
"expected_effects": [ "expected_effects": [