Align qwen27b baseline proposal to TP1 run script
This commit is contained in:
@@ -1,26 +1,14 @@
|
|||||||
{
|
{
|
||||||
"observation": "The incumbent should start from the known launch-safe qwen3.5-27b serving recipe on dash0 before asking the LLM to optimize throughput above that baseline.",
|
"observation": "The incumbent should start from the exact known launch-safe qwen3.5-27b serving recipe on dash0 before asking the LLM to optimize throughput per GPU above that baseline.",
|
||||||
"diagnosis": "This model uses a long-context hybrid stack and fp8 quantization. The safest first measurement is to preserve the existing warmup, hybrid-model, chunked-prefill, and prefix-caching behavior from run_qwen27b.sh, while keeping a conservative sequence cap.",
|
"diagnosis": "This model uses a long-context hybrid stack and fp8 quantization. The correct first measurement is to preserve the TP1 run_qwen27b.sh baseline exactly, then let later trials explore TP/DP and runtime knobs from that anchor.",
|
||||||
"config_patch": {
|
"config_patch": {
|
||||||
"env_patch": {
|
"env_patch": {},
|
||||||
"VLLM_ENABLE_TORCH_COMPILE": "1"
|
"flag_patch": {}
|
||||||
},
|
|
||||||
"flag_patch": {
|
|
||||||
"tensor-parallel-size": 4,
|
|
||||||
"data-parallel-size": 1,
|
|
||||||
"expert-parallel-size": 1,
|
|
||||||
"gpu-memory-utilization": 0.9,
|
|
||||||
"block-size": 64,
|
|
||||||
"max-num-batched-tokens": 8192,
|
|
||||||
"max-num-seqs": 16,
|
|
||||||
"enable-prefix-caching": true,
|
|
||||||
"enable-chunked-prefill": true
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
"expected_effects": [
|
"expected_effects": [
|
||||||
"Launch-safe baseline aligned with the current hand-tuned qwen27b recipe while using all 4 visible H20 GPUs",
|
"Launch-safe TP1 baseline exactly aligned with run_qwen27b.sh",
|
||||||
"Reliable first incumbent under the tighter TTFT and TPOT SLO",
|
"Reliable first incumbent for per-GPU comparison under the tighter TTFT and TPOT SLO",
|
||||||
"Clear trial history for the LLM to propose a higher-throughput follow-up patch"
|
"Clean anchor before topology exploration across TP*DP groups"
|
||||||
],
|
],
|
||||||
"why_not_previous_failures": "This baseline intentionally avoids speculative new kernels or batching spikes before we have an incumbent under the new SLO."
|
"why_not_previous_failures": "This baseline intentionally introduces no patch at all, so it cannot repeat previous tuning-only failures."
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user