Files
kvcache-simulator/configs/qwen2.5-coder-32b-h800.yaml
Gahow Wang ec73a95e05 KVCache simulator for LLM serving cluster routing research
Discrete-event simulator for evaluating KV cache-aware routing policies
in prefill-disaggregated LLM serving clusters. Models a two-tier KV cache
hierarchy (L0 GPU HBM + L1 CPU DRAM) with RDMA/PCIe link contention,
architecture-derived roofline compute (MoE, MLA, DSA), and a cluster-wide
meta-store for prefix-aware routing decisions.

Includes 11 routing policies (random, round_robin, least_loaded,
least_tokens, ttl_aware, precise, min_pd, cache_load, cache_score,
estimated_ttft, prefix_affinity), HuggingFace config.json auto-parsing,
built-in GPU hardware presets (H100/H800/H20/A100/B200), and ablation
tooling for systematic policy comparison across real Alibaba serving traces.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-04-14 01:16:02 +08:00

43 lines
1015 B
YAML

# Qwen2.5-Coder-32B (dense, GQA) on H800 SXM (80GB).
# Architecture from HuggingFace config.json — roofline auto-derived.
model:
name: qwen2.5-coder-32b
num_layers: 64
hidden_size: 5120
num_attention_heads: 40
num_kv_heads: 8 # GQA
head_dim: 128
intermediate_size: 27648 # SwiGLU FFN
dtype_bytes: 2 # BF16
block_size_tokens: 16
hardware:
gpu_flops: 9.89e14
gpu_mem_bw: 3.35e12
hbm_bytes: 20.0e9 # smaller budget: 32B weights are large
dram_bytes: 512.0e9
pcie_bw: 64.0e9
pcie_latency_us: 5.0
rdma_bw: 25.0e9
rdma_latency_us: 8.0
max_batch_slots: 128
prefill_chunk_tokens: 1024
cluster:
num_instances: 16
meta_store:
ttl_seconds: 60.0
router:
mode: ttl_aware
precise_probe_latency_us: 50.0
precise_probe_topk: 4
load_alpha: 1.0
sim:
trace_path: traces/qwen_coder_blksz_16.jsonl
max_requests: null
output_dir: runs/qwen32b
sample_interval_s: 1.0
seed: 42