Constrain harness topology by visible GPUs
This commit is contained in:
@@ -104,3 +104,29 @@ Started on `dash0` (`11.73.2.172`) at commit `e3ed775`.
|
|||||||
- monitor: read-only subagent `Wegener`
|
- monitor: read-only subagent `Wegener`
|
||||||
|
|
||||||
Acceptance for this run is based on end-to-end trial results, not unit tests. If the first four trials lag the min-prompt no-harness baseline (`0.0650`, `0.1992`, `0.2696`, then failed/NA), the run should be treated as a failed harness iteration and the harness should be optimized again.
|
Acceptance for this run is based on end-to-end trial results, not unit tests. If the first four trials lag the min-prompt no-harness baseline (`0.0650`, `0.1992`, `0.2696`, then failed/NA), the run should be treated as a failed harness iteration and the harness should be optimized again.
|
||||||
|
|
||||||
|
## V2 Result And Failure
|
||||||
|
|
||||||
|
V2 was stopped early after four trials because it did not improve the no-harness baseline and made a preventable launch-risk proposal.
|
||||||
|
|
||||||
|
Raw `request_rate/GPU`:
|
||||||
|
|
||||||
|
| Variant | iter1 | iter2 | iter3 | iter4 |
|
||||||
|
| --- | ---: | ---: | ---: | --- |
|
||||||
|
| no-harness min-prompt | 0.0650 | 0.1992 | 0.2696 | 0.2696 |
|
||||||
|
| harness v2 | 0.0650 | 0.1992 | 0.2696 | failed |
|
||||||
|
|
||||||
|
Harness v2 did correctly diagnose the first bottleneck and proposed:
|
||||||
|
|
||||||
|
- iter2: `tensor-parallel-size=2`, raw `0.1992 req/s/GPU`;
|
||||||
|
- iter3: `tensor-parallel-size=4`, raw `0.2696 req/s/GPU`.
|
||||||
|
|
||||||
|
However, iter4 proposed `tensor-parallel-size=8` and failed at engine launch. The study's `hardware.gpu_count` is 8, but the launch environment sets `CUDA_VISIBLE_DEVICES=0,1,2,4,5,6,7`, which exposes only 7 GPUs. Therefore TP=8 should not have been considered launch-safe.
|
||||||
|
|
||||||
|
This is a general harness bug: topology planning must use the effective visible GPU count from the execution profile, not just the nominal hardware count.
|
||||||
|
|
||||||
|
Fix:
|
||||||
|
|
||||||
|
- parse `engine.base_envs.CUDA_VISIBLE_DEVICES`;
|
||||||
|
- compute effective GPU count as `min(hardware.gpu_count, visible_device_count)`;
|
||||||
|
- filter topology candidates and adjacent TP frontier candidates by the effective GPU count.
|
||||||
|
|||||||
@@ -1210,17 +1210,18 @@ def _legal_topology_points(
|
|||||||
) -> list[dict[str, Any]]:
|
) -> list[dict[str, Any]]:
|
||||||
constraints = study.engine.topology_constraints
|
constraints = study.engine.topology_constraints
|
||||||
tunable = set(study.engine.tunable_flags)
|
tunable = set(study.engine.tunable_flags)
|
||||||
|
effective_gpu_count = _effective_gpu_count(study)
|
||||||
if constraints is not None and constraints.allowed_tensor_parallel_sizes:
|
if constraints is not None and constraints.allowed_tensor_parallel_sizes:
|
||||||
tp_values = sorted(set(constraints.allowed_tensor_parallel_sizes))
|
tp_values = sorted(set(constraints.allowed_tensor_parallel_sizes))
|
||||||
elif "tensor-parallel-size" in tunable:
|
elif "tensor-parallel-size" in tunable:
|
||||||
tp_values = [value for value in [1, 2, 4, 8] if value <= study.hardware.gpu_count]
|
tp_values = [value for value in [1, 2, 4, 8] if value <= effective_gpu_count]
|
||||||
else:
|
else:
|
||||||
tp_values = [current_tp]
|
tp_values = [current_tp]
|
||||||
|
|
||||||
if constraints is not None and constraints.allowed_data_parallel_sizes:
|
if constraints is not None and constraints.allowed_data_parallel_sizes:
|
||||||
dp_values = sorted(set(constraints.allowed_data_parallel_sizes))
|
dp_values = sorted(set(constraints.allowed_data_parallel_sizes))
|
||||||
elif "data-parallel-size" in tunable:
|
elif "data-parallel-size" in tunable:
|
||||||
dp_values = [value for value in [1, 2, 4, 8] if value <= study.hardware.gpu_count]
|
dp_values = [value for value in [1, 2, 4, 8] if value <= effective_gpu_count]
|
||||||
else:
|
else:
|
||||||
dp_values = [current_dp]
|
dp_values = [current_dp]
|
||||||
|
|
||||||
@@ -1243,14 +1244,16 @@ def _legal_topology_points(
|
|||||||
continue
|
continue
|
||||||
if (
|
if (
|
||||||
constraints.require_tp_dp_product_equals_gpu_count
|
constraints.require_tp_dp_product_equals_gpu_count
|
||||||
and tp_dp_product != study.hardware.gpu_count
|
and tp_dp_product != effective_gpu_count
|
||||||
):
|
):
|
||||||
continue
|
continue
|
||||||
elif tp_dp_product > study.hardware.gpu_count:
|
elif tp_dp_product > effective_gpu_count:
|
||||||
continue
|
continue
|
||||||
if constraints is not None and not constraints.allowed_tp_dp_products:
|
if constraints is not None and not constraints.allowed_tp_dp_products:
|
||||||
if tp_dp_product > study.hardware.gpu_count:
|
if tp_dp_product > effective_gpu_count:
|
||||||
continue
|
continue
|
||||||
|
if tp_dp_product > effective_gpu_count:
|
||||||
|
continue
|
||||||
for ep in ep_values:
|
for ep in ep_values:
|
||||||
enable_ep = current_enable_ep or ep > 1
|
enable_ep = current_enable_ep or ep > 1
|
||||||
if constraints is not None:
|
if constraints is not None:
|
||||||
@@ -1330,6 +1333,18 @@ def _normalized_topology_flags(flags: dict[str, Any]) -> dict[str, Any]:
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
def _effective_gpu_count(study: StudySpec) -> int:
|
||||||
|
visible = str(study.engine.base_envs.get("CUDA_VISIBLE_DEVICES") or "").strip()
|
||||||
|
if not visible:
|
||||||
|
return study.hardware.gpu_count
|
||||||
|
if visible.lower() in {"none", "void", "-1"}:
|
||||||
|
return 0
|
||||||
|
devices = [item.strip() for item in visible.split(",") if item.strip()]
|
||||||
|
if not devices:
|
||||||
|
return study.hardware.gpu_count
|
||||||
|
return min(study.hardware.gpu_count, len(devices))
|
||||||
|
|
||||||
|
|
||||||
def _score_topology_candidate(
|
def _score_topology_candidate(
|
||||||
top_bottleneck: str,
|
top_bottleneck: str,
|
||||||
bottleneck_hypotheses: list[dict[str, Any]],
|
bottleneck_hypotheses: list[dict[str, Any]],
|
||||||
@@ -1621,6 +1636,7 @@ def _round_up_to_multiple(value: int, multiple: int) -> int:
|
|||||||
|
|
||||||
def _next_allowed_tp(study: StudySpec, *, current_tp: int, current_dp: int) -> int | None:
|
def _next_allowed_tp(study: StudySpec, *, current_tp: int, current_dp: int) -> int | None:
|
||||||
constraints = study.engine.topology_constraints
|
constraints = study.engine.topology_constraints
|
||||||
|
effective_gpu_count = _effective_gpu_count(study)
|
||||||
if constraints is not None and constraints.allowed_tensor_parallel_sizes:
|
if constraints is not None and constraints.allowed_tensor_parallel_sizes:
|
||||||
candidates = sorted({int(item) for item in constraints.allowed_tensor_parallel_sizes})
|
candidates = sorted({int(item) for item in constraints.allowed_tensor_parallel_sizes})
|
||||||
else:
|
else:
|
||||||
@@ -1629,7 +1645,7 @@ def _next_allowed_tp(study: StudySpec, *, current_tp: int, current_dp: int) -> i
|
|||||||
if candidate <= current_tp:
|
if candidate <= current_tp:
|
||||||
continue
|
continue
|
||||||
tp_dp_product = candidate * current_dp
|
tp_dp_product = candidate * current_dp
|
||||||
if tp_dp_product > study.hardware.gpu_count:
|
if tp_dp_product > effective_gpu_count:
|
||||||
continue
|
continue
|
||||||
if constraints is not None:
|
if constraints is not None:
|
||||||
if (
|
if (
|
||||||
@@ -1639,7 +1655,7 @@ def _next_allowed_tp(study: StudySpec, *, current_tp: int, current_dp: int) -> i
|
|||||||
continue
|
continue
|
||||||
if (
|
if (
|
||||||
constraints.require_tp_dp_product_equals_gpu_count
|
constraints.require_tp_dp_product_equals_gpu_count
|
||||||
and tp_dp_product != study.hardware.gpu_count
|
and tp_dp_product != effective_gpu_count
|
||||||
):
|
):
|
||||||
continue
|
continue
|
||||||
return candidate
|
return candidate
|
||||||
|
|||||||
@@ -1175,6 +1175,100 @@ class CoreFlowTests(unittest.TestCase):
|
|||||||
self.assertIsNotNone(proposal)
|
self.assertIsNotNone(proposal)
|
||||||
self.assertEqual(proposal.config_patch.flag_patch, {"tensor-parallel-size": 2})
|
self.assertEqual(proposal.config_patch.flag_patch, {"tensor-parallel-size": 2})
|
||||||
|
|
||||||
|
def test_harness_excludes_topology_above_visible_gpu_count(self) -> None:
|
||||||
|
with tempfile.TemporaryDirectory() as tmp:
|
||||||
|
tmp_path = Path(tmp)
|
||||||
|
study_path = _write_study_assets(
|
||||||
|
tmp_path,
|
||||||
|
engine_overrides={
|
||||||
|
"base_envs": {"CUDA_VISIBLE_DEVICES": "0,1,2,4,5,6,7"},
|
||||||
|
"tunable_flags": ["tensor-parallel-size"],
|
||||||
|
"topology_constraints": {
|
||||||
|
"allowed_tensor_parallel_sizes": [1, 2, 4, 8],
|
||||||
|
"allowed_tp_dp_products": [1, 2, 4, 8],
|
||||||
|
},
|
||||||
|
},
|
||||||
|
)
|
||||||
|
result_path = tmp_path / "trial-0003.json"
|
||||||
|
result_path.write_text(
|
||||||
|
json.dumps(
|
||||||
|
{
|
||||||
|
"status": "completed",
|
||||||
|
"best_request_rate": 1.078,
|
||||||
|
"best_pass_rate": 0.958,
|
||||||
|
"probes": [
|
||||||
|
{
|
||||||
|
"threshold": 0.039,
|
||||||
|
"feasible": False,
|
||||||
|
"payload": {
|
||||||
|
"request_count": 100,
|
||||||
|
"pass_rate": 0.8,
|
||||||
|
"request_rate": 1.10,
|
||||||
|
"early_stop_reason": "slo_pass_rate_unrecoverable",
|
||||||
|
"latency_summary": {
|
||||||
|
"failed_reason_counts": {"tpot_ms>25.0": 20}
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
],
|
||||||
|
}
|
||||||
|
),
|
||||||
|
encoding="utf-8",
|
||||||
|
)
|
||||||
|
study = load_study_spec(study_path)
|
||||||
|
context = build_harness_context(
|
||||||
|
study=study,
|
||||||
|
window_summary={"prompt_tokens_p95": 7628, "prompt_tail_ratio_p95_p50": 3.8},
|
||||||
|
state=StudyState(
|
||||||
|
study_id=study.study_id,
|
||||||
|
best_trial_id="trial-0003",
|
||||||
|
best_request_rate=1.078,
|
||||||
|
best_request_rate_per_gpu=0.2695,
|
||||||
|
trials=[
|
||||||
|
TrialSummary(
|
||||||
|
trial_id="trial-0001",
|
||||||
|
status="completed",
|
||||||
|
best_request_rate=0.065,
|
||||||
|
best_request_rate_per_gpu=0.065,
|
||||||
|
config_patch={"env_patch": {}, "flag_patch": {}},
|
||||||
|
),
|
||||||
|
TrialSummary(
|
||||||
|
trial_id="trial-0002",
|
||||||
|
status="completed",
|
||||||
|
best_request_rate=0.398,
|
||||||
|
best_request_rate_per_gpu=0.199,
|
||||||
|
config_patch={
|
||||||
|
"env_patch": {},
|
||||||
|
"flag_patch": {"tensor-parallel-size": 2},
|
||||||
|
},
|
||||||
|
),
|
||||||
|
TrialSummary(
|
||||||
|
trial_id="trial-0003",
|
||||||
|
status="completed",
|
||||||
|
best_request_rate=1.078,
|
||||||
|
best_request_rate_per_gpu=0.2695,
|
||||||
|
result_path=str(result_path),
|
||||||
|
config_patch={
|
||||||
|
"env_patch": {},
|
||||||
|
"flag_patch": {"tensor-parallel-size": 4},
|
||||||
|
},
|
||||||
|
),
|
||||||
|
],
|
||||||
|
),
|
||||||
|
)
|
||||||
|
candidates = context["candidate_actions"]
|
||||||
|
self.assertFalse(
|
||||||
|
any(
|
||||||
|
action["config_patch"]["flag_patch"].get("tensor-parallel-size") == 8
|
||||||
|
for action in candidates
|
||||||
|
)
|
||||||
|
)
|
||||||
|
proposal = build_harness_guided_proposal(context)
|
||||||
|
self.assertTrue(
|
||||||
|
proposal is None
|
||||||
|
or proposal.config_patch.flag_patch.get("tensor-parallel-size") != 8
|
||||||
|
)
|
||||||
|
|
||||||
def test_harness_stop_blocked_until_slo_driven_topology_frontier_is_measured(self) -> None:
|
def test_harness_stop_blocked_until_slo_driven_topology_frontier_is_measured(self) -> None:
|
||||||
with tempfile.TemporaryDirectory() as tmp:
|
with tempfile.TemporaryDirectory() as tmp:
|
||||||
tmp_path = Path(tmp)
|
tmp_path = Path(tmp)
|
||||||
|
|||||||
Reference in New Issue
Block a user