Add codex and bailian LLM provider presets
This commit is contained in:
@@ -142,7 +142,7 @@
|
||||
"system_prompt": "Propose a single engine config patch that increases the maximum feasible sampling_u under the SLO target. Favor launch-safe changes grounded in the incumbent result and only propose knobs that plausibly improve throughput above the incumbent request rate.",
|
||||
"max_history_trials": 8,
|
||||
"endpoint": {
|
||||
"base_url": "http://tianx.ipads-lab.se.sjtu.edu.cn:8317/v1",
|
||||
"provider": "codex",
|
||||
"model": "gpt-5.4",
|
||||
"api_key_env": "OPENAI_API_KEY",
|
||||
"timeout_s": 180
|
||||
|
||||
Reference in New Issue
Block a user