Skip to content

Commit 45240d8

Browse files
authored
refactor(config): remove models field from vLLM endpoints (#413)
* refactor(config): remove models field from vLLM endpoints Remove the redundant models field from VLLMEndpoint configuration. Model-to-endpoint mapping is now solely determined by the preferred_endpoints field in model_config, eliminating the need for bidirectional association. Changes: - Remove Models field from VLLMEndpoint struct - Update GetEndpointsForModel to use only preferred_endpoints - Update GetAllModels to retrieve models from model_config keys - Update all configuration files to remove models field - Update all tests to reflect the new configuration structure - Update TypeScript interface in dashboard frontend This simplifies the configuration and removes potential inconsistencies between models and preferred_endpoints. Signed-off-by: bitliu <bitliu@tencent.com> * more Signed-off-by: bitliu <bitliu@tencent.com> * more Signed-off-by: bitliu <bitliu@tencent.com> --------- Signed-off-by: bitliu <bitliu@tencent.com>
1 parent c35e98a commit 45240d8

26 files changed

+105
-148
lines changed

config/config-mcp-classifier-example.yaml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -75,8 +75,6 @@ vllm_endpoints:
7575
- name: endpoint1
7676
address: 127.0.0.1
7777
port: 8000
78-
models:
79-
- openai/gpt-oss-20b
8078
weight: 1
8179
health_check_path: /health
8280

config/config.development.yaml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -29,8 +29,6 @@ vllm_endpoints:
2929
- name: "local-endpoint"
3030
address: "127.0.0.1"
3131
port: 8000
32-
models:
33-
- "test-model"
3432
weight: 1
3533

3634
model_config:

config/config.e2e.yaml

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -42,15 +42,11 @@ vllm_endpoints:
4242
- name: "qwen-endpoint"
4343
address: "127.0.0.1"
4444
port: 8000
45-
models:
46-
- "Model-A"
4745
weight: 1
4846
health_check_path: "/health"
4947
- name: "tinyllama-endpoint"
5048
address: "127.0.0.1"
5149
port: 8001
52-
models:
53-
- "Model-B"
5450
weight: 1
5551
health_check_path: "/health"
5652

config/config.production.yaml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -34,8 +34,6 @@ vllm_endpoints:
3434
- name: "endpoint1"
3535
address: "127.0.0.1"
3636
port: 8000
37-
models:
38-
- "openai/gpt-oss-20b"
3937
weight: 1
4038

4139
model_config:

config/config.recipe-accuracy.yaml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -44,8 +44,6 @@ vllm_endpoints:
4444
- name: "endpoint1"
4545
address: "127.0.0.1"
4646
port: 8000
47-
models:
48-
- "openai/gpt-oss-20b"
4947
weight: 1
5048

5149
model_config:

config/config.recipe-latency.yaml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -39,8 +39,6 @@ vllm_endpoints:
3939
- name: "endpoint1"
4040
address: "127.0.0.1"
4141
port: 8000
42-
models:
43-
- "openai/gpt-oss-20b"
4442
weight: 1
4543

4644
model_config:

config/config.recipe-token-efficiency.yaml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -44,8 +44,6 @@ vllm_endpoints:
4444
- name: "endpoint1"
4545
address: "127.0.0.1"
4646
port: 8000
47-
models:
48-
- "openai/gpt-oss-20b"
4947
weight: 1
5048

5149
model_config:

config/config.testing.yaml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -30,8 +30,6 @@ vllm_endpoints:
3030
- name: "mock"
3131
address: "172.28.0.10"
3232
port: 8000
33-
models:
34-
- "openai/gpt-oss-20b"
3533
weight: 1
3634
health_check_path: "/health"
3735

config/config.yaml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -34,8 +34,6 @@ vllm_endpoints:
3434
- name: "endpoint1"
3535
address: "127.0.0.1" # IPv4 address - REQUIRED format
3636
port: 8000
37-
models:
38-
- "openai/gpt-oss-20b"
3937
weight: 1
4038

4139
model_config:

config/examples/system_prompt_example.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -100,9 +100,9 @@ default_reasoning_effort: medium
100100
# vLLM endpoints configuration
101101
vllm_endpoints:
102102
- name: "mock"
103-
address: "http://127.0.0.1:8000"
104-
models:
105-
- "openai/gpt-oss-20b"
103+
address: "127.0.0.1"
104+
port: 8000
105+
weight: 1
106106

107107
# Usage Notes:
108108
# 1. System prompts are automatically injected based on query classification

0 commit comments

Comments
 (0)