Skip to content

Commit d2eeec1

Browse files
committed
more
Signed-off-by: bitliu <[email protected]>
1 parent d26e4d5 commit d2eeec1

File tree

5 files changed

+19
-25
lines changed

5 files changed

+19
-25
lines changed

website/docs/installation/configuration.md

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -159,10 +159,13 @@ Configure your LLM servers:
159159
vllm_endpoints:
160160
- name: "my_endpoint"
161161
address: "127.0.0.1" # Your server IP - MUST be IP address format
162-
port: 8000 # Your server port
163-
models:
164-
- "llama2-7b" # Model name - must match vLLM --served-model-name
165-
weight: 1 # Load balancing weight
162+
port: 8000 # Your server port
163+
weight: 1 # Load balancing weight
164+
165+
# Model configuration - maps models to endpoints
166+
model_config:
167+
"llama2-7b": # Model name - must match vLLM --served-model-name
168+
preferred_endpoints: ["my_endpoint"]
166169
```
167170
168171
#### Address Format Requirements
@@ -204,11 +207,12 @@ The model names in the `models` array must **exactly match** the `--served-model
204207
# vLLM server command:
205208
vllm serve meta-llama/Llama-2-7b-hf --served-model-name llama2-7b
206209
207-
# config.yaml must use the same name:
208-
vllm_endpoints:
209-
- models: ["llama2-7b"] # ✅ Matches --served-model-name
210-
210+
# config.yaml must reference the model in model_config:
211211
model_config:
212+
"llama2-7b": # ✅ Matches --served-model-name
213+
preferred_endpoints: ["your-endpoint"]
214+
215+
vllm_endpoints:
212216
"llama2-7b": # ✅ Matches --served-model-name
213217
# ... configuration
214218
```
@@ -683,12 +687,10 @@ vllm_endpoints:
683687
- name: "math_endpoint"
684688
address: "192.168.1.10" # Math server IP
685689
port: 8000
686-
models: ["math-model"]
687690
weight: 1
688691
- name: "general_endpoint"
689692
address: "192.168.1.20" # General server IP
690693
port: 8000
691-
models: ["general-model"]
692694
weight: 1
693695
694696
categories:
@@ -711,12 +713,10 @@ vllm_endpoints:
711713
- name: "endpoint1"
712714
address: "192.168.1.30" # Primary server IP
713715
port: 8000
714-
models: ["my-model"]
715716
weight: 2 # Higher weight = more traffic
716717
- name: "endpoint2"
717718
address: "192.168.1.31" # Secondary server IP
718719
port: 8000
719-
models: ["my-model"]
720720
weight: 1
721721
```
722722

website/docs/installation/installation.md

Lines changed: 7 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -109,13 +109,11 @@ Edit `config/config.yaml` to point to your LLM endpoints:
109109
vllm_endpoints:
110110
- name: "your-endpoint"
111111
address: "127.0.0.1" # MUST be IP address (IPv4 or IPv6)
112-
port: 11434 # Replace with your port
113-
models:
114-
- "your-model-name" # Replace with your model
112+
port: 11434 # Replace with your port
115113
weight: 1
116114

117115
model_config:
118-
"your-model-name":
116+
"your-model-name": # Replace with your model name
119117
pii_policy:
120118
allow_by_default: false # Deny all PII by default
121119
pii_types_allowed: ["EMAIL_ADDRESS", "PERSON", "GPE", "PHONE_NUMBER"] # Only allow these specific PII types
@@ -146,11 +144,12 @@ The model name in your configuration **must exactly match** the `--served-model-
146144
# When starting vLLM server:
147145
vllm serve microsoft/phi-4 --port 11434 --served-model-name your-model-name
148146
149-
# The config.yaml must use the same name:
150-
vllm_endpoints:
151-
- models: ["your-model-name"] # ✅ Must match --served-model-name
152-
147+
# The config.yaml must reference the model in model_config:
153148
model_config:
149+
"your-model-name": # ✅ Must match --served-model-name
150+
preferred_endpoints: ["your-endpoint"]
151+
152+
vllm_endpoints:
154153
"your-model-name": # ✅ Must match --served-model-name
155154
# ... configuration
156155
```

website/docs/training/model-performance-eval.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -69,11 +69,9 @@ vllm_endpoints:
6969
- name: "endpoint1"
7070
address: "127.0.0.1"
7171
port: 11434
72-
models: ["phi4"] # ✅ Matches --served_model_name phi4
7372
- name: "endpoint2"
7473
address: "127.0.0.1"
7574
port: 11435
76-
models: ["qwen3-0.6B"] # ✅ Matches --served_model_name qwen3-0.6B
7775

7876
model_config:
7977
"phi4": # ✅ Matches --served_model_name phi4

website/docs/tutorials/content-safety/pii-detection.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -61,11 +61,9 @@ vllm_endpoints:
6161
- name: secure-model
6262
address: "127.0.0.1"
6363
port: 8080
64-
models: ["secure-llm"]
6564
- name: general-model
6665
address: "127.0.0.1"
6766
port: 8081
68-
models: ["general-llm"]
6967

7068
# Model-specific configurations
7169
model_config:

website/docs/tutorials/intelligent-route/reasoning.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,6 @@ vllm_endpoints:
3434
- name: "endpoint1"
3535
address: "127.0.0.1"
3636
port: 8000
37-
models: ["deepseek-v31", "qwen3-30b", "openai/gpt-oss-20b"] # Must match --served-model-name
3837
weight: 1
3938

4039
# Reasoning family configurations (how to express reasoning for a family)

0 commit comments

Comments
 (0)