Skip to content

Commit ff4f6fb

Browse files
committed
Revert name change to make pool name far more descriptive.
1 parent e014105 commit ff4f6fb

File tree

5 files changed

+13
-13
lines changed

5 files changed

+13
-13
lines changed

config/manifests/ext_proc.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -44,11 +44,11 @@ apiVersion: inference.networking.x-k8s.io/v1alpha2
4444
kind: InferencePool
4545
metadata:
4646
labels:
47-
name: my-pool
47+
name: vllm-llama2-7b-pool
4848
spec:
4949
targetPortNumber: 8000
5050
selector:
51-
app: my-pool
51+
app: vllm-llama2-7b-pool
5252
extensionRef:
5353
name: inference-gateway-ext-proc
5454
---
@@ -75,7 +75,7 @@ spec:
7575
imagePullPolicy: Always
7676
args:
7777
- -poolName
78-
- "my-pool"
78+
- "vllm-llama2-7b-pool"
7979
- -v
8080
- "4"
8181
- -grpcPort

config/manifests/inferencemodel.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ spec:
66
modelName: tweet-summary
77
criticality: Critical
88
poolRef:
9-
name: my-pool
9+
name: vllm-llama2-7b-pool
1010
targetModels:
1111
- name: tweet-summary-1
1212
weight: 100
@@ -20,7 +20,7 @@ spec:
2020
modelName: meta-llama/Llama-2-7b-hf
2121
criticality: Critical
2222
poolRef:
23-
name: my-pool
23+
name: vllm-llama2-7b-pool
2424

2525
---
2626
apiVersion: inference.networking.x-k8s.io/v1alpha2
@@ -31,4 +31,4 @@ spec:
3131
modelName: Qwen/Qwen2.5-1.5B-Instruct
3232
criticality: Critical
3333
poolRef:
34-
name: my-pool
34+
name: vllm-llama2-7b-pool

config/manifests/vllm/cpu-deployment.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,16 +1,16 @@
11
apiVersion: apps/v1
22
kind: Deployment
33
metadata:
4-
name: my-pool
4+
name: vllm-llama2-7b-pool
55
spec:
66
replicas: 3
77
selector:
88
matchLabels:
9-
app: my-pool
9+
app: vllm-llama2-7b-pool
1010
template:
1111
metadata:
1212
labels:
13-
app: my-pool
13+
app: vllm-llama2-7b-pool
1414
spec:
1515
containers:
1616
- name: lora

config/manifests/vllm/gpu-deployment.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,16 +1,16 @@
11
apiVersion: apps/v1
22
kind: Deployment
33
metadata:
4-
name: my-pool
4+
name: vllm-llama2-7b-pool
55
spec:
66
replicas: 3
77
selector:
88
matchLabels:
9-
app: my-pool
9+
app: vllm-llama2-7b-pool
1010
template:
1111
metadata:
1212
labels:
13-
app: my-pool
13+
app: vllm-llama2-7b-pool
1414
spec:
1515
containers:
1616
- name: lora

test/e2e/epp/e2e_suite_test.go

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -57,7 +57,7 @@ const (
5757
// TODO [danehans]: Must be "default" until https://github.com/kubernetes-sigs/gateway-api-inference-extension/issues/227 is fixed
5858
nsName = "default"
5959
// modelServerName is the name of the model server test resources.
60-
modelServerName = "my-pool"
60+
modelServerName = "vllm-llama2-7b-pool"
6161
// modelName is the test model name.
6262
modelName = "tweet-summary"
6363
// envoyName is the name of the envoy proxy test resources.

0 commit comments

Comments
 (0)