diff --git a/config/manifests/inferencemodel.yaml b/config/manifests/inferencemodel.yaml index 4c7824ca3..9d2c9bbd5 100644 --- a/config/manifests/inferencemodel.yaml +++ b/config/manifests/inferencemodel.yaml @@ -1,7 +1,7 @@ apiVersion: inference.networking.x-k8s.io/v1alpha2 kind: InferenceModel metadata: - name: inferencemodel-sample + name: tweet-summarizer spec: modelName: tweet-summary criticality: Critical @@ -15,7 +15,7 @@ spec: apiVersion: inference.networking.x-k8s.io/v1alpha2 kind: InferenceModel metadata: - name: inferencemodel-base-model + name: base-model spec: modelName: meta-llama/Llama-2-7b-hf criticality: Critical @@ -26,7 +26,7 @@ spec: apiVersion: inference.networking.x-k8s.io/v1alpha2 kind: InferenceModel metadata: - name: inferencemodel-base-model-cpu + name: base-model-cpu spec: modelName: Qwen/Qwen2.5-1.5B-Instruct criticality: Critical diff --git a/config/samples/gateway_v1alpha1_inferencemodel.yaml b/config/samples/gateway_v1alpha1_inferencemodel.yaml index f1f46a2f1..34ea06803 100644 --- a/config/samples/gateway_v1alpha1_inferencemodel.yaml +++ b/config/samples/gateway_v1alpha1_inferencemodel.yaml @@ -4,12 +4,12 @@ metadata: labels: app.kubernetes.io/name: api app.kubernetes.io/managed-by: kustomize - name: inferencemodel-sample + name: sample-sql-assist spec: criticality: Critical modelName: sql-code-assist poolRef: - name: inferencepool-sample + name: vllm-llama-31-8b-sample-pool targetModels: - name: npc-bot-v1 weight: 50 diff --git a/config/samples/gateway_v1alpha1_inferencepool.yaml b/config/samples/gateway_v1alpha1_inferencepool.yaml index 42ac62965..4993d7864 100644 --- a/config/samples/gateway_v1alpha1_inferencepool.yaml +++ b/config/samples/gateway_v1alpha1_inferencepool.yaml @@ -4,7 +4,7 @@ metadata: labels: app.kubernetes.io/name: api app.kubernetes.io/managed-by: kustomize - name: inferencepool-sample + name: vllm-llama-31-8b-sample-pool spec: selector: app: npc-bot diff --git a/test/testdata/inferencepool-with-model-hermetic.yaml b/test/testdata/inferencepool-with-model-hermetic.yaml index 36b6e539d..26c057c24 100644 --- a/test/testdata/inferencepool-with-model-hermetic.yaml +++ b/test/testdata/inferencepool-with-model-hermetic.yaml @@ -13,7 +13,7 @@ spec: apiVersion: inference.networking.x-k8s.io/v1alpha2 kind: InferenceModel metadata: - name: inferencemodel-sample + name: sample namespace: default spec: modelName: sql-lora @@ -27,7 +27,7 @@ spec: apiVersion: inference.networking.x-k8s.io/v1alpha2 kind: InferenceModel metadata: - name: inferencemodel-sheddable + name: sheddable namespace: default spec: modelName: sql-lora-sheddable @@ -40,7 +40,7 @@ spec: apiVersion: inference.networking.x-k8s.io/v1alpha2 kind: InferenceModel metadata: - name: inferencemodel-generic + name: generic namespace: default spec: modelName: my-model @@ -54,7 +54,7 @@ spec: apiVersion: inference.networking.x-k8s.io/v1alpha2 kind: InferenceModel metadata: - name: inferencemodel-direct-model-name + name: direct-model-name namespace: default spec: modelName: direct-model