Skip to content

Commit 057d176

Browse files
committed
updated config map and lora syncer init container
Signed-off-by: Nir Rozenbaum <[email protected]>
1 parent 1d2bedc commit 057d176

File tree

2 files changed

+31
-1
lines changed

2 files changed

+31
-1
lines changed

config/manifests/vllm/cpu-deployment.yaml

Lines changed: 31 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -73,6 +73,19 @@ spec:
7373
name: shm
7474
- name: adapters
7575
mountPath: "/adapters"
76+
initContainers:
77+
- name: lora-adapter-syncer
78+
tty: true
79+
stdin: true
80+
image: us-central1-docker.pkg.dev/k8s-staging-images/gateway-api-inference-extension/lora-syncer:main
81+
restartPolicy: Always
82+
imagePullPolicy: Always
83+
env:
84+
- name: DYNAMIC_LORA_ROLLOUT_CONFIG
85+
value: "/config/configmap.yaml"
86+
volumeMounts: # DO NOT USE subPath, dynamic configmap updates don't work on subPaths
87+
- name: config-volume
88+
mountPath: /config
7689
restartPolicy: Always
7790
schedulerName: default-scheduler
7891
terminationGracePeriodSeconds: 30
@@ -84,3 +97,21 @@ spec:
8497
medium: Memory
8598
- name: adapters
8699
emptyDir: {}
100+
- name: config-volume
101+
configMap:
102+
name: vllm-qwen-adapters
103+
---
104+
apiVersion: v1
105+
kind: ConfigMap
106+
metadata:
107+
name: vllm-qwen-adapters
108+
data:
109+
configmap.yaml: |
110+
vLLMLoRAConfig:
111+
name: vllm-llama2-7b
112+
port: 8000
113+
ensureExist:
114+
models:
115+
- base-model: Qwen/Qwen2.5-1.5B
116+
id: tweet-summary-1
117+
source: SriSanth2345/Qwen-1.5B-Tweet-Generations

site-src/guides/index.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,6 @@ This quickstart guide is intended for engineers familiar with k8s and model serv
3535
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/config/manifests/vllm/gpu-deployment.yaml
3636
```
3737

38-
3938
=== "CPU-Based Model Server"
4039

4140
This setup is using the formal `vllm-cpu` image, which according to the documentation can run vLLM on x86 CPU platform.

0 commit comments

Comments
 (0)