Skip to content

Commit 25a61d8

Browse files
liu-congahg-g
andauthored
Update site-src/guides/model-server.md
Co-authored-by: Abdullah Gharaibeh <[email protected]>
1 parent dd40239 commit 25a61d8

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

site-src/guides/model-server.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ Triton(TensorRT-LLM)| TODO| Pending [PR](https://github.com/triton-inference-ser
1616

1717
vLLM is configured as the default in the [endpoint picker extension](https://github.com/kubernetes-sigs/gateway-api-inference-extension/tree/main/pkg/epp). No further configuration is required.
1818

19-
## Use Triton with TensorRT-LLM Backend
19+
## Triton with TensorRT-LLM Backend
2020

2121
You need to specify the metric names when starting the EPP container. Add the following to the `args` of the [EPP deployment](https://github.com/kubernetes-sigs/gateway-api-inference-extension/blob/296247b07feed430458b8e0e3f496055a88f5e89/config/manifests/inferencepool.yaml#L48).
2222
```

0 commit comments

Comments
 (0)