Skip to content

Commit

Permalink
Updates artifacts for v0.1.0 release
Browse files Browse the repository at this point in the history
Signed-off-by: Kellen Swain <[email protected]>
  • Loading branch information
kfswain committed Feb 6, 2025
1 parent 056adfd commit 0b6b6eb
Show file tree
Hide file tree
Showing 3 changed files with 3 additions and 3 deletions.
2 changes: 1 addition & 1 deletion pkg/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@ This quickstart guide is intended for engineers familiar with k8s and model serv
1. **Install the Inference Extension CRDs:**

```sh
kubectl apply -k https://github.com/kubernetes-sigs/gateway-api-inference-extension/config/crd
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/releases/download/v0.1.0/manifests.yaml
```

1. **Deploy InferenceModel**
Expand Down
2 changes: 1 addition & 1 deletion pkg/manifests/ext_proc.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -71,7 +71,7 @@ spec:
spec:
containers:
- name: inference-gateway-ext-proc
image: us-central1-docker.pkg.dev/k8s-staging-images/gateway-api-inference-extension/epp:main
image: us-central1-docker.pkg.dev/k8s-staging-images/gateway-api-inference-extension/epp:v0.1.0
args:
- -poolName
- "vllm-llama2-7b-pool"
Expand Down
2 changes: 1 addition & 1 deletion pkg/manifests/vllm/deployment.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@ spec:
spec:
containers:
- name: lora
image: "vllm/vllm-openai:latest"
image: "vllm/vllm-openai:0.7.1"
imagePullPolicy: Always
command: ["python3", "-m", "vllm.entrypoints.openai.api_server"]
args:
Expand Down

0 comments on commit 0b6b6eb

Please sign in to comment.