Skip to content

Commit ca2ffc0

Browse files
authored
Merge pull request #142 from stackhpc/vllm-omni-2
Add vllm-omni backend support
2 parents 79e24d0 + 24a9fea commit ca2ffc0

File tree

8 files changed

+43
-4
lines changed

8 files changed

+43
-4
lines changed

charts/azimuth-chat/azimuth-ui.schema.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,7 @@ sortOrder:
2424
- /azimuth-llm/api/azimuthNodeGroupSelector
2525
- /azimuth-llm/ui/appSettings/model_instruction
2626
- /azimuth-llm/ui/appSettings/page_title
27+
- /azimuth-llm/api/image/containerImage
2728
- /azimuth-llm/api/image/version
2829
- /azimuth-llm/ui/appSettings/llm_params/temperature
2930
- /azimuth-llm/ui/appSettings/llm_params/max_tokens

charts/azimuth-chat/values.schema.json

Lines changed: 11 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,10 +36,20 @@
3636
"image": {
3737
"type": "object",
3838
"properties": {
39+
"containerImage": {
40+
"type": "string",
41+
"title": "vLLM Container Image",
42+
"description": "Container to use as API backend. Currently only 'vllm/vllm-openai' and 'vllm/vllm-omni' are supported. Default is 'vllm/vllm-openai'.",
43+
"default": "vllm/vllm-openai",
44+
"enum": [
45+
"vllm/vllm-openai",
46+
"vllm/vllm-omni"
47+
]
48+
},
3949
"version": {
4050
"type": "string",
4151
"title": "Backend vLLM version",
42-
"description": "The vLLM version to use as a backend. Must be a version tag from [this list](https://github.com/vllm-project/vllm/tags)",
52+
"description": "The vLLM version to use as a backend. Must be a version tag from [this list](https://github.com/vllm-project/vllm/tags) when using vllm-openai or [this list](https://github.com/vllm-project/vllm-omni/tags) when using vllm-omni.",
4353
"default": "v0.11.0"
4454
}
4555
}

charts/azimuth-image-analysis/azimuth-ui.schema.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,7 @@ sortOrder:
2323
- /azimuth-llm/huggingface/token
2424
- /azimuth-llm/api/azimuthNodeGroupSelector
2525
- /azimuth-llm/ui/appSettings/page_title
26+
- /azimuth-llm/api/image/containerImage
2627
- /azimuth-llm/api/image/version
2728
- /azimuth-llm/ui/appSettings/llm_params/temperature
2829
- /azimuth-llm/ui/appSettings/llm_params/max_tokens

charts/azimuth-image-analysis/values.schema.json

Lines changed: 11 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -32,10 +32,20 @@
3232
"image": {
3333
"type": "object",
3434
"properties": {
35+
"containerImage": {
36+
"type": "string",
37+
"title": "vLLM Container Image",
38+
"description": "Container to use as API backend. Currently only 'vllm/vllm-openai' and 'vllm/vllm-omni' are supported. Default is 'vllm/vllm-openai'.",
39+
"default": "vllm/vllm-openai",
40+
"enum": [
41+
"vllm/vllm-openai",
42+
"vllm/vllm-omni"
43+
]
44+
},
3545
"version": {
3646
"type": "string",
3747
"title": "Backend vLLM version",
38-
"description": "The vLLM version to use as a backend. Must be a version tag from [this list](https://github.com/vllm-project/vllm/tags)",
48+
"description": "The vLLM version to use as a backend. Must be a version tag from [this list](https://github.com/vllm-project/vllm/tags) when using vllm-openai or [this list](https://github.com/vllm-project/vllm-omni/tags) when using vllm-omni.",
3949
"default": "v0.11.0"
4050
}
4151
}

charts/azimuth-llm/azimuth-ui.schema.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,7 @@ sortOrder:
2424
- /api/azimuthNodeGroupSelector
2525
- /ui/appSettings/model_instruction
2626
- /ui/appSettings/page_title
27+
- /api/image/containerImage
2728
- /api/image/version
2829
- /ui/appSettings/llm_params/temperature
2930
- /ui/appSettings/llm_params/max_tokens

charts/azimuth-llm/templates/api/deployment.yml

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,12 @@ spec:
2424
{{- else if .Values.api.intelXPUsEnabled }}
2525
image: "ghcr.io/stackhpc/vllm-xpu:{{ .Values.api.image.version }}"
2626
{{- else }}
27-
image: "vllm/vllm-openai:{{ .Values.api.image.version }}"
27+
image: "{{ .Values.api.image.containerImage }}:{{ .Values.api.image.version }}"
28+
{{- end }}
29+
{{- if eq .Values.api.image.containerImage "vllm/vllm-omni" }}
30+
command:
31+
- vllm
32+
- serve
2833
{{- end }}
2934
ports:
3035
- name: api

charts/azimuth-llm/values.schema.json

Lines changed: 11 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,10 +33,20 @@
3333
"image": {
3434
"type": "object",
3535
"properties": {
36+
"containerImage": {
37+
"type": "string",
38+
"title": "vLLM Container Image",
39+
"description": "Container to use as API backend. Currently only 'vllm/vllm-openai' and 'vllm/vllm-omni' are supported. Default is 'vllm/vllm-openai'.",
40+
"default": "vllm/vllm-openai",
41+
"enum": [
42+
"vllm/vllm-openai",
43+
"vllm/vllm-omni"
44+
]
45+
},
3646
"version": {
3747
"type": "string",
3848
"title": "Backend vLLM version",
39-
"description": "The vLLM version to use as a backend. Must be a version tag from [this list](https://github.com/vllm-project/vllm/tags)",
49+
"description": "The vLLM version to use as a backend. Must be a version tag from [this list](https://github.com/vllm-project/vllm/tags) when using vllm-openai or [this list](https://github.com/vllm-project/vllm-omni/tags) when using vllm-omni.",
4050
"default": "v0.11.0"
4151
}
4252
}

charts/azimuth-llm/values.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -37,6 +37,7 @@ api:
3737
# ghcr.io/stackhpc/vllm-xpu when api.gpus > 0 and intelXPUsEnabled is true,
3838
# or ghcr.io/stackhpc/vllm-cpu when api.gpus == 0
3939
repository:
40+
containerImage: vllm/vllm-openai
4041
version: v0.11.0
4142
monitoring:
4243
enabled: true

0 commit comments

Comments
 (0)