1
0
Fork 0

add ingress for ollama and remote litellm

This commit is contained in:
Massaki Archambault 2024-08-07 21:08:02 -04:00
parent c175883985
commit 2806142962
7 changed files with 61 additions and 19 deletions

View File

@ -1,7 +1,26 @@
resources: resources:
- ollama-deployment.yaml - ollama-deployment.yaml
- ollama-ingress.yaml
namePrefix: ollama- namePrefix: ollama-
commonLabels: commonLabels:
app.kubernetes.io/name: ollama app.kubernetes.io/name: ollama
configMapGenerator:
- name: kustomize-generated-config
literals:
- OLLAMA_EXTERNAL_HOST=ollama.badjnet.home
- OLLAMA_EXTERNAL_URL=http://ollama.badjnet.home
replacements:
- source:
kind: ConfigMap
name: kustomize-generated-config
fieldPath: data.OLLAMA_EXTERNAL_HOST
targets:
- select:
kind: Ingress
name: server
fieldPaths:
- spec.rules.0.host

View File

@ -30,7 +30,9 @@ spec:
- name: server - name: server
image: ollama/ollama image: ollama/ollama
# image: badjware/ollama-tweak # image: badjware/ollama-tweak
# env: env:
- name: OLLAMA_KEEP_ALIVE
value: 36h
# - name: HSA_OVERRIDE_GFX_VERSION # - name: HSA_OVERRIDE_GFX_VERSION
# value: 10.1.0 # value: 10.1.0
# - name: HSA_ENABLE_SDMA # - name: HSA_ENABLE_SDMA

View File

@ -0,0 +1,19 @@
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:
name: server
labels:
app.kubernetes.io/name: ollama
probe: blackbox-http
spec:
rules:
- host: ${OLLAMA_EXTERNAL_HOST}
http:
paths:
- path: /
pathType: Prefix
backend:
service:
name: server
port:
name: http

View File

@ -17,6 +17,7 @@ spec:
containers: containers:
- name: server - name: server
image: ghcr.io/open-webui/open-webui:main image: ghcr.io/open-webui/open-webui:main
tty: true
env: env:
- name: NAMESPACE - name: NAMESPACE
valueFrom: valueFrom:
@ -24,10 +25,16 @@ spec:
fieldPath: metadata.namespace fieldPath: metadata.namespace
- name: OLLAMA_BASE_URL - name: OLLAMA_BASE_URL
value: http://ollama-server.$(NAMESPACE).svc:11434 value: http://ollama-server.$(NAMESPACE).svc:11434
- name: RAG_EMBEDDING_ENGINE
value: ollama
- name: AUDIO_STT_ENGINE
value: openai
# - name: GLOBAL_LOG_LEVEL
# value: DEBUG
resources: resources:
requests: requests:
cpu: 250m cpu: 250m
memory: 1Gi memory: 500Mi
limits: limits:
cpu: 500m cpu: 500m
memory: 1Gi memory: 1Gi

View File

@ -1,13 +1,5 @@
model_list: model_list:
- model_name: llama3-8b - model_name: "*"
litellm_params: litellm_params:
model: ollama_chat/llama3 model: "ollama_chat/*"
api_base: http://ollama-server:11434 api_base: http://ollama-server:11434
- model_name: gemma2-9b
litellm_params:
model: ollama_chat/gemma2
api_base: http://ollama-server:11434
- model_name: phi3-3b
litellm_params:
model: ollama_chat/phi3:mini
api_base: http://ollama-server:11434

View File

@ -54,15 +54,18 @@ images:
newTag: "1.29" newTag: "1.29"
- name: ollama/ollama - name: ollama/ollama
newTag: 0.3.3 newTag: 0.3.3
# newTag: 0.3.3-rocm - name: ghcr.io/berriai/litellm
newTag: main-v1.43.1
- name: ghcr.io/open-webui/open-webui
newTag: v0.3.10 # https://github.com/open-webui/open-webui/discussions/4331
configMapGenerator: configMapGenerator:
- name: litellm-proxy-config # - name: litellm-proxy-config
namespace: llm # namespace: llm
behavior: replace # behavior: replace
files: # files:
- config.yaml=./configurations/litellm/config.yaml # - config.yaml=./configurations/litellm/config.yaml
# - name: home-assistant-server-config # - name: home-assistant-server-config
# namespace: home-assistant # namespace: home-assistant
# behavior: replace # behavior: replace

View File

@ -1,7 +1,7 @@
resources: resources:
- namespace.yaml - namespace.yaml
- ../../bases/ollama - ../../bases/ollama
- ../../bases/litellm # - ../../bases/litellm
- ../../bases/openwebui - ../../bases/openwebui
- ../../bases/sillytavern - ../../bases/sillytavern