1
0
Fork 0

fix issue with http_proxy

This commit is contained in:
Massaki Archambault 2024-02-21 22:45:35 -05:00
parent b5f804ce04
commit e2bbf8585e
3 changed files with 14 additions and 5 deletions

5
.env
View File

@ -1,3 +1,8 @@
# If set, HTTP_PROXY messes with inter-container communication in the deployment.
# Ollama downloads the models via https anyway so it should be safe to unset it
HTTP_PROXY=
http_proxy=
#=============================================================#
# LibreChat Configuration #
#=============================================================#

View File

@ -5,6 +5,11 @@ services:
command: --config /config.yaml
ports:
- 8000:8000
env_file:
- .env
environment:
- HOST=0.0.0.0
- PORT=8000
volumes:
- ./litellm/config.yaml:/config.yaml:ro
@ -33,6 +38,8 @@ services:
volumes:
- librechat_mongodb_data:/data/db
command: mongod --noauth
env_file:
- .env
meilisearch:
image: getmeili/meilisearch:v1.5
restart: unless-stopped
@ -40,6 +47,8 @@ services:
environment:
- MEILI_HOST=http://meilisearch:7700
- MEILI_NO_ANALYTICS=true
env_file:
- .env
volumes:
- librechat_meilisearch_data:/meili_data

View File

@ -1,10 +1,5 @@
#!/bin/bash -x
# Ollama has trouble handling HTTP_PROXY
# https://github.com/ollama/ollama/issues/2168
unset HTTP_PROXY
unset http_proxy
ollama serve &
sleep 1