1
0
Fork 0

Compare commits

..

2 Commits

Author SHA1 Message Date
Massaki Archambault 3b41d81357 add some comments to config files 2024-02-06 19:27:03 -05:00
Massaki Archambault 61e7c94537 fix typos 2024-02-06 19:16:13 -05:00
3 changed files with 11 additions and 4 deletions

6
.env
View File

@ -1,6 +1,8 @@
# HF_TOKEN= # https://github.com/ROCm/ROCm/issues/2625
PYTORCH_HIP_ALLOC_CONF=garbage_collection_threshold:0.9,max_split_size_mb:512 GPU_MAX_HW_QUEUES=1
# https://github.com/ROCm/ROCm/issues/2788#issuecomment-1915765846
HSA_OVERRIDE_GFX_VERSION=11.0.0 # GFX1101/Navi32 (RX 7800XT) HSA_OVERRIDE_GFX_VERSION=11.0.0 # GFX1101/Navi32 (RX 7800XT)
# PYTORCH_HIP_ALLOC_CONF=garbage_collection_threshold:0.9,max_split_size_mb:512
#=============================================================# #=============================================================#
# LibreChat Configuration # # LibreChat Configuration #

View File

@ -14,7 +14,7 @@ A quick prototype to self-host [LibreChat](https://github.com/danny-avila/LibreC
### Prerequisites ### Prerequisites
* Linux (WSL2 untested) * Linux (WSL2 is untested)
* An AMD 7xxx series GPU (technically optional, Ollama will fallback to using the CPU but it will be very slow. Other GPUS are supported but the deployment must be modified to use them) * An AMD 7xxx series GPU (technically optional, Ollama will fallback to using the CPU but it will be very slow. Other GPUS are supported but the deployment must be modified to use them)
* docker * docker
* docker-compose * docker-compose
@ -81,7 +81,7 @@ becomes:
## Architecture components ## Architecture components
* [LibreChat](https://github.com/danny-avila/LibreChat) is a ChatGPT clone with support. It is deployed alongside a [MongoDB](https://github.com/mongodb/mongo) database and [Meillisearch](https://github.com/meilisearch/meilisearch) for search. It is exposed on http://localhost:3080/. * [LibreChat](https://github.com/danny-avila/LibreChat) is a ChatGPT clone with support for multiple AI endpoints. It's deployed alongside a [MongoDB](https://github.com/mongodb/mongo) database and [Meillisearch](https://github.com/meilisearch/meilisearch) for search. It's exposed on http://localhost:3080/.
* [LiteLLM](https://github.com/BerriAI/litellm) is an OpenAI-like API. It is exposed on http://localhost:8000/ without any authentication by default. * [LiteLLM](https://github.com/BerriAI/litellm) is an OpenAI-like API. It is exposed on http://localhost:8000/ without any authentication by default.
* [Ollama](https://github.com/ollama/ollama) manages and serve the local models. * [Ollama](https://github.com/ollama/ollama) manages and serve the local models.

View File

@ -1,5 +1,6 @@
version: "3" version: "3"
services: services:
# Begin Ollama service
ollama: ollama:
image: ollama/ollama:0.1.22-rocm image: ollama/ollama:0.1.22-rocm
restart: unless-stopped restart: unless-stopped
@ -9,6 +10,7 @@ services:
- .env - .env
ports: ports:
- 11434:11434 - 11434:11434
# begin for AMD GPU support
devices: devices:
- /dev/kfd - /dev/kfd
- /dev/dri - /dev/dri
@ -19,10 +21,12 @@ services:
- SYS_PTRACE - SYS_PTRACE
security_opt: security_opt:
- seccomp=unconfined - seccomp=unconfined
# end of section for AMD GPU support
volumes: volumes:
- ./ollama/bootstrap.sh:/bootstrap.sh:ro - ./ollama/bootstrap.sh:/bootstrap.sh:ro
- ollama:/root/.ollama - ollama:/root/.ollama
# Begin LiteLLM service
litellm: litellm:
image: ghcr.io/berriai/litellm:main-latest image: ghcr.io/berriai/litellm:main-latest
command: --config /config.yaml command: --config /config.yaml
@ -31,6 +35,7 @@ services:
volumes: volumes:
- ./litellm/config.yaml:/config.yaml:ro - ./litellm/config.yaml:/config.yaml:ro
# Begin LibreChat
librechat: librechat:
image: ghcr.io/danny-avila/librechat image: ghcr.io/danny-avila/librechat
ports: ports: