From 3bb11173615fc69c9eee482e414137d13c8ca73f Mon Sep 17 00:00:00 2001 From: Massaki Archambault Date: Wed, 21 Feb 2024 22:01:24 -0500 Subject: [PATCH] add CPU version of the deployment --- README.md | 11 ++++++++++- docker-compose.cpu.yml | 17 +++++++++++++++++ 2 files changed, 27 insertions(+), 1 deletion(-) create mode 100644 docker-compose.cpu.yml diff --git a/README.md b/README.md index 925075f..c9a99e1 100644 --- a/README.md +++ b/README.md @@ -38,7 +38,16 @@ A quick prototype to self-host [LibreChat](https://github.com/danny-avila/LibreC 6. Browse http://localhost:3080/ 7. Create an admin account and start chatting! -The API along with the APIDoc will be available at http://localhost:8000/ +### Steps for NO GPU (use CPU) + +**Warning: This may be very slow depending on your CPU and may us a lot of RAM depending on the model** + +1. Make sure your drivers are up to date. +2. Clone the repo. +3. Copy the CPU compose spec to select it. `cp docker-compose.cpu.yml docker.compose.yml` +4. Run `docker compose up`. Wait for a few minutes for the model to be downloaded and served. +5. Browse http://localhost:3080/ +6. Create an admin account and start chatting! ## Configuring additional models diff --git a/docker-compose.cpu.yml b/docker-compose.cpu.yml new file mode 100644 index 0000000..35b9f18 --- /dev/null +++ b/docker-compose.cpu.yml @@ -0,0 +1,17 @@ +include: + - docker-compose.base.yml + +services: + # Begin Ollama service + ollama: + image: ollama/ollama:0.1.23 + restart: unless-stopped + entrypoint: /bootstrap.sh + command: mistral + env_file: + - .env + ports: + - 11434:11434 + volumes: + - ./ollama/bootstrap.sh:/bootstrap.sh:ro + - ./ollama:/root/.ollama \ No newline at end of file