include: - docker-compose.base.yml services: # Begin Ollama service ollama: image: ollama/ollama:latest restart: unless-stopped entrypoint: /bootstrap.sh command: mistral network_mode: service:open-webui environment: OLLAMA_HOST: http://localhost:11434 # begin for NVIDIA GPU support deploy: resources: reservations: devices: - driver: nvidia count: 1 capabilities: [gpu] # end of section for NVIDIA GPU support volumes: - ./ollama/bootstrap.sh:/bootstrap.sh:ro - ./ollama:/root/.ollama