You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
37 lines
1.1 KiB
37 lines
1.1 KiB
version: '3.6'
|
|
|
|
services:
|
|
api:
|
|
image: quay.io/go-skynet/local-ai:latest
|
|
# As initially LocalAI will download the models defined in PRELOAD_MODELS
|
|
# you might need to tweak the healthcheck values here according to your network connection.
|
|
# Here we give a timespan of 20m to download all the required files.
|
|
healthcheck:
|
|
test: ["CMD", "curl", "-f", "http://localhost:8080/readyz"]
|
|
interval: 1m
|
|
timeout: 20m
|
|
retries: 20
|
|
build:
|
|
context: ../../
|
|
dockerfile: Dockerfile
|
|
ports:
|
|
- 8080:8080
|
|
environment:
|
|
- DEBUG=true
|
|
- MODELS_PATH=/models
|
|
# You can preload different models here as well.
|
|
# See: https://github.com/go-skynet/model-gallery
|
|
- 'PRELOAD_MODELS=[{"url": "github:go-skynet/model-gallery/gpt4all-j.yaml", "name": "gpt-3.5-turbo"}]'
|
|
volumes:
|
|
- ./models:/models:cached
|
|
command: ["/usr/bin/local-ai" ]
|
|
chatgpt:
|
|
depends_on:
|
|
api:
|
|
condition: service_healthy
|
|
image: ghcr.io/mckaywrigley/chatbot-ui:main
|
|
ports:
|
|
- 3000:3000
|
|
environment:
|
|
- 'OPENAI_API_KEY=sk-XXXXXXXXXXXXXXXXXXXX'
|
|
- 'OPENAI_API_HOST=http://api:8080' |