From 3e0b75b5e20afec390db78c91ef5efabd4944773 Mon Sep 17 00:00:00 2001 From: mudler Date: Tue, 30 May 2023 18:15:43 +0200 Subject: [PATCH] examples: use gallery in chatbot-ui, add flowise Signed-off-by: mudler --- README.md | 1 + examples/README.md | 8 ++++++++ examples/chatbot-ui/README.md | 12 ++++-------- examples/chatbot-ui/docker-compose.yaml | 15 ++++++++++++++- examples/chatbot-ui/models/completion.tmpl | 1 - examples/chatbot-ui/models/gpt-3.5-turbo.yaml | 16 ---------------- examples/chatbot-ui/models/gpt4all.tmpl | 4 ---- 7 files changed, 27 insertions(+), 30 deletions(-) delete mode 100644 examples/chatbot-ui/models/completion.tmpl delete mode 100644 examples/chatbot-ui/models/gpt-3.5-turbo.yaml delete mode 100644 examples/chatbot-ui/models/gpt4all.tmpl diff --git a/README.md b/README.md index f3777af..ae312d4 100644 --- a/README.md +++ b/README.md @@ -169,6 +169,7 @@ Feel free to open up a PR to get your project listed! - [Spark](https://github.com/cedriking/spark) - [autogpt4all](https://github.com/aorumbayev/autogpt4all) - [Mods](https://github.com/charmbracelet/mods) +- [Flowise](https://github.com/FlowiseAI/Flowise) ## Short-term roadmap diff --git a/examples/README.md b/examples/README.md index 0229441..e3257cd 100644 --- a/examples/README.md +++ b/examples/README.md @@ -22,6 +22,14 @@ This integration shows how to use LocalAI with [mckaywrigley/chatbot-ui](https:/ [Check it out here](https://github.com/go-skynet/LocalAI/tree/master/examples/chatbot-ui/) +### Flowise + +_by [@mudler](https://github.com/mudler)_ + +This example shows how to use [FlowiseAI/Flowise](https://github.com/FlowiseAI/Flowise) with LocalAI. + +[Check it out here](https://github.com/go-skynet/LocalAI/tree/master/examples/flowise/) + ### Discord bot _by [@mudler](https://github.com/mudler)_ diff --git a/examples/chatbot-ui/README.md b/examples/chatbot-ui/README.md index 7cf4bbb..fa7ee36 100644 --- a/examples/chatbot-ui/README.md +++ b/examples/chatbot-ui/README.md @@ -4,22 +4,18 @@ Example of integration with [mckaywrigley/chatbot-ui](https://github.com/mckaywr ![Screenshot from 2023-04-26 23-59-55](https://user-images.githubusercontent.com/2420543/234715439-98d12e03-d3ce-4f94-ab54-2b256808e05e.png) -## Setup +## Run +In this example LocalAI will download the gpt4all model and set it up as "gpt-3.5-turbo". See the `docker-compose.yaml` ```bash # Clone LocalAI git clone https://github.com/go-skynet/LocalAI cd LocalAI/examples/chatbot-ui -# (optional) Checkout a specific LocalAI tag -# git checkout -b build - -# Download gpt4all-j to models/ -wget https://gpt4all.io/models/ggml-gpt4all-j.bin -O models/ggml-gpt4all-j - # start with docker-compose -docker-compose up -d --pull always +docker-compose up --pull always + # or you can build the images with: # docker-compose up -d --build ``` diff --git a/examples/chatbot-ui/docker-compose.yaml b/examples/chatbot-ui/docker-compose.yaml index c7782c3..27b4f4e 100644 --- a/examples/chatbot-ui/docker-compose.yaml +++ b/examples/chatbot-ui/docker-compose.yaml @@ -3,6 +3,14 @@ version: '3.6' services: api: image: quay.io/go-skynet/local-ai:latest + # As initially LocalAI will download the models defined in PRELOAD_MODELS + # you might need to tweak the healthcheck values here according to your network connection. + # Here we give a timespan of 20m to download all the required files. + healthcheck: + test: ["CMD", "curl", "-f", "http://localhost:8080/readyz"] + interval: 1m + timeout: 20m + retries: 20 build: context: ../../ dockerfile: Dockerfile @@ -11,11 +19,16 @@ services: environment: - DEBUG=true - MODELS_PATH=/models + # You can preload different models here as well. + # See: https://github.com/go-skynet/model-gallery + - 'PRELOAD_MODELS=[{"url": "github:go-skynet/model-gallery/gpt4all-j.yaml", "name": "gpt-3.5-turbo"}]' volumes: - ./models:/models:cached command: ["/usr/bin/local-ai" ] - chatgpt: + depends_on: + api: + condition: service_healthy image: ghcr.io/mckaywrigley/chatbot-ui:main ports: - 3000:3000 diff --git a/examples/chatbot-ui/models/completion.tmpl b/examples/chatbot-ui/models/completion.tmpl deleted file mode 100644 index 9867cfc..0000000 --- a/examples/chatbot-ui/models/completion.tmpl +++ /dev/null @@ -1 +0,0 @@ -{{.Input}} \ No newline at end of file diff --git a/examples/chatbot-ui/models/gpt-3.5-turbo.yaml b/examples/chatbot-ui/models/gpt-3.5-turbo.yaml deleted file mode 100644 index 5c192f5..0000000 --- a/examples/chatbot-ui/models/gpt-3.5-turbo.yaml +++ /dev/null @@ -1,16 +0,0 @@ -name: gpt-3.5-turbo -parameters: - model: ggml-gpt4all-j - top_k: 80 - temperature: 0.2 - top_p: 0.7 -context_size: 1024 -stopwords: -- "HUMAN:" -- "GPT:" -roles: - user: " " - system: " " -template: - completion: completion - chat: gpt4all \ No newline at end of file diff --git a/examples/chatbot-ui/models/gpt4all.tmpl b/examples/chatbot-ui/models/gpt4all.tmpl deleted file mode 100644 index f76b080..0000000 --- a/examples/chatbot-ui/models/gpt4all.tmpl +++ /dev/null @@ -1,4 +0,0 @@ -The prompt below is a question to answer, a task to complete, or a conversation to respond to; decide which and write an appropriate response. -### Prompt: -{{.Input}} -### Response: