diff --git a/README.md b/README.md index f3777af..ae312d4 100644 --- a/README.md +++ b/README.md @@ -169,6 +169,7 @@ Feel free to open up a PR to get your project listed! - [Spark](https://github.com/cedriking/spark) - [autogpt4all](https://github.com/aorumbayev/autogpt4all) - [Mods](https://github.com/charmbracelet/mods) +- [Flowise](https://github.com/FlowiseAI/Flowise) ## Short-term roadmap diff --git a/examples/README.md b/examples/README.md index 0229441..e3257cd 100644 --- a/examples/README.md +++ b/examples/README.md @@ -22,6 +22,14 @@ This integration shows how to use LocalAI with [mckaywrigley/chatbot-ui](https:/ [Check it out here](https://github.com/go-skynet/LocalAI/tree/master/examples/chatbot-ui/) +### Flowise + +_by [@mudler](https://github.com/mudler)_ + +This example shows how to use [FlowiseAI/Flowise](https://github.com/FlowiseAI/Flowise) with LocalAI. + +[Check it out here](https://github.com/go-skynet/LocalAI/tree/master/examples/flowise/) + ### Discord bot _by [@mudler](https://github.com/mudler)_ diff --git a/examples/chatbot-ui/README.md b/examples/chatbot-ui/README.md index 7cf4bbb..fa7ee36 100644 --- a/examples/chatbot-ui/README.md +++ b/examples/chatbot-ui/README.md @@ -4,22 +4,18 @@ Example of integration with [mckaywrigley/chatbot-ui](https://github.com/mckaywr ![Screenshot from 2023-04-26 23-59-55](https://user-images.githubusercontent.com/2420543/234715439-98d12e03-d3ce-4f94-ab54-2b256808e05e.png) -## Setup +## Run +In this example LocalAI will download the gpt4all model and set it up as "gpt-3.5-turbo". See the `docker-compose.yaml` ```bash # Clone LocalAI git clone https://github.com/go-skynet/LocalAI cd LocalAI/examples/chatbot-ui -# (optional) Checkout a specific LocalAI tag -# git checkout -b build - -# Download gpt4all-j to models/ -wget https://gpt4all.io/models/ggml-gpt4all-j.bin -O models/ggml-gpt4all-j - # start with docker-compose -docker-compose up -d --pull always +docker-compose up --pull always + # or you can build the images with: # docker-compose up -d --build ``` diff --git a/examples/chatbot-ui/docker-compose.yaml b/examples/chatbot-ui/docker-compose.yaml index c7782c3..27b4f4e 100644 --- a/examples/chatbot-ui/docker-compose.yaml +++ b/examples/chatbot-ui/docker-compose.yaml @@ -3,6 +3,14 @@ version: '3.6' services: api: image: quay.io/go-skynet/local-ai:latest + # As initially LocalAI will download the models defined in PRELOAD_MODELS + # you might need to tweak the healthcheck values here according to your network connection. + # Here we give a timespan of 20m to download all the required files. + healthcheck: + test: ["CMD", "curl", "-f", "http://localhost:8080/readyz"] + interval: 1m + timeout: 20m + retries: 20 build: context: ../../ dockerfile: Dockerfile @@ -11,11 +19,16 @@ services: environment: - DEBUG=true - MODELS_PATH=/models + # You can preload different models here as well. + # See: https://github.com/go-skynet/model-gallery + - 'PRELOAD_MODELS=[{"url": "github:go-skynet/model-gallery/gpt4all-j.yaml", "name": "gpt-3.5-turbo"}]' volumes: - ./models:/models:cached command: ["/usr/bin/local-ai" ] - chatgpt: + depends_on: + api: + condition: service_healthy image: ghcr.io/mckaywrigley/chatbot-ui:main ports: - 3000:3000 diff --git a/examples/chatbot-ui/models/completion.tmpl b/examples/chatbot-ui/models/completion.tmpl deleted file mode 100644 index 9867cfc..0000000 --- a/examples/chatbot-ui/models/completion.tmpl +++ /dev/null @@ -1 +0,0 @@ -{{.Input}} \ No newline at end of file diff --git a/examples/chatbot-ui/models/gpt-3.5-turbo.yaml b/examples/chatbot-ui/models/gpt-3.5-turbo.yaml deleted file mode 100644 index 5c192f5..0000000 --- a/examples/chatbot-ui/models/gpt-3.5-turbo.yaml +++ /dev/null @@ -1,16 +0,0 @@ -name: gpt-3.5-turbo -parameters: - model: ggml-gpt4all-j - top_k: 80 - temperature: 0.2 - top_p: 0.7 -context_size: 1024 -stopwords: -- "HUMAN:" -- "GPT:" -roles: - user: " " - system: " " -template: - completion: completion - chat: gpt4all \ No newline at end of file diff --git a/examples/chatbot-ui/models/gpt4all.tmpl b/examples/chatbot-ui/models/gpt4all.tmpl deleted file mode 100644 index f76b080..0000000 --- a/examples/chatbot-ui/models/gpt4all.tmpl +++ /dev/null @@ -1,4 +0,0 @@ -The prompt below is a question to answer, a task to complete, or a conversation to respond to; decide which and write an appropriate response. -### Prompt: -{{.Input}} -### Response: diff --git a/examples/flowise/README.md b/examples/flowise/README.md new file mode 100644 index 0000000..d8bb13d --- /dev/null +++ b/examples/flowise/README.md @@ -0,0 +1,26 @@ +# flowise + +Example of integration with [FlowiseAI/Flowise](https://github.com/FlowiseAI/Flowise). + +![Screenshot from 2023-05-30 18-01-03](https://github.com/go-skynet/LocalAI/assets/2420543/02458782-0549-4131-971c-95ee56ec1af8) + +You can check a demo video in the Flowise PR: https://github.com/FlowiseAI/Flowise/pull/123 + +## Run + +In this example LocalAI will download the gpt4all model and set it up as "gpt-3.5-turbo". See the `docker-compose.yaml` +```bash +# Clone LocalAI +git clone https://github.com/go-skynet/LocalAI + +cd LocalAI/examples/flowise + +# start with docker-compose +docker-compose up --pull always + +``` + +## Accessing flowise + +Open http://localhost:3000. + diff --git a/examples/flowise/docker-compose.yaml b/examples/flowise/docker-compose.yaml new file mode 100644 index 0000000..a39269c --- /dev/null +++ b/examples/flowise/docker-compose.yaml @@ -0,0 +1,37 @@ +version: '3.6' + +services: + api: + image: quay.io/go-skynet/local-ai:latest + # As initially LocalAI will download the models defined in PRELOAD_MODELS + # you might need to tweak the healthcheck values here according to your network connection. + # Here we give a timespan of 20m to download all the required files. + healthcheck: + test: ["CMD", "curl", "-f", "http://localhost:8080/readyz"] + interval: 1m + timeout: 20m + retries: 20 + build: + context: ../../ + dockerfile: Dockerfile + ports: + - 8080:8080 + environment: + - DEBUG=true + - MODELS_PATH=/models + # You can preload different models here as well. + # See: https://github.com/go-skynet/model-gallery + - 'PRELOAD_MODELS=[{"url": "github:go-skynet/model-gallery/gpt4all-j.yaml", "name": "gpt-3.5-turbo"}]' + volumes: + - ./models:/models:cached + command: ["/usr/bin/local-ai" ] + flowise: + depends_on: + api: + condition: service_healthy + image: flowiseai/flowise + ports: + - 3000:3000 + volumes: + - ~/.flowise:/root/.flowise + command: /bin/sh -c "sleep 3; flowise start" \ No newline at end of file