parent
4e2061636e
commit
325bc78acc
@ -0,0 +1,11 @@ |
||||
# Examples |
||||
|
||||
Here is a list of projects that can easily be integrated with the LocalAI backend. |
||||
|
||||
## Projects |
||||
|
||||
- [chatbot-ui](https://github.com/go-skynet/LocalAI/tree/master/examples/chatbot-ui/) (by [@mkellerman](https://github.com/mkellerman)) |
||||
|
||||
## Want to contribute? |
||||
|
||||
Create an issue, and put `Example: <description>` in the title! We will post your examples here. |
@ -0,0 +1,18 @@ |
||||
# chatbot-ui |
||||
|
||||
|
||||
## Setup |
||||
- Set the `services > api > volumes` parameter in the docker-compose.yaml file. This should point to your local folder with your models. |
||||
- Copy/Rename your model file to `gpt-3.5-turbo` (without any .bin file extension). |
||||
- Type `docker compose up` to run the api and the Web UI. |
||||
- Open http://localhost:3000 for the Web UI. |
||||
|
||||
|
||||
## Known issues |
||||
- Can't select the model from the UI. Seems hardcoded to `gpt-3.5-turbo`. |
||||
- If your machine is slow, the UI will timeout on the request to the API. |
||||
|
||||
|
||||
### Links |
||||
|
||||
- [mckaywrigley/chatbot-ui](https://github.com/mckaywrigley/chatbot-ui) |
@ -0,0 +1,23 @@ |
||||
version: '3.6' |
||||
|
||||
services: |
||||
api: |
||||
image: quay.io/go-skynet/local-ai:latest |
||||
ports: |
||||
- 8080:8080 |
||||
environment: |
||||
- THREADS=4 |
||||
- CONTEXT_SIZE=512 |
||||
- DEBUG=true |
||||
- MODELS_PATH=/models |
||||
volumes: |
||||
- ./models:/models:cached |
||||
command: ["/usr/bin/local-ai" ] |
||||
|
||||
chatgpt: |
||||
image: ghcr.io/mckaywrigley/chatbot-ui:main |
||||
ports: |
||||
- 3000:3000 |
||||
environment: |
||||
- 'OPENAI_API_KEY=sk-XXXXXXXXXXXXXXXXXXXX' |
||||
- 'OPENAI_API_HOST=http://host.docker.internal:8080' |
Loading…
Reference in new issue