From a18ff9c9b38c5415614870787295ab66195e66dd Mon Sep 17 00:00:00 2001 From: Ettore Di Giacinto Date: Thu, 27 Apr 2023 10:42:50 +0200 Subject: [PATCH] docs: move api docs (#96) --- README.md | 109 ++++++++++++++++++++++++++++-------------------------- 1 file changed, 56 insertions(+), 53 deletions(-) diff --git a/README.md b/README.md index f83364c..9133c80 100644 --- a/README.md +++ b/README.md @@ -215,6 +215,62 @@ Once the server is running, you can start making requests to it using HTTP, usin +### Supported OpenAI API endpoints + +You can check out the [OpenAI API reference](https://platform.openai.com/docs/api-reference/chat/create). + +Following the list of endpoints/parameters supported. + +Note: + +- You can also specify the model as part of the OpenAI token. +- If only one model is available, the API will use it for all the requests. + +#### Chat completions + +
+For example, to generate a chat completion, you can send a POST request to the `/v1/chat/completions` endpoint with the instruction as the request body: + +``` +curl http://localhost:8080/v1/chat/completions -H "Content-Type: application/json" -d '{ + "model": "ggml-koala-7b-model-q4_0-r2.bin", + "messages": [{"role": "user", "content": "Say this is a test!"}], + "temperature": 0.7 + }' +``` + +Available additional parameters: `top_p`, `top_k`, `max_tokens` +
+ +#### Completions + +
+ +To generate a completion, you can send a POST request to the `/v1/completions` endpoint with the instruction as per the request body: + +``` +curl http://localhost:8080/v1/completions -H "Content-Type: application/json" -d '{ + "model": "ggml-koala-7b-model-q4_0-r2.bin", + "prompt": "A long time ago in a galaxy far, far away", + "temperature": 0.7 + }' +``` + +Available additional parameters: `top_p`, `top_k`, `max_tokens` + +
+ +#### List models + +
+You can list all the models available with: + +``` +curl http://localhost:8080/v1/models +``` + +
+ ## Advanced configuration LocalAI can be configured to serve user-defined models with a set of default parameters and templates. @@ -278,59 +334,6 @@ See also [chatbot-ui](https://github.com/go-skynet/LocalAI/tree/master/examples/ -## Supported OpenAI API endpoints - -You can check out the [OpenAI API reference](https://platform.openai.com/docs/api-reference/chat/create). - -Following the list of endpoints/parameters supported. - -Note: - -- You can also specify the model as part of the OpenAI token. -- If only one model is available, the API will use it for all the requests. - -### Chat completions - -
-For example, to generate a chat completion, you can send a POST request to the `/v1/chat/completions` endpoint with the instruction as the request body: - -``` -curl http://localhost:8080/v1/chat/completions -H "Content-Type: application/json" -d '{ - "model": "ggml-koala-7b-model-q4_0-r2.bin", - "messages": [{"role": "user", "content": "Say this is a test!"}], - "temperature": 0.7 - }' -``` - -Available additional parameters: `top_p`, `top_k`, `max_tokens` -
- -### Completions - -
-To generate a completion, you can send a POST request to the `/v1/completions` endpoint with the instruction as per the request body: -``` -curl http://localhost:8080/v1/completions -H "Content-Type: application/json" -d '{ - "model": "ggml-koala-7b-model-q4_0-r2.bin", - "prompt": "A long time ago in a galaxy far, far away", - "temperature": 0.7 - }' -``` - -Available additional parameters: `top_p`, `top_k`, `max_tokens` - -
- -### List models - -
-You can list all the models available with: - -``` -curl http://localhost:8080/v1/models -``` - -
## Blog posts