Chats
caution
This is currently under development.
Overview
In Jan, chats
are LLM responses in the form of OpenAI compatible chat completion objects
.
- Models take a list of messages and return a model-generated response as output.
- An OpenAI Chat API compatible endpoint at
localhost:1337/v1/chats
.
Folder Structure
Chats are stateless, thus are not saved in janroot
. Any content and relevant metadata from calling this endpoint is extracted and persisted through Messages.
API Reference
Jan's Chat API is compatible with OpenAI's Chat API.
See Jan Chat API
Implementation
Under the hood, the /chat
endpoint simply reroutes an existing endpoint from Nitro server. Nitro is a lightweight & local inference server, written in C++ and embedded into the Jan app. See Nitro documentation.