This is a webui for any OpenAI-compatible chat API server. I use it mainly for interacting with my locally hosted LLMs. You can also hook it to any compatible chat server by setting the NEXT_PUBLIC_CHAT_ENDPOINT
environment variable to point to the chat completion endpoint (for example by creating a .env
file and put it in there).
v0.0.1 demo:
demo.mp4
First, install the required modules
npm install
# or
yarn install
# or
pnpm install
# or
bun install
Run the development server
npm run dev
# or
yarn dev
# or
pnpm dev
# or
bun --bun run dev
Open http://localhost:3000 with your browser to access the webui
How do I setup a local AI chat server?
Check out llama.cpp, that's what I'm using