Documentation Index
Fetch the complete documentation index at: https://docs.typingmind.com/llms.txt
Use this file to discover all available pages before exploring further.
Overview
Typing Mind allows you to connect the app with any local model you want.
- The model must be served via an OpenAI-compatible API endpoint.
- You must have some relevant technical skills to setup a custom model on your own server/endpoint.
Setup LocalAI on your device
👉 If you already have another setup for the local AI model endpoint, you can skip this step.
💡 Note that we added the
--cors true parameter to the command to make sure the local server is accessible from the browser. Typing Mind will send requests to the local model directly from the browser.

Setup Custom Model on TypingMind
Open TypingMind → Models → Add Custom Model:
Popular problems at this step
| Issue | Resolution |
|---|---|
| CORS related issues | Make sure your server configuration allows the endpoint to be accessible from the browser. Open the Network tab in the browser console to see more details. |
| Long waiting time | In the first request, your model can take a long time to respond. Check the terminal log of the Docker process to see if anything goes wrong. |
| API Key Missing | Typing Mind does not support API key authentication for custom model yet. Please reconfigurate your custom model to remove API key requirement. |
Chat with the new Custom Model
Once the model is tested and added successfully, you can select the custom model and chat with it normally.