LM Studio Setup
LM Studio is a desktop application that lets you run large language models locally on your computer.
Configuration
- Go to the Developer tab in LM Studio
- Click Start Server to start the local server (default port is 1234)
- Enable CORS in the server settings - this is required for Hyprnote to connect

Connecting to Hyprnote
- Open Hyprnote and go to Settings > Intelligence
- Expand the LM Studio provider card
- The default base URL
http://127.0.0.1:1234/v1should work if you haven't changed the port - Select LM Studio as your provider and choose a model from the dropdown
Troubleshooting
If Hyprnote cannot connect to LM Studio:
- Ensure the LM Studio server is running (check the Developer tab)
- Verify CORS is enabled in LM Studio settings
- Check that the port matches (default is 1234)
- Make sure no firewall is blocking the connection
Context Length Error
If you see an error like "Cannot truncate prompt with n_keep >= n_ctx" or "request exceeds the available context size", this means the model's context length is too small for your conversation or transcript.
To fix this:
- Open LM Studio and go to the Context tab in the model settings (on the right panel)
- Increase the Context Length value - try setting it to at least 16384 or higher depending on your needs
- Note that higher context lengths require more memory, so adjust based on your system's capabilities

Ollama Setup
Ollama is a command-line tool for running large language models locally.
Connecting to Hyprnote
- Open Hyprnote and go to Settings > Intelligence
- Expand the Ollama provider card
- The default base URL
http://127.0.0.1:11434/v1should work - Select Ollama as your provider and choose a model from the dropdown
Troubleshooting
If Hyprnote cannot connect to Ollama:
- Ensure Ollama is running (
ollama serve) - Check that you have at least one model pulled (
ollama list) - Verify the port is correct (default is 11434)
- On macOS, Ollama may already be running as a background service
