Skip to content

Lemonade Server Configuration

Overview

Lemonade Server starts automatically with the OS after installation. Configuration is managed through a single config.json file stored in the lemonade cache directory.

config.json

If you used an installer from the Lemonade release your config.json will be at these locations depending on your OS:

  • Linux (systemd): /var/lib/lemonade/.cache/lemonade/config.json
  • Windows: %USERPROFILE%\.cache\lemonade\config.json
  • macOS: /Library/Application Support/lemonade/.cache/config.json

If you are using a standalone lemond exectable, the default location is ~/.cache/lemonade/config.json.

Note: If config.json doesn't exist, it's created automatically with default values on first run.

Example config.json

{
  "config_version": 1,
  "port": 13305,
  "host": "localhost",
  "log_level": "info",
  "global_timeout": 300,
  "max_loaded_models": 1,
  "no_broadcast": false,
  "extra_models_dir": "",
  "models_dir": "auto",
  "ctx_size": 4096,
  "offline": false,
  "no_fetch_executables": false,
  "disable_model_filtering": false,
  "enable_dgpu_gtt": false,
  "rocm_channel": "preview",
  "llamacpp": {
    "backend": "auto",
    "args": "",
    "prefer_system": false,
    "rocm_bin": "builtin",
    "vulkan_bin": "builtin",
    "cpu_bin": "builtin"
  },
  "whispercpp": {
    "backend": "auto",
    "args": "",
    "cpu_bin": "builtin",
    "npu_bin": "builtin"
  },
  "sdcpp": {
    "backend": "auto",
    "args": "",
    "steps": 20,
    "cfg_scale": 7.0,
    "width": 512,
    "height": 512,
    "cpu_bin": "builtin",
    "rocm_bin": "builtin",
    "vulkan_bin": "builtin"
  },
  "flm": {
    "args": "",
  },
  "ryzenai": {
    "server_bin": "builtin"
  },
  "kokoro": {
    "cpu_bin": "builtin"
  }
}

Settings Reference

Key Type Default Description
port int 13305 Port number for the HTTP server
host string "localhost" Address to bind for connections
log_level string "info" Logging level (trace, debug, info, warning, error, fatal, none)
global_timeout int 300 Timeout in seconds for HTTP, inference, and readiness checks
max_loaded_models int 1 Max models per type slot. Use -1 for unlimited
no_broadcast bool false Disable UDP broadcasting for server discovery
extra_models_dir string "" Secondary directory to scan for GGUF model files
models_dir string "auto" Directory for cached model files. "auto" follows HF_HUB_CACHE / HF_HOME / platform default
ctx_size int 4096 Default context size for LLM models
offline bool false Skip model downloads
no_fetch_executables bool false Prevent downloading backend executable artifacts; backends must already be installed or use the system backend
disable_model_filtering bool false Show all models regardless of hardware capabilities
enable_dgpu_gtt bool false Include GTT for hardware-based model filtering
rocm_channel string "preview" ROCm backend channel: "preview" (default), "stable", or "nightly". See llama.cpp Backend for details

Backend Configuration

Backend-specific settings are nested under their backend name:

llamacpp — LLM inference via llama.cpp: | Key | Default | Description | |-----|---------|-------------| | backend | "auto" | Backend to use: "auto" means "choose for me" | | args | "" | Custom arguments to pass to llama-server | | prefer_system | false | Prefer system-installed llama.cpp over bundled | | *_bin | "builtin" | Backend binary selection — see Backend binary selection |

whispercpp — Audio transcription: | Key | Default | Description | |-----|---------|-------------| | backend | "auto" | Backend to use: "auto" means "choose for me" | | args | "" | Custom arguments to pass to whisper-server | | *_bin | "builtin" | Backend binary selection — see Backend binary selection |

sdcpp — Image generation: | Key | Default | Description | |-----|---------|-------------| | backend | "auto" | Backend to use: "auto" means "choose for me" | | args | "" | Custom arguments to pass to sd-server | | steps | 20 | Number of inference steps | | cfg_scale | 7.0 | Classifier-free guidance scale | | width | 512 | Image width in pixels | | height | 512 | Image height in pixels | | *_bin | "builtin" | Backend binary selection — see Backend binary selection |

flm — FastFlowLM NPU inference: | Key | Default | Description | |-----|---------|-------------| | args | "" | Custom arguments to pass to flm serve |

ryzenai — RyzenAI NPU inference: | Key | Default | Description | |-----|---------|-------------| | server_bin | "builtin" | Backend binary selection — see Backend binary selection |

kokoro — Text-to-speech: | Key | Default | Description | |-----|---------|-------------| | cpu_bin | "builtin" | Backend binary selection — see Backend binary selection |

Backend binary selection

Every *_bin key (e.g. llamacpp.vulkan_bin, whispercpp.cpu_bin, sdcpp.rocm_bin) accepts the same set of values:

Value Meaning
"builtin" (default) Use the version of the upstream backend that lemonade pins in its release. Recommended for most users — these versions are tested with this lemonade build.
"" Same as "builtin".
"latest" Resolve to the most-recent upstream GitHub release on first install or first status query for that backend, then install on demand. The resolved tag is recorded in <lemonade-home>/bin/<recipe>/<backend>/version.txt.
"b8664" / "v1.8.2" / etc. A specific upstream release tag. Lemonade downloads that exact version from GitHub.
"/path/to/bin" A directory you populated yourself (e.g. a local build). Lemonade uses the executable inside this directory and never downloads. The path must exist when set.

Note: the latest setting is experimental.

Important — llamacpp.rocm_bin version tags are channel-specific. Each ROCm channel downloads from a different GitHub repository, so you must set the correct rocm_channel before pinning rocm_bin to a specific tag. See Pinning to a Specific Version Tag for details.

Examples:

# Track upstream llama.cpp Vulkan releases (auto-resolve at lemond start)
lemonade config set llamacpp.vulkan_bin=latest

# Pin to a specific llama.cpp build
lemonade config set llamacpp.vulkan_bin=b8664

# Use your own llama.cpp build
lemonade config set llamacpp.vulkan_bin=/home/me/llama.cpp/build/bin

# Revert to the version lemonade ships
lemonade config set llamacpp.vulkan_bin=builtin

Behavior when *_bin changes

Changing a *_bin value applies live: lemonade unloads any model currently using that backend, downloads the new binary if needed, and reloads the model on the new binary. No lemond restart is required.

latest re-resolution

"latest" is resolved once per lemond process. The first install or status query for a latest-pinned backend hits the GitHub API; the resolved tag is then cached in memory for the rest of the process lifetime. Subsequent installs and status queries (including manual lemonade backends install) reuse the cached tag and do not re-query GitHub. Restart lemond to pick up a newer upstream release.

Upgrade signals in lemonade backends

The lemonade backends listing surfaces two upgrade signals for backends pinned to "latest":

  • update_available — A newer upstream release exists than what's installed. The backend keeps running on the installed version; the listed action is the install command to apply the upgrade when you're ready.
  • update_required — The installed version is older than the version lemonade ships in this release. This forces an upgrade prompt because running below the lemonade-shipped baseline is not supported.

Backends pinned to a specific tag (e.g. b8664) do not get either signal — they're treated as an explicit user choice.

Interactions with other config

  • offline: true blocks the GitHub call for "latest". If a previously-installed version.txt exists in the install directory, lemonade reuses that version with a warning. Otherwise the install fails.
  • no_fetch_executables: true blocks all downloads, including resolving and installing "latest" and any version-tag pin. Existing installs continue to work.

Editing Configuration

Use the lemonade config CLI to view and modify settings while the server is running. Changes are applied immediately and persisted to config.json.

# View all current settings
lemonade config

# Set one or more values
lemonade config set key=value [key=value ...]

Top-level settings use their JSON key name directly. Nested backend settings use dot notation (section.key=value):

# Change the server port and log level
lemonade config set port=9000 log_level=debug

# Change a backend setting
lemonade config set llamacpp.backend=rocm

# Set multiple values at once
lemonade config set port=9000 llamacpp.backend=rocm sdcpp.steps=30

lemond CLI arguments (fallback)

If the server cannot start (e.g., invalid port in config.json), lemond accepts --port and --host as CLI arguments to override config.json. These overrides are persisted so the server can start normally next time:

lemond --port 9000 --host 0.0.0.0

Edit config.json manually (last resort)

If the server won't start and CLI arguments aren't sufficient, you can edit config.json directly. Restart the server after making changes:

# Linux
sudo nano /var/lib/lemonade/.cache/lemonade/config.json
sudo systemctl restart lemonade-server

# Windows — edit with your preferred text editor:
# %USERPROFILE%\.cache\lemonade\config.json
# Then quit and relaunch from the Start Menu

lemond CLI

lemond [cache_dir] [--port PORT] [--host HOST]
  • cache_dir — Path to the lemonade cache directory containing config.json and model data. Optional; defaults to platform-specific location.
  • --port — Port to serve on (overrides config.json, persisted). Use as a fallback if the server cannot start.
  • --host — Address to bind (overrides config.json, persisted). Use as a fallback if the server cannot start.

API Key and Security

Regular API Key

The LEMONADE_API_KEY environment variable sets an API key for authentication on regular API endpoints (/api/*, /v0/*, /v1/*). On Linux with systemd, set it in the service environment (e.g., via a systemd override or drop-in file). On Windows, set it as a system environment variable.

Admin API Key

The LEMONADE_ADMIN_API_KEY environment variable provides elevated access to both regular API endpoints and internal endpoints (/internal/*). When set, it takes precedence over LEMONADE_API_KEY for client authentication.

Authentication Hierarchy:

Scenario LEMONADE_API_KEY LEMONADE_ADMIN_API_KEY Internal Endpoints Regular API Endpoints
No keys set (not set) (not set) No auth required No auth required
Only API key "secret" (not set) Requires key Requires key
Only admin key (not set) "admin" Requires admin key No auth required
Both keys different "regular" "admin" Requires admin key Either key accepted

Client Behavior: Clients (CLI, tray app) automatically prefer LEMONADE_ADMIN_API_KEY if set, otherwise fall back to LEMONADE_API_KEY.

Remote Server Connection

To make Lemonade Server accessible from other machines on your network, set the host to 0.0.0.0:

lemonade config set host=0.0.0.0

Note: Using host: "0.0.0.0" allows connections from any machine on the network. Only do this on trusted networks. Set LEMONADE_API_KEY or LEMONADE_ADMIN_API_KEY to manage access.

Next Steps

The Server Specification provides more information about how to integrate Lemonade Server into an application.