Open WebUI provides a highly polished chat interface in your browser for LLM interaction.
This guide walks through how to connect Lemonade Server to Open WebUI and highlights some great features you can start using right away:
Looking for a house-wide setup? If you want to set up Open WebUI + Lemonade for your whole home’s network, including mobile access, multi-user accounts, persistent chat history, and a custom telemetry dashboard plugin, check out this In-Depth Community Guide.
We recommend installing Open WebUI into a dedicated Python environment using the following commands:
pip install open-webui
Note: Open WebUI also provides a variety of other installation options, such as Docker, on their GitHub.
Run this command to launch the Open WebUI HTTP server:
open-webui serve
In a browser, navigate to http://localhost:8080/
Open WebUI will ask you to create a local administrator account. You can fill any username, password, and email you like. Once you are signed in, you will see the chat interface:

Install and run Lemonade Server. Download here.
http://localhost:8000/api/v1 (unless you’re using a different port), API key (this is unused but required, suggest just putting a -), and then click “Save”.
Now that everything is configured, you are ready to interact with an LLM!
Enter a message to the LLM and click send (or hit enter). The LLM will take a few seconds to load into memory and then you will see the response stream in.
Vision Language Models (VLMs) can take images as part of their input.
Scroll down until you see a model with the blue VISION label and click the “+” button to install it.
Paste an image into the chat box and type a prompt or question about your image. You can also use the “+” button in the chat box to upload images.
Open WebUI allows you to run Python code generated by an LLM directly within the interface.
Note: only certain Python modules are enabled in Open WebUI.
matplotlibis one of our favorites.
Ask the LLM to write some Python, then click the Run button at the top of the Python code block.
If all goes well, the result of running the Python code will appear below the code block.
Note: LLMs often produce incorrect code, so it might take a few chat iterations to fix any bugs. Copy-pasting the Python error message is usually enough to move things along.
Open WebUI has a built-in rendering engine for HTML, CSS, and JavaScript pages.
Smaller LLMs can produce simple pages with tasteful styling and basic interactivity, while larger LLMs can accomplish tasks like 3D rendering in 3js.
Ask a small LLM to write a simple HTML+CSS page. The preview may pop up automatically, but if it doesn’t you can click the Preview button above the HTML code block:
Ask a large LLM to create a 3D shape using 3js.
Open WebUI supports image generation using Stable Diffusion models through Lemonade Server.
Configuring Image Generation
Image Generation on.Standard (Open AI) as the Image Generation Engine.Prompt Generation on.OpenAI-API-Basis-URL, fill in http://localhost:8000/api/v1 (unless you’re using a different port).- for OpenAI-API-Key.{ "steps": 4, "cfg_scale": 1 }. See available parameters at Image Generation (Stable Diffusion CPP).Model, e.g., SDXL-Turbo.Save.Allow Image Generation for Model
Enable Image Generation as a capability for your model:
1. Go to Admin > Settings > Models and choose your model.
2. Turn on Image Generation. If you want start chat always with image generation, also toggle the default option.
Option 1: Using Image Generation Switch
To generate an image:
1. Toggle the Image Generation switch in the chat on.
2. Enter your image generation prompt.
3. Click Send.
Option 2: Native Tool-Based Generation (Agentic)
This mode uses tool calling for image generation and is recommended for high-quality models with tool calling capabilities. Normally the models will alter and improve your prompt.
Advanced Parameters and toggle Function Calling to Native.Note: Open WebUI recommends using native mode only for high-quality models. See Tool Calling Modes for more information. (try out >30B models like GPT-OSS-120B, GLM-4.7-Flash or Qwen-3-Next-80B-A3B)
These are just a few of our favorite ways to try out LLMs in Open WebUI. There are a lot more features to explore, such as voice interaction and chatting with documents, so be sure to check out the Open WebUI documentation and YouTube content.