Open WebUI#
Open WebUI is an extensible, self-hosted interface for AI that adapts to your workflow, all while operating entirely offline. The ‘Open WebUI’ app also launches an Ollama server inside an isolated network namespace, providing a private and offline interactive environment for running large language models. As a tradeoff you will not be able to interact with Open WebUI’s online elements as there is no connection to the internet.
VSC clusters that support the Open WebUI app:
Tier-2 Anansi
Open WebUI runs in a graphical desktop environment, similar to the Desktop app, with Firefox opened in kiosk mode. Keep in mind that for seemless copy/pasting, you should launch the Open WebUI app from a Chromium-based webbrowser (e.g., Google Chrome or Microsoft Edge).
By default the Open WebUI app will use the pre-installed Ollama models located in /databases/ollama. You can change this directory and use your custom models by setting the environment variable OLLAMA_MODELS.
By default the Open WebUI app stores its data and chat history in $VSC_HOME/.openwebui. You can change this directory by setting the environment variable OPEN_WEBUI_DATA.
Note
At the moment this app only works with GPU-enabled Ollama. In order to run it you will need to select a GPU or fraction of GPU.