Can connect to several existing backends like Anthropic, Cohere, OpenAI, NvidiaNIM, MistralAI etc,
and host models on its own - see the Cortex section on the screenshot below - showing Jan downloaded and hosting locally Llama3 8b q4 and Phi3 medium (q4).
Vane is one of the more pragmatic entries in the “AI search with citations” space: a self-hosted answering engine that mixes live web retrieval with local or cloud LLMs, while keeping the whole stack under your control.
Quick overview of most prominent UIs for Ollama in 2025
Locally hosted Ollama allows to run large language models on your own machine, but using it via command-line isn’t user-friendly.
Here are several open-source projects provide ChatGPT-style interfaces that connect to a local Ollama.
That’s very exciting!
Instead of calling copilot or perplexity.ai and telling all the world what you are after,
you can now host similar service on your own PC or laptop!