LLM

Ollama Enshittification - the Early Signs

Ollama Enshittification - the Early Signs

My view on current state of Ollama development

Ollama has quickly become one of the most popular tools for running LLMs locally. Its simple CLI, and streamlined model management have made it a go-to option for developers who want to work with AI models outside the cloud.

Chat UIs for Local Ollama Instances

Chat UIs for Local Ollama Instances

Quick overview of most prominent UIs for Ollama in 2025

Locally hosted Ollama allows to run large language models on your own machine, but using it via command-line isn’t user-friendly. Here are several open-source projects provide ChatGPT-style interfaces that connect to a local Ollama.

Search vs Deepsearch vs Deep Research

Search vs Deepsearch vs Deep Research

How different are they?

  • Search is best for quick, straightforward information retrieval using keywords.
  • Deep Search excels at understanding context and intent, delivering more relevant and comprehensive results for complex queries.
How Ollama Handles Parallel Requests

How Ollama Handles Parallel Requests

Understand Ollama concurrency, queueing, and how to tune OLLAMA_NUM_PARALLEL for stable parallel requests.

This guide explains how Ollama handles parallel requests (concurrency, queuing, and resource limits), and how to tune it using the OLLAMA_NUM_PARALLEL environment variable (and related knobs).