Self-Hosting
LLM Self-Hosting and AI Sovereignty
Control data and models with self-hosted LLMs
Self-hosting LLMs keeps data, models, and inference under your control-a practical path to AI sovereignty for teams, enterprises, nations. Here: what sovereign AI is, which aspects and methods are used to build it, how LLM self-hosting fits in, how countries are addressing the challenge.
Comparing LLMs performance on Ollama on 16GB VRAM GPU
LLM speed test on RTX 4080 with 16GB VRAM
Running large language models locally gives you privacy, offline capability, and zero API costs. This benchmark reveals exactly what one can expect from 9 popular LLMs on Ollama on an RTX 4080.
Top 19 Trending Go Projects on GitHub - January 2026
January 2026 trending Go repos
The Go ecosystem continues to thrive with innovative projects spanning AI tooling, self-hosted applications, and developer infrastructure. This overview analyzes the top trending Go repositories on GitHub this month.
GPU and RAM Prices Surge in Australia: RTX 5090 Up 15%, RAM Up 38% - January 2026
January 2025 GPU and RAM price check
Today we are looking at the top-level consumer GPUs, and RAM modules. Specifically I’m looking at RTX-5080 and RTX-5090 prices, and 32GB (2x16GB) DDR5 6000.
Open WebUI: Self-Hosted LLM Interface
Self-hosted ChatGPT alternative for local LLMs
Open WebUI is a powerful, extensible, and feature-rich self-hosted web interface for interacting with large language models.
vLLM Quickstart: High-Performance LLM Serving - in 2026
Fast LLM inference with OpenAI API
vLLM is a high-throughput, memory-efficient inference and serving engine for Large Language Models (LLMs) developed by UC Berkeley’s Sky Computing Lab.
DGX Spark AU Pricing: $6,249-$7,999 at Major Retailers
Real AUD pricing from Aussie retailers now
The NVIDIA DGX Spark (GB10 Grace Blackwell) is now available in Australia at major PC retailers with local stock. If you’ve been following the global DGX Spark pricing and availability, you’ll be interested to know that Australian pricing ranges from $6,249 to $7,999 AUD depending on storage configuration and retailer.
Self-Hosting Cognee: Choosing LLM on Ollama
Testing Cognee with local LLMs - real results
Cognee is a Python framework for building knowledge graphs from documents using LLMs. But does it work with self-hosted models?
Choosing the Right LLM for Cognee: Local Ollama Setup
Thoughts on LLMs for self-hosted Cognee
Choosing the Best LLM for Cognee demands balancing graph-building quality, hallucination rates, and hardware constraints. Cognee excels with larger, low-hallucination models (32B+) via Ollama but mid-size options work for lighter setups.
Using Ollama Web Search API in Python
Build AI search agents with Python and Ollama
Ollama’s Python library now includes native OLlama web search capabilities. With just a few lines of code, you can augment your local LLMs with real-time information from the web, reducing hallucinations and improving accuracy.
Vector Stores for RAG Comparison
Pick the right vector DB for your RAG stack
Choosing the right vector store can make or break your RAG application’s performance, cost, and scalability. This comprehensive comparison covers the most popular options in 2024-2025.
Using Ollama Web Search API in Go
Build AI search agents with Go and Ollama
Ollama’s Web Search API lets you augment local LLMs with real-time web information. This guide shows you how to implement web search capabilities in Go, from simple API calls to full-featured search agents.
RAM Price Surge: Up to 619% in 2025
RAM prices surge 163-619% as AI demand strains supply
The memory market is experiencing unprecedented price volatility in late 2025, with RAM prices surging dramatically across all segments.
Local LLM Hosting: Complete 2026 Guide - Ollama, vLLM, LocalAI, Jan, LM Studio & More
Master local LLM deployment with 12+ tools compared
Local deployment of LLMs has become increasingly popular as developers and organizations seek enhanced privacy, reduced latency, and greater control over their AI infrastructure.
NVidia RTX 5080 and RTX 5090 prices in Australia - November 2025
AI-suitable Consumer GPU' Prices - RTX 5080 and RTX 5090
Let’s compare prices for top-level consumer GPUs, that are suitable for LLMs in particular and AI in general. Specifically I’m looking at RTX-5080 and RTX-5090 prices.