AI
AI and machine learning on Linux — deploy LLMs, GPU setup, self-hosted AI tools, and intelligent automation for sysadmins and DevOps engineers.
Model Context Protocol (MCP) on Linux with Ollama: Connect AI to Your Tools
Implement Model Context Protocol (MCP) on Linux to connect Ollama LLMs to external tools, databases, and APIs. Covers...
Piper TTS on Linux: Build a Self-Hosted Text-to-Speech Server
Deploy Piper text-to-speech on Linux with no cloud dependencies. Build a fast, private TTS server using systemd, a REST...
Open WebUI Custom Pipelines and Functions on Linux
Build custom pipelines and functions in Open WebUI on Linux. Create filter pipelines, RAG integrations, API-connected...
Flux Image Generation on Linux: Self-Hosted AI Art Server
Set up a self-hosted Flux image generation server on Linux. Covers ComfyUI and API-based workflows, GPU requirements,...
Ollama and LangChain on Linux: Build AI Agents with Local Models
Build autonomous AI agents on Linux using LangChain and Ollama. Covers tool-calling, ReAct patterns, memory chains, and...
Ollama Docker Compose: Complete GPU Stack for Linux
Deploy Ollama with Docker Compose on Linux with full NVIDIA GPU passthrough. Covers multi-container stacks, persistent...
Generate Ansible Playbooks with Local LLMs: AI-Assisted Infrastructure as Code
Build a local CLI tool that generates Ansible playbooks from natural language using Ollama. No vendor lock-in, full...
AI-Powered Log Analysis on Linux: Use Ollama to Parse Syslog, Journald, and Application Logs
Use Ollama and local LLMs to analyze Linux logs from journalctl, auth.log, and nginx. Includes a Python log analyzer,...
ComfyUI on a Headless Linux Server: Stable Diffusion with API Access
Install and run ComfyUI on a headless Linux server for Stable Diffusion image generation. Covers Python venv setup,...
Private LLM for Enterprise: Linux Deployment Architecture and Security Guide
Enterprise reference architecture for self-hosted LLMs on Linux. Covers data sovereignty, hardware sizing, network...