AI Hardware
Chronicles
Running AI locally — real benchmarks on consumer hardware. VRAM math, GPU picks, Ollama setups, and model recommendations.
Most guides in English. Leer en español →
Latest guides
NVIDIA DGX Spark — Is This the Ultimate Local AI Workstation?
Grace Blackwell chip, 128 GB unified memory, 1 PFLOP AI compute. What it runs, who needs it, and what to buy right now while you wait.
Run Gemma 4 Locally — Complete Setup Guide with Ollama (2026)
Google Gemma 4 in 12B and 27B. 6.6 GB VRAM at Q4, 256K context, vision support. Step-by-step Ollama install, benchmarks, and compatible GPU list.
Running AI Locally Without a GPU: CPU-Only LLMs in 2026
phi-3-mini at 14 tok/s on a standard i7. Full model table for CPU inference with real benchmarks.
Local AI on Mac with Apple Silicon: What Models Can You Run?
An M4 Pro with 24 GB runs Llama 3.1 8B at 45 tok/s. Real M2/M3/M4 benchmarks, model tier table, and install guide.
Llama vs Mistral vs DeepSeek: Which Model to Download for Your GPU
Real comparison of three most-searched open-source models. VRAM table by tier, GPU benchmarks.
Best GPUs for Local AI in 2026: Real Comparison by Budget
RTX 3060, 4060 Ti 16GB, 4070 Ti Super, used 3090 and 4090. Real tok/s benchmarks and current prices.
Build a Local AI PC on a Budget — Complete 2026 Guide
RTX 3060 12GB + Ryzen 5 7600 + 32GB DDR5. 30 tok/s with Llama 8B Q4. Components table with price ranges.
RTX 3060 for AI: What Models Can You Run in 2026?
12 GB VRAM at entry price. Real benchmarks (30 tok/s Llama 8B Q4), compatible model table, Ollama setup.
Ollama vs LM Studio: Which to Choose for Home AI? (2026)
Ollama for developers and headless servers, LM Studio for GUI users. Real data, comparison table.
How Much VRAM Do I Need to Run AI Locally? (2026)
Real VRAM tables: Llama 3.1 8B needs 5 GB in Q4, DeepSeek R1 32B needs 19.2 GB. Quantization comparison.
How to Install Ollama on Windows: Step-by-Step Guide (2026)
Complete guide with NVIDIA, AMD, or CPU-only setup. Configuration, first models and common error fixes.
DeepSeek R1 Locally: VRAM Requirements, Distillations and Install
Full 671B requires clusters. 8B/14B/32B distillations run at home from 4.8 GB. GPU table and Ollama install.
Run Phi-4 Locally — Setup with Ollama and LM Studio
Microsoft Phi-4 (14B) in Q4: 8.4 GB VRAM, quality 88/100, MIT license. Step-by-step install with benchmarks.
Best LLMs for Coding Locally — VRAM and Performance Compared
Technical comparison of top coding LLMs. Exact VRAM requirements and which to choose for your hardware.
Whisper Local Transcription — GPU Benchmarks and Setup Guide
Full setup guide for Whisper. GPU benchmarks by card, tool comparison, and real-world use cases.