Set Up an Always-On AI Server
Configure a mini PC as a dedicated AI machine. Auto-start Ollama on boot, expose the API to your LAN, and access it from any device.
What You'll Need
- A mini PC, NUC, or old laptop you can leave running 24/7
- 16GB RAM minimum (32GB to run multiple or larger models)
- An Ethernet connection to your router (Wi-Fi works but adds latency)
- A USB drive or SD card for installing the OS
1 Choose Your Hardware
| Option | Price | RAM | Best For |
|---|---|---|---|
| Old laptop | $0 (reuse) | 8-16GB | Getting started, testing the concept |
| Beelink Mini S12 Pro | ~$160 | 16GB | Budget pick, quiet, low power (~15W) |
| Intel NUC / ASUS NUC | ~$300 | 32GB | Reliable, compact, good for 8B-13B models |
| Mac Mini (M-series) | ~$500+ | 16-24GB unified | Best performance per watt, runs large models well |
| Mini PC + NVIDIA GPU | ~$600+ | 32GB + VRAM | Fastest inference, handles 30B+ models |
2 Install the OS
Linux (Ubuntu Server):
- Download Ubuntu Server 24.04 LTS
- Flash to USB with balenaEtcher or
dd - Boot from USB, follow the installer
- Enable SSH during install (or
sudo apt install openssh-serverafter) - Set a static IP (we'll need this later)
Mac Mini: Just use macOS — Ollama runs great natively. Set a static IP in System Settings → Network → Ethernet → Details → TCP/IP → Manually.
3 Install Ollama & Configure for LAN
By default, Ollama only listens on 127.0.0.1 (localhost). To access it from other devices on your network, you need to tell it to listen on all interfaces.
Linux (systemd):
Mac (launchd):
Test from another device on your network:
OLLAMA_HOST=0.0.0.0 exposes Ollama to your local network, but your router's firewall keeps it off the internet. Do not port-forward 11434 to the internet — there's no authentication on the Ollama API by default.
4 Auto-Start on Boot
The server needs to come back up after power outages and reboots without you touching it.
Linux: Ollama's installer already creates a systemd service. Verify it's enabled:
Mac: Ollama auto-starts by default (it's a menu bar app). For a headless Mac Mini, enable auto-login:
BIOS settings (important for unattended operation):
- Restore on AC Power Loss: Power On — so it reboots after power cuts
- Wake on LAN: Enabled — wake it remotely if it sleeps
- Disable sleep/hibernate — the whole point is always-on
5 Use It From Anywhere in Your Home
Now any device on your network can talk to your AI server. Here's how to point different tools at it.
Ollama CLI from another machine:
Open WebUI (browser-based chat):
From Python scripts:
From AI OS:
ai.local instead of the IP. In Pi-hole admin → Local DNS → DNS Records, add ai.local → 192.168.1.50.
✅ What You've Set Up
- A dedicated AI server that starts automatically on boot — no monitor or keyboard needed
- Ollama listening on your LAN so any device can use it
- Remote access via SSH for maintenance
- Optional web UI for ChatGPT-like browser access from any device
Next Steps
- Add your model library — follow our Model Library guide to organize models on NVMe storage attached to the server.
- Set up Tailscale — access your AI server from outside your home. Tailscale creates a secure mesh VPN with zero port forwarding.
- Monitor with Grafana — track CPU, RAM, and GPU usage over time. Know when you're pushing the limits.
- Run AI OS on the server — put the full AI OS stack on your always-on machine so background loops, memory, and feeds run 24/7.
sudo apt update && sudo apt upgrade -y weekly (Linux) or enable automatic security updates. An always-on machine on your network should always be patched.