NAS for AI: Training Data & Model Storage
Set up a NAS as central storage for your AI lab. Share model files, training data, and backups across machines over your LAN.
What You'll Need
- A 2-bay (or more) NAS — Synology, TerraMaster, or UGREEN
- 2x hard drives — 4TB+ each recommended (HDD for bulk, SSD for speed)
- An Ethernet cable — NAS over WiFi is painfully slow; wire it to your router
- A computer on the same local network
1 Choose Your NAS
| NAS | Bays | RAM | Network | Price | Best For |
|---|---|---|---|---|---|
| UGREEN NASync DXP2800 | 2 | 4GB | 2.5GbE | ~$200 | Budget-friendly starter NAS |
| TerraMaster F2-223 | 2 | 4GB | 2.5GbE | ~$230 | Good value, Docker support |
| Synology DS224+ | 2 | 2GB | 1GbE | ~$300 | Best software ecosystem (DSM) |
| Synology DS423+ | 4 | 2GB | 2x 1GbE | ~$500 | Room to grow, expandable |
| QNAP TS-264 | 2 | 8GB | 2.5GbE | ~$400 | Power user, HDMI out, more RAM |
2 Initial Setup & Storage Pool
After inserting your drives and connecting Ethernet:
- Find your NAS on the network — most brands have a discovery tool:
# Synology: find.synology.com # TerraMaster: tnas.online # UGREEN: ugreenlink.com # Or just scan your network: arp -a | grep -i "unknown"
- Run the setup wizard — create your admin account and let it initialize the drives
- Create a storage pool — choose RAID 1 (mirror) for redundancy, or SHR (Synology Hybrid RAID) if available
- Create shared folders — set up folders for different purposes:
# Recommended folder structure: /ai-models/ # Ollama models, GGUF files /training-data/ # JSONL, datasets, corpora /backups/ # Machine backups, configs /projects/ # Shared project files
3 Mount the NAS from Your AI Machines
Once your NAS has shared folders, mount them on every machine that needs access.
macOS — mount via Finder or terminal:
Linux — mount via fstab for auto-mount on boot:
Windows — map as a network drive:
4 Point Ollama & Training Data at the NAS
Now redirect your AI tools to use the NAS for storage. This means any machine on your network pulls models from the same place.
Ollama — change the model storage path:
Symlink approach — keep tools unchanged, redirect the folder:
Training data — organize on the NAS:
5 Backups & Maintenance
A NAS isn't just storage — it's your safety net. Set up automated backups and health monitoring.
Automated snapshots (Synology):
Scrub schedule — check drive health:
Backup the NAS itself:
- External USB drive — plug into the NAS USB port, schedule nightly copy of critical folders
- Second NAS — rsync between two NAS devices for full redundancy
- Cloud (optional) — Synology Hyper Backup supports Backblaze B2 (~$5/TB/month) for offsite
Monitor drive health from the command line:
| Storage Task | Size Estimate | 2TB NAS | 4TB NAS | 8TB NAS |
|---|---|---|---|---|
| Small models (1B–3B) | ~2GB each | ~1000 | ~2000 | ~4000 |
| Medium models (7B–8B) | ~5GB each | ~400 | ~800 | ~1600 |
| Large models (70B) | ~40GB each | ~50 | ~100 | ~200 |
| Fine-tune JSONL datasets | ~10MB–1GB | Thousands | Thousands | Thousands |
| Chat history exports | ~1MB/1000 msgs | Millions | Millions | Millions |
✅ What You've Set Up
- A NAS with a RAID-protected storage pool and organized shared folders
- Network mounts on Mac, Linux, or Windows for seamless access
- Ollama models stored centrally — pull once, use from any machine
- Training data and exports organized and accessible over LAN
- Automated snapshots, scrub schedules, and drive health monitoring
Next Steps
- Run Docker on the NAS — some NAS devices (Synology, QNAP) can run containers directly. Host Open WebUI right on the NAS.
- Set up a Time Machine target — most NAS devices can act as a Mac backup destination. Protect your whole machine, not just AI files.
- Add an SSD cache — if your NAS has M.2 slots, adding an SSD cache dramatically speeds up small file reads (metadata, JSONL files).
- Pair with your AI server — the always-on AI server loads models from the NAS, serves them to everything else on your network.
- Centralize training data for fine-tuning — keep all your JSONL files on the NAS and point your model library at it.