Install Ollama & Run Your First LLM
Download Ollama, pull a model, and have a conversation with a local AI in under 5 minutes. No GPU required.
What You'll Need
- A computer with at least 8GB RAM (16GB+ recommended for 7B+ models)
- macOS, Linux, or Windows 10/11
- A terminal app (Terminal on Mac, bash on Linux, PowerShell on Windows)
- ~5GB free disk space for your first model
1 Install Ollama
Mac — Download from the official site or use Homebrew:
Linux — One-line install script:
Windows — Download from ollama.com/download and run the installer.
Verify the install:
2 Pull Your First Model
Ollama works like Docker for AI models — you pull a model once, and it stays on your machine.
3 Start Chatting
That's it. You're running a language model locally on your own hardware. No API keys, no cloud, no data leaving your machine.
Press Ctrl+D or type /bye to exit the chat.
4 Try More Models
Ollama has a library of models. Here are the best ones to start with:
| Model | Size | RAM Needed | Best For |
|---|---|---|---|
llama3.2 |
2GB | 8GB | Fast chat, Q&A, quick tasks |
llama3.1:8b |
4.7GB | 16GB | Conversations, writing, coding |
mistral |
4.1GB | 16GB | Balanced quality, good reasoning |
codellama |
3.8GB | 16GB | Code generation and debugging |
phi3 |
2.2GB | 8GB | Small but surprisingly capable |
deepseek-coder-v2 |
8.9GB | 32GB | Serious code + math |
5 Use Ollama as an API
Ollama runs a local server on port 11434. You can hit it from any app, script, or tool:
http://localhost:11434 instead — instant local AI for your existing tools.
✅ What You've Set Up
- Ollama installed and running on your machine
- At least one local LLM downloaded and ready to chat
- A local API server you can integrate with any tool
- Zero cloud dependency — everything runs offline
Next Steps
- Add voice input — pair Ollama with Whisper for speech-to-text (guide coming soon)
- Build a model library — organize models on external NVMe storage (guide coming soon)
- Make it always-on — set up Ollama to auto-start on boot and serve your LAN (guide coming soon)
- Try AI OS — our project adds memory, identity, and learning loops on top of Ollama