Local AI models with Ollama
Why local AI models?
Running AI models locally on your desktop provides several important advantages:
| Aspect | Local Models | Cloud API |
|---|---|---|
| Privacy | Your data stays on your computer | Data sent to cloud servers |
| Cost | Free after installation | Pay per API call |
| Speed | No internet latency | Depends on connection |
| Offline | Works without internet | Requires internet connection |
| Control | Full control over your data | Data handled by third parties |
Why Ollama?
Ollama is the leading platform for running open-source AI models locally. Key features:
- ✅ Easy installation and setup
- ✅ Thousands of available models
- ✅ Lightweight and fast
- ✅ Cross-platform (Windows, macOS, Linux)
- ✅ Simple model management
- ✅ OpenAI-compatible API
Installation
1. Download and install Ollama
Visit ollama.ai and download the installer for your operating system.
2. Verify installation
After installation, verify Ollama is working:
ollama --version
For Windows, restart your terminal after installation.
3. Pull a model
Download a model (example with qwen2.5):
ollama pull qwen2.5:7b
This will download the model. Depending on your internet speed, this may take a few minutes.
Model installation
Quick start
For the best balance between quality and performance, we recommend:
ollama pull qwen3-vl:4b
Recommended model: qwen3-vl:4b
This is the recommended model because it:
- ✅ Requires only 4GB of RAM
- ✅ Includes vision capabilities (see images)
- ✅ Offers good speed and quality balance