Overview
Ollama allows you to run large language models locally on your Mac. Perfect for privacy-focused work and offline usage.Supported Models
Any model available in Ollama:| Model | Size | Best For |
|---|---|---|
| Llama 3.1 70B | 40GB | Complex tasks |
| Llama 3.1 8B | 4.7GB | General purpose |
| Mistral 7B | 4.1GB | Fast responses |
| CodeLlama | 4.7GB | Programming |
| Phi-3 | 2.2GB | Lightweight |
| Gemma 2 | 5.4GB | Balanced |
| Qwen 2 | Various | Multilingual |
Setup
1
Install Ollama
Download from ollama.ai or:
2
Pull a Model
3
Verify Installation
4
Configure in EnConvo
- Open Settings → AI Provider
- Select Ollama
- Go to Credentials module
- Set endpoint:
http://localhost:11434
5
Select Model
Choose from your installed models
Configuration
| Setting | Description | Default |
|---|---|---|
| Credentials | Endpoint configuration | localhost:11434 |
| Model Name | Installed model | llama2:latest |
| Temperature | Creativity (0-2) | Medium (1) |
Reasoning Mode
Enable thinking for compatible models:| Option | Description |
|---|---|
| Disabled | Standard responses |
| Thinking | Enable reasoning |
Popular Models
System Requirements
| RAM | Recommended Models |
|---|---|
| 8GB | 7B models (Llama 3.1 8B, Mistral 7B) |
| 16GB | Larger 7B models, some 13B |
| 32GB | 13B-30B models |
| 64GB+ | 70B models |
Privacy Benefits
Complete Privacy
Data never leaves your Mac
Offline Access
Works without internet
No Usage Limits
Unlimited local queries
Full Control
Choose exactly which models to run
Troubleshooting
Connection refused
Connection refused
- Ensure Ollama is running:
ollama serve - Check port 11434 is available
- Verify endpoint in settings
Slow responses
Slow responses
- Use smaller models
- Close memory-intensive apps
- Consider quantized models
Out of memory
Out of memory
- Use smaller model
- Reduce context length
- Restart Ollama