What Should I Run?
Stop guessing. Pick the right model for what you actually need to do — with RAM requirements and Mac compatibility for each recommendation.
Classify, tag, and sort emails locally without sending your data to the cloud. Small models excel at this task.
Write, complete, and refactor code locally. Larger models significantly outperform smaller ones for complex coding tasks.
Analyze server logs, detect anomalies, and summarize incidents locally. Runs 24/7 without API costs.
Extract text from scanned documents, receipts, and handwritten notes. Vision models run entirely on your Mac.
Run a local AI assistant for smart home control, voice commands, and automation rules. No cloud dependency.
Build retrieval-augmented generation pipelines over your private documents. Keep your data local.
Running OpenClaw with local models on a Mac Mini? Here's which models actually fit — and which ones will tank your performance.
Summarize articles, meeting notes, research papers, and long documents locally. Fast and private.
DevPulse checks your actual available RAM and tells you which models will run right now — not just on paper.
Download for macOS