Ollama is great for running LLMs locally, but it doesn’t ask where you want to store your 5GB–10GB+ models. On Macs with limited internal SSD space, that’s a problem.

Here’s how I tracked down Ollama’s hidden model cache, moved it to an external drive, and kept everything running smoothly — even on my M1 Mac.

The Ultimate Guide: Move Ollama Model Storage to an External Drive on macOS

TL;DR

Models are secretly stored in ~/.ollama/models

Why This Guide?

Other guides (like this gist) touch on this, but:

What You’ll Need

Step-by-Step Instructions

1. Locate Ollama’s Storage Folder

Ollama doesn’t mention this anywhere, but it saves models to:

~/.ollama/models

Check if it exists:

ls -lah ~/.ollama/models

You’ll likely see folders and files totaling 5–15 GB, depending on how many models you’ve pulled.

2. Create a Destination on Your External Drive

Plug in your external and make a folder for Ollama:

mkdir -p /Volumes/_EXTERNAL_DRIVE_NAME/_PATH_TO_/ollama

3. Move the Models

mv ~/.ollama/models /Volumes/_EXTERNAL_DRIVE_NAME/_PATH_TO_/ollama

You now have freed up internal storage.

Ollama expects the folder to be in ~/.ollama/models, so we trick it:

ln -s /Volumes/_EXTERNAL_DRIVE_NAME/_PATH_TO_/ollama ~/.ollama/models

This way, the CLI keeps working as if nothing changed.

5. Test It

List models:

ollama list

Run a model:

ollama run deepseek-r1:7b

If it responds — your symlink works.

6. Optional: Set the OLLAMA_MODELS Environment Variable

You can use this instead of a symlink, but it’s less reliable in current builds:

echo ‘export OLLAMA_MODELS="/Volumes/2USBEXT/DEV/ollama">> ~/.zshrc
source ~/.zshrc

Use it only if you want to switch between different drives easily.

Bonus Section: Best Models for M1 Macs

I tested a bunch, and these are fast + stable:

Model Use Case Size Notes
phi3 Chat + reasoning ~2.2 GB Fast, small, ideal for M1
mistral Chat + code ~4.1 GB Great all‑rounder
llama3:8b Smarter conversation ~4.7 GB Best Meta model < 10 GB
codellama Code gen & fill ~5.6 GB If you’re building code tools
gemma:2b Lightweight anything ~1.4 GB Tiny model for light tasks

Pull with:

ollama pull phi3
ollama pull mistral
ollama pull llama3

All of them will now download straight to your external drive thanks to the symlink.

Cleanup Tip

Want to find where that space went?

du -sh ~/.ollama/*

Or if you’re really unsure where Ollama is hiding files:

sudo find /Users -name "*.gguf" 2>/dev/null

Cleanup Tip

Ollama is fast and powerful, but its CLI assumes you’re fine giving up 10–20 GB of SSD space without asking.

This guide puts that control back in your hands — and hopefully saves a few MacBook lives in the process.


🧠 Frequently Asked Questions: Ollama Storage

Quick answers for common storage and performance concerns.

1) Core Storage & Path FAQs

Where does Ollama store models on macOS by default?

By default, Ollama stores models in the hidden directory ~/.ollama/models within your user home folder.

Can I just move the folder through Finder?

It’s safer to use the Terminal for this to ensure hidden files are handled correctly and to create the symbolic link (symlink) that Ollama needs to find the new location.

What is a symlink and why do I need it?

A symbolic link (symlink) acts like a shortcut. By creating one, you tell macOS that whenever Ollama looks for ~/.ollama/models, it should actually go to your external drive.

2) Troubleshooting & Performance FAQs

Will models run slower from an external drive?

If you are using a USB 3.0 or SSD external drive, the performance impact is usually negligible for inference. However, initial model loading might take a few seconds longer compared to an internal NVMe SSD.

What happens if I open Ollama without the drive plugged in?

Ollama will see an empty directory or a broken link and won’t find your models. Simply plug the drive back in and restart Ollama to restore access.

Can I use the OLLAMA_MODELS environment variable instead?

Yes, you can set OLLAMA_MODELS=/Volumes/YourDrive/ollama in your .zshrc, but the symlink method is often more ‘set and forget’ across different terminal sessions and apps.

3) Maintenance FAQs

How do I move models back to internal storage?

Delete the symlink (rm ~/.ollama/models), move the folder back from your external drive to ~/.ollama/, and restart the Ollama service.

Does this work for Linux or Windows too?

The concept is the same, but the paths differ. On Linux, the default path is usually /usr/share/ollama/.ollama/models. On Windows, it’s typically in your user AppData folder.

How can I see which models are taking up the most space?

Run ollama list to see sizes, or use du -sh ~/.ollama/models/* in your terminal for a detailed breakdown per model file.


System Update: Thinking Modes

Explore the Framework

These concepts are part of a broader framework for building intent-aware AI systems. I've distilled these strategies into a short, practical guide called Thinking Modes.

View the Book →

END_OF_REPORT 🌿✨