How to Run Ollama on a Synology NAS — Australia Setup Guide

You can run Ollama on a Synology NAS using Docker via Container Manager. But only on x86-based models with enough RAM. This guide covers compatible hardware, setup steps, model selection, and what to expect from local AI inference on a NAS in Australia.

This page contains affiliate links. If you purchase via our links we may earn a small commission, at no extra cost to you. Editorial independence policy.

Running Ollama on a Synology NAS is practical on x86-based models. The DS425+, DS925+, DS1525+, and above. Using Container Manager (Docker) to deploy the Ollama container with a persistent volume for model storage. This is genuinely useful for private AI inference: your queries stay on your own hardware, no API keys, no usage costs, no data leaving your network. This guide covers the complete setup. Container Manager deployment, model download, Open WebUI integration for a ChatGPT-like interface, performance expectations on NAS hardware, and which models are practical at each CPU RAM tier. It also answers the honest question: is a NAS actually the right place for Ollama? Australian model pricing is in the AU section below.

In short: Ollama runs on Synology NAS units with an Intel or AMD x86 CPU via Container Manager (Docker). You need at least 8GB RAM for small models like Llama 3.2 3B or Phi-3 Mini. ARM-based models (DS124, DS223, DS223J) are not compatible. CPU-only inference is slow. Expect 3-8 tokens per second on typical prosumer hardware. This is suitable for personal, private AI use, not production workloads.

What Is Ollama and Why Run It on a NAS?

Ollama is an open-source runtime that makes it straightforward to download, manage, and serve large language models locally. Instead of sending queries to OpenAI or another cloud provider, everything runs on your own hardware. For privacy-conscious users, those with sensitive data, or anyone who wants to avoid per-token API costs, local inference is appealing.

A NAS is an interesting host for Ollama because it's already on 24/7, it's connected to your local network, and it has enough CPU headroom to run inference workloads in the background. On a capable Synology model, you can expose Ollama's REST API to other devices on your LAN. So your laptop, phone, or home automation system can all make requests without anything leaving your network.

The practical use cases for Australians running this at home or in a small office include: summarising documents, drafting text, answering questions about local files, or integrating with tools like Open WebUI for a ChatGPT-style interface running entirely on local hardware.

Which Synology Models Can Run Ollama?

This is the most important question to answer before you invest time in this setup. Ollama requires an x86-64 CPU. That rules out every Synology model built on a Realtek ARM processor. Including the DS124, DS223, and DS223J, which are some of the most popular entry-level units sold in Australia.

The models that work are the Plus-series and higher, all of which use Intel Celeron or AMD Ryzen processors. RAM is the other critical constraint. Ollama needs enough free RAM to load a model into memory. If you run out, it either crashes or falls back to extremely slow behaviour. As a practical rule, budget at least 6-8GB of available RAM for a 3B parameter model, and 12-16GB for a 7B model.

Synology Models Suitable for Ollama (AU Retail, March 2026)

DS225+ DS225+ DS425+ DS425+ DS725+ DS725+ DS925+ DS925+ DS1525+ DS1525+ DS1825+ DS1825+
CPU Intel Celeron J4125 (4-core)Intel Celeron J4125 (4-core)AMD Ryzen R1600 (2-core)AMD Ryzen R-series (4-core)AMD Ryzen V1500B (4-core)AMD Ryzen V1500B (4-core)
Base RAM 2GB DDR42GB DDR44GB DDR4 ECC4GB DDR4 ECC8GB DDR4 ECC8GB DDR4 ECC
Max RAM 6GB6GB32GB32GB32GB32GB
Ollama Ready (base RAM)? No. Upgrade requiredNo. Upgrade requiredMarginal. Upgrade recommendedMarginal. Upgrade recommendedYes (small models)Yes (small models)
AU Price from $599 (PLE Computers)$819 (Scorptec)$869 (Mwave)$995 (Scorptec)$1,285 (Mwave)$1,765 (Mwave)

Prices last verified: 18 March 2026. Always check retailer before purchasing.

RAM upgrade required for most models: The DS225+ and DS425+ ship with only 2GB of RAM. Far too little for any useful LLM. You'll need to add a compatible SO-DIMM module before Ollama is practical. Synology's official 4GB non-ECC SO-DIMM (D4NESO-2666-4G) is $168 at Mwave. The DS725+, DS925+, and higher models come with 4GB or 8GB but benefit from an upgrade if you plan to run 7B parameter models.

Models That Will Not Work

Don't attempt this on ARM-based Synology units. The DS224+, DS423, DS223j, and any unit with a Realtek or Marvell processor. These cannot run x86 Docker containers. Only models with an Intel or AMD x86 CPU support the Ollama container. For a full list of current models with processor types, see the Synology NAS Australia guide.

Prerequisites Before You Start

Before installing Ollama, make sure your Synology is ready:

  • DSM 7.2 or later. Container Manager replaced Docker Package in DSM 7.2. If you're on an older DSM, update first.
  • Container Manager installed. Open Package Center and install Container Manager if it isn't already present. It's free.
  • Sufficient free RAM. Check Resource Monitor. You need at least 6GB free after DSM's own overhead (roughly 1-2GB) to run a 3B model. 8-10GB free is comfortable for 7B models.
  • Sufficient storage space. LLM model files are large. A 3B parameter model in Q4 quantisation is approximately 2GB. A 7B model is 4-5GB. Make sure your NAS volume has room, and consider putting model storage on an NVMe SSD cache volume if available for faster model loading.
  • A static IP or DHCP reservation for your NAS. You'll be connecting other devices to Ollama's API on port 11434. A consistent IP address makes this reliable.

Step 1. Create a Shared Folder for Ollama Data

Ollama stores downloaded models in a data directory. Rather than letting them land inside the container's ephemeral storage (which would be lost if the container is recreated), map a persistent folder on your NAS volume.

  1. Open File Station or Control Panel → Shared Folders.
  2. Create a new shared folder. For example, docker if you don't already have one, then inside it create a subfolder: docker/ollama.
  3. Note the full path. It will be something like /volume1/docker/ollama. You'll use this in the container configuration.

Step 2. Deploy Ollama via Container Manager

Synology's Container Manager provides a GUI for running Docker containers. You can deploy Ollama through it without needing to touch the command line. Though SSH gives you more control if you're comfortable with it.

Using the Container Manager GUI:

  1. Open Container Manager from the DSM main menu.
  2. Go to Registry and search for ollama/ollama. Select the official image and click Download. Choose the latest tag.
  3. Once downloaded, go to Container and click Create.
  4. Select the ollama/ollama image.
  5. Give the container a name. E.g., ollama.
  6. Under Advanced Settings → Volume, add a volume mount: map /volume1/docker/ollama (host path) to /root/.ollama (container path).
  7. Under Port Settings, map host port 11434 to container port 11434 (TCP).
  8. Under Environment, add the variable OLLAMA_HOST with value 0.0.0.0. This allows connections from other devices on your network, not just localhost.
  9. Leave the network mode as bridge unless you have a specific reason to change it.
  10. Click Apply and start the container.
💡

Prefer the command line? SSH into your Synology (enable SSH in Control Panel → Terminal & SNMP first) and run: docker run -d --name ollama -p 11434:11434 -v /volume1/docker/ollama:/root/.ollama -e OLLAMA_HOST=0.0.0.0 --restart unless-stopped ollama/ollama. This gives you the same result with more control over restart behaviour.

Step 3. Download a Model

With the container running, you need to pull a model. You do this by executing a command inside the running container. In Container Manager, select the ollama container, click Details, then go to the Terminal tab and open a bash session. Alternatively, use SSH and run:

docker exec -it ollama ollama pull llama3.2:3b

This downloads the Llama 3.2 3B model in Q4 quantisation. Approximately 2GB. It's a practical starting point for CPU-only inference on a Synology NAS. The download goes directly to your mapped volume at /volume1/docker/ollama, so it persists across container restarts.

Recommended models for Synology NAS hardware (CPU-only):

  • llama3.2:3b. Good general-purpose 3B model. Needs ~6GB RAM free. Manageable on a DS925+ or DS1525+.
  • phi3:mini. Microsoft's Phi-3 Mini (3.8B). Efficient and capable for its size. Good on 8GB RAM systems.
  • mistral:7b. Stronger responses but needs ~10GB RAM free. Only practical on the DS1525+, DS1825+, or units with upgraded RAM.
  • qwen2.5:3b. Strong multilingual and coding capability at 3B scale. Practical on the same hardware as Llama 3.2 3B.

Avoid 13B models or larger. They'll thrash RAM and swap heavily on any Synology unit, making inference unusably slow or causing the container to crash.

Step 4. Verify Ollama Is Running

From any device on your local network, open a browser or use curl to test the Ollama API. Replace 192.168.1.x with your NAS's actual IP address:

curl http://192.168.1.x:11434/api/tags

You should see a JSON response listing the models you've downloaded. If you get a connection refused error, check that the container is running in Container Manager and that the port mapping is correct.

You can also run a quick inference test from the terminal:

docker exec -it ollama ollama run llama3.2:3b "Summarise the benefits of local AI inference in two sentences."

The first run will be slightly slower as the model loads into RAM. Subsequent queries in the same session are faster.

Step 5. Add a Web Interface with Open WebUI (Optional)

Ollama's API is useful for programmatic access, but for day-to-day use a chat interface is much more practical. Open WebUI is the most popular option. It gives you a ChatGPT-style interface that connects to your local Ollama instance.

Deploy it as a second container in Container Manager:

  1. Search the Registry for ghcr.io/open-webui/open-webui and download the main tag.
  2. Create a new container with the following settings:
    • Volume: map /volume1/docker/open-webui to /app/backend/data
    • Port: map host port 3000 to container port 8080
    • Environment variable: OLLAMA_BASE_URL = http://[your-NAS-IP]:11434
  3. Start the container and access Open WebUI at http://[your-NAS-IP]:3000.

On first launch, Open WebUI asks you to create an admin account. All data stays local. There's no cloud component unless you explicitly add one. This is the setup that makes local AI genuinely usable for non-technical household members or staff in a small office.

Performance Expectations on Synology Hardware

Set realistic expectations. CPU-only inference is slow compared to a desktop PC with a discrete GPU. On a DS925+ (AMD Ryzen quad-core) running a 3B model, expect roughly 4-8 tokens per second. That's readable in real-time but noticeably slower than cloud AI services. On a DS1525+ or DS1825+ with the same model, performance is similar. The V1500B is a capable processor but inference is still CPU-bound.

For most text tasks. Drafting emails, summarising documents, answering questions. This speed is entirely acceptable. Where it becomes frustrating is long-form generation: asking the model to write a 1,000-word document at 5 tokens per second means waiting several minutes. Know the use case before committing to this setup.

NVMe SSD cache can help with model load times (getting the model into RAM from storage is faster on NVMe), but once the model is loaded, inference speed is entirely CPU and RAM bandwidth. Storage speed doesn't help after that point.

Network Access and Remote Use

On your local network, Ollama is accessible at port 11434 on your NAS's IP address. For remote access. Accessing your local Ollama instance from outside your home. You have a few options, but there are important caveats for Australian users.

Many Australian NBN services, particularly those on CGNAT (Carrier Grade NAT), cannot receive inbound connections from the internet at all. CGNAT is common on mobile broadband, some fixed wireless NBN services, and some NBN 25/50 plans depending on your RSP. If your connection is behind CGNAT, port forwarding won't work and you'll need a VPN tunnel solution (like Tailscale or Synology's own VPN server) to bridge the gap.

Even on a standard NBN connection without CGNAT, typical upload speeds on an NBN 100 plan average around 17-20Mbps in practice (the plan cap is 20Mbps upload). For Ollama's REST API traffic. Which is text-based. This is more than sufficient. The bottleneck for remote access will be inference speed on the NAS, not your upload speed.

The recommended approach for remote access is Tailscale: install Tailscale on your NAS (it's available as a Synology package or can run as a Docker container) and on your remote device. This creates a private mesh network that bypasses CGNAT entirely and doesn't require port forwarding or exposing Ollama to the public internet.

Do not expose Ollama directly to the internet without authentication. Ollama's default configuration has no authentication. Anyone who can reach port 11434 can query your models and consume your NAS resources. If you must expose it beyond your LAN, put it behind a reverse proxy (Nginx Proxy Manager works well on Synology) with authentication, or use a VPN. The Tailscale approach keeps it off the public internet entirely.

Hardware Recommendations by Use Case

If you're buying a Synology NAS specifically with Ollama in mind (or want to ensure the one you're buying will support it), here's how the current AU retail lineup stacks up:

Minimum viable for Ollama (3B models only, with RAM upgrade): The DS225+ suits small households who want local AI inference alongside standard NAS duties. It needs a RAM upgrade. The base 2GB is inadequate. And at $585 from Mwave or $599 from Scorptec, the total investment including a RAM module is roughly $750-770. Inference will be slow but functional for light use.

Better balance for mixed workloads: The DS925+ at $995 from Scorptec ships with 4GB RAM and an AMD Ryzen quad-core processor. With a RAM upgrade to 16GB or 32GB, this handles 7B models comfortably alongside normal NAS workloads. The DS725+ at $869 (Mwave, Scorptec) offers a similar CPU profile in a 2-bay enclosure.

Best performance for Ollama on Synology: The DS1525+ at $1,285 from Mwave or $1,399 from Scorptec, or the DS1825+ at $1,799 from Scorptec, offer the AMD Ryzen V1500B with base 8GB RAM that can be expanded to 32GB. These suit users who want Ollama running persistently alongside other services. Synology Drive, Surveillance Station, Hyper Backup. Without resource contention.

Note on drive compatibility: Synology reversed its controversial third-party drive restrictions with DSM 7.3 in October 2025, restoring support for Seagate and WD drives on desktop Plus series models. However, M.2 NVMe SSDs still require drives from Synology's official Hardware Compatibility List for cache or storage pool creation. If you're buying NVMe for model storage, check compatibility before purchasing. Australian Consumer Law protections apply when purchasing from Australian retailers including Mwave, Scorptec, and PLE Computers.

Troubleshooting Common Issues

Container exits immediately after starting: Check Container Manager logs. The most common cause is insufficient RAM. The model can't be loaded and the process crashes. Check Resource Monitor for available RAM before starting the container.

Slow first response, then crashes: The NAS is running out of RAM and swapping to disk. Either switch to a smaller model, add more RAM, or both. Swap on a spinning HDD will make inference effectively unusable.

Can't connect from other devices on the LAN: Confirm the OLLAMA_HOST=0.0.0.0 environment variable is set in the container. Without this, Ollama only listens on localhost inside the container. Also check that port 11434 is mapped correctly in the container's port settings.

Model pull fails or times out: Ollama pulls models from the Ollama model registry over the internet. Check your NAS has internet access (test with a ping from the DSM terminal) and that DNS resolution is working. Large model files can time out on slower connections. Re-run the pull command and it will resume from where it stopped.

Open WebUI can't connect to Ollama: Verify the OLLAMA_BASE_URL environment variable in the Open WebUI container uses the NAS's LAN IP address, not localhost or 127.0.0.1. These two containers are separate processes and don't share a localhost context unless you use Docker's host networking mode.

Is a NAS the Right Place for Ollama?

This is worth answering honestly. A NAS is a reasonable host for Ollama if you already have a capable Synology unit and you want private, always-on AI access without buying additional hardware. It's a low-cost way to experiment with local LLMs if the NAS is already part of your setup.

It's not the right choice if you need fast inference, want to run larger models, or plan to use AI heavily throughout the day. A mini PC with 32GB RAM and a modern Intel or AMD processor. Or any machine with a mid-range NVIDIA GPU. Will give you significantly faster inference for less money than a comparable Synology NAS. The NAS wins on consolidation: one device doing storage, backup, media serving, and AI, rather than running separate hardware for each.

Synology's strength has always been DSM's ease of use and the reliability of its ecosystem for straightforward workloads. Container Manager makes deploying Ollama approachable for users who wouldn't want to manage a Linux server, and that's a genuine advantage. The Need to Know IT team's view is that Ollama on Synology is a solid fit for the "curious home user" segment. People who want to try local AI without building a dedicated machine. For anything more demanding, dedicated hardware is the honest recommendation.

Neither Synology nor any Australian retailer will walk you through this setup. Synology has no local support presence in Australia, and resellers generally won't provide post-sale technical guidance beyond the basics. That's the gap this guide is designed to fill.


🇦🇺 Australian Buyers: NAS Models for Ollama

The most practical Synology models for running Ollama in Australia in 2026:

  • DS425+ (~$819 at Scorptec/Mwave). Intel Celeron J4125, 2GB RAM upgradeable to 6GB. Handles small models (Gemma 2B, Phi-3 Mini) for personal use.
  • DS925+ (~$995 at Scorptec). AMD Ryzen R1600, 4GB ECC RAM upgradeable to 32GB. Best value for Llama 3.2 3B and similar models at multiple concurrent users.
  • DS1525+ (~$1,285 Mwave). AMD Ryzen V1500B, 8GB ECC upgradeable to 32GB. Handles 7B quantised models for light household use.

All available from Scorptec, Mwave, and PLE. For remote API access over NBN: Ollama's API can be accessed via Tailscale or DDNS. CGNAT-affected connections cannot use direct port forwarding. Use the NBN Remote Access Checker to confirm your connection supports inbound connections.

Related reading: our NAS buyer's guide, our NAS vs cloud storage comparison, and our NAS explainer.

Free tools: NAS Sizing Wizard and AI Hardware Requirements Calculator. No signup required.

Can I run Ollama on a Synology DS223 or DS124?

No. The DS223 and DS124 both use Realtek ARM processors (RTD1619B). Ollama requires an x86-64 CPU. These models are not compatible. You need a Plus-series or higher Synology NAS with an Intel or AMD processor. Starting with the DS225+ (Intel Celeron J4125) at $585 from Mwave.

How much RAM does Ollama need on a Synology NAS?

At minimum, you need enough free RAM after DSM's own overhead (roughly 1-2GB) to load the model. For a 3B parameter model in Q4 quantisation, budget 6GB free. Meaning your NAS should have at least 8GB total RAM. For a 7B model, aim for 10-12GB free, so 16GB total is a comfortable target. The DS225+ and DS425+ ship with only 2GB RAM and need an upgrade before Ollama is practical. Synology's 4GB non-ECC SO-DIMM (D4NESO-2666-4G) is $168 at Mwave and compatible with the J4125-based models.

Will Ollama slow down my NAS's normal file sharing and backup functions?

It depends on the workload. Ollama inference is CPU and RAM intensive. During an active inference request, you may see elevated CPU usage (80-100% on all cores) and reduced headroom for other tasks. In practice, for a home or small office NAS that isn't under constant heavy load, the impact is manageable. Inference requests are typically short bursts, not sustained. If you run Hyper Backup or Active Backup jobs at the same time as Ollama inference, expect both to slow down. Schedule backup jobs for off-peak hours if you plan to use Ollama regularly.

Can I access Ollama running on my Synology from outside my home network?

Yes, but not by simply opening a port on your router. First, check whether your NBN connection is behind CGNAT. This is common on some Australian RSPs and blocks inbound connections entirely. The recommended approach for remote access is Tailscale: install it on your NAS and your remote devices to create a private mesh network that works through CGNAT and doesn't expose Ollama to the public internet. Never expose port 11434 directly to the internet without authentication, as Ollama has no built-in access control.

Which Ollama models work best on Synology NAS hardware?

For CPU-only inference on most Synology prosumer units, 3B parameter models in Q4 quantisation are the practical sweet spot. Llama 3.2 3B, Phi-3 Mini, and Qwen 2.5 3B all deliver useful responses at manageable speeds (4-8 tokens/second on a Ryzen-based NAS). Mistral 7B is viable on units with 16GB or more RAM but is noticeably slower. Avoid anything larger than 7B. 13B and above will cause RAM pressure and may crash the container on all current Synology desktop models.

Does Ollama on Synology support GPU acceleration?

No. Synology NAS units do not have discrete GPUs, and DSM does not support GPU passthrough to Docker containers even for the integrated graphics present in Intel Celeron processors. All inference runs on the CPU only. If GPU acceleration is important to you, Ollama on a NAS is not the right solution. Consider a dedicated mini PC or a machine with an NVIDIA GPU running Ollama natively on Linux or Windows.

Will Ollama data survive a container update or recreation?

Yes, provided you've mapped the model storage directory correctly. When you map /volume1/docker/ollama (or your chosen path) to /root/.ollama inside the container, all downloaded models are stored on your NAS volume. Not inside the container's filesystem. If you delete and recreate the container, or pull a new Ollama image, your models remain in place and are immediately available without re-downloading. This is why the volume mapping step is critical. Skip it and all model downloads are lost when the container is updated.

Looking for the right Synology NAS for your home or office? The Need to Know IT buying guide covers every current AU-available model with real prices from Mwave, Scorptec, and PLE Computers.

Synology NAS Buying Guide →