Running AI locally means queries never leave your hardware. No cloud vendor. No terms of service that change next quarter. No federal subpoena to a third party that holds your community's data. That is the sovereignty case for local AI. The practical case is that the tools to do this are now accessible enough that a single person with moderate technical experience can set one up in an afternoon.

The core stack

Three tools do most of the work:

  • Ollama handles downloading and running open-source models on your hardware. It runs on Mac, Windows, and Linux. One command installs a model. It can also serve a local network so multiple users share one machine.
  • Open WebUI gives non-technical staff a familiar chat interface -- similar to ChatGPT -- that connects to Ollama running on the same machine or on a local server. No command line required for end users.
  • qwen3:14b is a strong general-purpose model at the 14-billion-parameter tier. It runs well on a standard laptop with 16GB of RAM, handles multilingual content including some Indigenous languages, and performs reliably on document analysis, drafting, and summarization tasks.

Hardware guidance

  • Entry level (single user, 14B models): Any laptop or mini PC with 16GB of RAM. A Mac Mini M2 or M4 base model works well. Expected cost: $600 to $800 used or $800 to $1,000 new.
  • Shared departmental use (multiple users, 14B to 32B models): A Mac Mini M4 Pro with 24GB unified memory or a Windows workstation with 32GB RAM and a mid-range GPU. Expected cost: $1,200 to $2,000.
  • Org-wide server (70B models, RAG pipeline): A Mac Mini M4 Pro with 48GB unified memory is the most accessible entry point at this tier -- roughly $2,000. A workstation with an Nvidia GPU and 48GB or more VRAM works as well but requires more configuration.
  • Models are free. The only ongoing costs are hardware and electricity.

Video: setting up a local AI stack

This walkthrough covers installing Ollama, pulling a model, and connecting Open WebUI for a shared local interface.

What local infrastructure enables

  • Queries processed on hardware you own -- no data leaves your network.
  • Staff can use AI tools for drafting, summarizing, and analyzing documents without submitting community data to cloud vendors.
  • A single machine can serve an entire office over a local network.
  • No ongoing subscription cost once hardware is purchased.
  • The setup works offline -- useful in areas with unreliable connectivity.

What local infrastructure does not solve

  • Physical security. If the machine is physically accessible to unauthorized parties, the data sovereignty benefit is partially undermined. Standard endpoint security applies.
  • Network security. If Open WebUI is accessible on a network with poor access controls, queries from any connected device can reach the model. Limit access to trusted devices on the local network.
  • Data governance. A local AI tool does not automatically know which data is culturally restricted, legally sensitive, or subject to tribal data sovereignty policies. Staff still need guidance on what to input and what to keep out.
  • Model capability gaps. Local models at the 14B tier are capable but not equivalent to the largest cloud models for complex reasoning or specialized domains. For most drafting, summarization, and Q&A tasks tribal staff perform, the gap is acceptable. For high-stakes legal, medical, or regulatory analysis, no AI -- local or cloud -- replaces expert review.
  • Updates and maintenance. Someone needs to keep the software updated and monitor for security advisories on Ollama and Open WebUI. This is light work but requires assigning responsibility.

Local AI is a floor, not a ceiling. It gives your team a safe baseline for AI use -- a tool that works, does not transmit data externally, and does not require cloud accounts or subscriptions. From that baseline, you can expand capability as your data governance policy matures.