Osaurus: The Native macOS AI Harness That Unifies Local and Cloud Intelligence

For years, using AI meant surrendering your data to someone else’s server. Every conversation, every file you uploaded, every preference you expressed stored in a cloud you don’t control. Then local AI arrived, but it came with a catch: terminal commands, compatibility nightmares, and the nagging feeling that your Mac was being treated as an afterthought.

Osaurus changes the equation.

Born from Dinoki Labs and co-founded by Terence Pae (formerly of Tesla and Netflix) and Sam Yoo, Osaurus is an open-source, Apple-only AI harness built purely in Swift for Apple Silicon. It sits between you and any AI model local or cloud and keeps the irreplaceable stuff on your machine: your context, your memory, your tools, your identity.

Since launching nearly a year ago, Osaurus has been downloaded over 112,000 times and has earned 5.2k GitHub stars and it’s not hard to see why.


What Exactly Is Osaurus?

At its core, Osaurus is what the industry calls a “harness” a control layer that connects different AI models, tools, and workflows through a single interface. Unlike developer-centric alternatives such as OpenClaw or Hermes, Osaurus presents a consumer-friendly native Mac app. It addresses security concerns by running operations in a hardware-isolated, virtual sandbox, limiting AI to a defined scope and keeping your computer and data safe.

The philosophy is simple: models are getting cheaper and more interchangeable by the day. What’s irreplaceable is the layer around them. Others keep that layer on their servers. Osaurus keeps it on your Mac.


The Model Ecosystem: Local Power Meets Cloud Flexibility

Local Models (MLX Optimized)

Osaurus runs local models through MLX, Apple’s own machine learning framework with first-class GPU support via unified memory. It maintains a curated, optimized model library on Hugging Face with quantizations specifically tuned for Apple Silicon.

Supported local models include:

  • Gemma 4 (2B to 31B variants)
  • Qwen3.6 (up to 122B MoE)
  • MiniMax M2.5/M2.7
  • DeepSeek V4
  • GPT-OSS
  • Llama
  • Mistral Medium 3.5
  • NVIDIA Nemotron Omni
  • Liquid AI’s LFM family (non-transformer architecture)
  • Apple’s on-device Foundation Models (macOS 26+, zero configuration)

Cloud Providers

When you need more power, Osaurus connects seamlessly to:

  • OpenAI (GPT-4o, o-series)
  • Anthropic (Claude family)
  • Google Gemini
  • xAI / Grok
  • Venice AI (privacy-focused, no data retention)
  • OpenRouter (one key, many providers)
  • Ollama and LM Studio (remote/local servers)

The best part? Your agents, memory, and tools stay intact when you switch from a local Gemma model to Claude or GPT-4o.


Feature Breakdown at a Glance

FeatureWhat It DoesWhy It Matters
Agent SystemCustom AI agents with unique prompts, memory, and themesOne agent for coding, another for research each specialized and persistent
Multi-Layer MemoryIdentity, pinned facts, and per-session episodes distilled automaticallyAgents remember what matters without bloating context windows (~800 tokens or less per turn)
Agent LoopModel writes a markdown todo list, executes it, and verifies resultsTrue autonomous task completion in a single chat window
Linux SandboxIsolated VM (Alpine Linux) for code executionRun shell, Python, Node.js with zero risk to your Mac (macOS 26+)
MCP Server + ClientFull Model Context Protocol supportShare tools with Cursor, Claude Desktop, and other MCP-compatible apps
20+ Native PluginsMail, Calendar, Vision, Browser, Git, Filesystem, Music, Search, FetchYour AI can actually do things on your Mac
Voice InputOn-device transcription via FluidAudioDictate in chat or use a global hotkey to transcribe into any app — no audio leaves your Mac
Cryptographic Identitysecp256k1 addresses for you and each agentVerifiable chain of trust, revocable access keys
RelaySecure WebSocket tunnels via agent.osaurus.aiPublic agent URLs without port forwarding or ngrok
Schedules & WatchersTimer-based or file-change-triggered agent runsAutomate daily journals, screenshot organizers, end-of-day commits
API CompatibilityDrop-in OpenAI, Anthropic, and Ollama endpointsAny SDK you already use just works at localhost:1337

System Requirements & Performance

SpecificationMinimumRecommended
macOS Version15.5+26+ (Tahoe) for Sandbox & Apple Foundation Models
ChipApple Silicon (M1, M2, M3, or newer)M3 Pro/Max or M4 for larger models
RAM (Local Models)64 GB128 GB for DeepSeek V4-class models
StorageVaries by modelExternal drive supported via OSU_MODELS_DIR

As Terence Pae notes, local AI’s “intelligence per wattage” is on its own innovation curve. “Last year, local AI could barely finish sentences, but today it can actually run tools, write code, access your browser, and order stuff from Amazon.”


Installation (Under a Minute)

Osaurus respects your time. Install via Homebrew:

brew install --cask osaurus

Or grab the .dmg from osaurus.ai. Launch with ⌘ Space → “Osaurus”, or use the CLI:

osaurus ui       # Open the chat UI
osaurus serve    # Start the server
osaurus status   # Check status

A five-step onboarding walks you through creating your first agent, picking a model, and setting up your cryptographic identity. No config files.


Who Is Osaurus For?

  • Privacy-conscious professionals in legal, healthcare, or finance who can’t risk data leaving their machine
  • Developers who want a native, API-compatible local server that isn’t another Electron app
  • Creatives and knowledge workers who want AI that remembers their preferences across sessions
  • AI enthusiasts who want to experiment with local models without wrestling with the command line
  • Businesses considering on-prem AI deployments (a direction the Osaurus team is actively exploring)

The Bigger Picture: Why Local-First AI Matters

Osaurus arrives at a pivotal moment. As cloud AI providers race to build massive data centers, the Osaurus team sees a future where organizations deploy a Mac Studio on-prem instead. “Instead of relying on the cloud, they can actually deploy a Mac Studio on-prem, and it should use substantially less power. You still have the capabilities of the cloud, but you will not be dependent on a data center,” Pae explained.

This isn’t just about privacy it’s about sovereignty. Your memory, your files, your tools, your identity. Owned by you. Encrypted at rest. Signed at every boundary. Nothing leaves your Mac unless you explicitly choose a cloud provider.


Frequently Asked Questions (FAQ)

Q: Is Osaurus really free and open source?
A: Yes. Osaurus is MIT licensed, built in public on GitHub, and free to use. You can read it, fork it, and ship with it.

Q: Do I need an internet connection to use Osaurus?
A: No, if you’re running local models, Osaurus works fully offline. Cloud providers are entirely optional and only connect when you choose.

Q: Can I use Osaurus with my existing AI tools and SDKs?
A: Absolutely. Osaurus speaks OpenAI, Anthropic, and Ollama APIs from the same local port (127.0.0.1:1337). Drop it into your existing workflow without rewrites.

Q: How much RAM do I actually need?
A: For smaller models like Gemma 4 2B (4-bit), ~1.5 GB. For serious local work, 64 GB is the practical minimum. For frontier-class models like DeepSeek V4, 128 GB is recommended. MoE (Mixture of Experts) models are more memory-efficient than their parameter count suggests.

Q: Is my data safe with Osaurus?
A: Osaurus runs in a hardware-isolated sandbox and stores everything locally. API keys live in the macOS Keychain. The app uses cryptographic identity (secp256k1) for verifiable trust chains. No backdoors, no data mining.

Q: Can I build my own plugins?
A: Yes. Osaurus supports a v3 plugin API with hot reload, and older v1/v2 plugins still load unchanged. You can create plugins in Swift or use simple JSON recipes no Xcode or code signing required for basic extensions.

Q: What makes Osaurus different from Ollama or LM Studio?
A: While Ollama and LM Studio focus primarily on model serving, Osaurus is a complete harness: persistent memory, autonomous agents, native Mac integration, MCP support, sandboxed code execution, and a consumer-friendly UI. It’s Apple-only and Swift-native, meaning no Electron bloat.

Q: Does Osaurus support voice?
A: Yes. Recent updates added on-device voice transcription via FluidAudio on Apple’s Neural Engine. You can dictate in chat, use wake-word activation, or press a global hotkey to transcribe into any app. No audio ever leaves your Mac.

Q: What’s next for Osaurus?
A: The team is currently participating in the Alliance accelerator in New York and exploring enterprise offerings for industries like legal and healthcare where local LLMs address critical privacy concerns.


Final Thoughts

Osaurus represents a growing movement in AI: the shift from rented intelligence to owned intelligence. It acknowledges that while models will continue to proliferate and improve, the real competitive advantage lies in the personal layer you build around them your workflows, your memory, your tools.

If you’re on Apple Silicon and you’ve been waiting for an AI experience that feels native, private, and actually useful, Osaurus might be the harness you’ve been looking for.

Ready to own your AI? Download Osaurus at osaurus.ai or install it via brew install --cask osaurus. The future of personal AI runs locally and it runs on your Mac.

Leave a Reply