OrionPod
🚀 Alpha — Early Access

OrionPod

Run AI Locally

No cloud dependencies.
No heavy infrastructure.
Just your machine.

Features

Everything you need to run AI locally

No internet needed
No cloud dependency
Yours, forever

Your Personal AI

AI that lives on your machine and works for you. Offline, uncensored, and always available. No subscriptions, no rate limits, no one watching. Just your own intelligence, running locally.

LM Studio
500+ MB
Ollama
~100 MB
OrionPod
~30 MB
App Size

Lightweight

17x smaller than LM Studio. No bloated Electron shell, no Python runtime, just native Rust + Metal.

Built from the ground up in Rust, the same language powering Firefox, Cloudflare, and Deno. Zero garbage collection overhead, minimal runtime, and direct Metal GPU bindings mean OrionPod stays fast and lean even under heavy inference loads.

Starts in under 2 seconds. Uses <50 MB RAM idle.

Try it out

Completely Private

100% local. Your data never leaves your machine. No Cloud, No Telemetry, No API keys.

GPU Accelerated

Auto-detects Metal on Apple Silicon and CUDA on NVIDIA for native GPU inference.

GGUF Compatible
Mistral-7B
4.1 GB · Q4_K_M
GGUF Risky
DeepSeek-R1-14B
8.5 GB · Q4_K_M
GGUF Compatible
Qwen-2.5-7B
4.7 GB · Q4_K_M
GGUF Compatible
Kimi-VL-A3B
2.0 GB · Q4_K_M
GGUF Not Compatible
Llama-3.3-70B
42 GB · Q4_K_M
GGUF Risky
Gemma-3-12B
7.3 GB · Q4_K_M

HuggingFace Built-in

Browse, search, and download GGUF models from HuggingFace directly inside the app. Run Mistral, DeepSeek, Qwen, Kimi, Llama, Gemma and more, all locally. Automatically filters out models that aren't compatible with your system, so you only see what actually runs on your hardware.

tokens/s
42.8
latency
23ms
memory
2.1G
gpu
Metal
ctx len
4096
temp
0.7

Real-time Observability

Monitor tokens/s, memory, latency, and GPU usage as your model runs. All live!

Whether you're choosing between a 7B and a 14B model, or fine-tuning inference parameters, live observability helps you get the best performance out of every token.

See it in Action

A clean, native desktop experience

Get Started

Up and running in seconds

System Requirements

  • macOS 10.15 (Catalina) or later
  • Apple Silicon (M-series) recommended for Metal GPU acceleration

Install

Download DMG Open the DMG → drag OrionPod to Applications → launch.

FAQ

Common questions & troubleshooting

macOS says OrionPod "can't be opened" — what do I do?

Because OrionPod isn't signed with an Apple Developer certificate (yet), macOS Gatekeeper blocks it on first launch. This is normal for indie/open-source apps. Here's how to install it:

  1. 1.
    Open the DMG: double-click OrionPod_0.1.5_universal.dmg
  2. 2.
    Drag OrionPod to Applications: standard DMG install.
  3. 3.
    First launch will be blocked: macOS shows: "OrionPod can't be opened because Apple cannot check it for malicious software."
  4. 4.
    Open it anyway: pick one of these:

    OPTION A Easiest

    • → Right-click (or Control-click) OrionPod.app → click Open
    • → Click Open in the dialog that appears
    • → Only needed once — future launches work normally

    OPTION B If Option A doesn't work

    • → Go to System Settings → Privacy & Security
    • → Scroll to the Security section
    • → You'll see: "OrionPod was blocked from use because it is not from an identified developer"
    • → Click Open Anyway
    • → Enter your password
    • → Click Open in the final confirmation
  5. 5.
    Done!: OrionPod launches and works normally from now on. 🎉
What models can I run?

OrionPod supports GGUF-format models. You can browse and download them directly from HuggingFace inside the app. Popular models like Mistral, DeepSeek, Qwen, Kimi, Llama, and Gemma all work. The app automatically filters out models that won't run on your hardware.

Is it really 100% private?

Yes. OrionPod runs entirely on your machine. There's no telemetry, no analytics, no cloud calls, no API keys. Your prompts and model outputs never leave your device. Read our privacy policy, it's short, because there's nothing to collect.

What hardware do I need?

macOS 10.15 (Catalina) or later. Apple Silicon (M1/M2/M3/M4) is recommended for Metal GPU acceleration, but Intel Macs work too. For 7B models, 8 GB of RAM is a comfortable minimum; larger models need more.

Is OrionPod free?

Yes, OrionPod is free and open source. It's an individual project built with love. No subscriptions, no paywalls, no premium tiers.