home →
v2.0 — Next Release

Monthly Plans

Free developer key forever. Paid tiers bill monthly — no annual lock-in, cancel anytime. Every tier keeps your memory, files, and dashboard local.

Free · v1.0
LIVE
now
Developer
$0
free forever · no sign-up
Everything you need to make any local LLM remember you. Runs fully offline.
5
files/mo
10 MB
per file
50 MB
total
  • Memory & Intelligence
  • Persistent memory — 5,000 facts total
  • NLP fact extraction (identity, work, prefs, health)
  • Smart context compression (long chats stay coherent)
  • Basic knowledge graph — entities auto-extracted
  • Routing & Agents
  • Multi-backend routing (Ollama, llama.cpp, vLLM, any)
  • Agent mode — tool use, memory search, URL fetch
  • Local image generation (sd.cpp backend)
  • Security & Access
  • Secret Shield — API keys & tokens auto-redacted
  • OAuth sign-in — one cloud provider (read-only)
  • Up to 5 isolated users with private memory
  • Workspace & Files
  • File upload & RAG (5 files/mo, 33+ formats)
  • Live dashboard at localhost:9099
  • Memory export (plaintext)
  • Works with Open WebUI, SillyTavern, any OpenAI app
Download on GitHub →
Local · v2.0
cloud
free!
Standard
$5
per month · full local experience
Unlock Ghost Agent, memory lifecycle, persona learning, and the brain dashboard. Your local LLM finally feels like it knows you.
20
files/mo
25 MB
per file
100 MB
total
  • ↓ Everything in Developer, plus: ↓
  • Intelligence Layer
  • Ghost Agent — silent background context enrichment
  • Persona Engine — AI learns your tone & style
  • Intent classification (routes requests by goal)
  • Brain dashboard — live pipeline view
  • Memory System
  • Memory Lifecycle — active / fading / dormant / auto-expire
  • Memory consolidation (strengthens with use, fades when stale)
  • Semantic Recall — bundled embeddings catch paraphrases
  • Conversation branching & version history
  • Access & Security
  • Remote HTTPS access (Cloudflare Tunnel built-in)
  • Team & parental controls (caps, rate limits, trust modes)
  • Encrypted backup (AES-256)
  • Extensibility & Limits
  • TrimScript plugins (10 max)
  • 25K facts · 20 files/mo · 100 MB
Join the waitlist →
Most Popular
Hybrid · v2.0
Pro
$15
per month · where hybrid coding lives
Code with your local LLM and cloud models hand-in-hand. Draft local, review with Claude, debug with GPT — all with the same persistent memory.
100
files/mo
500 MB
per file
uncapped
total
  • ↓ Everything in Standard, plus: ↓
  • Hybrid Cloud + Local
  • Hybrid cloud+local routing (Claude & OpenAI)
  • Hybrid workflow planning — auto-splits tasks between models
  • Full OAuth — no raw API keys anywhere
  • Model aliases (fast/smart/cheap shortcuts)
  • Hybrid cost tracker & budgets
  • Autonomous Agents
  • Ghost Pre-Search — predicts follow-ups & pre-fetches
  • Sub-agent swarm (analyst + planner + researcher in parallel)
  • Prompt enhancement (local expands short asks before cloud)
  • Context Control
  • History search (every past chat, indexed)
  • Context selector (fine-tune what injects per message)
  • Extensibility & Limits
  • TrimScript plugins (50) · unlimited users
  • 100K facts · 100 files/mo · 500 MB
  • Priority support
Join the waitlist →
All-in · v2.0+
👑
every
thing
Premium
$30
per month · teams & production
The full production stack: visual knowledge-graph explorer, n8n workflow integration, Ghost Mesh, and every experimental feature as it lands.
files/mo
1 GB
per file
uncapped
total
  • ↓ Everything in Pro, plus: ↓
  • Production Tools
  • Visual Knowledge Graph explorer (zoomable, clickable)
  • MCP & n8n integration — 4 memory tools exposed to workflows
  • Claude Code subprocess offload (full coding sessions)
  • Multi-Agent & Insights
  • Ghost Mesh — multi-agent collaboration (analyst + predictor + planner)
  • Memory staleness transparency (see why facts made it in)
  • Ambient Task Tracker (deadline detection from chat)
  • Ghost Mailbox admin · unlimited plugins
  • Enterprise & Unlimited
  • Unlimited facts, files, storage
  • SLA support (48hr priority response)
  • On-prem licensing available
  • Early access to v2.1 features
Join the waitlist →
100% Local
Your Data Stays Put
Memory, files, dashboard — all on your box. Even the cloud providers can't see your context.
Hybrid
Local + Cloud, Hand in Hand
Draft locally. Escalate to Claude or GPT when you need the firepower. Same memory, both worlds.
Monthly
No Lock-in, Cancel Anytime
Downgrade → free tier still works forever. Upgrade → every v1.x feature ships to you as it lands.
🚀

Upgrading Takes 2 Minutes

Start on the free key. Upgrade when you need more.

1

Start Free

Grab UPtrim from GitHub today. Point it at Ollama or llama.cpp. The free developer key unlocks 5K memories, RAG, multi-user, and the dashboard — no sign-up.

2

Feel the Ceiling

You want Ghost Agent, hybrid cloud routing, sub-agent swarms, or the visual knowledge graph. Pick the tier that gives you what's next.

3

Paste & Go

Drop the new key in the dashboard. Features activate immediately, no restart. Cancel next month if it isn't for you. Free tier keeps working forever.

License FAQ

Pricing and billing, explained.

How does monthly billing work?
Pick a tier, pay monthly, use it. Your license activates immediately and renews on the same day next month. Cancel anytime from your dashboard — no prorated refunds, but you keep full tier access through the end of your billing cycle. If you don't renew, your proxy drops back to the free developer key features. Nothing breaks.
What's in each tier, plainly?
Developer (Free): persistent memory, basic knowledge graph, local agent mode, local image gen, 5-user isolation, 5K facts, read-only cloud OAuth. Standard ($5/mo): Ghost Agent, memory consolidation, embedding re-ranking, persona learning, conversation branching, brain dashboard, remote HTTPS access, parental controls, encrypted backups, TrimScript plugins, 25K facts. Pro ($15/mo): full hybrid cloud+local routing, Web Scout, sub-agent swarm, hybrid workflow planning, cost tracker, model aliases, context selector, 100K facts. Premium ($30/mo): visual knowledge graph explorer, n8n/MCP integration, Claude Code subprocess, Ghost Mesh (multi-agent collab), ambient task tracking, memory staleness transparency, unlimited everything, SLA support.
What's the hybrid coding workflow about?
This is why Pro exists. Your local LLM is fast and free but weaker on hard problems; Claude and GPT are expensive but precise. Pro's hybrid router keeps them working hand-in-hand: you draft and iterate with your local model, and when you hit something tricky — a gnarly refactor, an unfamiliar API, a performance question — the request quietly escalates to the cloud. Both sides see the same memory (your project history, preferences, past decisions), so the cloud response picks up mid-thought. Commit messages locally, code review in Claude, both feeding the same knowledge graph.
Why should I pay if I can just plug in my own Claude key?
You can, and Developer already gives you read-only OAuth for one provider. What Pro adds is the orchestration: smart routing decides which model for which request, cost tracking per-call, your persistent memory automatically injected into both sides, prompt enhancement that makes short asks worth 2-3x more, and conversation history you can grep months later. Raw API access is a chatbox; Pro makes it a personal AI that knows you.
Is v2.0 available right now?
v1.0 is live on GitHub today with the free developer key. v2.0 is in active development — a much bigger release with hybrid cloud routing, the full intelligence stack, visual knowledge graph, and all the features described in Pro and Premium. Early adopters who subscribe to any paid tier at launch get every v2.0 feature rolled out as it ships, no price change. Watch the GitHub repo for the v2.0 release notes.
Can teams get custom licensing?
Yes. Use the contact page for volume discounts, custom deployment, on-prem licensing, or onboarding help. Teams of 10+ get discounted seats.
What if I only want the free version forever?
That's fine — and it's fully functional. The free developer key on GitHub gives you persistent memory, context trimming, up to 5 users, file uploads (15 MB), and the dashboard. No sign-up, no credit card, no expiration. Paid tiers just add the extras when you want them.
Does my data go to Anthropic/OpenAI if I use hybrid mode?
Only the specific request you route there — and only if you've opted in via OAuth. Your memory, files, history, and identity all stay on your machine. Hybrid mode sends the minimum prompt to the cloud provider, gets the response back, and stores nothing with them. You can inspect every request in the brain dashboard before it goes out.

Start Free. Upgrade when you're ready.

v1.0 is live on GitHub today. v2.0 drops soon — early paid subscribers get every new feature as it ships.

bored? try typing “pong” or “tetris”

Got your license? Download & Install