AI sessions break. Corrupted tool chains return 400s. Content filters block legitimate output. Rate limits lock you out mid-task. Runaway agents burn your budget in minutes. And the vendors close the bugs as stale.
NeuroRouter sits between your tools and the API. It blocks secrets, strips wasted tokens, repairs broken requests, and keeps your work alive when everything else fails.
Runs locally. Your API key never leaves your machine except to the provider.
The community edition is published at obstalabs/neurorouter. Install it with Homebrew, Scoop, or download the release binaries directly.
GitHub repo Latest releasemacOS or Linux via the public Obsta Labs tap.
brew tap obstalabs/tap brew install obstalabs/tap/neurorouter
Windows via the public Obsta Labs bucket.
scoop bucket add obstalabs https://github.com/obstalabs/scoop-bucket scoop install obstalabs/neurorouter
Download tarballs and zip archives from the release page.
https://github.com/obstalabs/neurorouter/releases/latest
Everything happens before the request leaves your machine. No latency. No infra. No key exposure. No lock-in.
Block secrets
Secrets never reach the API. Credentials, tokens, and connection strings are intercepted inline. Detected leaks are flagged for rotation — not just blocked, but reported with specific remediation steps so you know exactly what to fix.
Strip waste
Six filters strip structural noise from every request: stale file reads, thinking blocks, orphaned tool results, failed retries, duplicate system reminders, and oversized content blocks. The API only sees clean context. Real sessions show ~40% noise reduction (see example below).
Reduce cost
Tracks how much noise was removed before billing. A session at OPS 75% means 25% of what would have been sent was waste — caught and stripped. Real sessions show ~40% noise reduction. You pay for clean context, not accumulated junk.
Drop-in replacement for your API endpoint.
# Claude Code ANTHROPIC_BASE_URL=http://localhost:9120 claude # Codex / OpenAI tools codex -c 'openai_base_url="http://127.0.0.1:9120"' # Any tool that supports base URL override OPENAI_API_BASE=http://localhost:9120 aider
One environment variable. Works with Claude Code, Codex, OpenClaw, Cursor, Aider, Continue.dev, and any tool that lets you set an API base URL. Your API key is forwarded as-is at the HTTP layer — never parsed, never stored, never logged.
# start the proxy neurorouter --port 9120 # see what would be filtered without sending neurorouter --port 9120 --dry-run
A session sends the same file content 3 times, includes a 120KB thinking block, and repeats system instructions in every request. NeuroRouter strips it before the API call. Less wasted context. Lower cost. Same workflow.
OPS shows how much unnecessary traffic was removed before billing. A session at OPS 75% means 25% of what would have been sent was noise — caught and removed. Track it in your status line or your logs.
ContextSpectre measures IPS — how clean your context stays over time. Without filtering, raw tool output (git diffs, test logs, build output) enters context and gets re-read on every turn. Typical unfiltered sessions score IPS 50-60.
With NeuroRouter Pro active, IPS stays at 90+. The proxy strips noise before it enters context, so the model only sees clean tool results. Less noise in, less re-read tax out.
OPS and IPS are complementary. OPS measures what NeuroRouter removes from each request (output purity). IPS measures what remains in context across turns (input purity). Together they show the full picture — waste prevented at the request layer and waste prevented from accumulating over the session.
IPS metric definition: contextspectre docs/concepts.md
Your API keys never leave your machine. NeuroRouter runs locally, uses environment variables or client passthrough auth, and never stores, transmits, or logs credentials. Keys exist only in process memory for the duration of the upstream request.
This is a structural difference from cloud LLM proxies. A 2026 study found 26 LLM proxy services collecting user credentials. The LiteLLM supply-chain breach (March 2026) compromised thousands of organizations. NeuroRouter eliminates this class of risk entirely — there is no server to breach and no database to leak.
lsof -i -P | grep neurorouter — only your upstreamgo build ./cmd/neurorouterFree
$0 — AGPL v3, self-hosted
Cleans AI requests before they hit the model. Removes wasted context, blocks secrets, and strips structural noise — locally, with zero setup. One protocol per instance, one session at a time.
Pro
$29 / month
Keep going no matter what. Session corrupted? Repaired before the 400. Credits burning? Mechanical work routed to cheaper models automatically. Cooldown incoming? Work rescued. Content filter kills your output? Retried and recovered. One daemon, all sessions, no lost work.
Team
$49 / seat / month
Enforce clean, safe AI usage across your team. Shared rules, consistent routing, and guardrails that prevent costly mistakes before they happen.
Enterprise
Custom pricing
Control AI usage at scale without losing speed. Org-wide policies, secure routing, and protection against data leaks, runaway cost, and workflow breakdowns.