English | 中文 | 한국어 | 日本語 | Français | Deutsch | Español | Português
CheetahClaws (Nano Claude Code) : A Fast, Easy-to-Use, Python-Native Personal AI Assistant for Any Model, Inspired by OpenClaw and Claude Code, Built to Work for You Autonomously 24/7
Website · Brief Intro · Issue · The newest source of Claude Code
curl -fsSL https://raw.githubusercontent.com/SafeRL-Lab/cheetahclaws/main/scripts/install.sh | bashAfter installation:
source ~/.zshrc # macOS
# or: source ~/.bashrc # Linux
cheetahclaws # start chatting!Other install methods: pip install | uv install | run from source | full details
-
Apr 15, 2026 (v3.05.72): Trading agent, error classifier, parallel tools, prompt injection detection, SQLite sessions, tool cache, auxiliary model, safe stdio
- Trading agent module (
modular/trading/) — AI-powered multi-agent trading analysis and backtesting system. 5-phase analysis pipeline: data collection (technical indicators, fundamentals, news) → Bull/Bear researcher debate with BM25 memory → research judge recommendation → risk management panel (aggressive/conservative/neutral 3-way debate) → portfolio manager final decision (BUY/OVERWEIGHT/HOLD/UNDERWEIGHT/SELL). 4 built-in backtest strategies (dual MA, RSI mean reversion, Bollinger breakout, MACD crossover) with equity and crypto engines. 7 AI tools (GetMarketData,GetPrice,GetTechnicalIndicators,GetFundamentals,GetNews,RunBacktest,TradingMemory). 11 pure-Python technical indicators. Data source fallback chains (yfinance → coingecko → akshare). Post-trade reflection mechanism feeds lessons back into BM25 memory. SSJ integration as option 14 with guided sub-menu. Supports US/HK/A-share stocks and 20+ cryptos. Install:pip install "cheetahclaws[trading]". - Error classifier (
error_classifier.py) — centralized API error taxonomy (auth, billing, rate_limit, context_overflow, model_not_found, overloaded, connection, timeout) with per-category recovery hints, retryability, and backoff multipliers. Replaces fragile string matching inagent.pyandcheetahclaws.py. - Parallel tool execution (
agent.py) — when the LLM returns multiple tool calls,concurrent_safe=Truetools (Read, Glob, Grep, WebSearch, etc.) now run in parallel via ThreadPoolExecutor (up to 8 workers). Write tools remain sequential. Permission checks are still serial. - Prompt injection detection (
context.py) — CLAUDE.md files are scanned for 8 threat patterns (e.g., "ignore previous instructions", "system prompt override", credential exfiltration via curl/echo) before injection into the system prompt. Detected files are excluded with a security warning. - SQLite session store + full-text search (
session_store.py) — sessions are now saved to SQLite (WAL mode) alongside JSON files. FTS5 index enables/search <query>to find past conversations by content. Auto-imports legacyhistory.jsonon first search. - Tool result cache (
tool_registry.py) — read-only tools cache results bysha256(name + params), LRU eviction at 64 entries. Write tools (Write, Edit, Bash, NotebookEdit) invalidate the cache automatically. Eliminates redundant file reads in agent loops. - Auxiliary model routing (
auxiliary.py) — side tasks (context compression, summarization) now route to a fast/cheap model (Gemini Flash, GPT-4o-mini, etc.) instead of the primary model. Auto-detects from available API keys. Configurable viaauxiliary_modelin config. - Auto-discovery tool loading (
tools/__init__.py) — extension modules loaded via_EXTENSION_MODULESlist +__import__()loop instead of manual import statements. Adding a new extension is one line. - Safe stdio wrapper (
cheetahclaws.py) —sys.stdout/sys.stderrwrapped with_SafeWriterthat silently handlesBrokenPipeErrorand closed file descriptors. Prevents crashes when terminal disconnects during bridge/daemon operation. - One-line installer (
scripts/install.sh) —curl -fsSL .../install.sh | bashhandles platform detection (Linux/macOS/WSL2/Termux), Python/git/pip checks, clone, install, and PATH setup. First run triggers the setup wizard automatically. - Contributing section in README with quick-start commands for contributors, linking to CONTRIBUTING.md and Plugin Authoring Guide.
- Browser tool (
tools/browser.py) —WebBrowserenders JavaScript pages with headless Chromium (via playwright). Supports extract, screenshot, and click actions with CSS selectors. Solves dynamic/SPA pages thatWebFetchcan't handle. Optional:pip install cheetahclaws[browser]. - Email tools (
tools/email.py) —ReadEmail(IMAP) reads inbox with search by sender/subject;SendEmail(SMTP) sends emails with threading support. Zero external deps (Python stdlib). Configure with/config email_address=.... - File tools (
tools/files.py) —ReadPDFextracts text from PDFs (pymupdf);ReadImagedoes OCR on images (pytesseract, 99 languages);ReadSpreadsheetreads Excel/CSV/TSV with formatted table output. Optional:pip install cheetahclaws[files]. [all]extra —pip install cheetahclaws[all]installs every optional dependency (voice, vision, autosuggest, browser, files, OCR).- Version bumped to 3.05.72.
- Trading agent module (
-
Apr 15, 2026 (v3.05.71): Plugin docs, example template, config namespace fix, typing-time autosuggest
- Plugin authoring guide (
docs/guides/plugin-authoring.md) — full guide for building third-party plugins: tools (TOOL_DEFS), commands (COMMAND_DEFS), skills, MCP servers, manifest format, testing, publishing checklist, and common mistakes. - Example plugin template (
examples/example-plugin/) — copy-and-edit starter with working tools (ExampleSearch,ExampleStatus), command (/examplewith subcommands), skill, andplugin.jsonmanifest. - Fix
confignamespace collision — renamedconfig.pytocc_config.pyto avoid conflict with systemconfignamespace packages.pip install -e .followed bycheetahclawsfrom outside the project directory no longer crashes withImportError. - Typing-time autosuggest (PR #38 by @honghua) — optional
prompt_toolkitintegration for inline ghost suggestions and keyboard-selectable completion menu while typing slash commands. Install withpip install cheetahclaws[autosuggest]. Falls back to readline when not installed. Env varCHEETAH_PT_INPUT=0to opt out. - Python 3.10-3.13 compat fix (PR #38) —
Path.read_text(newline=)intools/fs.pyreplaced with portableopen()helper (thenewline=kwarg is 3.14+ only). - Version bumped to 3.05.71.
- Plugin authoring guide (
-
Apr 14, 2026 (v3.05.70): Setup wizard, Ollama UX, context indicator, and session robustness
- Interactive setup wizard (
commands/core.py,cheetahclaws.py) —cheetahclaws --setupor/setuplaunches a guided setup: pick from 6 providers (Ollama, Anthropic, OpenAI, Gemini, DeepSeek, custom), auto-detect env vars, set API key, verify connection. Auto-triggers on first run (noconfig.json). API key missing warning now suggests--setup. - Ollama UX improvements —
/modelnow shows live local Ollama models (via/api/tags) instead of a hardcoded list./model ollamatriggers the interactive model picker. Connection failures and 404 errors now give actionable messages ("Is Ollama running?", "Pull it with: ollama pull ..."). Tool-calling fallback message clarified. - Context usage in prompt — the REPL prompt now shows context window usage as a percentage: dim when <40%, yellow at 40-70%, red at >=70%. Users can see when compaction is approaching without running
/context. - Session save/resume robustness — atomic writes (write-to-temp + rename) prevent corruption on crash.
/loadand/resumenow catch corrupted JSON with friendly error messages and suggest daily backups. History file corruption no longer blocks auto-save. - Version from pyproject.toml —
VERSIONis now read dynamically frompyproject.toml(single source of truth), no more hardcoded version drift. Falls back toimportlib.metadatawhen installed as a package. /doctorenhanced — added internet connectivity check andpytedependency check; optional vs required deps now distinguished ([FAIL]for missing required deps).- Fix
mcpnamespace collision — renamed internalmcp/package tocc_mcp/to avoid conflict with the officialmcppip package (Anthropic MCP SDK). Previously,pip install .followed bycheetahclawscrashed withImportError: cannot import name 'MCPClient'. - Version bumped to 3.05.70.
- Interactive setup wizard (
-
Apr 14, 2026 (v3.05.69): Actionable error messages, dependency sync, and contributor guide
- Actionable API error messages (
cheetahclaws.py) — the REPL error handler now detects 6 common failure modes (invalid API key, network timeout, Ollama not running, rate limit, model not found, insufficient credits) and prints a specific hint alongside the error instead of a generic message. The proactive watcher background thread no longer dumps raw Python tracebacks to stdout — errors are routed throughlogging_utilsinstead. - Dependency sync (
pyproject.toml,requirements.txt) —pyte>=0.8.0added topyproject.tomlcore dependencies (was only inrequirements.txt, causing import failures afterpip install .).requirements.txtrewritten to mirrorpyproject.tomlas single source of truth, with optional deps (sounddevice,Pillow) clearly marked. CONTRIBUTING.md— new contributor guide covering project structure, architecture (config vs RuntimeContext, tool/plugin/hooks systems), development conventions, and a PR checklist. Addresses recurring PR issues where contributors misunderstood the plugin loader (TOOL_DEFSvsregister_tool()), hooks system (no event-based hooks), and runtime state management.- Version bumped to 3.05.69.
- Actionable API error messages (
For more news, see here
CheetahClaws: A Lightweight and Easy-to-Use Python Reimplementation of Claude Code Supporting Any Model, such as Claude, GPT, Gemini, Kimi, Qwen, Zhipu, DeepSeek, MiniMax, and local open-source models via Ollama or any OpenAI-compatible endpoint.
- Why CheetahClaws
- CheetahClaws vs OpenClaw
- Features
- Supported Models
- Installation
- Usage: Closed-Source API Models
- Usage: Open-Source Models (Local)
- Model Name Format
- Trading Agent (multi-agent analysis, backtesting, memory)
- Documentation (guides for all features)
- Contributing
- FAQ
- Citation
Claude Code is a powerful, production-grade AI coding assistant — but its source code is a compiled, 12 MB TypeScript/Node.js bundle (~1,300 files, ~283K lines). It is tightly coupled to the Anthropic API, hard to modify, and impossible to run against a local or alternative model.
CheetahClaws reimplements the same core loop in ~10K lines of readable Python, keeping everything you need and dropping what you don't. See here for more detailed analysis (CheetahClaws v3.03), English version and Chinese version
| Dimension | Claude Code (TypeScript) | CheetahClaws (Python) |
|---|---|---|
| Language | TypeScript + React/Ink | Python 3.8+ |
| Source files | ~1,332 TS/TSX files | ~85 Python files |
| Lines of code | ~283K | ~40K |
| Built-in tools | 44+ | 27 |
| Slash commands | 88 | 36 |
| Voice input | Proprietary Anthropic WebSocket (OAuth required) | Local Whisper / OpenAI API — works offline, no subscription |
| Model providers | Anthropic only | 8+ (Anthropic · OpenAI · Gemini · Kimi · Qwen · DeepSeek · MiniMax · Ollama · …) |
| Local models | No | Yes — Ollama, LM Studio, vLLM, any OpenAI-compatible endpoint |
| Build step required | Yes (Bun + esbuild) | No — run directly with python cheetahclaws.py (or install to use cheetahclaws) |
| Runtime extensibility | Closed (compile-time) | Open — register_tool() at runtime, Markdown skills, git plugins |
| Task dependency graph | No | Yes — blocks / blocked_by edges in task/ package |
- UI quality — React/Ink component tree with streaming rendering, fine-grained diff visualization, and dialog systems.
- Tool breadth — 44 tools including
RemoteTrigger,EnterWorktree, and more UI-integrated tools. - Enterprise features — MDM-managed config, team permission sync, OAuth, keychain storage, GrowthBook feature flags.
- AI-driven memory extraction —
extractMemoriesservice proactively extracts knowledge from conversations without explicit tool calls. - Production reliability — single distributable
cli.js, comprehensive test coverage, version-locked releases.
- Multi-provider — switch between Claude, GPT-4o, Gemini 2.5 Pro, DeepSeek, Qwen, MiniMax, or a local Llama model with
--modelor/model— no recompile needed. - Local model support — run entirely offline with Ollama, LM Studio, or any vLLM-hosted model.
- Readable source — the full agent loop is 174 lines (
agent.py). Any Python developer can read, fork, and extend it in minutes. - Zero build —
pip install -r requirements.txtand you're running. Changes take effect immediately. - Dynamic extensibility — register new tools at runtime with
register_tool(ToolDef(...)), install skill packs from git URLs, or wire in any MCP server. - Task dependency graph —
TaskCreate/TaskUpdatesupportblocks/blocked_byedges for structured multi-step planning (not available in Claude Code). - Two-layer context compression — rule-based snip + AI summarization, configurable via
preserve_last_n_turns. - Notebook editing —
NotebookEditdirectly manipulates.ipynbJSON (replace/insert/delete cells) with no kernel required. - Diagnostics without LSP server —
GetDiagnosticschains pyright → mypy → flake8 → py_compile for Python and tsc/shellcheck for other languages, with zero configuration. - Offline voice input —
/voicerecords viasounddevice/arecord/SoX, transcribes with localfaster-whisper(no API key, no subscription), and auto-submits. Keyterms from your git branch and project files boost coding-term accuracy. - Cloud session sync —
/cloudsavebacks up conversations to private GitHub Gists with zero extra dependencies; restore any past session on any machine with/cloudsave load <id>. - SSJ Developer Mode —
/ssjopens a persistent power menu with 10 workflow shortcuts: Brainstorm → TODO → Worker pipeline, expert debate, code review, README generation, commit helper, and more. Stays open between actions; supports/commandpassthrough. - Telegram Bot Bridge —
/telegram <token> <chat_id>turns cheetahclaws into a Telegram bot: receive user messages, run the model, and send back responses — all from your phone. Slash commands pass through, and a typing indicator keeps the chat feeling live. - WeChat Bridge —
/wechat loginauthenticates with WeChat via a QR code scan (the same iLink Bot API used by the official WeixinClawBot /openclaw-weixinplugin), then starts a long-poll bridge. Slash command passthrough, interactive menu routing, typing indicator, session auto-recovery, and per-peercontext_tokenmanagement all work out of the box. - Slack Bridge —
/slack <xoxb-token> <channel_id>connects cheetahclaws to a Slack channel using the Slack Web API (stdlib only — noslack_sdkrequired). Pollsconversations.historyevery 2 seconds; replies update an in-place "Thinking…" placeholder. Slash command passthrough, interactive menu routing, and auto-start on launch. - Worker command —
/workerauto-implements pending tasks frombrainstorm_outputs/todo_list.txt, marks each one done after completion, and supports task selection by number (e.g.1,4,6). - Force quit — 3× Ctrl+C within 2 seconds triggers immediate
os._exit(1), unblocking any frozen I/O. - Proactive background monitoring —
/proactive 5mactivates a sentinel daemon that wakes the agent automatically after a period of inactivity, enabling continuous monitoring loops, scheduled checks, or trading bots without user prompts. - Rich Live streaming rendering — When
richis installed, responses stream as live-updating Markdown in place (no duplicate raw text), with clean tool-call interleaving. - Native Ollama reasoning — Local reasoning models (deepseek-r1, qwen3, gemma4) stream their
<think>tokens directly to the terminal viaThinkingChunkevents; enable with/verboseand/thinking. - Native Ollama vision —
/image [prompt]captures the clipboard and sends it to local vision models (llava, gemma4, llama3.2-vision) via Ollama's native image API. No cloud required. - Reliable multi-line paste — Bracketed Paste Mode (
ESC[?2004h) collects any pasted text — code blocks, multi-paragraph prompts, long diffs — as a single turn with zero latency and no blank-line artifacts. - Rich Tab completion — Tab after
/shows all commands with one-line descriptions and subcommand hints; subcommand Tab-complete works for/mcp,/plugin,/tasks,/cloudsave, and more. - Checkpoint & rewind —
/checkpointlists all auto-snapshots of conversation + file state;/checkpoint <id>rewinds both files and history to any earlier point in the session. - Plan mode —
/plan <desc>(or theEnterPlanModetool) puts Claude into a structured read-only analysis phase; only the plan file is writable. Claude writes a detailed plan, then/plan donerestores full write permissions for implementation.
OpenClaw is another popular open-source AI assistant built on TypeScript/Node.js. The two projects have different primary goals — here is how they compare.
| Dimension | OpenClaw (TypeScript) | CheetahClaws (Python) |
|---|---|---|
| Language | TypeScript + Node.js | Python 3.8+ |
| Source files | ~10,349 TS/JS files | ~85 Python files |
| Lines of code | ~245K | ~12K |
| Primary focus | Personal life assistant across messaging channels | AI coding assistant / developer tool |
| Architecture | Always-on Gateway daemon + companion apps | Zero-install terminal REPL |
| Messaging channels | 20+ (WhatsApp · Telegram · Slack · Discord · Signal · iMessage · Matrix · WeChat · …) | Terminal + Telegram bridge + WeChat bridge (iLink) + Slack bridge (Web API) |
| Model providers | Multiple (cloud-first) | 7+ including full local support (Ollama · vLLM · LM Studio · …) |
| Local / offline models | Limited | Full — Ollama, vLLM, any OpenAI-compatible endpoint |
| Voice | Wake word · PTT · Talk Mode (macOS/iOS/Android) | Offline Whisper STT (local, no API key) |
| Code editing tools | Browser control, Canvas workspace | Read · Write · Edit · Bash · Glob · Grep · NotebookEdit · GetDiagnostics |
| Build step required | Yes (pnpm install + daemon setup) |
No — pip install and run |
| Mobile companion | macOS menu bar + iOS/Android apps | — |
| Live Canvas / UI | Yes (A2UI agent-driven visual workspace) | — |
| MCP support | — | Yes (stdio/SSE/HTTP) |
| Runtime extensibility | Skills platform (bundled/managed/workspace) | register_tool() at runtime, MCP, git plugins, Markdown skills |
| Hackability | Large codebase (245K lines), harder to modify | ~12K lines — full agent loop visible in one file |
- Omni-channel inbox — connects to 20+ messaging platforms (WhatsApp, Signal, iMessage, Discord, Teams, Matrix, WeChat…); users interact from wherever they already are.
- Always-on daemon — Gateway runs as a background service (launchd/systemd); no terminal required for day-to-day use.
- Mobile-first — macOS menu bar, iOS Voice Wake / Talk Mode, Android camera/screen recording — feels like a native app, not a CLI tool.
- Live Canvas — agent-driven visual workspace rendered in the browser; supports A2UI push/eval/snapshot.
- Browser automation — dedicated Chrome/Chromium profile with snapshot, actions, and upload tools.
- Production reliability — versioned npm releases, comprehensive CI, onboarding wizard,
openclaw doctordiagnostics.
- Coding toolset — Read/Write/Edit/Bash/Glob/Grep/NotebookEdit/GetDiagnostics are purpose-built for software development; CheetahClaws understands diffs, file trees, and code structure.
- True local model support — full Ollama/vLLM/LM Studio integration with streaming, tool-calling, and vision — no cloud required.
- 8+ model providers — switch between Claude, GPT-4o, Gemini, DeepSeek, Qwen, MiniMax, and local models with a single
--modelflag. - Hackable in minutes — 12K lines of readable Python; the entire agent loop is in
agent.py; extend withregister_tool()at runtime without rebuilding. - Zero setup —
pip install cheetahclawsand runcheetahclaws; no daemon, no pairing, no onboarding wizard. - MCP support — connect any MCP server (stdio/SSE/HTTP); tools auto-registered.
- SSJ Developer Mode —
/ssjpower menu chains Brainstorm → TODO → Worker → Debate in a persistent interactive session; automates entire dev workflows. - Offline voice —
/voicetranscribes locally withfaster-whisper; no subscription, no OAuth, works without internet. - Session cloud sync —
/cloudsavebacks up full conversations to private GitHub Gists with zero extra dependencies.
| If you want… | Use |
|---|---|
| A personal assistant you can message on WhatsApp/Signal/Discord | OpenClaw |
| An AI coding assistant in your terminal | CheetahClaws |
| Full offline / local model support | CheetahClaws |
| A mobile-friendly always-on experience | OpenClaw |
| To read and modify the source in an afternoon | CheetahClaws |
| Browser automation and a visual Canvas | OpenClaw |
| Multi-provider LLM switching without rebuilding | CheetahClaws |
Agent loop — CheetahClaws uses a Python generator that yields typed events (TextChunk, ToolStart, ToolEnd, TurnDone). The entire loop is visible in one file, making it easy to add hooks, custom renderers, or logging.
Tool registration — every tool is a ToolDef(name, schema, func, read_only, concurrent_safe) dataclass. Any module can call register_tool() at import time; MCP servers, plugins, and skills all use the same mechanism.
Context compression
| Claude Code | CheetahClaws | |
|---|---|---|
| Trigger | Exact token count | len / 3.5 estimate, fires at 70 % |
| Layer 1 | — | Snip: truncate old tool outputs (no API cost) |
| Layer 2 | AI summarization | AI summarization of older turns |
| Control | System-managed | preserve_last_n_turns parameter |
Memory — Claude Code's extractMemories service has the model proactively surface facts. CheetahClaws's memory/ package is tool-driven: the model calls MemorySave explicitly, which is more predictable and auditable. Each memory now carries confidence, source, last_used_at, and conflict_group metadata; search re-ranks by confidence × recency; and /memory consolidate offers a manual consolidation pass without silently modifying memories in the background.
- Developers who want to use a local or non-Anthropic model as their coding assistant.
- Researchers studying how agentic coding assistants work — the entire system fits in one screen.
- Teams who need a hackable baseline to add proprietary tools, custom permission policies, or specialised agent types.
- Anyone who wants Claude Code-style productivity without a Node.js build chain.
| Feature | Details |
|---|---|
| Multi-provider | Anthropic · OpenAI · Gemini · Kimi · Qwen · Zhipu · DeepSeek · MiniMax · Ollama · LM Studio · Custom endpoint |
| Interactive REPL | readline history, Tab-complete slash commands with descriptions + subcommand hints; Bracketed Paste Mode for reliable multi-line paste |
| Agent loop | Streaming API + automatic tool-use loop |
| 27 built-in tools | Read · Write · Edit · Bash · Glob · Grep · WebFetch · WebSearch · NotebookEdit · GetDiagnostics · MemorySave · MemoryDelete · MemorySearch · MemoryList · Agent · SendMessage · CheckAgentResult · ListAgentTasks · ListAgentTypes · Skill · SkillList · AskUserQuestion · TaskCreate/Update/Get/List · SleepTimer · EnterPlanMode · ExitPlanMode · (MCP + plugin tools auto-added at startup) |
| MCP integration | Connect any MCP server (stdio/SSE/HTTP), tools auto-registered and callable by Claude |
| Plugin system | Install/uninstall/enable/disable/update plugins from git URLs or local paths; multi-scope (user/project); recommendation engine |
| AskUserQuestion | Claude can pause and ask the user a clarifying question mid-task, with optional numbered choices |
| Task management | TaskCreate/Update/Get/List tools; sequential IDs; dependency edges; metadata; persisted to .cheetahclaws/tasks.json; /tasks REPL command |
| Diff view | Git-style red/green diff display for Edit and Write |
| Context compression | Auto-compact long conversations to stay within model limits |
| Persistent memory | Dual-scope memory (user + project) with 4 types, confidence/source metadata, conflict detection, recency-weighted search, last_used_at tracking, and /memory consolidate for auto-extraction |
| Multi-agent | Spawn typed sub-agents (coder/reviewer/researcher/…), git worktree isolation, background mode |
| Skills | Built-in /commit · /review + custom markdown skills with argument substitution and fork/inline execution |
| Plugin tools | Register custom tools via tool_registry.py |
| Permission system | auto / accept-all / manual / plan modes |
| Checkpoints | Auto-snapshot conversation + file state after each turn; /checkpoint to list, /checkpoint <id> to rewind; /rewind alias; 100-snapshot sliding window |
| Plan mode | /plan <desc> enters read-only analysis mode; Claude writes only to the plan file; EnterPlanMode / ExitPlanMode agent tools for autonomous planning |
| 36 slash commands | /model · /config · /save · /cost · /memory · /skills · /agents · /voice · /proactive · /checkpoint · /plan · /compact · /status · /doctor · … |
| Voice input | Record → transcribe → auto-submit. Backends: sounddevice / arecord / SoX + faster-whisper / openai-whisper / OpenAI API. Works fully offline. |
| Brainstorm | /brainstorm [topic] generates N expert personas suited to the topic (2–100, default 5, chosen interactively), runs an iterative debate, saves results to brainstorm_outputs/, and synthesizes a Master Plan + auto-generates brainstorm_outputs/todo_list.txt. |
| SSJ Developer Mode | /ssj opens a persistent interactive power menu with 15 shortcuts: Brainstorm, TODO viewer, Worker, Expert Debate, Propose, Review, Readme, Commit, Scan, Promote, Video factory, TTS factory, Monitor, Trading, Agent. Stays open between actions; /command passthrough supported. |
| Trading agent | /trading analyze <SYMBOL> runs a full multi-agent pipeline: data collection → Bull/Bear researcher debate → research judge → risk management panel (aggressive/conservative/neutral) → portfolio manager final decision (BUY/OVERWEIGHT/HOLD/UNDERWEIGHT/SELL). /trading backtest runs strategy backtests with 4 built-in strategies. BM25 memory system learns from past trades. Supports US/HK/A-share stocks and 20+ cryptos. |
| Monitor | /monitor (no args → wizard) subscribes to AI-monitored topics on a schedule and pushes reports to Telegram/Slack/console. Topics: ai_research (arxiv), stock_<TICKER>, crypto_<SYMBOL>, world_news (Reuters/BBC/AP), custom:<query>. Schedules: 15m to weekly. Background scheduler daemon with /monitor start/stop/status. |
| Autonomous Agents | /agent (no args → wizard) launches autonomous background agent loops driven by Markdown task templates. 4 built-in templates: research_assistant, auto_bug_fixer, paper_writer, auto_coder. Iteration summaries pushed via bridge. Custom templates: drop a .md file into ~/.cheetahclaws/agent_templates/. |
| Remote Control job queue | All three bridges (Telegram/Slack/WeChat) maintain a per-bridge FIFO job queue when the AI is busy. !jobs / !j — dashboard; !job <id> — detail; !retry <id> — re-run a failed job; !cancel [id] — stop current job. Tool step tracking with on_tool_start/on_tool_end hooks. Persistent log at ~/.cheetahclaws/jobs.json. |
| Worker | /worker [task#s] reads brainstorm_outputs/todo_list.txt, implements each pending task with a dedicated model prompt, and marks it done (- [x]). Supports task selection (/worker 1,4,6), custom path (--path), and worker count limit (--workers). Detects and redirects accidental brainstorm .md paths. |
| Telegram bridge | /telegram <token> <chat_id> starts a bot bridge: receive messages from Telegram, run the model, and reply — all from your phone. Typing indicator, slash command passthrough (including interactive menus), and auto-start on launch if configured. |
| WeChat bridge | /wechat login authenticates via QR code scan (same as WeixinClawBot / openclaw-weixin plugin), then starts the iLink long-poll bridge. context_token echoed per peer, typing indicator, slash command passthrough, session expiry auto-recovery. Credentials saved for auto-start on next launch. |
| Slack bridge | /slack <xoxb-token> <channel_id> connects to a Slack channel via the Web API (no external packages). Polls conversations.history every 2 s; replies update an in-place "Thinking…" placeholder. Slash command passthrough, interactive menu routing, auth validation on start, auto-start on next launch. |
| Video factory | /video [topic] runs the full AI video pipeline: story generation (active model) → TTS narration (Edge/Gemini/ElevenLabs) → AI images (Gemini Web free or placeholders) → subtitle burn (Whisper) → FFmpeg assembly → final .mp4. 10 viral content niches, landscape or short format, zero-cost path available. |
| TTS factory | /tts interactive wizard: AI writes script (or paste your own) → synthesize to MP3 in any voice style (narrator, newsreader, storyteller, ASMR, motivational, documentary, children, podcast, meditation, custom). Engine auto-selects: Gemini TTS → ElevenLabs → Edge TTS (always-free). CJK text auto-switches to a matching voice. |
| Vision input | /image (or /img) captures the clipboard image and sends it to any vision-capable model — Ollama (llava, gemma4, llama3.2-vision) via native format, or cloud models (GPT-4o, Gemini 2.0 Flash, …) via OpenAI image_url multipart format. Requires pip install cheetahclaws[vision]; Linux also needs xclip. |
| Tmux integration | 11 tmux tools for direct terminal control: create sessions/windows/panes, send commands, capture output. Auto-detected; zero impact if tmux is absent. Enables long-running tasks that outlive Bash tool timeouts. Cross-platform (tmux on Unix, psmux on Windows). |
| Shell escape | Type !command in the REPL to execute any shell command directly without AI involvement (!git status, !ls, !python --version). Output prints inline. |
| Proactive monitoring | /proactive [duration] starts a background sentinel daemon; agent wakes automatically after inactivity, enabling continuous monitoring loops without user prompts |
| Force quit | 3× Ctrl+C within 2 seconds triggers os._exit(1) — kills the process immediately regardless of blocking I/O |
| Rich Live streaming | When rich is installed, responses render as live-updating Markdown in place. Auto-disabled in SSH sessions to prevent repeated output; override with /config rich_live=false. |
| Context injection | Auto-loads CLAUDE.md, git status, cwd, persistent memory |
| Session persistence | Autosave on exit to daily/YYYY-MM-DD/ (per-day limit) + history.json (master, all sessions) + session_latest.json (/resume); sessions include session_id and saved_at metadata; /load grouped by date |
| Cloud sync | /cloudsave syncs sessions to private GitHub Gists; auto-sync on exit; load from cloud by Gist ID. No new dependencies (stdlib urllib). |
| Extended Thinking | Toggle on/off for Claude models; native <think> block streaming for local Ollama reasoning models (deepseek-r1, qwen3, gemma4) |
| Cost tracking | Token usage + estimated USD cost |
| Non-interactive mode | --print flag for scripting / CI |
| Provider | Model | Context | Strengths | API Key Env |
|---|---|---|---|---|
| Anthropic | claude-opus-4-6 |
200k | Most capable, best for complex reasoning | ANTHROPIC_API_KEY |
| Anthropic | claude-sonnet-4-6 |
200k | Balanced speed & quality | ANTHROPIC_API_KEY |
| Anthropic | claude-haiku-4-5-20251001 |
200k | Fast, cost-efficient | ANTHROPIC_API_KEY |
| OpenAI | gpt-4o |
128k | Strong multimodal & coding | OPENAI_API_KEY |
| OpenAI | gpt-4o-mini |
128k | Fast, cheap | OPENAI_API_KEY |
| OpenAI | gpt-4.1 |
128k | Latest GPT-4 generation | OPENAI_API_KEY |
| OpenAI | gpt-4.1-mini |
128k | Fast GPT-4.1 | OPENAI_API_KEY |
| OpenAI | gpt-5 |
128k | Next-gen flagship | OPENAI_API_KEY |
| OpenAI | gpt-5-nano |
128k | Fastest GPT-5 variant | OPENAI_API_KEY |
| OpenAI | gpt-5-mini |
128k | Balanced GPT-5 variant | OPENAI_API_KEY |
| OpenAI | o4-mini |
200k | Fast reasoning | OPENAI_API_KEY |
| OpenAI | o3 |
200k | Strong reasoning | OPENAI_API_KEY |
| OpenAI | o3-mini |
200k | Compact reasoning | OPENAI_API_KEY |
| OpenAI | o1 |
200k | Advanced reasoning | OPENAI_API_KEY |
gemini-2.5-pro-preview-03-25 |
1M | Long context, multimodal | GEMINI_API_KEY |
|
gemini-2.0-flash |
1M | Fast, large context | GEMINI_API_KEY |
|
gemini-1.5-pro |
2M | Largest context window | GEMINI_API_KEY |
|
| Moonshot (Kimi) | moonshot-v1-8k |
8k | Chinese & English | MOONSHOT_API_KEY |
| Moonshot (Kimi) | moonshot-v1-32k |
32k | Chinese & English | MOONSHOT_API_KEY |
| Moonshot (Kimi) | moonshot-v1-128k |
128k | Long context | MOONSHOT_API_KEY |
| Alibaba (Qwen) | qwen-max |
32k | Best Qwen quality | DASHSCOPE_API_KEY |
| Alibaba (Qwen) | qwen-plus |
128k | Balanced | DASHSCOPE_API_KEY |
| Alibaba (Qwen) | qwen-turbo |
1M | Fast, cheap | DASHSCOPE_API_KEY |
| Alibaba (Qwen) | qwq-32b |
32k | Strong reasoning | DASHSCOPE_API_KEY |
| Zhipu (GLM) | glm-4-plus |
128k | Best GLM quality | ZHIPU_API_KEY |
| Zhipu (GLM) | glm-4 |
128k | General purpose | ZHIPU_API_KEY |
| Zhipu (GLM) | glm-4-flash |
128k | Free tier available | ZHIPU_API_KEY |
| DeepSeek | deepseek-chat |
64k | Strong coding | DEEPSEEK_API_KEY |
| DeepSeek | deepseek-reasoner |
64k | Chain-of-thought reasoning | DEEPSEEK_API_KEY |
| MiniMax | MiniMax-Text-01 |
1M | Long context, strong reasoning | MINIMAX_API_KEY |
| MiniMax | MiniMax-VL-01 |
1M | Vision + language | MINIMAX_API_KEY |
| MiniMax | abab6.5s-chat |
256k | Fast, cost-efficient | MINIMAX_API_KEY |
| MiniMax | abab6.5-chat |
256k | Balanced quality | MINIMAX_API_KEY |
| Model | Size | Strengths | Pull Command |
|---|---|---|---|
llama3.3 |
70B | General purpose, strong reasoning | ollama pull llama3.3 |
llama3.2 |
3B / 11B | Lightweight | ollama pull llama3.2 |
qwen2.5-coder |
7B / 32B | Best for coding tasks | ollama pull qwen2.5-coder |
qwen2.5 |
7B / 72B | Chinese & English | ollama pull qwen2.5 |
deepseek-r1 |
7B–70B | Reasoning, math | ollama pull deepseek-r1 |
deepseek-coder-v2 |
16B | Coding | ollama pull deepseek-coder-v2 |
mistral |
7B | Fast, efficient | ollama pull mistral |
mixtral |
8x7B | Strong MoE model | ollama pull mixtral |
phi4 |
14B | Microsoft, strong reasoning | ollama pull phi4 |
gemma3 |
4B / 12B / 27B | Google open model | ollama pull gemma3 |
codellama |
7B / 34B | Code generation | ollama pull codellama |
llava |
7B / 13B | Vision — image understanding | ollama pull llava |
llama3.2-vision |
11B | Vision — multimodal reasoning | ollama pull llama3.2-vision |
Note: Tool calling requires a model that supports function calling. Recommended local models:
qwen2.5-coder,llama3.3,mistral,phi4.
OpenAI newer models (gpt-5 / o3 / o4 family): These models require
max_completion_tokensinstead of the legacymax_tokensparameter. CheetahClaws handles this automatically — no configuration needed.
Reasoning models:
deepseek-r1,qwen3, andgemma4stream native<think>blocks. Enable with/verboseand/thinkingto see thoughts in the terminal. Note: models fed a large system prompt (like cheetahclaws's 25 tool schemas) may suppress their thinking phase to avoid breaking the expected JSON format — this is model behavior, not a bug.
curl -fsSL https://raw.githubusercontent.com/SafeRL-Lab/cheetahclaws/main/scripts/install.sh | bashWorks on Linux, macOS, WSL2, and Android (Termux). The installer handles everything: checks Python 3.10+, clones the repo, installs via pip, and adds cheetahclaws to your PATH.
After installation:
source ~/.zshrc # macOS (zsh)
# or: source ~/.bashrc # Linux (bash)
cheetahclaws # start chatting!First run will guide you through setup (pick provider, set API key). Or run cheetahclaws --setup anytime.
Windows: Native Windows is not supported. Install WSL2 and run the command above inside WSL.
Android / Termux: The installer auto-detects Termux and skips incompatible optional dependencies. Manual install:
pkg install python git && pip install cheetahclaws.
git clone https://github.com/SafeRL-Lab/cheetahclaws.git
cd cheetahclaws
pip install .After that, cheetahclaws is available as a global command:
cheetahclaws # start REPL
cheetahclaws --model gpt-4o # choose a model
cheetahclaws -p "explain this" # non-interactive
cheetahclaws --setup # re-run setup wizardTo update after pulling new code:
cd cheetahclaws
git pull
pip install .pip install ".[voice]" # voice input (sounddevice)
pip install ".[vision]" # clipboard image capture (Pillow)
pip install ".[autosuggest]" # typing-time slash command autosuggest (prompt_toolkit)
pip install ".[browser]" # headless browser for JS-rendered pages (playwright)
pip install ".[files]" # PDF + Excel reading (pymupdf, openpyxl)
pip install ".[ocr]" # image OCR (pytesseract, Pillow)
pip install ".[trading]" # trading agent (yfinance, rank-bm25)
pip install ".[all]" # everything aboveNote: After installing
[browser], runplaywright install chromiumto download the browser binary.
uv installs cheetahclaws into an isolated environment and puts it on your PATH:
# Install uv (if not already installed)
curl -LsSf https://astral.sh/uv/install.sh | sh
# Clone and install
git clone https://github.com/SafeRL-Lab/cheetahclaws.git
cd cheetahclaws
uv tool install .To update: uv tool install . --reinstall
To uninstall: uv tool uninstall cheetahclaws
git clone https://github.com/SafeRL-Lab/cheetahclaws.git
cd cheetahclaws
pip install -r requirements.txt
python cheetahclaws.pyThis is useful for development — changes take effect immediately without reinstalling.
Get your API key at console.anthropic.com.
export ANTHROPIC_API_KEY=sk-ant-api03-...
# Default model (claude-opus-4-6)
cheetahclaws
# Choose a specific model
cheetahclaws --model claude-sonnet-4-6
cheetahclaws --model claude-haiku-4-5-20251001
# Enable Extended Thinking
cheetahclaws --model claude-opus-4-6 --thinking --verboseGet your API key at platform.openai.com.
export OPENAI_API_KEY=sk-...
cheetahclaws --model gpt-4o
cheetahclaws --model gpt-4o-mini
cheetahclaws --model gpt-4.1-mini
cheetahclaws --model o3-miniGet your API key at aistudio.google.com.
export GEMINI_API_KEY=AIza...
cheetahclaws --model gemini/gemini-2.0-flash
cheetahclaws --model gemini/gemini-1.5-pro
cheetahclaws --model gemini/gemini-2.5-pro-preview-03-25Get your API key at platform.moonshot.cn.
export MOONSHOT_API_KEY=sk-...
cheetahclaws --model kimi/moonshot-v1-32k
cheetahclaws --model kimi/moonshot-v1-128kGet your API key at dashscope.aliyun.com.
export DASHSCOPE_API_KEY=sk-...
cheetahclaws --model qwen/Qwen3.5-Plus
cheetahclaws --model qwen/Qwen3-MAX
cheetahclaws --model qwen/Qwen3.5-FlashGet your API key at open.bigmodel.cn.
export ZHIPU_API_KEY=...
cheetahclaws --model zhipu/glm-4-plus
cheetahclaws --model zhipu/glm-4-flash # free tierGet your API key at platform.deepseek.com.
export DEEPSEEK_API_KEY=sk-...
cheetahclaws --model deepseek/deepseek-chat
cheetahclaws --model deepseek/deepseek-reasonerGet your API key at platform.minimaxi.chat.
export MINIMAX_API_KEY=...
cheetahclaws --model minimax/MiniMax-Text-01
cheetahclaws --model minimax/MiniMax-VL-01
cheetahclaws --model minimax/abab6.5s-chatOllama runs models locally with zero configuration. No API key required.
Step 1: Install Ollama
# macOS / Linux
curl -fsSL https://ollama.com/install.sh | sh
# Or download from https://ollama.com/downloadStep 2: Pull a model
# Best for coding (recommended)
ollama pull qwen2.5-coder # 4.7 GB (7B)
ollama pull qwen2.5-coder:32b # 19 GB (32B)
# General purpose
ollama pull llama3.3 # 42 GB (70B)
ollama pull llama3.2 # 2.0 GB (3B)
# Reasoning
ollama pull deepseek-r1 # 4.7 GB (7B)
ollama pull deepseek-r1:32b # 19 GB (32B)
# Other
ollama pull phi4 # 9.1 GB (14B)
ollama pull mistral # 4.1 GB (7B)Step 3: Start Ollama server (runs automatically on macOS; on Linux run manually)
ollama serve # starts on http://localhost:11434Step 4: Run cheetahclaws
cheetahclaws --model ollama/qwen2.5-coder
cheetahclaws --model ollama/llama3.3
cheetahclaws --model ollama/deepseek-r1Or
python cheetahclaws.py --model ollama/qwen2.5-coder
python cheetahclaws.py --model ollama/llama3.3
python cheetahclaws.py --model ollama/deepseek-r1
python cheetahclaws.py --model ollama/qwen3.5:35bList your locally available models:
ollama listThen use any model from the list:
cheetahclaws --model ollama/<model-name>LM Studio provides a GUI to download and run models, with a built-in OpenAI-compatible server.
Step 1: Download LM Studio and install it.
Step 2: Search and download a model inside LM Studio (GGUF format).
Step 3: Go to Local Server tab → click Start Server (default port: 1234).
Step 4:
cheetahclaws --model lmstudio/<model-name>
# e.g.:
cheetahclaws --model lmstudio/phi-4-GGUF
cheetahclaws --model lmstudio/qwen2.5-coder-7bThe model name should match what LM Studio shows in the server status bar.
For self-hosted inference servers (vLLM, TGI, llama.cpp server, etc.) that expose an OpenAI-compatible API:
Quick Start for option C: Step 1: Start vllm:
CUDA_VISIBLE_DEVICES=7 python -m vllm.entrypoints.openai.api_server \
--model Qwen/Qwen2.5-Coder-7B-Instruct \
--host 0.0.0.0 \
--port 8000 \
--enable-auto-tool-choice \
--tool-call-parser hermes
Step 2: Start cheetahclaws:
export CUSTOM_BASE_URL=http://localhost:8000/v1
export CUSTOM_API_KEY=none
cheetahclaws --model custom/Qwen/Qwen2.5-Coder-7B-Instruct
# Example: vLLM serving Qwen2.5-Coder-32B
python -m vllm.entrypoints.openai.api_server \
--model Qwen/Qwen2.5-Coder-32B-Instruct \
--port 8000 \
--enable-auto-tool-choice \
--tool-call-parser hermes
# Then run cheetahclaws pointing to your server:
cheetahclawsInside the REPL:
/config custom_base_url=http://localhost:8000/v1
/config custom_api_key=token-abc123 # skip if no auth
/model custom/Qwen2.5-Coder-32B-Instruct
Or set via environment:
export CUSTOM_BASE_URL=http://localhost:8000/v1
export CUSTOM_API_KEY=token-abc123
cheetahclaws --model custom/Qwen2.5-Coder-32B-InstructFor a remote GPU server:
/config custom_base_url=http://192.168.1.100:8000/v1
/model custom/your-model-nameThree equivalent formats are supported:
# 1. Auto-detect by prefix (works for well-known models)
cheetahclaws --model gpt-4o
cheetahclaws --model gemini-2.0-flash
cheetahclaws --model deepseek-chat
# 2. Explicit provider prefix with slash
cheetahclaws --model ollama/qwen2.5-coder
cheetahclaws --model kimi/moonshot-v1-128k
# 3. Explicit provider prefix with colon (also works)
cheetahclaws --model kimi:moonshot-v1-32k
cheetahclaws --model qwen:qwen-maxAuto-detection rules:
| Model prefix | Detected provider |
|---|---|
claude- |
anthropic |
gpt-, o1, o3 |
openai |
gemini- |
gemini |
moonshot-, kimi- |
kimi |
qwen, qwq- |
qwen |
glm- |
zhipu |
deepseek- |
deepseek |
MiniMax-, minimax-, abab |
minimax |
llama, mistral, phi, gemma, mixtral, codellama |
ollama |
CheetahClaws includes a built-in AI-powered trading analysis and backtesting module. Install trading dependencies:
pip install "cheetahclaws[trading]"/trading analyze NVDARuns a 5-phase pipeline: data collection (technical indicators, fundamentals, news) → Bull/Bear researcher debate → research judge recommendation → risk management panel (aggressive / conservative / neutral) → portfolio manager final decision with a 5-tier rating: BUY / OVERWEIGHT / HOLD / UNDERWEIGHT / SELL.
Each agent uses BM25 memory to recall similar past situations and learns from outcomes via post-trade reflection.
/trading backtest AAPL dual_ma # single strategy
/trading backtest TSLA # AI picks best strategy4 built-in strategies: dual_ma (SMA crossover), rsi_mean_reversion, bollinger_breakout, macd_crossover. Engines for US/HK equities and crypto. Reports Sharpe, Sortino, Calmar, max drawdown, win rate, profit factor.
/ssj → 14. 📈 Trading opens a guided sub-menu:
| Option | Action |
|---|---|
| a. Quick Analyze | Full multi-agent analysis for any symbol |
| b. Backtest | Pick strategy or compare all 4 |
| c. Price Check | Current price + key metrics |
| d. Indicators | 11 technical indicators report |
| e. Trading Bot | Autonomous multi-symbol analysis |
| f. History | Past trading decisions |
| g. Memory | Trading memory status |
US stocks (AAPL), HK stocks (0700.HK), A-shares (000001.SZ), crypto (BTC, ETH, + 18 more). Data sources with automatic fallback chains — no API keys required.
Full guide: docs/guides/trading.md
Detailed guides have been moved to docs/guides/ to keep this README focused. Click any link below:
| Guide | What's Inside |
|---|---|
| Reference | CLI, 36+ commands, 33 built-in tools (incl. WebBrowse, ReadEmail, SendEmail, ReadPDF, ReadImage, ReadSpreadsheet), session search, auxiliary model, error classification, prompt injection detection, tool cache, parallel tools |
| Extensions | Memory system, Skills, Sub-Agents, MCP servers, Plugin system, Monitor subscriptions, Autonomous Agents |
| Bridges | Telegram, WeChat, Slack setup and remote control from your phone |
| Voice & Video | Voice input (offline Whisper), Video Content Factory, TTS Content Factory |
| Trading | Multi-agent analysis (Bull/Bear debate, Risk panel, PM), backtesting (4 strategies, equity + crypto engines), BM25 memory, data fallback chains, SSJ integration |
| Advanced | Brainstorm, SSJ Developer Mode, Tmux, Proactive monitoring, Checkpoints, Plan mode, Session management, Cloud sync |
| Recipes | 12 step-by-step examples: code review, Telegram remote control, autonomous research, bug fix, brainstorm, session search, browse web pages, email, PDF/Excel analysis, and more |
| Plugin Authoring | Build your own plugin: tools, commands, skills, MCP servers, publishing checklist |
| Example Plugin | Copy-and-edit starter template with working tools, commands, and skills |
| Contributing | Project structure, architecture guide, PR checklist |
cheetahclaws [OPTIONS] [PROMPT]
Options:
-p, --print Non-interactive: run prompt and exit
-m, --model MODEL Override model (e.g. gpt-4o, ollama/llama3.3)
--accept-all Auto-approve all operations (no permission prompts)
--verbose Show thinking blocks and per-turn token counts
--thinking Enable Extended Thinking (Claude only)
--version Print version and exit
-h, --help Show help
Examples:
# Interactive REPL with default model
cheetahclaws
# Switch model at startup
cheetahclaws --model gpt-4o
cheetahclaws -m ollama/deepseek-r1:32b
# Non-interactive / scripting
cheetahclaws --print "Write a Python fibonacci function"
cheetahclaws -p "Explain the Rust borrow checker in 3 sentences" -m gemini/gemini-2.0-flash
# CI / automation (no permission prompts)
cheetahclaws --accept-all --print "Initialize a Python project with pyproject.toml"
# Debug mode (see tokens + thinking)
cheetahclaws --thinking --verboseSee Reference Guide for the full list of 36+ slash commands, tool descriptions, and configuration options.
We welcome contributions! See the Contributing Guide for project architecture, code conventions, and PR checklist.
Quick start for contributors:
git clone https://github.com/SafeRL-Lab/cheetahclaws.git
cd cheetahclaws
pip install -r requirements.txt
pip install pytest
python -m pytest tests/ -x -q # 341+ tests should pass
python cheetahclaws.py # run the REPLBuilding a plugin? See the Plugin Authoring Guide and the example plugin template.
Q: How do I add an MCP server?
Option 1 — via REPL (stdio server):
/mcp add git uvx mcp-server-git
Option 2 — create .mcp.json in your project:
{
"mcpServers": {
"git": {"type": "stdio", "command": "uvx", "args": ["mcp-server-git"]}
}
}Then run /mcp reload or restart. Use /mcp to check connection status.
Q: An MCP server is showing an error. How do I debug it?
/mcp # shows error message per server
/mcp reload git # try reconnecting
If the server uses stdio, make sure the command is in your $PATH:
which uvx # should print a path
uvx mcp-server-git # run manually to see errorsQ: Can I use MCP servers that require authentication?
For HTTP/SSE servers with a Bearer token:
{
"mcpServers": {
"my-api": {
"type": "sse",
"url": "https://myserver.example.com/sse",
"headers": {"Authorization": "Bearer sk-my-token"}
}
}
}For stdio servers with env-based auth:
{
"mcpServers": {
"brave": {
"type": "stdio",
"command": "uvx",
"args": ["mcp-server-brave-search"],
"env": {"BRAVE_API_KEY": "your-key"}
}
}
}Q: Tool calls don't work with my local Ollama model.
Not all models support function calling. Use one of the recommended tool-calling models: qwen2.5-coder, llama3.3, mistral, or phi4.
ollama pull qwen2.5-coder
cheetahclaws --model ollama/qwen2.5-coderQ: How do I connect to a remote GPU server running vLLM?
/config custom_base_url=http://your-server-ip:8000/v1
/config custom_api_key=your-token
/model custom/your-model-name
Q: How do I check my API cost?
/cost
Input tokens: 3,421
Output tokens: 892
Est. cost: $0.0648 USD
Q: Can I use multiple API keys in the same session?
Yes. Set all the keys you need upfront (via env vars or /config). Then switch models freely — each call uses the key for the active provider.
Q: How do I make a model available across all projects?
Add keys to ~/.bashrc or ~/.zshrc. Set the default model in ~/.cheetahclaws/config.json:
{ "model": "claude-sonnet-4-6" }Q: Qwen / Zhipu returns garbled text.
Ensure your DASHSCOPE_API_KEY / ZHIPU_API_KEY is correct and the account has sufficient quota. Both providers use UTF-8 and handle Chinese well.
Q: Can I pipe input to cheetahclaws?
echo "Explain this file" | cheetahclaws --print --accept-all
cat error.log | cheetahclaws -p "What is causing this error?"Q: How do I run it as a CLI tool from anywhere?
Use uv tool install — it creates an isolated environment and puts cheetahclaws on your PATH:
cd cheetahclaws
uv tool install .After that, just run cheetahclaws from any directory. To update after pulling changes, run uv tool install . --reinstall.
Q: How do I set up voice input?
# Minimal setup (local, offline, no API key):
pip install sounddevice faster-whisper numpy
# Then in the REPL:
/voice status # verify backends are detected
/voice # speak your promptOn first use, faster-whisper downloads the base model (~150 MB) automatically.
Use a larger model for better accuracy: export NANO_CLAUDE_WHISPER_MODEL=small
Q: Voice input transcribes my words wrong (misses coding terms).
The keyterm booster already injects coding vocabulary from your git branch and project files.
For persistent domain terms, put them in a .cheetahclaws/voice_keyterms.txt file (one term per line) — this is checked automatically on each recording.
Q: Can I use voice input in Chinese / Japanese / other languages?
Yes. Set the language before recording:
/voice lang zh # Mandarin Chinese
/voice lang ja # Japanese
/voice lang auto # reset to auto-detect (default)
Whisper supports 99 languages. auto detection works well but explicit codes improve accuracy for short utterances.
If you find the repository useful, please cite the study
@article{cheetahclaws2026,
title={CheetahClaws: An Extensible, Python-Native Agent System for Autonomous Multi-Model Workflows},
author={CheetahClaws Team},
journal={github},
year={2026}
}






