Skip to content

Getting Started

Prerequisites

Option Requirement
Build from source Rust 1.78 or later (rustup update stable)
Pre-built binary None — download and run
Docker Docker Engine 24 or later

You also need at least one LLM API key. Both agents default to claude-sonnet-4-6 (Anthropic), but each agent can be configured independently with any provider that supports the Anthropic or OpenAI chat completions API — including OpenAI, OpenRouter, Ollama, vLLM, and LiteLLM.


Install

Option 1 — cargo install

cargo install phalus

Verify the installation:

phalus --version

Option 2 — Pre-built binary

Download the appropriate archive from the releases page:

# Linux x86_64 example
curl -L https://github.com/phalus-sh/phalus/releases/latest/download/phalus-linux-x86_64.tar.gz \
  | tar xz
sudo mv phalus /usr/local/bin/
phalus --version

Option 3 — Docker

docker pull ghcr.io/phalus-sh/phalus:latest

For convenience, add a shell alias:

alias phalus='docker run --rm \
  -e PHALUS_LLM__AGENT_A_API_KEY \
  -e PHALUS_LLM__AGENT_B_API_KEY \
  -v "$PWD":/work -w /work \
  ghcr.io/phalus-sh/phalus:latest'

Configure API Keys

PHALUS requires separate API keys for Agent A (Analyzer) and Agent B (Builder). Using separate keys provides stronger isolation evidence, though the same key works.

Environment variables (recommended for getting started):

export PHALUS_LLM__AGENT_A_API_KEY="sk-ant-..."
export PHALUS_LLM__AGENT_B_API_KEY="sk-ant-..."

Config file (recommended for regular use):

mkdir -p ~/.phalus
cat > ~/.phalus/config.toml <<'EOF'
[llm]
agent_a_provider = "anthropic"
agent_a_model    = "claude-sonnet-4-6"
agent_a_api_key  = "sk-ant-..."

agent_b_provider = "anthropic"
agent_b_model    = "claude-sonnet-4-6"
agent_b_api_key  = "sk-ant-..."
EOF

Verify that the configuration is loaded correctly (API keys are always redacted in this output):

phalus config

Using alternative providers — set the provider and base URL for OpenAI-compatible endpoints:

# OpenAI
export PHALUS_LLM__AGENT_A_PROVIDER=openai
export PHALUS_LLM__AGENT_A_BASE_URL=https://api.openai.com
export PHALUS_LLM__AGENT_A_MODEL=gpt-4o
export PHALUS_LLM__AGENT_A_API_KEY="sk-..."

# Ollama (local)
export PHALUS_LLM__AGENT_A_PROVIDER=ollama
export PHALUS_LLM__AGENT_A_BASE_URL=http://localhost:11434
export PHALUS_LLM__AGENT_A_MODEL=llama3
export PHALUS_LLM__AGENT_A_API_KEY=unused

See Configuration — OpenAI-Compatible Providers for full examples including OpenRouter, vLLM, and mixed-provider setups.


Quick Start

Run a single package

The run-one command is the fastest way to try PHALUS. It does not require a manifest file.

phalus run-one npm/[email protected] --license mit

The format is ecosystem/name@version. Supported ecosystems: npm, pypi, crates, go.

You should see output similar to:

Inspect the output

phalus inspect ./phalus-output --csp --similarity --audit

CSP section lists the ten specification documents Agent A produced:

=== CSP Specs ===
  [email protected] (10 documents)
    - 01-overview.json
    - 02-api-surface.json
    - 03-behavior-spec.json
    - 04-edge-cases.json
    - 05-configuration.json
    - 06-type-definitions.json
    - 07-error-catalog.json
    - 08-compatibility-notes.json
    - 09-test-scenarios.json
    - 10-metadata.json

Similarity section shows how close the generated code is to the original (lower is better for the clean room claim):

=== Similarity Reports ===
  [email protected]:
    token_similarity: 0.1800
    name_overlap:     0.9000
    string_overlap:   0.1200
    overall_score:    0.2500
    verdict:          PASS

Note: name overlap is intentionally high — the public API names must match by design.

Audit section shows the event log for the run:

=== Audit Log ===
  [2026-03-26T10:00:00Z] seq=0 type=manifest_parsed
  [2026-03-26T10:00:01Z] seq=1 type=docs_fetched
  [2026-03-26T10:00:03Z] seq=2 type=spec_generated
  [2026-03-26T10:00:03Z] seq=3 type=firewall_crossing
  [2026-03-26T10:00:08Z] seq=4 type=implementation_generated
  [2026-03-26T10:00:09Z] seq=5 type=validation_completed

Run from a manifest

# Preview what would be processed
phalus plan package.json

# Run the full pipeline
phalus run package.json --license apache-2.0 --output ./output/

Output Structure

After a successful run, the output directory contains:

phalus-output/
├── left-pad/
│   ├── package.json
│   ├── LICENSE
│   ├── README.md
│   ├── src/
│   │   └── index.js
│   ├── test/
│   │   └── index.test.js
│   ├── validation.json          # similarity + verdict
│   └── .cleanroom/
│       └── csp/
│           ├── 01-overview.json
│           ├── 02-api-surface.json
│           └── ...
└── audit.jsonl                  # job-level audit trail

Split pipeline: Agent A and Agent B separately

You can run Agent A (spec generation) and Agent B (code implementation) as separate steps. This allows you to review, edit, or programmatically modify the specification before building.

# Step 1: Generate CSP only (Agent A)
phalus run-one npm/[email protected] --dry-run

# Step 2: Review the specification
phalus inspect ./phalus-output --csp
cat ./phalus-output/lodash/.cleanroom/csp/03-behavior-spec.json

# Step 3: Build from the CSP (Agent B)
phalus build ./phalus-output/lodash/.cleanroom/csp/

See the Cookbook for advanced workflows including injecting custom security constraints and batch processing with review gates.


Scan Dependencies for Licenses

Before reimplementing, you can audit your dependency licenses:

phalus scan ./my-project

This walks the project for manifests and SBOMs, resolves license metadata from registries, and classifies each dependency. Use --save to persist results and query them through the API later. See the CLI reference for full options.


Next Steps