claude-review vs Anthropic Code Review¶
Anthropic offers a built-in Code Review feature as part of their Claude platform. This page explains what it is, how it differs from claude-review, and why you might choose one over the other.
The fundamental difference¶
Anthropic Code Review gives you access to Claude. claude-review gives you access to Claude plus a model that learns your codebase.
Every time you run claude-review with --memory, it stores findings locally. A daemon consolidates them into cross-PR patterns every 30 minutes. The next review is informed by everything the tool has seen before. After a month of use, your instance knows which files in your repo are chronic hotspots, which categories of bugs keep appearing, and which findings your team consistently rejects as noise.
Anthropic's managed service resets to zero on every PR. That's not a criticism — it's a design tradeoff. For claude-review, persistent local memory is the core design.
What is Anthropic Code Review?¶
Anthropic Code Review is a managed service integrated into GitHub. When enabled, it automatically adds a reviewer to your pull requests, posts inline comments, and provides a summary of issues. It is configured through Anthropic's dashboard and requires no self-hosted infrastructure.
As of early 2026, it is available only to GitHub users on Team or Enterprise Anthropic plans.
Feature comparison¶
| Feature | Anthropic Code Review | claude-review |
|---|---|---|
| Cost per review | $15–25 (managed pricing) | $0.02–3.00 (your API tokens) |
| Platforms | GitHub only | GitHub, GitLab, Bitbucket, local git |
| Plan required | Team or Enterprise | Any Anthropic API key |
| Self-hosted | No — runs on Anthropic's infrastructure | Yes — runs on your machine |
| Open source | No | MIT licensed |
| Source code leaves your system | Yes | No — directly to api.anthropic.com only |
| Memory across PRs | No — resets to zero every PR | Yes — SQLite, per-repo, always local |
| Cross-PR pattern detection | No | Yes — consolidation daemon every 30 min |
| False positive suppression | No | Yes — rejected findings remembered |
| Pre-commit hook | No | Yes (install-hook command) |
| Local diff review | No | Yes (diff command) |
| Cost estimate before running | No | Yes (--estimate flag) |
| Custom focus areas | No | Yes (--focus flag) |
| JSON/annotations output | Limited | Full structured output for CI |
| Configurable agent count | No | Yes (1–10 parallel agents) |
| Works in air-gapped environments | No | Yes (binary + SQLite, no external deps except Anthropic API) |
| CI pipeline integration | GitHub only | Any CI system |
| Model selection | Anthropic-controlled | Your choice (Haiku/Sonnet/Opus) |
When to use Anthropic Code Review¶
Anthropic Code Review is a good fit if:
- You use GitHub exclusively and want zero setup
- Your team is already on a Team/Enterprise Anthropic plan
- You want inline PR comments posted directly to GitHub without any CLI tooling
- You don't need cross-PR memory or custom configuration
When to use claude-review¶
claude-review is a better fit if:
- You use GitLab, Bitbucket, or local git repos
- You're on a Starter/individual Anthropic API plan
- Cost is a concern — a typical PR costs 10–50× less
- You want reviews before pushing (local diffs, pre-commit hook)
- You need CI-integrated structured output (JSON, GitHub Checks annotations)
- You want persistent memory and cross-PR pattern detection
- You work in a regulated environment where code cannot transit through external managed services
- You want to customize which concern areas are reviewed, the model, or the review pipeline
Architecture differences¶
Anthropic Code Review¶
The diff is sent to Anthropic's infrastructure, processed by their pipeline, and returned as GitHub PR comments. You have no visibility into or control over the pipeline.claude-review¶
Local git diff / GitHub PR / GitLab MR / Bitbucket PR
│
▼
claude-review (running on your machine)
│
├── Query memory DB (if --memory)
│
├── Parallel finder agents (your API key → api.anthropic.com)
├── Verifier agent
├── Ranker agent
│
├── Write REVIEW.md / review.json / annotations.json
│
└── Ingest findings into local memory.db (if --memory)
The diff is sent from your machine directly to api.anthropic.com — the same endpoint you use for any Claude API call. No additional data handling by an intermediary service.
Cost breakdown example¶
For a medium-sized PR (500 changed lines):
Anthropic Code Review: - Fixed fee: $15–25 per review - Monthly for a 5-engineer team (10 PRs/week): ~$600–1,000/month
claude-review with Haiku (default): - Typical cost: $0.05–0.30 per review - Monthly for a 5-engineer team (10 PRs/week): ~$2–12/month
claude-review with Sonnet (for critical PRs): - Typical cost: $1–3 per review - Monthly for occasional critical reviews: ~$20–60/month
Combining both tools¶
Some teams use both:
claude-reviewruns locally and in CI for every PR (fast, cheap, catches the bulk of issues)- Anthropic Code Review runs for production releases or security-critical branches (managed, posts GitHub comments)
This gives the best of both: continuous coverage at low cost, with managed deep reviews when stakes are highest.
Summary¶
claude-review was built specifically for teams that:
- Can't or don't want to use a managed service
- Need multi-platform support
- Want lower costs, more control, and persistent cross-PR intelligence
If you're already on GitHub and happy with the managed experience, Anthropic Code Review is convenient. If you need more flexibility, claude-review is the open-source alternative.