AI Code Review

We Submitted CodeAnt AI to an Independent Benchmark: Here Is What Happened

Amartya | CodeAnt AI Code Review Platform
Sonali Sood

Founding GTM, CodeAnt AI

CODEANT AI · MARCH 2026 · BENCHMARK

AI code review tools have exploded in the past two years. Every major AI vendor now offers some version of automated pull request review. Tools promise to detect bugs, security issues, performance problems, and code quality regressions before code merges.

But until recently there was a fundamental problem:

There was no independent benchmark measuring how well these tools actually perform.

Most published comparisons were produced by the vendors themselves. Predictably, every benchmark declared its own tool the winner. That changed in March 2026 with the release of Martian’s Code Review Bench, the first independent evaluation framework designed specifically for AI code review systems.

The benchmark evaluates 17 AI code review tools across more than 200,000 real pull requests from open-source repositories. Instead of relying only on curated datasets, the benchmark measures something much closer to reality:

Which review comments developers actually act on.

In this first release, CodeAnt AI ranked #3 globally, achieving a 51.7% F1 score across real-world pull requests.

This article explains:

  • How the benchmark works

  • What precision, recall, and F1 mean for code review

  • Where CodeAnt AI ranked

  • Why different tools score differently

  • What the results reveal about the future of AI code review

The First Independent AI Code Review Benchmark

For years, evaluating AI code review tools was surprisingly difficult. Every AI code review vendor publishes benchmarks where they win. Greptile publishes benchmarks where Greptile wins. CodeRabbit publishes benchmarks where CodeRabbit wins. This is the state of the category.

In February 2026, a research lab called Martian, built by researchers from DeepMind, Anthropic, and Meta, and notably not in the business of selling code review tools, published what is arguably the first genuinely independent benchmark for AI code review agents. They open-sourced the dataset, the judge prompts, the evaluation pipeline, and the methodology. Anyone can reproduce the results.

We submitted CodeAnt AI. Here is what we found. Martian does not build a competing code review product.

This independence is what makes the benchmark credible.

The benchmark analyzes:

  • 200,000+ real pull requests

  • thousands of GitHub repositories

  • 17 AI code review tools

Instead of manually labeling bugs, the benchmark observes developer behavior directly. If developers fix code after a review comment, that comment is treated as a meaningful signal.

If developers ignore or dismiss the comment, that signal is recorded as well. This creates one of the largest behavioral datasets ever used to evaluate automated code review.

What Martian's Benchmark Actually Measures

Most vendor benchmarks measure what they want to measure. Martian's benchmark measures two things that actually matter to engineering teams: precision and recall.

Precision is the percentage of a tool's review comments that lead to an actual code change. If a tool leaves 100 comments and developers act on 52 of them, the precision is 52%. The rest are noise.

Recall is the percentage of real issues in a PR that the tool actually catches. A tool can achieve high precision by commenting very rarely, but it will miss most of the bugs. The benchmark measures both.

The offline benchmark runs every tool against 50 curated PRs from five major open-source repositories, Sentry, Grafana, Cal.com, Discourse, and Keycloak, each with human-verified lists of real issues a reviewer should catch.

The online benchmark continuously samples fresh real-world PRs from GitHub, tracking which AI review comments developers actually act on versus ignore. Because the PRs are recent, tools cannot have memorised them during training.

How Martian’s Benchmark Works

Martian designed the benchmark using two complementary evaluation systems. This architecture solves many problems that have historically made AI evaluation unreliable.

1. Online Benchmark: Real Developer Behavior

The online benchmark tracks how developers respond to automated review comments.

Martian monitored open-source pull requests across GitHub repositories during January–February 2026. For each review comment generated by a tool, the benchmark asks:

Did the developer modify the code after the comment? If yes, the comment counts as a true positive. If the comment was ignored, dismissed, or resulted in no change, it contributes to the evaluation differently depending on the scenario.

This method removes a common source of bias in AI benchmarks.

There is:

  • no curated answer key

  • no manual annotation

  • no static list of bugs

Only real developer decisions. Because the dataset includes over 200,000 pull requests, it provides an unusually strong signal about how useful review comments actually are in practice.

2. Offline Benchmark: Controlled Comparison

While the online benchmark measures real-world behavior, Martian also runs a traditional offline evaluation.

In this setup:

  • All tools review the same 50 pull requests

  • Results are compared against a curated gold dataset

This allows researchers to compare tools under identical conditions.

The offline benchmark answers questions like:

  • Which tool identifies known bugs?

  • Which tool misses issues?

  • Which tool generates excessive false positives?

While smaller than the online dataset, this controlled environment helps validate the tools’ technical detection capabilities.

Where CodeAnt AI Landed

The results below are sorted by F1 score, the benchmark’s primary metric.

Rank

Tool

F1

Precision

Recall

#1

Qodo Extended

64.3%

62.3%

66.4%

#2

Augment

53.8%

47.0%

62.8%

#3

CodeAnt AI

51.7%

52.2%

51.1%

#4

Qodo

47.9%

42.6%

54.7%

#5

Cursor Bugbot

44.9%

46.2%

43.8%

#6

Devin

42.9%

50.5%

37.2%

#7

Cubic Dev

41.8%

29.9%

69.3%

#8

Propel

41.6%

46.0%

38.0%

#9

Greptile

40.6%

31.6%

56.9%

#10

Claude Code Reviewer

39.0%

37.3%

40.9%

#11

GitHub Copilot

35.5%

26.6%

53.3%

CodeAnt AI ranked #3 globally, achieving:

  • 51.7% F1 score

  • 52.2% precision

  • 51.1% recall

Across thousands of repositories and hundreds of thousands of pull requests.

See the full benchmark results at codereview.withmartian.com

The practical meaning of a 52.2% precision score: when CodeAnt AI leaves a review comment on your team's PRs, more than half the time a developer will act on it. The rest of the category is closer to one in four.

Alert fatigue is real in code review. A tool that comments constantly but is wrong most of the time trains developers to ignore it. Over time, teams stop reading the bot's comments entirely. Precision is not a vanity metric, it is the difference between a code review tool that engineers trust and one they mute.

What Precision, Recall, and F1 Mean in Code Review

Understanding these metrics is essential for interpreting the benchmark.

Precision

Precision measures how often a tool’s comments identify real problems. A precision of 52.2% means: More than 1 in 2 comments generated by CodeAnt resulted in a real code change.

In practice this reflects signal-to-noise ratio. Low precision tools produce many false alarms. Developers eventually ignore them. High precision tools generate fewer, more trusted comments.

Why precision alone is not the whole story

A tool can game precision by commenting almost never. Leave one comment per PR, make it a genuinely critical bug, and your precision approaches 100%. But you will miss everything else.

This is why the benchmark measures both precision and recall together. CodeAnt AI's 51.1% recall means it is also catching a substantial proportion of the real issues in each PR — not just cherry-picking one obvious problem to protect its score.

The combination is what matters. Catching real issues (recall) without creating noise (precision) is the actual problem code review tools are trying to solve.

Recall

Recall measures how many of the real issues in a pull request the tool actually detects. CodeAnt AI’s recall of 51.1% means the system surfaces more than half of the issues present in the pull requests it analyzes.

Recall reflects coverage. A tool with low recall may stay quiet most of the time but miss critical bugs.

F1 Score

F1 combines precision and recall into a single balanced metric. It penalizes tools that optimize only one side of the tradeoff. A tool cannot achieve a high F1 score by:

  • detecting everything but generating noise

  • commenting rarely but missing bugs

The metric rewards systems that maintain both accuracy and coverage.

But benchmarks can be gamed

Martian's own blog post makes this explicit. They designed the benchmark with this in mind, which is why they run both an offline benchmark (fixed dataset, reproducible) and an online benchmark (fresh real-world PRs, no training data leakage) as a check on each other.

We are sharing these results because we think they are accurate. But we also think the strongest proof that a code review tool works is not a benchmark score at all.

It is finding real bugs in real production code before attackers do.

Why AI Code Review Tools Struggle With This Balance

Automated code review has always faced a structural challenge. Developers dislike tools that generate too many false positives. But they also dislike tools that miss real problems. This creates a difficult optimization problem.

High Recall Strategy

Some tools aim to detect as many issues as possible. They comment frequently. The result is high recall but low precision. Developers begin ignoring the comments.

High Precision Strategy

Other tools comment only when extremely confident. This produces high precision. But recall drops sharply. Important issues remain undetected.

Balanced Systems

The most useful tools operate in the middle. They detect a large portion of issues while maintaining a high signal-to-noise ratio. This balance is what the F1 score attempts to measure.

Why the Offline Benchmark May Underestimate Some Tools

Martian also highlights an important limitation in the current offline evaluation. The curated gold dataset used for the offline benchmark was initially built using datasets from two existing tools:

  • Augment

  • Greptile

While this allowed the benchmark to launch quickly, it introduces a structural bias. The dataset naturally reflects the categories of bugs those tools were originally designed to detect.

If another tool identifies a real issue not present in the gold set, the benchmark may classify that comment as a false positive.

In other words:

  • The tool may be correct.

  • But the dataset does not yet contain the issue.

Martian has already observed cases where comments initially marked as false positives were later confirmed to be legitimate bugs.

This is one reason the benchmark will continue evolving.

Why the Rankings May Change Over Time

The benchmark is designed to update regularly. Each monthly iteration introduces:

  • additional pull requests

  • expanded gold datasets

  • improved evaluation coverage

As the dataset becomes more representative of real-world code, offline and online results are expected to converge.

This means the leaderboard will likely shift over time.

Tools that detect broader categories of issues may benefit as the dataset expands.

Where CodeAnt AI Performs Particularly Well

Beyond the overall leaderboard, the benchmark also evaluates specific code review domains. In several categories, CodeAnt ranks at the top of the benchmark.

These include:

  • Security patch analysis

  • Testing issues in pull requests

  • Logging and PII leak detection

  • Large pull request review

Each of these areas is analyzed in detail in the following benchmark breakdowns:

  • Security Patch Detection Benchmark

  • Testing Issue Detection Benchmark

  • Logging and PII Detection Benchmark

  • Large Pull Request Review Benchmark

Each analysis explores how different AI systems perform on specific categories of engineering problems.

What the Benchmark Reveals About AI Code Review

The Martian benchmark provides an early look at how modern AI systems perform in real engineering environments.

Several patterns emerge from the results.

  • First, AI code review is still a rapidly evolving field. No tool currently achieves perfect precision or recall.

  • Second, tools differ significantly in how they balance signal and coverage.

  • Some systems prioritize detection breadth, while others prioritize comment accuracy.

  • Third, real developer behavior is a powerful evaluation signal.

Traditional benchmarks often struggle to capture how tools behave in real workflows. By observing developer responses to comments, Martian’s benchmark measures the real impact of automated review systems. This approach may become the standard for evaluating developer tools going forward.

Conclusion: A New Standard for Evaluating AI Code Review

The Martian Code Review Bench is the first large-scale independent attempt to measure AI code review tools using real developer behavior. Across more than 200,000 pull requests, the benchmark reveals both the progress and the limitations of modern AI-assisted code review.

In this first release:

CodeAnt AI ranked #3 globally with a 51.7% F1 score, balancing both precision and recall across thousands of repositories and engineering teams. More importantly, the benchmark introduces a transparent methodology that can evolve over time.

As datasets expand and evaluation methods improve, the leaderboard will continue to change. What will remain constant is the core objective:

High-signal code review that developers trust and act on. If you want to see what CodeAnt AI surfaces in your own repositories, you can install it in minutes and start reviewing pull requests today.

FAQs

What is the Martian Code Review Benchmark?

How does the benchmark measure AI code review performance?

Why does the benchmark include both online and offline evaluations?

What does CodeAnt AI’s 51.7% F1 score mean in practice?

How can teams start using AI for automated code review?

Table of Contents

Start Your 14-Day Free Trial

AI code reviews, security, and quality trusted by modern engineering teams. No credit card required!

Share blog: