New code ships fast.
Now it ships optimal too.

Every pull request your team opens, Codeflash benchmarks the changed code, catches regressions before they merge, and posts a faster rewrite with the numbers to prove it. Your engineers see it in the same PR they're already reviewing. No new workflow.

"After installing Codeflash in the GitHub Pull Request code review stage, it tries to optimize every new code we write. With that, I can be more confident that our engineers are shipping more optimized code every time."
Brad Dwyer · Founder & CTO, Roboflow

Performance wins decay. This stops that.

A one-time optimization engagement cuts your bill. But new code ships every week, and most of it has never been profiled. Within a year, the gains are gone. Continuous Optimization closes that loop. The same agent that found your bottlenecks now watches every PR going forward.

Regressions caught at the source
Spotted in the PR where they're introduced, not three sprints later in a postmortem.
No extra work for your team
The benchmark and rewrite show up as a comment in the PR your engineers are already reviewing.
AI code gets reviewed too
We've found 118 functions up to 446× slower in two AI-written PRs. The agent catches what code review can't.
Savings compound, not decay
Every optimized PR keeps the baseline lower. The bill bends down and stays there.

What happens on every PR.

01

Detect

The agent identifies which functions changed in the diff and selects representative inputs based on prior execution traces.

02

Benchmark

Runs the old and new version on isolated hardware. The result is only reported if it's statistically significant.

03

Rewrite

If a faster equivalent exists, the agent writes it and verifies correctness against your test suite before surfacing it.

04

Comment

Posts directly on the PR with before/after numbers and a one-click patch. Your engineers decide whether to apply it.

A PR comment that pays for itself.

codeflash-agent · PR #1204 · app/services/billing.py
// benchmarked 2 changed functions

compute_usage_tier()
  before: 412 µs   after: 38 µs   10.8× faster
  sorted dict lookup → precomputed bucket
  patch ready · apply suggestion →

_format_invoice_line()
  before: 88 µs   after: 91 µs   no regression (within noise)

// 354 regression tests · ok

Works where your team already works.

GitHub

On every PR, automatically

Check runs and inline comments on every pull request. You can configure it to block merges on regressions, or keep it advisory.

Claude Code

Inline while writing

The plugin surfaces optimization suggestions as your engineers write code, before a PR is even opened.

Cursor

Inside the editor

Regressions and rewrites surface without leaving the editor. The feedback loop tightens to the moment of authorship.

Codex

Same loop, native to Codex

If your team uses Codex as their primary agent, the Codeflash plugin runs the same benchmarking and rewrite loop inside it.

Common questions.

Will this slow down our CI pipeline?

No. Benchmarks run on our hardware, out-of-band. Your CI pipeline just reads the result from us. There's no compute overhead on your side.

Does it block merges?

Only if you configure it to. The default is advisory: it posts the benchmark and patch as a comment, and your engineers decide whether to apply it. You can enable merge-blocking on regressions for critical paths.

What languages are supported?

Python, Java, JavaScript, TypeScript, Go, and more.

Will you use our code to train models?

On Pro, never. Not ours, not any third party's. On the free plan your code is public by definition, so we make no training restriction on public code.

How is this different from just asking Claude or Cursor to optimize the code?

AI coding tools suggest changes based on what they can see in the file. Codeflash runs an actual benchmark before and after, verifies correctness against your test suite, and only surfaces a rewrite if the numbers prove it's faster. The difference is measurement versus intuition.

Ready to keep your bill down for good?

Continuous Optimization pairs with an Optimization Engagement. The engagement cuts the bill; this keeps it there.