Skip to main content
Software Engineering Daily

CodeRabbit and RAG for Code Review with Harjot Gill

48 min episode · 2 min read
·

Episode

48 min

Read time

2 min

AI-Generated Summary

Key Takeaways

  • Multi-model architecture: CodeRabbit uses seven to eight different LLMs simultaneously, matching workload to model capabilities—GPT-4o-mini for summarization, o3-mini for deep reasoning—rather than letting users choose models, achieving better price-to-performance ratios than single-model approaches.
  • Sandboxed code navigation: Instead of tool calls or MCPs, CodeRabbit clones repositories into cloud sandboxes where agents execute CLI commands, run AST queries, and perform web searches to validate bugs, pioneering this approach two years before similar tools emerged.
  • Dynamic task decomposition: A root agent breaks code reviews into subtasks delegated to specialized sub-agents, with judge LLMs filtering low-quality inferences based on context quality, preventing hallucinations from reaching users through multi-layer validation before surfacing insights.
  • Context preparation strategy: Reasoning models like Sonnet 3.7 require cleaned, re-ranked context rather than raw RAG stuffing—models overthink and derail with unfiltered data, so CodeRabbit spends significant compute on context cleanup before expensive reasoning model calls.

What It Covers

CodeRabbit CEO Harjot Gill explains how his AI code review platform uses multi-model LLM architecture, sandboxed CLI environments, and dynamic task graphs to review 100,000 developers' code daily with reasoning models like o3-mini.

Key Questions Answered

  • Multi-model architecture: CodeRabbit uses seven to eight different LLMs simultaneously, matching workload to model capabilities—GPT-4o-mini for summarization, o3-mini for deep reasoning—rather than letting users choose models, achieving better price-to-performance ratios than single-model approaches.
  • Sandboxed code navigation: Instead of tool calls or MCPs, CodeRabbit clones repositories into cloud sandboxes where agents execute CLI commands, run AST queries, and perform web searches to validate bugs, pioneering this approach two years before similar tools emerged.
  • Dynamic task decomposition: A root agent breaks code reviews into subtasks delegated to specialized sub-agents, with judge LLMs filtering low-quality inferences based on context quality, preventing hallucinations from reaching users through multi-layer validation before surfacing insights.
  • Context preparation strategy: Reasoning models like Sonnet 3.7 require cleaned, re-ranked context rather than raw RAG stuffing—models overthink and derail with unfiltered data, so CodeRabbit spends significant compute on context cleanup before expensive reasoning model calls.

Notable Moment

Gill reveals CodeRabbit deliberately avoids building features where model capabilities fall short, refusing to lower quality standards despite market demand, prioritizing reliability over feature expansion until technology advances sufficiently to maintain their accuracy reputation.

Know someone who'd find this useful?

You just read a 3-minute summary of a 45-minute episode.

Get Software Engineering Daily summarized like this every Monday — plus up to 2 more podcasts, free.

Pick Your Podcasts — Free

Keep Reading

More from Software Engineering Daily

We summarize every new episode. Want them in your inbox?

Similar Episodes

Related episodes from other podcasts

This podcast is featured in Best Cybersecurity Podcasts (2026) — ranked and reviewed with AI summaries.

You're clearly into Software Engineering Daily.

Every Monday, we deliver AI summaries of the latest episodes from Software Engineering Daily and 192+ other podcasts. Free for up to 3 shows.

Start My Monday Digest

No credit card · Unsubscribe anytime