Hydra
Autonomous Code Governance
The fix, not the flag.
───────────────────────────
GTM Research & Strategy  ·  May 2026
Multi-source verified: Exa + Perplexity + Jina
The Problem

AI made your team faster at writing code.
Not at shipping clean code.

Code output tripled. Review capacity stayed flat. Every tool your team bought made the list longer. Nobody closed the loop.

The AI Velocity Paradox

The data confirms what every engineering leader already feels.

41%
of GitHub commits are now AI-generated or AI-assisted
98%
increase in PR volume on teams using AI coding tools
91%
increase in PR review time on high-AI teams
1.7×
more issues in AI-generated code vs. human code
20% faster
How fast teams think they are with AI coding tools
19% slower
How fast they actually are — LinearB, 8.1M PRs
32.7%
AI-generated PR acceptance rate vs. 84.4% for manual PRs
$52K
Lost per year in a 10-person team from review bottleneck

Sources: LinearB 8.1M PR study · Faros AI 10K developer study · CodeRabbit 470 PR analysis · Sonar 2026 developer survey

The Bottleneck Nobody Names

AI review tools solved the first bottleneck. They created the second one.

Bottleneck 1 — solved (2023-2025)

PR review time was too long
CodeRabbit, Copilot Review, Qodo — AI assistants cut review time 30-50%. Every tool in the market solved this. It is now table stakes.

Bottleneck 2 — unsolved (today)

Review comments pile up
Someone still has to implement each finding, push the fix, pass CI, get another review, close the ticket.
43% require manual debugging in production
After passing QA — Lightrun 2026 report.
Even Anthropic didn't close the loop
Claude Code Review: $15-25/review. Comments only. No fix. No ticket close.
The Problem

Every tool in this market hands your team a list.

Find issue  →  Post comment  →  ■ STOP
The human triages, writes the fix, opens the PR, creates the ticket, closes the ticket.
GitHub Copilot. CodeRabbit. Qodo. Augment. SonarQube. All of them.
Find issue  →  Write ticket  →  Fix code  →  Open PR  →  Close ticket
Hydra. No human in the critical path.
Section 01

The Product

Four layers. One loop. Find it, fix it, improve it, govern it — with no human in the critical path.
How It Works

Find. Fix. Improve. Govern.

Layer 1
Find

Continuous codebase discovery — bugs, vulnerabilities, technical debt, convention drift. Not triggered by PR events. Runs on the system.

Layer 2
Fix

Baseline tests written first. Fix applied using your codebase's own conventions. PR opened. Linear ticket closed. No developer in the loop.

Layer 3
Improve

Continuous loops on accumulated technical debt. No PR trigger. No sprint allocation. The codebase gets cleaner every cycle.

Layer 4
Govern

Deterministic scanner rules generated from your failure modes. Codebase-specific. Your team does not write them. Hydra does.

All four layers run continuously. The codebase improves whether or not anyone opens a PR.

Layer 1 — Find

Three-step analysis. Not a diff reader.

1
Discovery
Reads the entire codebase. Builds a .hydra/ directory: architecture, conventions, how-to guides, application profile. ~16 minutes. A few dollars.
2
Deterministic Scanner
GREP patterns + rules engine. No LLM cost. No false positive risk. Runs first, fast, cheap. Replaces Semgrep/SonarQube linting as the foundation layer.
3
39 Audit Tools — 6 Domain Groups
Security · Code Quality · Reliability · API & Data · Ops · UX. 18+ languages. Structured checklists plus open-ended problem finding. Application profile weighting applies.
4
Opus Meta-Review
Final pass with Claude Opus and an open prompt: find what steps 1-3 missed. No checklist. No category constraint.

Application profiling

Hydra weights findings by what matters for your specific product. A customer-facing SaaS weights accessibility + security higher. An internal tool weights performance over accessibility. No competitor profiles the application.

Effort + risk scoring

Every finding rated on two axes: severity (how dangerous) and effort (how hard to fix). Built-in triage. No manual prioritization required.

Layer 2 — Fix

14 gates. The fix either passes every one or nothing ships.

1–2
Test baseline + synthesis
Baseline tests written first. Fix generated using your repo's own conventions — not generic best practices.
3–6
Quality gates — hard blocks
Diff-size guard (≤20 files, ≤2K lines). Test regression gate. Build. Lint. Any failure: nothing ships.
7–9
Scanner loop + reviewer agent + revision
Scanner quality loop (max 2 iterations). Independent reviewer agent. One revision pass; re-runs gates 4–6.
10–12
Post-fix audit + verification + PR
Scanner + LLM verify no regressions. Detection-query confirms original issue closed. PR opened, Linear ticket closed on merge.

BYOK

Bring Your Own Anthropic API Key. Compute costs go directly to Anthropic. Hydra never sees your key or marks up costs. Your code, your keys, your costs.

No code governance tool with autonomous fix execution has BYOK. Hydra is the first.

Fix cost

$1–3 for a standard fix. Worst case (fix + scanner revision + review + revision + re-review): $5–10. The entire gauntlet for under $10.

What fix handles

Bugs and security issues where functionality does not change. Fix runs in an isolated git worktree. Partial timeout: rejected, nothing ships.

Safety Architecture

Safe by design. The guardrails are the architecture.

The question every skeptical engineering leader asks: "What if it gets it wrong?"

Step 1:  Baseline tests written  →  must pass before execution begins
Step 2:  Agent synthesis  →  fix generated using your codebase's own conventions
Step 3:  Diff-size guard  →  ≤20 files, ≤2000 lines total, ≤500 lines per file
Step 4:  Test regression gate  →  hard block on new test failures
Step 5:  Build gate  →  per-project build must pass
Step 6:  Lint gate  →  per-project lint must pass
Step 7:  Scanner quality loop  →  max 2 iterations, cycle detection
Step 8:  Reviewer agent  →  independent LLM review of the fix
Step 9:  Revision loop  →  one shot; re-runs gates 4–6
Step 10:  Post-fix audit  →  scanner + LLM verify no regressions
Step 11:  Detection-query verify  →  confirm original issue is closed
Step 12:  Push + PR creation  →  every change reviewable and reversible

14 gates. Hard blocks on regressions. The fix either passes every gate or nothing ships. No competitor has published this architecture.

Layer 3 — Improve

The debt from before Hydra arrived. Gone.

What Improve does

Continuous loops on accumulated technical debt — dead code, naming conventions, structural refactors, documentation gaps, deprecated APIs. Not bug fixes. The things that work but make the codebase slower to operate in.

Runs without a PR trigger. The codebase improves between sprints. No sprint allocation required.

Why this is different

All competitors are event-driven. They run when a PR opens, stop when it closes. CodeRabbit reviews diffs. Qodo reviews the PR. Augment reviews the PR. None run on the codebase between events.

Hydra runs on the system, not the event. The codebase improves whether or not anyone opens a PR.

Kaizen — free standalone CLI

A separate free tool built on the same engine. Give it a focus area and a budget. It runs loops until the work is done. No Hydra subscription required. An entry point into the ecosystem.

Layer 4 — Govern

The governance layer that builds itself.

1
Pattern recognition
During LLM analysis, when an agent finds an issue detectable deterministically, it generates a suggested scanner pattern
2
Quality threshold filter
Patterns meeting quality thresholds added to the tenant's scanner library. False-positive patterns automatically deactivated.
3
Virtual patches
Rules applied at audit runtime without redeployment. Real-time protection before a rule is fully validated.
4
Global contribution (opt-in)
Patterns that generalize across codebases improve Hydra for all users.

vs. Qodo's approach

Qodo's Rules System requires humans to author and approve every rule. Hydra generates rules from observed failure modes. No rule authoring. No maintenance.

CLAUDE.md injection

Hydra patches your repo's CLAUDE.md to point to its documentation. Every developer using Claude Code gets Hydra's architecture context in every AI session — without installing anything.

Section 02

Competitive
Landscape

Four market categories. One gap nobody has filled.
The Market Map

The market split into four categories. Hydra spans all of them.

Category 1 — No fix, no loop
PR Review Bots

Flag issues as PR comments. Human must act. CodeRabbit · GitHub Copilot Review · Qodo · Augment · Claude Code Review

Category 3 — No fix, process controls
Governance / Policy Tools

Enforce rules, audit logs, agent behavior. Zero fix capability. Codesteward · Mault · Guardian · CoderOps · Pandorian

Category 2 — Partial fix, security-only
Autonomous Fix Tools

Open fix PRs for security findings. No quality/debt. No gauntlet. No ticket close. Snyk Agent Fix · AquilaX · Gitar · Mobb · DeepSource Autofix

Category 4 — General dev tasks, not quality
Full-Loop Dev Agents

Close issue-to-PR loop for general coding tasks. Not code quality governance. Codowave · Codegen · Devin

Full Capability Map

Every other tool stops at detection.

Capability Qodo Augment CodeRabbit GitHub CCR SonarQube Snyk Hydra
Full codebase discovery + documentationNoNoNoNoNoNoYes
Application profiling + context weightingNoNoNoNoNoNoYes
Deterministic scanner patternsPartialNoNoNoYesYesYes
Multi-agent parallel LLM analysis15+ agentsContext engineNoNoNoNo6 groups / 40+ dims
Opus meta-review passNoNoNoNoNoNoYes
Autonomous fix executionNoNoNoNoBetaNoYes
Safety gauntlet before mergeNoNoNoNoNoNoYes
Linear ticket lifecycle closureNoNoNoNoNoNoYes
Continuous improvement (not PR-triggered)NoNoNoNoNoNoYes
Self-generating governance rulesNo (human-authored)NoNoNoNoNoYes
BYOK (your Anthropic key)NoNoNoNoNoNoYes
PLG free tierNoNoYesBundledCommunityYes (open source)Yes
Section 03

The Market

Size, structure, timing, and why the window is measured in months, not years.
Market Size

$6.7B today. $25.7B by 2030. Growing 30-40% per year.

What the funding confirms

  • Cursor — raising at $50B valuation
  • Augment Code — $977M on $252M raised
  • CodeRabbit — $88M raised; 2M repos, 13M PRs reviewed
  • Qodo — $120M raised ($70M Series B, March 2026)
  • Snyk — $343M ARR, $8.5B valuation

This market is producing unicorns at an unusual rate. These are scale-stage companies, not early bets.

Three conditions are true simultaneously

1
The problem is acute
41% of GitHub code is AI-generated. Review bottlenecks worsen every sprint.
2
The technology is ready
Multi-agent autonomous fix execution crossed a reliability threshold in late 2024.
3
No one has closed the loop
The full-loop autonomous remediation governance category has no incumbent. Window is open.
The Category

Autonomous Code Governance.
No one owns it yet.

Not AI code review. Not security SAST. A new operational model where the loop between finding a problem and having a codebase that is permanently better closes automatically. No human in the critical path.

Section 04

Go-To-Market

ICP · Personas · PLG motion · Pricing · Paid acquisition · Content · Channel strategy
Ideal Customer Profile

20-200 developers. GitHub + Linear. Already using AI coding tools.

Primary ICP

  • Size: 20-200 developers
  • Industry: SaaS, AI-native, developer tools; secondary: fintech, security
  • Stack: GitHub for version control, Linear for issue tracking
  • AI adoption: Already using Cursor, Claude Code, or GitHub Copilot
  • Languages: Python, TypeScript, Go, or Rust primary

Why this profile

They feel the AI Velocity Paradox daily. Linear is installed — the ticket lifecycle loop is immediately visible value. They use Claude Code, so CLAUDE.md injection works from day one.

Trigger events — strong

  • Engineering leader hired to scale without adding headcount
  • PR cycle times increasing as team grows
  • Security audit surfacing vulnerability accumulation

Trigger events — moderate

  • AI adoption creates review bottleneck for the first time
  • Technical debt backlog too large to address manually

Avoid in year one

  • On-prem / FedRAMP requirements
  • Not using Linear (Jira is on the roadmap)
  • Not using GitHub (GitHub-native today)
Personas

Three buyers. Different entry points. One product.

Practitioner — installs it
Senior Engineer / Tech Lead

Problem: The PR pile grows faster than the team can work through it. Repetitive fixes take time that should go to architecture.

What wins them: First fix in under 10 minutes. Low false positive rate. Baseline tests before every change.

"It actually fixes things. And when it flags something, it's real."

Champion — builds the case
Engineering Manager

Problem: Senior engineers are spending half the week in review queues. Technical debt accumulates faster than the sprint can address it.

What wins them: Closed Linear tickets instead of open comments. Debt backlog shrinking without sprint allocation.

"The debt backlog is shrinking and no one is doing it manually."

Approver — signs the contract
VP Engineering / CTO

Problem: Code output tripled. Review capacity didn't. No governance layer across all repos.

What wins them: "Your team is generating 3x more code. Your review capacity has not scaled. Hydra closes that gap — and it gets better the more repos you run it on."

"It gets better on its own. We don't have to manage it."

PLG Motion + Pricing

Start free. The codebase gets better either way.

Install   GitHub App  →  no credit card
         Full Discovery + Audit runs
         First 5 fixes execute autonomously
         Linear tickets close
Refer    Share with another team  →  +25 fixes/mo
         Referred team installs, gets their own free tier
         BYOK means referral fixes cost Hydra nothing
Convert  Hits limit within first sprint
         Upgrade prompt: value-led, not punitive
Expand   More repos · More devs · Jira
         Enterprise at 500+ devs

PLG free-to-paid benchmark: 8-15% in 90 days (OpenView Partners) · PQL vs MQL: 5-6x higher conversion (Paddle) · Referral loop activates after retention loop is validated — users must hit the aha moment before they refer.

Free
$0

Full Discovery + Audit. 5 fixes/mo. No card.

Business
$40

Unlimited repos. Audit logs. Jira. Priority queue.

Enterprise
Custom

SSO · VPC · SLA · Compliance

Team at $20/dev/month is below Augment ($60-$200) and Sourcegraph Cody ($59). Priced to build the installed base.

PLG Activation + PQL

The aha moment is a closed ticket — not a comment, not a list.

4-step activation sequence

1
Install GitHub App — one click
No configuration required. Three clicks from "heard about Hydra" to "running on my repo." Install UX is the entire acquisition funnel. Optimize before everything else.
2
Hydra indexes the codebase — 2–5 minutes
Builds HYDRA.md + .hydra/ directory: architecture map, conventions, principles, how-to guides, profile. Shown as a progress indicator. First moment of realized value — before anything is fixed.
3
Full audit — findings dashboard
User sees how many issues exist, categorized by severity, domain, and effort. Discovery. This is why the free tier needs to be genuinely useful.
4
One finding autonomously fixed. PR opened. Linear ticket closed.
The aha moment. A closed ticket — not a comment, not a suggestion. The engineering manager gets evidence, not a list. This is categorically different from every other tool in the market.

Target: first autonomous fix in under 10 minutes of install. This single metric predicts everything else downstream.

PQL scoring — convert at peak value

SignalWeight
3+ users from same company domainHigh
Hit free tier fix limit 3 consecutive monthsHigh
Connected Linear with 10+ ticket closuresHigh
Viewed pricing page 3+ timesMedium
Generated documentation for 5+ reposMedium
Single user, one repoLow — monitor only

2 High signals OR 1 High + 3 Medium triggers outreach. Reach at peak perceived value — after 40 autonomous fixes and 20 closed tickets, not at first limit hit. PQL vs MQL: 5–6x higher conversion rate (Paddle research).

The Snyk playbook

Snyk reached 50K registered developers before $100K ARR. Identified developers who ran snyk test 3+ times in a week. Founder-led outreach directly via GitHub profile email. Free tier = distribution engine. Revenue follows the enterprise contract the free users made inevitable.

SEO

Three keyword tiers. The Tier 1 window is open now — it closes when competitors name the category.

TierKeywordsCompetitionAction
Tier 1 "autonomous code remediation governance" · "fix gauntlet code quality" · "self-improving code scanner" · "AI code governance autonomous fix" Near zero — verified Publish at launch. Own before anyone else names the category.
Tier 2 "how to audit Python / TypeScript / Go codebase automatically" · "CodeRabbit alternative fix code" · "technical debt remediation not detection" · "AI velocity paradox software" · "BYOK AI code review" Moderate — active competition from Gitar, SonarQube, gitautoreview.com Build M1–3. Differentiate on autonomous fix delivery, not detection.
Tier 3 "best AI code review tools" · "SonarQube alternative" · "automated code review" · "SAST tools comparison" · "autonomous code remediation" High — SonarQube, OpenText, established vendors dominate Build toward M6+. Comparison pages. Not the starting point.

Month 3+ — programmatic SEO

Language × problem matrix (30+ pages): "audit Python codebase for security issues automatically," "fix TypeScript code debt with AI," etc. Comparison pages: Hydra vs. CodeRabbit / SonarQube / AquilaX. Use-case pages: fintech, Claude Code teams, AI-native startups.

Eight priority content pieces — in this order

1. "What is autonomous code remediation governance?"
Hub page. Category definition. Everything links here. Ships at launch.
2. "The AI code review bottleneck everyone missed"
98% more PRs, 91% longer review time. Names the remediation queue as the second bottleneck. High-share potential.
3. "AI code review vs. fix vs. governance: what's the difference?"
Three categories explained. Positions Hydra as the synthesis.
4. "Why your code governance tool should never see your API key"
BYOK explainer. Developer trust, cost control, data residency. Targets privacy-first segment.
5–9. "How to audit your [Python / Go / TypeScript / Ruby / Java] codebase automatically"
Six language-specific pages. Highest long-tail volume in the cluster. Each ends with autonomous fix — no tool in the SERP does this.
10. "Why autonomous code fixes need a quality gauntlet before merge"
Fix gauntlet explainer. Hydra-invented concept. Zero competition.
11. "The AI velocity paradox: why your team is slower with AI tools"
LinearB 8.1M PR data. Sonar 1,100 dev survey. Expected to be the broadest-sharing piece.
Organic Social

Four platforms. Each one requires a different approach. None of them tolerate generic content.

PlatformAudienceWhat worksWhat failsHydra angle
LinkedIn VP Eng · CTO · EM — buyer persona Data-driven posts. AI velocity paradox stats land hard here. "Your team generates 3x more code. Review capacity didn't scale." Short-form with one striking number. Feature announcements without business context. Anything that reads as a product update. Buyer awareness. The VP Eng sees the problem framed in their language before they ever search for a solution.
X / Twitter Developers · technical founders · developer tool community Technical content. Screenshots of the fix gauntlet running. "We built 39 specialized agents" angle. Threads with real data. Engaging with AI coding tool conversations. Marketing copy. Anything that doesn't show the product working or teach something. Product credibility. The place where developer tools build reputation before launch. Monitor and engage every "AI code quality" thread.
Hacker News Senior engineers · technical founders · skeptics Show HN with live demo. Technical "how we built it" posts. Author engagement in top comments is essential — upvote patterns correlate with response quality in first 3 hours. Launch announcements without technical depth. Anything that reads as marketing. "Show HN: We built 39 specialized AI agents that find, fix, and close the ticket — no human in the path." Architecture deep-dive post week 3.
Reddit r/programming (6.3M) · r/LocalLLaMA (500K) · r/devops (1.1M) · r/ExperiencedDevs (350K) Data-first posts. "We analyzed 8M PRs" style. Community contributes to the research framing. Must have prior comment history before posting. Direct product pitches. Anything that doesn't lead with data or a novel insight. r/LocalLLaMA: BYOK-first — "your code never touches our servers." r/programming: AI velocity paradox data. r/devops: "no human in the critical path."
Channel Strategy

Seven channels. Organic compounds. Paid amplifies. Referral loops.

1
CLAUDE.md injection — the built-in growth loop
Hydra patches every connected repo's CLAUDE.md. Every developer using Claude Code in that repo gets Hydra's architecture context in every AI session — without installing anything. One install. Every session. Every developer. Referral with zero incremental cost. ~2,000 installs in Y1 from referral alone.
2
GitHub App Marketplace
Native discovery by developers already evaluating GitHub integrations. One-click install. No friction. No outbound. Works as a passive discovery engine from day one. ~5,000 installs in Y1.
3
Developer communities — HN, Reddit, Discord
Show HN on launch day. ProductHunt. r/programming, r/devops, r/LocalLLaMA. Claude Code + Cursor Discord. One well-placed Show HN drives 500–2,000 installs in 48 hours. ~3,000 installs at launch.
4
Content / SEO + LinkedIn
Zero existing content on the category keywords. Publish at launch; compound for 12 months. Every piece distributed on LinkedIn simultaneously — AI velocity paradox data travels fast to engineering managers and VPs. LinkedIn is where the buyer persona lives. ~4,000 installs from organic search + LinkedIn in Y1.
5
Paid acquisition — Google, Reddit Ads, Carbon Ads
$12K/mo targeting high-intent queries ("CodeRabbit alternative fix code," "BYOK code governance") and developer subreddits. Amplifies organic from day one. CAC target ≤$150. Detail on next slide. ~6,000 additional installs in Y1 from paid.
6
Anthropic + Linear + GitHub ecosystems
BYOK is Anthropic-native. Linear closes the ticket lifecycle. GitHub Marketplace featured placement. All three have warm developer communities with direct intent overlap. ~1,000 installs in Y1 from ecosystem channels.
7
Referral loop — Dropbox model, zero COGS
Refer a team → unlock 25 fixes/month. Referred team installs, gets their own free tier, hits their own aha moment. BYOK means every referral fix runs on the user's Anthropic key — Hydra's marginal cost is zero. Referral loop activates only after the retention loop is validated: users must close 3+ tickets before they are prompted to refer. Reforge-coined mechanic applied to PLG SaaS. ~3,000 additional installs in Y1 from referral loop.
Paid Acquisition

Four channels. Two audiences. Organic is the engine — paid accelerates it.

PlatformAudienceCreativeBudget
Google Search High-intent — searching for a solution now A: "Every other tool flags it. Hydra fixes it." Queries: "CodeRabbit alternative fix code," "BYOK code governance," "autonomous code remediation" $5K/mo
LinkedIn VP Eng · CTO · EM — decision makers B: "Your team is 19% slower with AI tools than without." Sponsored content. Lands on AI velocity paradox post. Buyer awareness play. $4K/mo
Reddit Ads Practitioners — r/programming · r/LocalLLaMA · r/devops A on r/programming · r/ExperiencedDevs. B on r/LocalLLaMA (BYOK angle). Amplifies organic posts already running in same communities. $3K/mo
Carbon + Retargeting Developers on docs pages — already in coding context "Find it. Fix it. Govern it. — install free." Text-only. Retarget site visitors. Highest ROI once site traffic is established (M2+). $2K/mo
Total $14K/mo

Two creative hypotheses — test in parallel

A — Contrast (Google + Reddit)
"Every other tool flags it. Hydra fixes it."

Targets developers already frustrated with comment-only review tools. High-intent. Lands on contrast-first homepage.

B — Data (LinkedIn + Reddit)
"Your team is 19% slower with AI tools than without."

AI Velocity Paradox stat. Lands on the bottleneck blog post. Reaches buyers who feel the problem but haven't named it yet.

Kill criteria

CAC target ≤$150. Kill any channel above $300 at 30 days. If organic CAC is beating paid, shift budget to content. At M3, if install-to-paid ≥8%, scale paid budget. Paid follows what works — it doesn't lead.

Launch Sequence

Three phases. One milestone each. Measured in PLG metrics.

Phase 1 — D0 to M1
Public launch + first installs
Show HN + ProductHunt live. Billing active. KB open. GitHub Marketplace listing live.
Email drip: day 1, 3, 7 sequences for free users.
Direct outreach to first 50 installs. Fix friction in real time.
First two blog posts live: category definition + second bottleneck.

Milestone: 500 free installs · <15 min time-to-first-fix · 30% W1 retention

Phase 2 — M1 to M4
Convert + compound
14-day trial expires. First Team plan revenue enters.
Linear ecosystem post. Reddit + HN follow-up threads.
SEO posts indexed. Category keywords begin ranking.
First case studies from early adopters. Programmatic comparison pages live.

Milestone: first paid conversion · 5% free-to-paid rate in PQL segment

Section 05

Roadmap

The engine is built. What's left is the commercial wrapper — billing, onboarding, and the first-run experience. 30 days.
Current State

The governance engine is production-grade. What's left is the commercial wrapper.

What's built — shipped and deployed

7 workflows · 200+ tests · 39 audit tools
Discovery · Audit · Scan · Fix · Task · Improve · PR Review · auto-Learn. Security, Code Quality, Reliability, API & Data, Ops, UX. 18+ languages.
14-step fix gauntlet — every gate is hard
Test baseline → synthesis → diff guard → regression gate → build → lint → scanner quality loop → reviewer agent → revision → post-fix audit → PR. Fail any gate, nothing ships.
Self-improvement system — virtual patches
Findings auto-generate scanner patterns + virtual patch overlays. Applied immediately, no redeployment. Tenant patterns stay private; mechanical fields contributed to global pool.
Discovery export — HYDRA.md + .hydra/ artifacts
Publishes architecture docs, conventions, how-to guides directly to the target repo. Claude Code, Cursor, and Copilot read these automatically. Your AI agents know your system.
Production infra · full-stack frontend · multi-tenancy
K8s + Datadog + OTEL at hydra.gateway.iru.dev. React 19 SPA (40+ pages). Per-schema PostgreSQL isolation. GitHub App + Linear fully integrated.

Commercial wrapper — 4 weeks

The engine is built. We deliberately sequenced product-market fit before monetization infrastructure. Billing is a 2-week Stripe integration. Neither changes the engine.

Billing / Stripe — Week 1
Stripe integration, plan gating, webhook handling, upgrade flow.
First-run onboarding — Week 2
Guided setup. Aha moment target: first fix merged within 15 minutes.
UX/UI redesign — Week 3
Brand identity from Neil applied to dashboard and homepage.
GitHub Marketplace listing — Week 4
Tyler executes. Passive discovery channel live from day one.
30-Day Launch Sprint

Four weeks. One constraint per week. Day 30: live.

Week 1 — Days 1–7
Billing
Stripe integration + subscription management
Free tier limits enforced (5 fixes/mo cap)
Upgrade flow + pricing page
Webhook handling: upgrades, cancellations, failures
Week 2 — Days 8–14
Onboarding
Guided setup wizard: GitHub App → API key → repo → audit
Empty state: public repo audit fallback for clean repos
14-day trial mechanics: no card until day 14
Email drip: day 1, 3, 7 activation sequences
Week 3 — Days 15–21
UX + Brand
Brand identity from Neil applied to dashboard
Homepage + landing page copy live
KB password removed for public access
First two blog posts published
Day 30
Target public launch date
<15 min
Target time-to-first-fix for new users
30%
Target W1 retention — developers who return after first session
5%
Target free-to-paid conversion in PQL segment (month 1+)
Section 06

Financial
Projections

Three scenarios. Bottom-up math. CodeRabbit as the anchor comp. No paid acquisition assumed in any scenario.
Model Assumptions

Bottom-up. Three scenarios. Every assumption labeled.

Assumption Conservative Base Aggressive
Y1 free installs (organic + paid)15,00030,00050,000
Free-to-paid conversion5%8%12%
Blended ARPA / mo$350$500$650
Monthly gross churn4% → 3%2.5% → 2%1.5% → 1%
Enterprise ARR (M5–M12)none~$500K~$1.5M
NRR target (Y2)108%118%132%
Paying accounts at M12~500~1,050~2,100
Y1 ARR at M12~$3M~$6.5M~$15M
Y2 ARR~$9M~$20M~$45M
Y3 ARR~$22M~$50M~$95M

Base case funnel math

30,000 free installs (organic + $12K/mo paid)
× 30% activation (complete first audit)
= 9,000 activated developers
× 8% free-to-paid conversion
= ~1,050 paying PLG accounts at M12
× $500 blended ARPA
= $525K MRR PLG · $6.3M ARR PLG
+ enterprise motion M5+ (~$500K ARR)
= ~$6.5M ARR total run rate at M12

Why this model is defensible

BYOK
90%+ gross margin from day one. No LLM COGS. Infra only.
7 WF
7 workflows vs competitors' single workflow — higher ARPA justification.
NRR
Expansion from repos + devs drives 118% NRR by Y2. Revenue compounds without new accounts.
ENT
Enterprise motion starts M5. VP Eng / CTO from PLG accounts. $40–100K ACV. SOC 2 in process.
ARR Trajectory

Year 1 to Year 3. Three scenarios. PLG base + enterprise motion from M5.

Conservative Base Aggressive
Y1 ARR $3M $6.5M $15M
Y2 ARR $9M $20M $45M
Y3 ARR $22M $50M $95M
Paying accounts (Y1 M12) ~500 ~1,050 ~2,100
YoY growth Y1→Y2 200% 208% 200%
YoY growth Y2→Y3 144% 150% 111%
Gross margin 90%+ 90%+ 90%+

All scenarios are illustrative. Near-term focus: validate conversion and retention with real customers before scaling paid acquisition.

Key inflection points — base case

M2
First revenue — $50K MRR
14-day trial expires. First Team and Business plan upgrades. PQL scoring active. Founder-led outreach begins.
M5
Enterprise motion starts — $150K MRR
First VP Eng / CTO identified from PLG account analytics. Enterprise pilot at $40–80K ACV. SOC 2 Type I in process. CLAUDE.md viral loop generating measurable referral installs.
M12
$542K MRR · $6.5M ARR run rate
1,050+ paying accounts. NRR approaching 118% as teams expand repo coverage. 3–5 enterprise accounts. Programmatic SEO driving organic install growth.
Y3
$50M ARR — Jira live, RBAC mature, enterprise cohort
7 workflows + 39 tools at full enterprise scale. BYOK model means no margin compression as LLM usage scales. Expansion revenue compounds without new account growth.
Unit Economics

BYOK changes the math. No LLM costs means no margin compression at any scale.

90%+
Gross margin — BYOK eliminates LLM COGS entirely. Infra only.
$15,000
LTV per account (base) — $500 ARPA × 90% margin ÷ 2.5% monthly churn
50–75×
LTV:CAC — PLG CAC ~$200 organic, ~$150 paid. Blended ~$175.
118%+
Target NRR at Y2 — expansion from repos + devs outpaces gross churn

Why BYOK rewrites the AI SaaS margin model

Company typeGross marginWhy
Traditional B2B SaaS75–85%Hosting + support only
AI SaaS (LLM-powered)50–70%LLM costs are 20–40% of revenue
Hydra (BYOK)90%+User pays Anthropic directly. Zero LLM COGS.

No code governance tool with autonomous fix execution has BYOK. Hydra is the only tool in the category with AI-native margins instead of AI-compressed margins.

NRR expansion path

Y1
110% NRR — repo expansion
Teams start with 1–2 repos, expand to 5+ as they trust the system. Team → Business upgrades begin. CLAUDE.md viral loop drives organic referral installs.
Y2
118%+ NRR — seat expansion + enterprise
Headcount grows → seat count grows → revenue grows without new sales motions. Enterprise accounts at $40–100K ACV convert from PLG base. Jira integration unlocks a new ICP segment.
At 118% NRR + 2.5% gross churn: net revenue from existing customers grows 18% YoY before a single new account is added. BYOK means this margin never compresses as AI usage scales — unlike every LLM-on-own-infra competitor.
Section 07

Launch
Status

What's built, what's blocking, what ships next.
Launch Readiness

Four blockers. All known. All solvable in 30 days.

Week 2 — Blocker
Onboarding

SetupWizard exists. First-run loop does not. Aha moment: first fix merged within 15 minutes of install.

Week 3 — Blocker
UX/UI

Full redesign planned. Blocked on Neil's brand identity delivery. Engineering ready to execute immediately.

Week 4 — Publishing step
Marketplace

GitHub Marketplace listing. Tyler executes. Not an engineering dependency.

What is not a launch blocker

Jira integration is an explicit stub (enterprise feature, post-launch). SSO/SAML and full RBAC are enterprise tier — not required for Free, Team, or Business launch. These are roadmap items, not gates.

Core product: done

7 workflows · 39 tools · 200+ tests · 14-step fix gauntlet · self-improvement system · GitHub App · Linear · multi-tenant PostgreSQL · K8s production infra · React 19 SPA (40+ pages). The engine is built. The wrapper is not.

The Position

Every other tool leaves a comment.
Hydra leaves it done.

Category

Autonomous Code Governance

Tagline

The fix, not the flag.

Pricing

Free → $20 → $40 → Enterprise

Status

Pre-launch · v3.11.17