Back to Development
developmentL1 Ad-hocCode Review & Quality

Manual review of 100% code

Why relying entirely on human code review creates a quality bottleneck that scales poorly and establishes the baseline every higher maturity level is designed to escape.

  • ·All code is reviewed by a human before merge
  • ·No automated review tooling beyond basic CI checks
  • ·Code review turnaround is tracked (even if slow)
  • ·Team is aware that AI-generated code has higher defect rates (1.7x issues, 2.74x security vulnerabilities)

Evidence

  • ·PR approval records showing human reviewer on every merged PR
  • ·Average review turnaround time in PR analytics

What It Is

Manual review of 100% of code is the baseline state for most engineering teams - every line merged to the main branch has been read and approved by at least one human reviewer. There is no automated quality gate beyond perhaps a basic linter, no AI reviewer, and no policy for selectively skipping or accelerating review based on change risk. Every pull request, regardless of size or complexity, waits in the same queue.

At L1 (Ad-hoc), this process is entirely dependent on reviewer availability and attention. A two-line config change waits behind a 400-line feature PR. A reviewer who is in meetings all morning becomes a bottleneck for three developers. The quality of review varies enormously: a tired reviewer at 5pm catches fewer bugs than the same reviewer fresh at 10am, but the process treats both reviews as equivalent.

The L1 review process typically has no documented standards. What one reviewer calls "bad naming" another ignores. Architecture decisions are enforced through review comments, which means they're only enforced when the reviewer knows about them and has time to write a comment. New team members learn standards from reviewer feedback, which is inconsistent by nature.

This is not a criticism of L1 teams - manual review is where every team starts, and it provides genuine value. Human reviewers catch business logic errors, subtle security issues, and design problems that automated tools still miss. The problem is that 100% manual review doesn't scale. As team size grows, review demand grows linearly while review capacity grows slowly. The bottleneck becomes structural.

Why It Matters

Understanding the limitations of L1 review is essential context for why the maturity matrix progresses the way it does. Every higher level is specifically designed to reduce review burden without sacrificing quality:

  • L2 introduces AI-assisted review suggestions and linting, so reviewers spend less time on obvious issues
  • L3 deploys an AI agent as a first-pass reviewer that handles routine checks automatically
  • L4 introduces auto-merge for Green-rated changes, removing human review from the critical path entirely
  • L5 has humans review only architectural exceptions - the rare Red cases

At L1, the review bottleneck has compounding costs beyond just wait time. Developers who submit PRs and wait hours for feedback context-switch to other work. When review comments eventually arrive, they have to reload the mental context of the PR, costing additional time. Meanwhile, other PRs have been merged to the branch they branched from, creating merge conflicts that require additional work to resolve.

The other significant cost of L1 is inconsistency. Without automated enforcement, standards exist only in reviewers' heads. This creates two failure modes: (1) important standards are missed in review, allowing problems into the codebase, and (2) reviewers over-enforce personal preferences, creating friction without quality benefit.

Tip

Measure your average PR review turnaround time now, before introducing any AI tooling. This baseline makes it possible to demonstrate the concrete improvement that comes from L2 and L3 investments. Even a rough estimate (median hours from "review requested" to "first review") is more useful than no data.

Getting Started

6 steps to get from here to the next level

Common Pitfalls

Mistakes teams actually make at this stage - and how to avoid them

How Different Roles See It

B
BobHead of Engineering

Bob's team of 50 engineers has a review SLA of "24 hours," but in practice PRs often wait 2-3 days. Senior engineers complain about review load. Junior developers complain about slow feedback. Deployment frequency is limited because code sits in review queues rather than flowing to production.

What Bob should do - role-specific action plan

S
SarahProductivity Lead

Sarah is tracking PR cycle time as part of her developer productivity metrics. The data shows that the median PR spends 18 hours in review before merging. Her dashboard shows this is the single biggest contributor to long cycle times. Her stakeholders want to know what she's going to do about it.

What Sarah should do - role-specific action plan

V
VictorStaff Engineer - AI Champion

Victor has been the primary reviewer for 60-70% of the team's PRs because he's the one who knows the codebase best. He's spending 3 hours a day reviewing code and is frustrated that he can't make progress on his own architectural work. He's seen AI review tools mentioned in conference talks but doesn't know which ones are worth the time to evaluate.

What Victor should do - role-specific action plan