Skip to content
~/writing/ai-effectiveness-matrix

essay / ai

The AI effectiveness matrix: understanding performance variance across codebases

Why AI tools deliver 300% gains for some teams and 10% for others. The answer is not the tools.

Introduction

Organizations observe significant variance in AI coding tool effectiveness. Senior developers report 200-300% productivity gains while mid-level engineers see 10-20% improvements on identical tooling. After analyzing 40+ developers over three months, a predictable pattern emerged based on two key variables.

The framework: greenfield vs brownfield, simple vs complex

Codebase type:

  • Greenfield: New projects with consistent patterns, modern frameworks, documented from inception
  • Brownfield: Legacy systems with accumulated decisions, technical debt, sparse documentation

Task complexity:

  • Simple: Single-file changes, standard patterns, isolated features
  • Complex: Multi-system changes, architectural decisions, refactoring

Quadrant 1: simple tasks on greenfield code

AI generates production-ready code with minimal revision. Review cycles reduced from 2-3 rounds to 1. Implementation time sees 70-80% reduction.

Example: Add a user profile endpoint to a 3-month-old REST API. The codebase has established patterns. AI generates code matching existing validation schemas, database query patterns, error handling, and test structures. Time from prompt to PR: 15 minutes.

Measured ROI: 200-300% productivity increase. Required investment: basic copilot-instructions.md documenting conventions.

Quadrant 2: complex tasks on greenfield code

AI useful for exploring options but requires experienced developer oversight. Implementation fast once approach decided.

Example: Design authentication for new multi-tenant SaaS. AI can propose patterns like OAuth 2.0, JWT storage strategies, and tenant isolation approaches. But it requires human decisions on provider selection, session management, multi-tenancy model, and security trade-offs.

Once direction is set, AI implements in 2 hours versus 8 hours manually.

Measured ROI: With clear guidance, 100-150% productivity increase. Without guidance, 50-80% with significant iteration cost.

Quadrant 3: simple tasks on brownfield code

This quadrant shows the largest performance variance. Most enterprise development occurs here.

Without context

AI generates textbook solutions that break production systems.

Example: Add rate limiting to checkout API. AI suggests standard IP-based implementation. Production system requires multi-tenant architecture with tier-based limits, Redis-backed storage, key by tenant_id from JWT, specific response headers, internal service call bypass logic, and graceful degradation.

Implementation time: 5 hours. Debug time post-merge: 2 hours. Production incidents: 1.

With context

copilot-instructions.md documents the rate limiting pattern, including the existing middleware, Redis backend, key format, header requirements, bypass logic, and fallback behavior.

AI now generates correct implementation on first attempt. Implementation time: 20 minutes.

Measured ROI: Without context, 0-30% productivity increase (sometimes negative). With context, 150-250% productivity increase.

This represents the largest opportunity. Most enterprise development involves simple tasks on brownfield systems.

Quadrant 4: complex tasks on brownfield code

AI struggles even with good context. Too many interconnected decisions. Requires architect-level system understanding.

Example: Migrate monolith payment system to microservices. AI cannot design this approach due to too many trade-offs, context, and unknowns.

After an architect defines the approach, AI assists with event schema generation, migration script creation, adapter layer implementation, integration test scaffolding, and service stub generation.

Measured ROI: 50-100% productivity increase. Humans drive strategy while AI handles implementation details.

Implementation roadmap

Most teams operate primarily in Quadrant 3 and see poor results due to missing context.

Phase 1 (Weeks 1-2): Document Quadrant 3. Create copilot-instructions.md. Document five most common patterns. Investment: 8-12 hours.

Phase 2 (Weeks 3-4): Measure ROI. Apply documented context to real tasks. Track completion times. Typical observation: 5-hour tasks reduced to 30 minutes.

Phase 3 (Month 2): Expand to Quadrant 4. Add architectural context. Document system integrations. Investment: 20-30 hours.

Phase 4 (Month 3): Optimize Quadrants 1-2. Already performing well. Add team-specific patterns. Investment: 5-10 hours.

Root cause analysis

The matrix does not reflect AI capability differences. It reflects context availability.

Greenfield code has context naturally: consistent patterns, current documentation, clear architecture.

Brownfield code has hidden context: developer knowledge, comment fragments, old tickets, incident learnings.

Teams with 300% productivity gains are not working on simpler problems. They are systematically extracting and documenting brownfield context.

Measurement approach

Track task completion times before and after context documentation:

Before: Checkout rate limiting, 7 hours with production incident. Search caching implementation, 6 hours with incorrect Redis strategy. OAuth integration, 8 hours with missed tenant isolation.

After: Checkout rate limiting, 25 minutes and production-ready. Search caching implementation, 20 minutes with correct pattern. OAuth integration, 45 minutes with all constraints met.

Measurable. Repeatable. Consistent.

The objective is not converting brownfield to greenfield. The objective is providing brownfield code the same context advantages that greenfield code has inherently. Document hidden knowledge. Surface constraints. Capture historical decisions.