Tim Varley Logo
Tim Varley Systems Engineer

Organizational Design

Part 2: Hiring for judgment, team structures that enable AI orchestration, and when to standardize.

Orchestrated by Tim Varley
Published
article

The AI Engineering Transformation - Part 2: Organizational Design

Hiring for Judgment, Not Speed

Stop asking candidates to implement a linked list on a whiteboard.

Start asking:

  • “Walk me through how you’d validate AI-generated code for a payment processing system. What would you check first? What would make you reject it outright?”
  • “Describe a time AI suggested an elegant solution that was fundamentally wrong. How did you catch it?”
  • “You’re reviewing a pull request where 80% of the code is AI-generated. It passes tests. How do you evaluate whether it should merge?”

You’re hiring for:

  • Strong mental models of systems, not implementation speed
  • Productive skepticism toward plausible-sounding answers
  • Ability to articulate tradeoffs clearly under uncertainty
  • Experience debugging complex, distributed systems where AI’s simple explanations miss subtle realities

The best AI engineers today aren’t necessarily the best traditional coders. They’re the ones who think clearly about systems, constraints, and failure modes—and can direct AI to explore solution spaces they specify.

“Good judgment comes from experience, and experience comes from bad judgment.” - Fred Brooks

Redesigning the Interview Loop

Your interview process probably optimizes for the wrong signals now.

Old Loop: “Can you implement this algorithm?”
Candidates demonstrate they can write a red-black tree from memory. Signal: They memorized common patterns.

New Loop: “Can you orchestrate AI to implement this system correctly?”
Candidates demonstrate ability to:

  1. Specify requirements with precision (including edge cases AI will miss)
  2. Evaluate AI-generated solutions critically
  3. Spot cargo cult patterns—code that looks sophisticated but misses domain requirements
  4. Explain why a solution works, not just that it works

The Reality Testing Exercise
Give candidates AI-generated code for a critical system (authentication, payment processing, data consistency). The code looks professional, passes basic tests, but contains 3-4 subtle issues. Ask them to review it as they would a team member’s PR.

Great candidates find the issues and articulate the reasoning failures that led to them. They explain what AI doesn’t understand about the domain. They propose specific tests that would catch the class of bug, not just this instance.

New Team Structures That Work

Theory is cheap. Here’s what functional AI-native teams actually do:

Prompt Guilds: Sharing Techniques, Not Prescriptions
Engineers meet biweekly (30-45 minutes) to share techniques and patterns that worked, not prescriptive prompts to copy-paste. Format: 2-3 engineers present a “technique teardown” (problem → approach → outcome → lessons learned). Techniques are added to a searchable guild wiki.

This is critical: shared AGENT.md files or team-wide system prompts are detrimental. They get changed constantly, try to serve everyone’s workflow, and end up serving no one well.

The guild shares:

  • Techniques: “When asking AI to refactor, describe the failure modes the current code handles. AI simplifies away edge cases.”
  • Anti-patterns: “Don’t ask AI to ‘make this faster’ without specifying constraints. You’ll get premature optimization.”
  • Role-specific approaches: “For debugging distributed systems, give AI the full trace plus invariants that should hold.”
  • Lessons learned: “Postgres query optimization prompts need actual table sizes and indexes. AI hallucinates performance characteristics.”

What the guild does NOT share: Personal AGENT.md files, cursor rules, or system prompts. Each engineer’s AI setup should be deeply personal, learning their workflow, thinking style, and preferences. As described in The AI Engineer Mindset series, different roles (architect, coder, debugger) require different tools and configurations. YOUR AI should learn YOUR workflow.

This creates compounding organizational learning. New engineers learn techniques from guild archives, then adapt them to their personal AI configuration.

Architecture Councils
Senior engineers meet weekly to make high-leverage decisions about system design, technology choices, and patterns AI should follow. They maintain a “judgment-required” document listing decisions AI cannot make reliably:

  • Performance tradeoffs requiring production telemetry
  • Security boundaries in distributed systems
  • Data model changes with migration implications
  • Abstractions that will shape six months of work

When AI proposes solutions in these spaces, the council reviews them. When decisions are routine, engineers direct AI with council-approved patterns and ship without bottlenecking.

Important distinction: The council defines what decisions require judgment and why, not how individual engineers should prompt their AI. Engineers take council guidance and translate it into their personal AI workflows however works best for them.

Reality Testing Pods
Two engineers per team rotate into “reality testing” roles for one-week sprints. They focus exclusively on reviewing AI-generated code using systematic frameworks:

  1. Resource analysis: What happens under load? Does this leak memory? Are there unbounded loops?
  2. Failure mode analysis: What happens if this fails halfway through? Are errors handled or ignored?
  3. Domain model check: Does this match actual business rules, or did AI simplify something critical?
  4. Security review: Are there injection vulnerabilities? Authorization bypasses? Timing attacks?

The pod doesn’t block work—they’re not gatekeepers. They coach. When they catch issues, they create new guild patterns so the whole team learns.

AI Pair Programming Rotations
Every engineer spends time as “conductor”—orchestrating AI to build something while a peer observes their prompting strategy, context management, and reality testing approach. The observer documents patterns worth sharing with the guild.

This builds organizational muscle memory. AI orchestration becomes a learned skill, not individual art.

When Standardization Is Appropriate: Headless Automation

There’s one domain where shared, prescriptive prompts make sense: headless automation.

Headless AI agents run without human interaction—CI/CD pipelines, automated testing, scheduled analysis, code generation for build systems. These need standardized, deterministic prompts because:

  • No human is there to course-correct
  • Consistency matters more than personalization
  • They’re serving the system, not adapting to an engineer’s workflow
  • Failures need to be reproducible and debuggable

Examples of appropriate standardization:

Automated Test Generation

Given this API endpoint specification and these business rules,
generate comprehensive test cases covering:
- Happy path with valid inputs
- Boundary conditions (empty, null, max values)
- Error cases (invalid auth, malformed requests)
- Concurrent access scenarios
Output format: pytest-compatible test functions with descriptive names.

Code Review Automation

Analyze this pull request for:
- Security vulnerabilities (SQL injection, XSS, auth bypass)
- Resource leaks (unclosed files, unbounded loops, memory growth)
- Error handling gaps (uncaught exceptions, silent failures)
- Inconsistency with codebase patterns in [architecture repo]
Flag issues with severity (critical/major/minor) and line numbers.

Documentation Generation

Given this API implementation, generate OpenAPI spec including:
- All endpoints with HTTP methods
- Request/response schemas with examples
- Error responses with status codes
- Authentication requirements
Validate output against OpenAPI 3.0 schema.

These prompts should be:

  • Version controlled
  • Tested for consistency
  • Reviewed by architecture councils
  • Updated when patterns change

But they’re fundamentally different from personal AI workflows. A headless agent optimizes for consistency and reliability. A personal AI setup optimizes for YOUR effectiveness.

Don’t confuse the two. Shared AGENT.md files for interactive development are harmful. Standardized prompts for automation pipelines are essential.

The Talent Market Shift

The market for AI-effective engineers is tightening fast.

What’s changing:

  • Engineers who can effectively orchestrate AI command significant salary premiums at top tier companies
  • Engineering leaders who’ve successfully transformed teams are being recruited aggressively
  • The “10x engineer” narrative is evolving: it’s no longer about raw coding output, it’s about judgment under uncertainty

Your Build vs. Hire Decision
You can train your existing senior engineers in AI orchestration. Most pick it up in weeks if given structured frameworks (see The AI Engineer Mindset series for technical depth on what this looks like from the practitioner perspective).

But you need at least one person who’s already excellent at this to bootstrap the culture. Hire one senior engineer who demonstrates mastery of AI orchestration. Let them run your first pilot, establish patterns, and train others. Their leverage is extraordinary—they essentially multiply the effectiveness of everyone they teach.

---