Tim Varley Logo
Tim Varley Systems Engineer

Dialoguing with the Unconscious

Discover how AI represents the collective unconscious of software development and develop metacognition to use AI as a thinking partner.

Orchestrated by Tim Varley
Published
article

The AI Engineer Mindset - Part 2: Dialoguing with the Unconscious

Building on Part 1’s orchestration techniques, this part examines the deeper psychological and philosophical aspects of working with AI.

TL;DR

  • AI provides access to software development’s collective unconscious
  • Develop metacognition to identify your own limiting assumptions
  • Use AI as a sparring partner to surface blind spots, not as an oracle
  • Maintain your own internal models to avoid cognitive dependency
  • Feed refined understanding back into prompts for iterative improvement

Once you accept the orchestration role, the true strangeness emerges. The feeling that you’re not talking to a program but dialoguing with the collective unconscious of our entire profession.

“The mind is not a vessel to be filled, but a fire to be kindled.” - Plutarch

Every LLM has ingested millions of codebases, forum threads, papers, and bug reports. When you prompt it, you’re not querying a database. You’re negotiating with a synthesized intelligence that has “read” more code than any human could in a lifetime. It doesn’t have original ideas, but it has unparalleled pattern recognition across the entire corpus of human software thought.

This creates a new form of literacy that is part technical, part psychological, and part philosophical. You must develop metacognition—the ability to watch your own thinking and notice when your framing is limiting the possibility space.

The AI excels at this because it has no ego invested in your initial approach. It will happily tell you that your elegant solution is over-engineered or that you’re solving the wrong problem. But only if you create the conditions for that honesty.

Debugging provides the clearest example. The old way: form a hypothesis, add logging, reproduce the issue, iterate. The new way: provide rich context about symptoms, invariants, business rules, and observed behaviors, then ask the AI to propose multiple mental models for what might be happening. Watch how it surfaces dimensions of the problem you were blind to because you were too close to the code.

The Dependency Trap

The danger, of course, is dependency. When the AI becomes your primary thinking partner, it’s easy to outsource too much of your own reasoning. The warning signs are subtle:

  • You can’t explain why a solution works, only that “the AI suggested it”
  • You’ve stopped mentally modeling system behavior before asking
  • You accept the first plausible answer without verification
  • Your debugging skills atrophy because you always ask AI first

The engineers who will thrive long-term are those who use AI to augment their cognition rather than replace it. They maintain their own internal models, using the AI as a sparring partner rather than an oracle. When the AI proposes a solution, they ask: “Does this align with my understanding of the domain? What assumptions is this making? Where might it break?”

The Hallucination Problem

AI will confidently generate plausible-sounding but fundamentally broken solutions. It “hallucinates” API methods that don’t exist, architectural patterns that look sophisticated but violate your constraints, and optimizations that introduce race conditions.

Your defense is skepticism grounded in knowledge. Always verify:

  • Does this library/method actually exist in this version?
  • Does this pattern fit our actual business rules?
  • What failure modes is this ignoring?
  • Can I trace the logic path myself?

The AI is a mirror of our profession’s accumulated knowledge, but mirrors can distort. Your judgment is the corrective lens.

This is where the D&D analogy extends. The best DMs don’t just narrate—they respond to player actions in ways that deepen the world. Similarly, the best AI engineers don’t just consume outputs. They use them to refine their own understanding, feeding improved mental models back into the next round of prompting.

In Part 3, we explore the creative tension between exploration and rigor.

Key Takeaways

  • AI allows dialoguing with the collective unconscious of software development
  • Develop metacognition to notice limiting frames in your own thinking
  • Use AI as sparring partner to surface blind spots in debugging and design
  • Avoid dependency by maintaining your own internal models
  • Refine understanding by feeding improved mental models back into prompts
---