portfolio Anshul Bisen
ask my work

What AI changes about engineering leadership

Code review matters more than code writing. Architectural taste matters more than implementation speed. Evaluating AI-generated code is the new critical skill.

AI coding tools have not changed the core job of engineering leadership. You still need to ship reliable software, grow capable teams, and make sound technical decisions under uncertainty. The fundamentals are the same. What has changed is where the leverage points are.

For fifteen years, the scarcest resource on an engineering team was implementation bandwidth. There was always more work than hands. Leaders spent most of their energy on prioritization: what gets built this sprint, what gets deferred, and how to squeeze more output from the same team size. AI has not eliminated the scarcity, but it has moved it. The bottleneck is shifting from code production to code evaluation.

Useful AI looks more like leverage than magic.

I write these AI posts from the far side of the honeymoon phase. It also builds on what I learned earlier in “Opus 4.6 has a million-token context window and I am still not sure what to do with it.” The interesting question is no longer whether the models are impressive. It is where they meaningfully improve decision quality across real systems like portfolio search, aigw, jarvis, and the review loops around everyday engineering work.

The workflow, not the hype.

Code Review Is the New Critical Path

When every engineer on the team can produce 3x more code with AI assistance, the volume of code entering review increases proportionally. If your review capacity does not scale with production capacity, review becomes the bottleneck. PRs stack up. Merge queues grow. The team produces more code but ships at the same rate because nobody can review it fast enough.

This means investing in review capacity is now more important than investing in production capacity. Practically, that looks like:

  • Pairing senior engineers on review rather than on implementation. Two senior engineers reviewing AI-generated PRs produces more value than two senior engineers writing code.
  • Building review tooling that highlights the most important parts of large diffs. AI-generated code often includes boilerplate that does not need careful review alongside critical logic that does.
  • Training mid-level engineers explicitly on review skills. Code review was always important. Now it is the differentiating skill.
  • Using AI for first-pass review to reduce the human review burden to architectural and business logic concerns.

Architectural Taste Over Implementation Speed

AI tools are fastest at implementation: given a clear specification, produce working code. They are weakest at architecture: given ambiguous requirements, design a system that will scale, maintain, and adapt to changing needs.

This inverts the traditional leverage model. Speed of implementation used to be a competitive advantage. Now it is commoditized. What remains scarce is the ability to make good architectural decisions: which patterns to use, which abstractions to create, which boundaries to draw, and which trade-offs to make.

Engineering leaders who grew up valuing speed of execution need to recalibrate toward valuing quality of decisions. The leaders who can evaluate an AI-generated implementation and say “this works but the abstraction is wrong for where we are headed” are more valuable than those who can implement it faster by hand.

Evaluating AI-Generated Code

The most important technical skill on engineering teams is increasingly the ability to evaluate AI-generated code. This is different from reading code written by a colleague. AI-generated code has specific failure modes:

  • It often looks correct at a surface level while encoding subtle misunderstandings of the domain. The code compiles, the tests pass, but the business logic is slightly wrong in ways that only someone with deep domain knowledge would catch.
  • It tends toward over-engineering. AI models have seen many codebases and will introduce patterns, abstractions, and layers that are not justified by the current requirements.
  • It hallucinates dependencies and APIs. The code references a library method that does not exist or passes arguments in the wrong order for a function signature the model has confused with a similar one.
  • It lacks context about team conventions. The generated code may be correct but inconsistent with the patterns the team has established, creating maintenance burden.

Training engineers to spot these failure modes efficiently is a leadership investment with immediate payoff. The teams that develop strong AI code evaluation skills move faster than those who blindly accept or blanket reject AI output.

Team Structure Implications

The long-term team structure implications are still emerging, but some patterns are already visible:

  • Senior-to-junior ratios will shift. If AI handles much of the implementation work that juniors used to do, teams need fewer juniors and more seniors for review, architecture, and mentorship.
  • Technical leads become more important. The role of someone who can evaluate AI output, guide architecture, and ensure quality becomes the critical scaling mechanism.
  • Specialization matters more. AI is good at general-purpose code. Humans add value through domain expertise, system knowledge, and organizational context that AI lacks.
  • On-call and incident response skills become more differentiating. AI cannot debug a production outage at 3 AM with partial information and time pressure. This remains a purely human skill.

The Leadership Adjustment

The point is judgment, not novelty.

Earlier in this story I was mostly trying to survive the blast radius myself. Here I was trying to design a system where the team did not need heroics in the first place. The same philosophy now shapes jarvis, alfred, and the portfolio RAG stack.

The leaders who treat AI as a junior engineer on every desk will outperform those who treat it as a threat. AI does not replace engineering leadership. It amplifies the gap between good leadership and bad leadership. Good leaders leverage it. Bad leaders are exposed by it.

The engineering leaders who will thrive in the AI era are those who shift their attention from “how do we write more code” to “how do we make better decisions about code.” The code is the easy part now. The decisions behind the code were always the hard part. AI just made that truth impossible to ignore.