8 minute read

A leaked internal memo from Red Hat’s CTO and SVP of Engineering, dated March 31, 2026, is making the rounds. The headline: all of Global Engineering is required to adopt an “Agentic Software Development Lifecycle.” Engineers must use AI tooling. Workflows will be measured by objective metrics: cycle time, defect rate. Roles are “evolving.”

The Register’s headline said it more plainly: “Sounds like an excellent time to start honing your Debian skills.”

The same week: Oracle cut 30,000 employees, 18% of its workforce, while posting $6 billion quarterly profit. The stated reason is funding $156 billion in AI infrastructure. Meanwhile, Anthropic quietly acknowledged that enterprise users of Claude Code are burning through token quotas “way faster than expected,” causing widespread outages in automated workflows.

These three stories are not separate events. They are the same event, seen from different angles.

The agentic software development lifecycle is no longer a slide in a conference deck. It is being operationalized, often badly, right now.

What “Agentic SDLC” Actually Means

Let’s be precise, because the term gets used loosely.

The traditional SDLC is a human-driven sequence: requirements, design, implementation, test, deploy, operate. AI tools in the last two years have augmented individual steps: copilot autocomplete, AI-assisted code review, LLM-generated test cases.

Agentic SDLC is different in kind, not just degree. An agent doesn’t assist a human writing a function. It takes a specification, plans a sequence of actions, writes code, runs tests, reads the failures, revises the implementation, and iterates autonomously, in a loop, often across hundreds of API calls per task.

The difference shows up in the numbers. Anthropic’s Claude Code customers ran over quota faster than anyone expected because agentic tasks don’t consume tokens like chat sessions. A single coding task might trigger 200-500 LLM calls. The token economics of agentic work are an order of magnitude different from the chat-mode assumptions baked into most AI contracts and infrastructure plans.

This is why the capacity crunch surprised Anthropic. It will surprise every engineering org that has priced or provisioned AI services based on chat-mode baselines.

The Problem With the Mandate Approach

Red Hat’s memo-mandate approach to this transition is understandable. It’s also likely to fail, or at minimum, produce compliance without competence.

Here’s why.

Mandating tooling adoption without restructuring workflows produces engineers who use AI tools the same way they used non-AI tools. They write a prompt like a function signature, accept the first output that compiles, and ship code they haven’t actually internalized. The productivity gain is real but shallow. The architectural decisions, the edge case handling, the operational assumptions: those are still being made by engineers who may or may not have developed the judgment to evaluate AI output critically.

Worse: it creates a new class of technical debt that’s harder to see. Code written by an agent and reviewed at the output level by a human who’s optimized for throughput is code that works today and surprises you in six months.

The more dangerous failure mode is organizational. When you mandate AI adoption and tie it to objective metrics, you incentivize engineers to optimize the metrics rather than the underlying work. Cycle time goes down. Defect rate, measured on a 30-day lag, looks fine. What you’ve actually done is moved risk downstream, into production, into customer experience, into the next major incident.

What the Transition Actually Requires

I’ve led engineering organizations through technology transitions: from on-premise to cloud-native, from monolith to microservices, from human-review-only to automated CI/CD gates. The pattern of what works is consistent.

The unit of change is the workflow, not the tool.

Engineers don’t need to be told to use Claude Code. They need their development workflow redesigned around what agentic tools are good at and where they break down. That’s a different conversation, and it requires someone who understands both the tools and the existing workflow well enough to know where the leverage points are.

For agentic coding tools, this means rethinking the specification layer. An agent doesn’t work well with vague requirements. It works well with precise inputs: acceptance criteria, edge cases, expected behaviors, test fixtures. If your team’s requirements are “build a user profile page,” the agent will hallucinate the rest and you’ll spend two hours correcting a misunderstanding that would have taken five minutes to clarify upfront.

This is not a coincidence. The discipline of writing specifications that are precise enough for an agent to act on is the same discipline that makes software engineering predictable at scale. The teams that do well with agentic tools are, almost without exception, the teams that already write good acceptance criteria, maintain strong test coverage, and treat the specification as a first-class artifact, not an afterthought before the sprint starts.

The transition requires new governance, not just new tools.

When a human writes code, the review process catches most of the important issues. The reviewer can ask “why did you do it this way?” and get a reasoned answer. They can infer intent from the structure of the code.

When an agent writes code, the review process has to be redesigned. You can’t ask the agent why it made a decision, at least not in a way that gives you reliable signal about the underlying reasoning. You’re reviewing output, not process. That requires different checklists, different review focus, and different mental models for what “looks right” means.

The engineering orgs that figure this out first will have a durable advantage. The ones that bolt AI tooling onto existing review processes and call it transformation will be slower and more fragile than the ones that didn’t adopt AI at all.

Measurement needs to evolve simultaneously.

Red Hat is right that objective metrics matter. They’re wrong, or at least incomplete, about which metrics. Cycle time and defect rate are lagging indicators that can be gamed. The leading indicators for a healthy agentic engineering org look more like: specification quality scores, test coverage at specification time (before agent invocation), agent task success rate on first pass, and rework rate on agent-generated code.

These metrics don’t exist as standard dashboard items in Jira or Linear. Building them requires intentional instrumentation and a clear model of what the agentic workflow actually looks like in your context.

The Org Design Question Nobody Is Asking

Most of the conversation about agentic AI and engineering orgs is framed as a staffing question: will we need fewer engineers? The answer is probably yes, in aggregate and over time, for some categories of work. But that’s not the useful question for the next 18 months.

The useful question is: what does your engineering org’s structure look like when agents handle a significant portion of implementation work?

The answer is not “the same structure, but with fewer people.” It’s a different structure. The ratio of people who can write precise specifications to people who write code shifts dramatically. The value of someone who can define acceptance criteria, understand edge cases, and evaluate agent output critically, without writing the implementation themselves, goes up significantly. The value of someone who writes fast, doesn’t ask questions, and ships features goes down relative to its current premium.

This is a different org design problem. It requires thinking about career paths, hiring profiles, and team structures that most engineering leaders haven’t had to consider before. The teams being built today that will thrive in an agentic workflow are not the teams that look like 2024’s high-performing engineering orgs.

What to Do Now

If you’re an engineering leader trying to navigate this transition, a few concrete things that actually matter:

Invest in your specification layer before your agent layer. The biggest leverage point in an agentic workflow is the quality of the input to the agent, not the sophistication of the agent itself. Teams that develop a disciplined approach to writing specifications, ones that are rigorous enough to act on without human clarification, will get dramatically better results from AI coding tools than teams that don’t, regardless of which model they use.

Audit your token economics. If you have AI tooling in production and you haven’t modeled what agentic workloads do to your token consumption, you are flying blind. Claude Code’s quota exhaustion issue is not unique to Anthropic. It will hit any team running multi-step agentic workflows at scale. Build the consumption model now.

Redesign review, don’t inherit it. Your current code review process was designed for human-written code. It needs to be redesigned for agent-generated code. The checklist is different. The focus is different. The time allocation is different. This is one of the highest-leverage investments you can make in the next 90 days.

Change management is the bottleneck, not the technology. The engineers most resistant to agentic tooling are often your best ones, because they have the most internalized judgment about how software should be built. Losing their buy-in is catastrophic. The transition has to be framed as “we’re raising the floor and changing what good engineering judgment means,” not “we’re automating your job.” Those are different conversations and they land very differently.

The Red Hat memo will probably accomplish something. Engineers will use more AI tooling. Metrics will move. Some of the movement will be real.

But the engineering organizations that come out of this transition with durable advantages won’t be the ones that mandated adoption fastest. They’ll be the ones that understood what actually changes when agents become part of the development loop, and redesigned their workflows, their governance, and their org structures accordingly.

That’s a harder problem than picking the right AI coding tool. It’s also the one worth solving.

I’m building ShipKode Arc, a governed agentic control plane that structures the journey from customer signal to engineering-ready specification. The specification layer is where most agentic SDLC transitions break down first. If you’re thinking through what this transition looks like for your organization, I’m happy to compare notes.