News Daily Nation Digital News & Media Platform

collapse
Home / Daily News Analysis / The question AI providers hope VPs of Engineering never ask

The question AI providers hope VPs of Engineering never ask

Apr 21, 2026  Twila Rosenbaum  10 views
The question AI providers hope VPs of Engineering never ask

The adoption of AI coding tools is surging, yet many engineering leaders remain focused on usage metrics rather than actual outcomes. This oversight creates a significant blind spot that could lead to unnecessary expenses. A critical question that remains unaddressed in the AI industry concerns the effectiveness of the code generated by these AI agents: how much of it actually makes it to production?

It is not about the volume of generated code or the number of prompts executed. The essential metric is how much of that code survives the review process, passes continuous integration (CI), gets merged, deployed, and ultimately reaches the end user. Unfortunately, many engineering leaders are unable to answer this question, and AI providers lack the incentive to assist them in finding the answer.

Significant Spending with Limited Visibility

According to recent data from the Stanford AI Spend Index, the average expenditure on AI coding tools is $86 per developer each month, across a sample of 140 companies and over 113,000 developers. The top quartile of spenders exceeds $195, with some organizations allocating over $28,000 per developer monthly. Notably, Anthropic has recently reported an annualized revenue exceeding $30 billion, up from $9 billion just four months prior. Furthermore, SemiAnalysis indicates that 4% of all public GitHub commits are now attributed to Claude Code, a figure expected to exceed 20% by the year's end.

Within Linear’s enterprise workspaces, over 75% have implemented AI coding agents. While the financial flow into these tools is substantial, the tracking of code that actually ships remains inadequate.

The Incentive Misalignment

AI providers typically charge based on token consumption. This means that the more tokens used by engineers, the more revenue is generated for the provider. Providers earn revenue when a token is consumed, not when the generated code passes review, gets merged, or is successfully deployed. This creates a misalignment in incentives; a developer who prompts an AI tool multiple times to generate a function that ultimately requires human intervention incurs significantly higher costs than a developer who achieves the desired outcome on the first attempt.

Currently, many engineering leaders cannot differentiate between these scenarios. They often only see a single line item on the AI invoice, oblivious to which tokens contributed to production code and which resulted in inefficiencies. This is not a conspiracy, but rather a structural challenge that the VP of Engineering must address, as the providers have no motivation to resolve it on their behalf.

A Familiar Pattern Emerges

Historically, the initial phase of cloud computing saw companies rapidly adopting services like AWS and Azure, leading to excessive expenditure without proper measurement of usage. It took years for organizations to realize that they were overspending by 30 to 40 percent on cloud infrastructure due to a lack of monitoring. The current trajectory of AI spending mirrors this pattern, only with a faster growth rate and a broader measurement gap.

Cloud providers eventually had to implement cost optimization tools in response to customer demand, and a similar shift is anticipated in the AI sector. Engineering leaders who prioritize measurement will gain a competitive advantage, optimize spending, negotiate better contracts, and discern which tools are truly beneficial. Those who neglect this will continue to incur costs, hoping that the outputs justify their investments.

Essential Measurement Metrics

The current landscape is saturated with dashboards displaying adoption metrics and seat utilization, but what is truly lacking is the ability to trace AI-generated code from inception to deployment. This entails commit-level attribution that identifies the AI agent responsible for code creation, the ratio of AI-generated to human-edited code, and whether the code successfully passed through the review and deployment stages.

By linking AI expenditure to tangible production outcomes, organizations can answer critical questions about their AI investments. They can ascertain which teams effectively leverage AI agents and which merely consume tokens without yielding results. Understanding whether rising AI costs stem from effective adoption or expensive failures will be crucial.

At Waydev, we have dedicated the past year to enhancing our measurement capabilities, leveraging nearly a decade of experience in assessing engineering behavior for major corporations. With AI altering the inputs, we have adapted our measurement framework accordingly.

This new platform enables tracking of AI adoption, impact, and return on investment (ROI) across the entire software development lifecycle, bridging the gap between AI spending and production outcomes.

Distinguishing Usage from Value

The AI sector has led engineering leaders to believe that increased usage equates to greater value. However, usage does not necessarily translate to impact. A team generating 10,000 lines of AI code weekly but only shipping 2,000 to production is not outperforming a team that produces 3,000 lines and successfully ships 2,500. Yet, adoption dashboards often paint a misleading picture, favoring the team with higher generation numbers.

This misconception represents a significant blind spot that is becoming increasingly costly. The era of unchecked AI expenditure is drawing to a close. Engineering leaders who proactively establish measurement frameworks will dominate discussions regarding AI ROI for years to come, while those who delay will find themselves justifying expenses they do not fully comprehend.


Source: TNW | Insider News


Share:

Your experience on this site will be improved by allowing cookies Cookie Policy