AI News‘Tokenmaxxing’ is making developers less productive than they think

‘Tokenmaxxing’ is making developers less productive than they think

2:01 AM IST · April 18, 2026

‘Tokenmaxxing’ is making developers less productive than they think

There’s an old saw in management: What you measure matters. And, typically, you get more of whatever you’re measuring. Software engineers have debated productivity metrics for decades, starting with lines of code. But as the new generation of AI coding agents delivers more code than ever, what their managers ought to be measuring is less clear. Enormous token budgets — essentially, the amount of AI processing power a developer is authorized to consume — have become a badge of honor among Silicon Valley developers, but that’s a very weird way to think about productivity. Measuring an input to the process makes little sense when you presumably care more about the output. It might make sense if you’re trying to encourage more AI adoption (or selling tokens), but not if you’re trying to become more efficient. Consider the evidence from a new class of companies operating in the “developer productivity insight” space. They’re finding that developers using tools like Claude Code, Cursor, and Codex generate a lot more accepted code than they did before. But they also find that engineers have to return to revise that accepted code far more often than before, undercutting claims of increased productivity. Alex Circei, the CEO and founder ofWaydev, is building an intelligence layer to track these dynamics; his firm works with 50 different customers that employ more than 10,000 software engineers. (Circei has contributed to TechCrunch in the past, but this reporter had never met him before.) He says that engineering managers are seeing code acceptance rates of 80% to 90% — meaning the share of AI-generated code that developers approve and keep — but they’re missing the churn that happens when engineers have to revise that code in the following weeks, which drives the real-world acceptance rate down between 10% and 30% of generated code. The rise of AI coding tools led Waydev, founded in 2017 to provide developer analytics, to totally rework its platform in the last six months to address the proliferation of rapid coding tools. Now, the company is releasing new tools that track the metadata generated by AI agents, offering analytics on the quality and cost of their code to provide engineering managers with more insight into both AI adoption and efficacy. While analytics companies have an incentive to highlight the problems they find, the evidence is mounting that large organizations are still figuring out how to use AI tools efficiently. Major companies are noticing — Atlassian acquired DX, another engineering intelligence startup, for $1 billion last year, to help its customers understand the return on investment on coding agents. The data from across the industry tells a consistent story: More code is being written, but a disproportionate amount of it isn’t sticking. GitClear, another company in this space,published a reportin January that found AI tools increased productivity, but also that its data showed “regular AI users averaged 9.4x higher code churn than their non-AI counterparts” — more than double the productivity gains the tools provided. Faros AI, an engineering analytics platform, drew on two years of customer data for itsMarch 2026 report. The finding: code churn — lines of code deleted versus lines added — had increased 861% under high AI adoption. Jellyfish, which bills itself as an intelligence platform for AI-integrated engineering,collected dataon 7,548 engineers in the first quarter of 2026. The firm found that the engineers with the largest token budgets produced the most pull requests (proposed changes to a shared codebase), but the productivity improvement didn’t scale. They achieved two times the throughput at 10 times the cost of tokens. In other words, the tools are generating volume, not value. These kinds of statistics ring true when you talk to developers, who are finding that code review and technical debt are stacking up, even as they revel in the freedom of the new tools. One common finding is the difference between senior and junior engineers, with the latter accepting far more AI-generated code, and dealing with a larger amount of rewriting as a consequence. Still, even as developers work to understand exactly what their agents are up to, they don’t anticipate turning back anytime soon. “This is a new era of software development, and you have to adapt, and you are forced to adapt as a company,” Circei told TechCrunch. “It’s not like it will be a cycle that will pass.”

read more

Latest AI News

View All News →
Are we tokenmaxxing our way to nowhere?

Are we tokenmaxxing our way to nowhere?

Loading the player… The gap between AI insiders and everyone elseis widening, and the spending, suspicion, and even new vocabulary are starting to show it. While OpenAI is busy buying up everything fromfinance appstotalk shows, a certain shoe company justrebranded as an AI infrastructure play, and Anthropic unveiled a model it says istoo powerful to release publicly …but apparently not too powerful to demo to Federal Reserve Chair Jerome Powell. Watch as hosts Kirsten Korosec, Anthony Ha, and Sean O’Kane dig into what’s actually being built in AI infrastructure, who’s winning the enterprise battle between OpenAI and Anthropic, and more of the week’s headlines on this episode of TechCrunch’sEquitypodcast. Subscribe to Equity onYouTube,Apple Podcasts,Overcast,Spotifyand all the casts. You also can follow Equity onXandThreads, at @EquityPod.

3 hours ago

View

‘Tokenmaxxing’ is making developers less productive than they think

‘Tokenmaxxing’ is making developers less productive than they think

There’s an old saw in management: What you measure matters. And, typically, you get more of whatever you’re measuring. Software engineers have debated productivity metrics for decades, starting with lines of code. But as the new generation of AI coding agents delivers more code than ever, what their managers ought to be measuring is less clear. Enormous token budgets — essentially, the amount of AI processing power a developer is authorized to consume — have become a badge of honor among Silicon Valley developers, but that’s a very weird way to think about productivity. Measuring an input to the process makes little sense when you presumably care more about the output. It might make sense if you’re trying to encourage more AI adoption (or selling tokens), but not if you’re trying to become more efficient. Consider the evidence from a new class of companies operating in the “developer productivity insight” space. They’re finding that developers using tools like Claude Code, Cursor, and Codex generate a lot more accepted code than they did before. But they also find that engineers have to return to revise that accepted code far more often than before, undercutting claims of increased productivity. Alex Circei, the CEO and founder ofWaydev, is building an intelligence layer to track these dynamics; his firm works with 50 different customers that employ more than 10,000 software engineers. (Circei has contributed to TechCrunch in the past, but this reporter had never met him before.) He says that engineering managers are seeing code acceptance rates of 80% to 90% — meaning the share of AI-generated code that developers approve and keep — but they’re missing the churn that happens when engineers have to revise that code in the following weeks, which drives the real-world acceptance rate down between 10% and 30% of generated code. The rise of AI coding tools led Waydev, founded in 2017 to provide developer analytics, to totally rework its platform in the last six months to address the proliferation of rapid coding tools. Now, the company is releasing new tools that track the metadata generated by AI agents, offering analytics on the quality and cost of their code to provide engineering managers with more insight into both AI adoption and efficacy. While analytics companies have an incentive to highlight the problems they find, the evidence is mounting that large organizations are still figuring out how to use AI tools efficiently. Major companies are noticing — Atlassian acquired DX, another engineering intelligence startup, for $1 billion last year, to help its customers understand the return on investment on coding agents. The data from across the industry tells a consistent story: More code is being written, but a disproportionate amount of it isn’t sticking. GitClear, another company in this space,published a reportin January that found AI tools increased productivity, but also that its data showed “regular AI users averaged 9.4x higher code churn than their non-AI counterparts” — more than double the productivity gains the tools provided. Faros AI, an engineering analytics platform, drew on two years of customer data for itsMarch 2026 report. The finding: code churn — lines of code deleted versus lines added — had increased 861% under high AI adoption. Jellyfish, which bills itself as an intelligence platform for AI-integrated engineering,collected dataon 7,548 engineers in the first quarter of 2026. The firm found that the engineers with the largest token budgets produced the most pull requests (proposed changes to a shared codebase), but the productivity improvement didn’t scale. They achieved two times the throughput at 10 times the cost of tokens. In other words, the tools are generating volume, not value. These kinds of statistics ring true when you talk to developers, who are finding that code review and technical debt are stacking up, even as they revel in the freedom of the new tools. One common finding is the difference between senior and junior engineers, with the latter accepting far more AI-generated code, and dealing with a larger amount of rewriting as a consequence. Still, even as developers work to understand exactly what their agents are up to, they don’t anticipate turning back anytime soon. “This is a new era of software development, and you have to adapt, and you are forced to adapt as a company,” Circei told TechCrunch. “It’s not like it will be a cycle that will pass.”

3 hours ago

View

Sources: Cursor in talks to raise $2B+ at $50B valuation as enterprise growth surges

Sources: Cursor in talks to raise $2B+ at $50B valuation as enterprise growth surges

AI coding startup Cursor is nearing new funding in which the four-year-old company would raise at least $2 billion in fresh capital, according to four sources familiar with the matter. Returning investors Thrive and Andreessen Horowitz are expected to lead the financing at a $50 billion valuation, prior to the new capital injection, the people said. Battery Ventures, a new investor, may also participate in the financing, according to two sources. Strategic investor Nvidia is also expected to write a check, one person said. Although the round is already oversubscribed, the deal terms are not final and may still change. The financing, if completed, would nearly double Cursor’s previous$29.3 billion post-money valuation, assigned to the company during its last fundraise six months ago. Despite fierce competition from other AI-coding offerings, such as Anthropic’s Claude Code and OpenAI’s revamped Codex, Cursor’s revenue continues to climb rapidly. Cursor forecasts ending 2026 with an annualized revenue run rate of more than $6 billion, two people said. This trajectory implies the company expects to at least triple its annualized revenue over the next 10 months. In February, Cursor reached $2 billion in annualized revenue, calculated by projecting its most recent monthly sales over a year,Bloomberg reported. Like many AI-coding startups reliant on third-party models, Cursor operatedatnegative gross marginsuntil recently, meaningit cost more to run the product than the startup could charge for it. The introduction of a proprietaryComposer modellast November, along with the ability to call on less expensive models like China’s Kimi, has helped the company achieve slight gross margin profitability, the people said. On a more granular level, the company has reached positive gross margins on its sales to large enterprises, but continues to lose money on individual developer accounts, according to one person. By relying less on outside providers, Cursor is trying to avoid being replaced by its own suppliers, most notably Anthropic, whose Claude Code has emerged as the startup’s main rival. Cursor and Battery Ventures declined comment. Thrive, a16z, and Nvidia didn’t respond to request for comment. Cursor, previously known as Anysphere, was co-founded in 2022 by Michael Truell, Sualeh Asif, Arvid Lunnemark, and Aman Sanger while they were students at MIT.

3 hours ago

View

Anthropic launches Claude Design, a new product for creating quick visuals

Anthropic launches Claude Design, a new product for creating quick visuals

Anthropicannouncedon Friday that it’s launching Claude Design, a new experimental product that lets users create visuals like prototypes, slides, one-pagers, and more using Claude. The company says Claude Design is intended to help people like founders and product managers without a design background share their ideas more easily. With Claude Design, users describe what they want, and Claude will create an initial version. From there, users can refine the visuals with direct edits or requests. For example, you could ask Claude to “prototype a serene mobile meditation app. It should have calming typography, subtle nature-inspired colors, and a clean layout.” You could then tweak the colors, the size of the typography, or ask Claude to add a dark mode toggle. While Claude Design may initially seem like it’s looking to compete with popular design app Canva, whichhas just expanded its own AI capabilities,Anthropic told TechCrunch that it’s intended to complement it rather than replace it. The company said its new product is built for people who aren’t starting from a design tool and need to get from an idea to something visual quickly. Once teams create presentation decks or prototypes, they can export them as PDFs, URLs, PPTX files, or send them to Canva. Once in Canva, they are fully editable and collaborative, Anthropic says. Claude Design can also apply a team’s design system to every project it creates so that the results are consistent with the company’s overall visual style. Anthropic says Claude Design is able to do this by reading a company’s codebase and design files. Additionally, teams can refine these components and maintain more than one design system. The new product is powered by Claude Opus 4.7 and is available in research preview for Claude Pro, Max, Team, and Enterprise subscribers. The launch highlights Anthropic’s ongoing push into the enterprise and prosumer categories, as competition intensifies around AI workplace tools. In January, Anthropic rolled outClaude Cowork, an agentic assistant built for complex tasks. A few weeks later, the company broughtagentic plug-ins to Coworkthat are designed to automate specialized tasks within a company’s various departments. Today’s announcement comes a few days afterBloombergreported that VCs have been offering the company a preemptive funding round that wouldvalue it at $800 billion or more, which would almost match or even surpass its rival OpenAI. But so far, Anthropic isn’t interested in the latest offers, according to the report.

7 hours ago

View