Does Anthropic Claude’s Enterprise, the first to move to usage-based pricing, really save money with AI employees?

ChainNewsAbmedia

When business owners balance the idea of “using AI to replace human labor to cut costs,” Anthropic has changed the rules of the game. This AI giant recently updated the billing structure for its Claude Enterprise offering, splitting the usage of Claude, Claude Code, and Cowork out of the previous $40 per month subscription fee and switching to separate charges based on the actual number of Tokens consumed. Now, the cost of AI employees doesn’t seem nearly as cheap as outsiders are claiming.

(Can using Classical Chinese and AI chat save Tokens? A screenshot sparks debate—engineers say the truth is that using English is the way to go)

The era of flat-rate pricing is over—Claude Enterprise billing has been revised: pay for what you use

The Information reports that Anthropic’s updated enterprise documentation states: “The monthly seat fee (seats) covers only platform access and does not include any usage; all usage is billed separately according to standard API rates.” What companies used to buy as “unlimited access” has now been changed to “a per-use model.”

Under the old plan, subscription fees per company account were about $40 to $200 per month, along with a 10% to 15% API discount. While the new plan lowers subscription fees to $20 per month, it simultaneously cancels all API discounts and requires enterprises to commit in advance and prepay estimated monthly Token usage. No matter whether actual usage is more or less, the committed amount is still paid, and committing to a higher volume does not result in a lower unit price.

This structure is predictable annual recurring revenue for Anthropic, but for enterprises it amounts to shifting usage costs and risks.

“Scarce compute resources” is the real trigger behind the revised pricing

Anthropic calls this adjustment “product optimization,” but the underlying driver is persistently high compute costs. Even if Anthropic’s annualized revenue jumped from $9 billion to $30 billion in just four months, what users are getting is not discounts—it’s a restructuring of its revenue model.

At the core is how AI agents (AI Agent) consume resources. Traditional chat usage is like “small sips,” but an agent workflow that includes multi-step task chaining, repeated execution, and even multi-agent collaboration is more like “a big drink.”

The supply side is tight as well. Blackwell GPU leasing prices rose 48% within two months; CoreWeave increased prices by more than 20% since late last year; and a forecast from U.S. banks predicts that compute demand contraction will continue through 2029. The income brought by fixed-rate pricing has long been unable to carry the burden for Anthropic.

Unstable service is the most real warning light for enterprise customers

In addition, service stability is another major problem. Retool founder David Hsu said in an interview via The Wall Street Journal that even though Claude Opus 4.6 performs better than OpenAI, he ultimately moved his workflows to the latter. The reason is that Claude’s service frequently goes down, which often prevents him from delivering code on time.

Within the 90 days ending April 8 of this year, the Anthropic API achieved only 98.95% uptime, far below the industry standard of 99.99%. Hsu’s decision illustrates one thing: when it comes to choosing between service reliability and model capability, enterprises need AI that is stable.

The real cost of AI employees is far more complex than what’s on the bill

Today, the traditional AI pricing model of “monthly subscription fees” is already gone. Total cost is now recalculated based on actual Token usage. Negotiating usage discounts or flexible adjustment clauses in the contract, or proactively controlling spending by optimizing prompts, batch processing, and caching strategies, has become a new challenge for enterprises moving toward AI adoption and transformation.

A few days ago, OpenAI also announced that Codex would be switched to Token-based billing. GitHub tightened Copilot usage limits on April 10, and Windsurf replaced the points-based system with daily quotas. The entire AI industry is simultaneously signaling the end of the fixed-rate era.

Before companies evaluate how much manpower cost they can save by “deploying AI,” they may also need to test whether users can produce consistent, high-quality output within constrained budgets.

This article, about Anthropic Claude Enterprise being the first to shift to usage-based billing—does it really save money with AI employees? It first appeared on Chain News ABMedia.

Disclaimer: The information on this page may come from third parties and does not represent the views or opinions of Gate. The content displayed on this page is for reference only and does not constitute any financial, investment, or legal advice. Gate does not guarantee the accuracy or completeness of the information and shall not be liable for any losses arising from the use of this information. Virtual asset investments carry high risks and are subject to significant price volatility. You may lose all of your invested principal. Please fully understand the relevant risks and make prudent decisions based on your own financial situation and risk tolerance. For details, please refer to Disclaimer.

Related Articles

DeepSeek V4-Pro API Gets 75% Discount Until May 5, Output Price Drops to $0.87 Per Million Tokens

Gate News message, April 26 — DeepSeek announced a limited-time 75% discount on V4-Pro API pricing, valid until May 5 at 15:59 UTC. After the discount, pricing per million tokens is: input cache hit $0.03625

GateNews2m ago

Anthropic Deploys Election Safeguards for Claude Ahead of 2026 Midterms

Anthropic announced Friday a set of election integrity measures designed to prevent its Claude AI chatbot from being weaponized to spread misinformation or manipulate voters ahead of the 2026 U.S. midterm elections and other major contests around the world this year. The San Francisco-based

CryptoFrontier5h ago

DeepRoute.ai Advanced Driver Assistance System breakthrough: over 300k vehicles deployed. 2026 target: 1 million City NOA fleet.

DeepRoute.ai announced that its advanced driver-assistance system has been deployed in China for a cumulative total of more than 300k vehicles. In the past year, it helped avoid more than 180k potential incidents. Its 2026 goal is for its city NOA vehicle fleet to reach 1 million vehicles, with utilization exceeding 50%, and it is seen as a key step toward large-scale commercial deployment of Robotaxis. This move shows that autonomous driving in China has entered routine usage, while also creating a divergence from the United States’ vertical integration pathway, affecting the timing of the Asia-Pacific supply chain.

ChainNewsAbmedia6h ago

DeepSeek Releases V4-Pro and V4-Flash Models at 98% Lower Cost Than OpenAI's GPT-5.5 Pro

Gate News message, April 25 — DeepSeek released preview versions of V4-Pro and V4-Flash on April 24, both open-weight models with one million token context windows. V4-Pro features 1.6 trillion total parameters but activates only 49 billion per inference pass using a Mixture-of-Experts architecture.

GateNews11h ago

Judge Dismisses Fraud Claims in Elon Musk's OpenAI Lawsuit; Case Advances to Trial with Two Remaining Allegations

Gate News message, April 24 — A federal judge has dismissed fraud claims from Elon Musk's lawsuit against OpenAI, Sam Altman, Greg Brockman, and Microsoft, clearing the way for the case to proceed to trial on two remaining allegations: breach of charitable trust and unjust enrichment. U.S.

GateNews15h ago

OpenAI CEO Sam Altman Apologizes for Failing to Report School Shooter's Banned Account to Police

Gate News message, April 25 — OpenAI Chief Executive Officer Sam Altman apologized to the Tamborine community in Canada for the company's failure to notify police about a banned account linked to Jesse Van Rootselaar, who killed eight people at a school in February before taking his own life. OpenAI

GateNews15h ago
Comment
0/400
No comments