2 Sources
2 Sources
[1]
Tokens may soon drive the AI economy
A new economic reality is starting to take hold in AI. It already underpins the industry's giant data centres and it will one day become an iron rule for all companies that use machine-generated intelligence. That, at least, is according to Jensen Huang, chief executive of Nvidia, who promoted the idea heavily at his company's main annual tech event this week. His theory helps to make a case for Nvidia's continued dominance in chips. But it also reveals how far the industry has to go to make a wider case for the technology. Huang's take on AI economics is based around the production, consumption and monetisation of tokens. These are the most basic units of output from large language models: it takes about 1,300 tokens to generate 1,000 words of text. The key metric, he argues, is the cost per token of output. And as the main input into AI-powered services, he adds, tokens translate directly into revenue. It is not hard to see why the Nvidia boss wants a nervous Wall Street to focus on token economics. Forget the gargantuan capital spending or the fact that so many competitors are lining up to eat into Nvidia's fat profit margins, he seems to be saying: as long as his company's chips keep pumping out tokens at the lowest cost and as long as demand for tokens continues to far outstrip supply, then all is well with the AI boom. As a theory of Nvidia's continued pre-eminence, it sounds compelling. But if token economics is ever to rule the AI world in the way that Huang predicts, some important gaps need to be filled in. One is the lack of a clear link between the production of tokens and the creation of value for customers. Just because the cost of tokens is falling doesn't mean the services created with AI suddenly become valuable or that this will automatically generate revenue across the industry, as Huang suggests. Complicating this picture is the fact that newer AI models consume far larger numbers of tokens. The "reasoning" models that emerged late in 2024, starting with OpenAI's o1, perform far more work to arrive at an answer. These are now being supplemented by agents, which promise to automate white-collar work and bring an explosion in token use -- and, by extension, hefty bills for companies that give workers unlimited use of AI. Nvidia and the rest of the AI industry have barely scratched the surface when it comes to showing how this will translate into revenue for their customers. In software engineering, which has seen the first widespread use of AI agents, there have been efforts to measure how token use is linked to output and to use this to apportion tokens to workers. Eventually, tech companies dream of AI becoming a core part of employment, with the cost of all white-collar workers coming to be seen as a salary plus a certain number of tokens per month. For now, that is still only a pipe dream. The second significant piece missing from Huang's narrative about an emerging token economy is how the companies that produce tokens, the raw commodity on which all of this depends, will make profits. If these "AI factories" all use Nvidia's latest chips, then it may be hard for any of them to gain a cost-per-token advantage or retain any pricing power. The big price declines that have accompanied the plunging cost of producing tokens seem to bear this out. When OpenAI launched GPT-4 two years ago, for instance, it charged $33 for 1mn tokens. Today, it charges only 9 cents for 1mn tokens produced by its cheapest model. That may be great for customers, but it has fed worries about commoditisation. Such worries are hardly new. It is the same argument that was heard in the early days of cloud computing, when Amazon Web Services charged for access to basic data storage and computing power. How could cloud companies ever make a decent profit if computing services were stripped back and sold as commodities like this? The answer was that these were only the first components of what became higher-value services -- full-scale computing platforms on which customers could run their businesses. Whether OpenAI and Anthropic will be able to work a similar trick is unclear, but the opportunity before them is clear. There may be other explanations for the healthy profit margins in cloud computing. The business is ruled by a small oligopoly. Cloud companies have also faced pressure from regulators to reduce switching costs that may help to pad their profits. For now, there is no shortage of competition among frontier AI companies. How that shakes out in future will go a long way to shaping the industry's profits.
[2]
AI Adoption Is Being Measured in Tokens, but the Metric Falls Short, Experts Say | PYMNTS.com
By completing this form, you agree to receive marketing communications from PYMNTS and to the sharing of your information with our sponsor, if applicable, in accordance with our Privacy Policy and Terms and Conditions. Tokens are the foundational unit through which AI models process all information. Tokens are tiny units of data that result from breaking larger chunks of information into smaller pieces. AI models process tokens to learn relationships between them and unlock capabilities such as prediction, generation and reasoning. For large language models, short words may be represented by a single token, while longer words may be split into two or more tokens. The word "darkness," for example, would be split into two tokens, "dark" and "ness," with each token bearing a numerical representation as explained by Nvidia. Every prompt a worker sends to an AI system and every response the system returns are measured and often billed in tokens. Every prompt and response consumes tokens and incurs charges. That direct relationship between usage and cost is what makes tokens attractive as a management tool. Unlike the seat-based pricing that defined earlier generations of enterprise software, token consumption is granular, real-time and tied directly to behavior. The shift from seat counts to token consumption mirrors how enterprise AI spending itself has changed. While the unit price of AI tokens is falling, overall enterprise spending on and scaling of AI systems is rising. The number of users, complexity of models, and intensity of workloads will likely drive greater token consumption and, consequently, higher costs. OpenAI's own data on its enterprise customer base illustrates how dramatically usage patterns have shifted. Average reasoning token consumption per organization has increased by approximately 320 times in the past 12 months, suggesting that more intelligent models are being systematically integrated into expanding products and services. That figure has become a headline metric in the company's internal reporting on adoption progress. Nvidia CEO Jensen Huang went further at the company's GTC conference this week, framing tokens as a new form of corporate currency. "I could totally imagine in the future every single engineer in our company will need an annual token budget," Huang said, estimating that employee token allocations could reach half of base salary in value. The appeal of token metrics runs into a fundamental problem: tokens measure volume, not outcome. Generating through packaged software abstracts tokens almost entirely, while consuming through APIs makes tokens explicit, but this can bring transparency and also volatility, as costs rise based on workload design, prompt length, and hidden choices of infrastructure providers. A poorly structured prompt that forces the model to iterate, rephrase or regenerate a response will consume more tokens than a concise, well-targeted query, yet both may or may not produce useful output. If an AI agent saves a customer service representative 15 minutes of work, but costs $4 in inference tokens to run, the ROI is negative, as explained by AnalyticsWeek. The kind unit-economics mismatch is more visible as companies move from pilots to production deployments. As companies move from experimental chatbots to thousands of autonomous "agentic" workflows running around the clock, the sheer volume of tokens consumed has created a massive budgetary leak. The dynamic draws comparisons to earlier enterprise metrics that proved easier to game than to interpret. Click-through rates once served as a proxy for advertising effectiveness; hours logged once functioned as a proxy for productivity. Both created incentives that diverged from the outcomes they were meant to track. If token consumption becomes a performance indicator tied to employee evaluations, workers may optimize for AI interaction frequency rather than task quality. Knowing that "AI spend is up 40%" is not enough. Organizations need a single pane of glass that links every workload, tenant and token to their owners or business outcomes.
Share
Share
Copy Link
Nvidia CEO Jensen Huang is championing tokens as the fundamental unit of the AI economy, predicting employees may need token budgets worth half their salary. But while token economics offers a compelling framework for measuring AI adoption and billing, critics argue it measures volume rather than value—creating potential budgetary challenges without clear links to business outcomes.
Nvidia CEO Jensen Huang used the company's annual GTC conference this week to promote a vision where tokens become the defining currency of artificial intelligence. These basic units of output from large language models—roughly 1,300 tokens generate 1,000 words of text—are being positioned as the key AI metric for measuring both production efficiency and economic value
1
. Huang's framework centers on cost per token as the critical measure, arguing that as long as Nvidia's chips produce tokens at the lowest cost and demand outstrips supply, the AI boom remains healthy1
.
Source: PYMNTS
The shift toward token economics reflects broader changes in how enterprise AI spending is structured. Unlike seat-based pricing that characterized earlier software generations, token consumption offers granular, real-time measurement tied directly to behavior
2
. OpenAI data reveals the scale of this transformation: average reasoning token consumption per organization has surged approximately 320 times over the past 12 months, signaling systematic integration of more intelligent models into expanding products and services2
. Huang went further, estimating that employee token allocations could eventually reach half of base salary in value, stating at GTC that he could "totally imagine in the future every single engineer in our company will need an annual token budget"2
.Despite the compelling narrative around measuring AI adoption through tokens, significant gaps exist in connecting token production to actual value creation. Just because the cost of tokens is falling doesn't automatically mean AI-powered services become valuable or generate revenue across the industry, as Huang suggests
1
. The fundamental problem is that tokens measure volume, not outcome2
.This disconnect becomes stark when examining unit economics. A poorly structured prompt forcing a model to iterate or regenerate consumes more tokens than a concise query, yet both may produce equally useful—or useless—output
2
. More troubling, if an AI agent saves a customer service representative 15 minutes but costs $4 in inference tokens to run, the ROI turns negative2
. Organizations attempting to link AI usage to business outcomes find themselves struggling to connect token consumption metrics with tangible value delivery.The situation grows more complex with newer AI models. Reasoning models that emerged late in 2024, starting with OpenAI's o1, consume far larger numbers of tokens to arrive at answers. These are now being supplemented by agents promising to automate white-collar work, bringing an explosion in token use and potentially hefty bills for companies offering workers unlimited AI access
1
. Software engineering has seen initial efforts to measure how token use links to output, with some attempting to apportion tokens to workers. Tech companies envision AI becoming core to employment, where white-collar worker costs equal salary plus a certain number of tokens per month—but that remains a pipe dream for now1
.The second major gap in token economics involves how companies producing tokens will maintain profitability. If AI factories all use Nvidia's latest chips, gaining a cost-per-token advantage or retaining pricing power becomes difficult
1
. Price declines accompanying plunging production costs tell a concerning story. When OpenAI launched GPT-4 two years ago, it charged $33 for 1 million tokens. Today, it charges only 9 cents for 1 million tokens from its cheapest model1
. While beneficial for customers, this feeds worries about commoditization.These concerns echo debates from cloud computing's early days, when skeptics questioned whether Amazon Web Services could profit from selling basic storage and computing as commodities. Cloud providers eventually built higher-value platforms on which customers could run entire businesses, though oligopoly dynamics and regulatory pressure reducing switching costs may also explain healthy profit margins
1
. Whether OpenAI and Anthropic can execute a similar transformation remains unclear, though the opportunity exists1
. For now, intense competition among frontier AI companies continues, and how this resolves will significantly shape industry profitability1
.Related Stories
As companies move from pilots to production deployments and from experimental chatbots to thousands of autonomous agentic workflows running continuously, token consumption has created what some describe as massive budgetary leaks
2
. While unit prices fall, overall enterprise AI spending and system scaling continue rising. The number of users, model complexity, and workload intensity drive greater token consumption and consequently higher costs2
.The dynamic invites comparisons to earlier enterprise metrics that proved easier to game than interpret. Click-through rates once proxied for advertising effectiveness; hours logged functioned as productivity measures. Both created incentives diverging from intended outcomes
2
. If token consumption becomes a performance indicator tied to employee evaluations, workers may optimize for AI interaction frequency rather than task quality2
. Knowing that "AI spend is up 40%" provides insufficient insight without systems linking every workload, tenant, and token to their owners or business outcomes2
. Organizations need comprehensive visibility connecting AI adoption and billing to measurable value, not just consumption volume, as they navigate this emerging economic reality.Summarized by
Navi
Yesterday•Technology

16 Aug 2025•Technology

29 Oct 2025•Business and Economy
