4 Sources
4 Sources
[1]
Why Bigger Isn't Always Better in AI
For years, artificial intelligence has been governed by a simple creed: Bigger is better. Feed a model more data, more chips and more electricity, and it will become smarter. That has been true to a certain extent, especially in an industry obsessed with building "superintelligence," or all-knowing computer systems. But this scaling approach has also produced a hefty environmental toll, one that is fast becoming a political liability as communities from Malaysia to the MAGA heartland push back against new data centers. But what if we're thinking about it all wrong? The more important question is whether most businesses actually need these God-like AI systems, and whether they are even the right tools for the tasks that matter most. For all the hype, much of AI's practical, commercial value will come from automating narrow, repetitive tasks, not from building an omniscient machine. For that kind of work, smaller, specialized tools may be the best option. They're cheaper to run, easier to secure, less demanding on water and energy resources, and often just as effective. If AI is to become genuinely useful without becoming politically and environmentally untenable, the future may belong not to the biggest models, but to the smartest uses of more modest ones. Take agentic AI, long touted as the next big breakthrough for productivity. Its promises are some of the most seductive for business leaders: software that can act on a users' behalf and handle routine tasks to free people up for more valuable work. But many of these chores are so narrow and mundane that they don't require supreme intelligence, or the energy appetites of giant, cloud-based systems. The way we have been thinking about AI for decades now has been shaped by same bigger-is-better approach, Daniela Rus, the director of the MIT Computer Science and Artificial Intelligence Laboratory, told me. "The end result," she said, "is that we have these huge models that have a very large energy costs and also very large water costs, and this translates into a big environmental footprint." That realization pushed Rus toward alternatives. Her academic experimentation eventually led to LiquidAI, an MIT spinout that builds small, task-specific models for enterprises and researchers and that can run on devices. The company has done this by using only 1,000 GPUs, or advanced AI processers. (OpenAI, by contrast, said last July that it expected to bring more than one million GPUs online by the end of the year.) Even simple courtesies expose the inefficiencies of the current paradigm. Saying "please" and "thank you" to ChatGPT is reportedly costing OpenAI tens of millions of dollars in energy and power costs. To Ramin Hasani, Rus' former student and the co-founder of LiquidAI, that's evidence of how much waste is baked into today's systems. Smaller, specialized AI, he told me, can match their larger cloud-based counterparts on specific tasks while using "orders of magnitude less amounts of energy consumption." Rus and Hasani are hardly alone. Researchers from Nvidia Corp. and the Georgia Institute of Technology argued in a paper last year that insisting on large models for agentic tasks "reflects a misallocation of computational resources" that is "economically inefficient and environmentally unsustainable at scale." In their view, shifting such workloads to smaller ones is not merely a technical refinement, but a "moral" obligation. So why are companies still pouring so much money into tapping giant systems for routine work? One answer is inertia. The same Nvidia-led researchers point to the enormous capital already committed to the existing, centralized system (US tech giants are collectively estimated to invest some $650 billion in AI infrastructure this year alone). Once that much money is on the table, it becomes harder to question whether the underlying approach still makes the most sense. Then there is hype. Small models do not attract the same marketing frenzy or media attention, the paper noted, even when they are better suited for many business uses. It's easier to sell the fantasy of an AI that knows everything than one that quietly processes documents or handles back-office work more cheaply and securely. Sign up for the Bloomberg Opinion bundle Sign up for the Bloomberg Opinion bundle Sign up for the Bloomberg Opinion bundle Get Matt Levine's Money Stuff, John Authers' Points of Return and Jessica Karl's Opinion Today. Get Matt Levine's Money Stuff, John Authers' Points of Return and Jessica Karl's Opinion Today. Get Matt Levine's Money Stuff, John Authers' Points of Return and Jessica Karl's Opinion Today. Bloomberg may send me offers and promotions. Plus Signed UpPlus Sign UpPlus Sign Up By submitting my information, I agree to the Privacy Policy and Terms of Service. This all matters especially here in Asia, where many middle powers are mulling how to keep pace in the AI race without the capital for vast data-center buildouts or the energy required to power them. If the future of AI depends entirely on massive infrastructure, most countries will be left buying access to someone else's system. But the real impact may be less about superintelligence and more about humble systems that do specific jobs well. Which makes now a good moment for a rethink. 2025 was supposed to be the year of AI agents. It didn't go as planned. The economics were too punishing and the cybersecurity risks too obvious. Smaller models -- including ones that can run entirely on a device -- could change that calculus. They promise lower costs, tighter security, and a much more compact environmental footprint. AI doesn't have to be inherently unsustainable. But getting to a different future will require more pluralism in research, more willingness to challenge Big Tech's incentives, and less concentration of technological power. The smartest thing policymakers and business leaders can do now may be to think smaller. For this tech revolution to scale, it may first have to shrink. More From Bloomberg Opinion: * Anthropic Is Very Different to OpenAI's 'Yes Man': Parmy Olson * Anthropic Isn't Exaggerating About an AI Panopticon: Dave Lee * China's AI Lobster Craze Comes With Claws: Catherine Thorbecke Want more Bloomberg Opinion? OPIN <GO> . Or you can subscribe to our daily newsletter .
[2]
Computing power is no longer the AI bottleneck -- it's energy production
For decades, AI was held back by slow, expensive computers. Today, the problem is simpler, but harder to fix: finding enough reliable electricity to keep data centers running as AI spreads into everyday life. For much of the 20th century, artificial intelligence (AI) struggled not because researchers lacked ambition, but because the hardware available to power it simply wasn't powerful enough. Early AI systems hit hard limits on processing speed and memory, contributing to repeated "AI winters" as progress stalled and funding dried up. That problem is mostly gone now. Today, AI models are trained on specialized chips in huge data centers, and they can scale up in weeks instead of years. Compute, which used to be the main bottleneck, is now something that can be bought with enough money. Companies like Nvidia or AMD are also mass-producing even more powerful graphics processing units (GPUs) -- components conventionally used for gaming or visualization but also well suited to processing AI calculations -- as each year goes by. So, beyond the fundamental architectures at the heart of these models, what's keeping AI from becoming even more advanced? The new limit is far more physical in nature -- and far harder to work around. It's electricity. Why AI's energy appetite is exploding Modern AI models don't just train once and then stop. They run all the time, powering things like chatbots, search tools, image generators and more autonomous agents. This change has made AI a constant, large-scale user of electricity. According to Sampsa Samila, academic director of the AI and the Future of Management Initiative at Barcelona's IESE Business School, the problem isn't a lack of energy in absolute terms. "It's not the overall supply of energy, but having reliable, firm capacity at the right place and the right time that is in short supply," he told Live Science. Predictions for AI energy consumption show this strain clearly. The International Energy Agency (IEA) expects data centers to consume more than twice as much electricity by the end of the decade, reaching levels similar to those in major industrial economies. In some parts of the U.S, data centers already use as much power as heavy industry. How AI is actually used matters just as much as how it's trained. Training large language models (LLMs) still consumes a lot of power, but it tends to occur in large, infrequent runs. What's growing faster is the everyday work -- models responding to users, over and over again. Samila notes that newer "reasoning" systems, which spend more time working out an answer, push energy use into normal operations rather than occasional training bursts. A grid built for a slower world Power grids were designed for gradual growth, not for city-sized loads appearing almost overnight. Juan Arismendi-Zambrano, an assistant professor at Ireland's University College Dublin (UCD) Michael Smurfit Graduate Business School, said the main issue is timing. Large AI campuses grow faster than grid upgrades or government approvals can keep up with. This creates a real bottleneck: getting enough power, when and where it's needed. "The 'short supply' of AI electricity is, in my view, less about an absolute global lack of electricity and more about local bottlenecks created by fast deployment of large data centres," Arismendi-Zambrano told Live Science. "These campuses scale quicker than electricity grid upgrades, or bureaucracy can respond. Especially when they land in rural areas chosen for cheap land and political 'lobbying' for states, but not engineered for sudden, concentrated load. The result is a very physical constraint: access to a lot of electricity power, on time, at the right node," he said. Clustering data centers in one area makes the problem worse. Jens Förderer, a professor at the University of Mannheim Business School in Germany, pointed to Northern Virginia's "Data Center Alley," where many facilities draw huge amounts of power from the same grid. Power plants, transmission lines and substations take years to build, but AI companies often start using compute much sooner, sometimes even before their buildings are finished. "When many city-scale loads draw from the same local grid, scaling electricity provision becomes far harder," Förderer said. How the industry is scrambling to respond There is no single fix for AI's energy problem. Instead, companies are pursuing several strategies at once. One is building power closer to the data centers themselves. Large tech firms have signed long-term contracts to support new power generation, including nuclear plants, and are exploring on-site power where grid upgrades move too slowly. Google, for example, has been doing this in Texas through its acquisition of energy developer Intersect, which builds large-scale solar and storage projects alongside data center demand rather than waiting for grid upgrades. Microsoft, meanwhile, has signed a long-term deal with Constellation Energy tied to the planned restart of a nuclear reactor at Pennsylvania's Three Mile Island site to supply power for its data centers. Another is choosing locations based on electricity, rather than users. As Förderer noted, data centers are increasingly sited where power is easiest to scale, even if that means moving further from major population centers. Then there is reuse -- including a surprising source. Former cryptocurrency mining facilities are emerging as candidates for AI workloads. Once criticized for their energy use, these sites already have what AI needs most: large grid connections, cooling systems and experience running power-hungry hardware around the clock. The crossover between Bitcoin and AI may look strange, but the underlying physics is the same. "These facilities already have large grid connections, and some former miners may pivot toward AI workloads," Förderer said. Canadian miner Bitfarms has recently announced plans to transition its facilities away from Bitcoin mining toward high-performance computing and AI data centers, while Hut 8 -- originally a Bitcoin mining company -- struck a major $7 billion lease deal in late 2025 to provide data-center capacity for AI computing Some ideas look even further afield. Space-based data centers are sometimes pitched as a way to sidestep Earth's grid entirely, using constant solar energy and the cold of space for cooling. Samila said the idea works on paper, but the numbers get intimidating fast. A single 5-gigawatt facility would require around 2.5 by 2.5 miles (4 by 4 kilometers) of solar panels in orbit. It's "in principle doable," he added, but only with some serious engineering. Latency, upkeep and launch logistics remain open questions. Efficiency may be the fastest lever of all. Förderer pointed out that advances in chips, model design and system architecture have already reduced the energy required per unit of intelligence. Some recent efforts include an MIT breakthrough that aims to cut energy use by stacking components vertically, as well as a "rainbow-on-a-chip" that uses lasers to transmit data in components. Such gains won't eliminate the need for more power, but they can slow the rate at which demand grows. Does unlocking energy unlock smarter AI? The growing demand placed upon the electricity grid by AI also raises environmental concerns. Engineer Aoife Foley, professor and chair in Net Zero Infrastructure at the University of Manchester in the U.K., pointed out that the wider IT sector already makes up about 1.4% of global carbon emissions. AI workloads use much more energy than regular cloud computing, and while big tech companies are investing in renewables and better cooling, Foley said these efforts alone are not enough."These impacts can be reduced through smarter model optimisation and a closer alignment between data centre strategy and regional renewable generation," she told Live Science. Despite the scale of the challenge, none of the experts see electricity as a shortcut to artificial general intelligence (AGI) -- a hypothetical form of AI that can simulate behaviour as intelligent as, or more intelligent than, that of a human being. More energy makes it easier to build and run bigger systems, but it doesn't solve the harder problems. Instead, Förderer argued that the real limits sit elsewhere -- in access to data, in new model architectures and in genuine advances in reasoning. "Energy is necessary but not sufficient," Samila said in agreement, adding that today's dominant approach to improving AI relies on massive amounts of power, but more electricity alone will not magically produce AGI. More energy doesn't guarantee smarter machines, but it does change who gets to participate. Access to power will shape where AI is built, who can afford to run it and how broadly it's deployed. The bottleneck has shifted away from silicon and toward the physical world, where grids, permits, and power plants move at a very different pace than code.
[3]
Inside the Dirty, Dystopian World of AI Data Centers
As we drove through southwest Memphis, KeShaun Pearson told me to keep my window down -- our destination was best tasted, not viewed. Along the way, we passed an abandoned coal plant to our right, then an active power plant to our left, equipped with enormous natural-gas turbines. Pearson, who directs the nonprofit Memphis Community Against Pollution, was bringing me to his hometown's latest industrial megaproject. Already, the air smelled of soot, gasoline, and asphalt. Then I felt a tickle sliding up my nostrils and down into my throat, like I was getting a cold. As we approached, I heard the rumble of cranes and trucks, and then from behind a patch of trees emerged a forest of electrical towers. Finally, I saw it -- a white-walled hangar, bigger than a dozen football fields, where Elon Musk intends to build a god. This is Colossus: a data center that Musk's artificial-intelligence company, xAI, is using as a training ground for Grok, one of the world's most advanced generative-AI models. Training these models takes a staggering amount of energy; if run at full strength for a year, Colossus would use as much electricity as 200,000 American homes. When fully operational, Musk has written on X, this facility and two other xAI data centers nearby will require nearly two gigawatts of power. Annually, those facilities could consume roughly twice as much electricity as the city of Seattle. To get Colossus up and running fast, xAI built its own power plant, setting up as many as 35 natural-gas turbines -- railcar-size engines that can be major sources of smog -- according to imagery obtained by the Southern Environmental Law Center. Pearson coughed as we drove by the facility. The scratch in my throat worsened, and I rolled up my window. xAI's rivals are all building similarly large data centers to develop their most powerful generative-AI models; a metropolis's worth of electricity will surge through facilities that occupy a few city blocks. These companies have primarily made their chatbots "smarter" not by writing niftier code but by making them bigger: ramming more data through more powerful computer chips that use more electricity. OpenAI has announced plans for facilities requiring more than 30 gigawatts of power in total -- more than the largest recorded demand for all of New England. Since ChatGPT's launch, in November 2022, the capital expenditures of Amazon, Microsoft, Meta, and Google have exceeded $600 billion, and much of that spending has gone toward data centers -- more, even after adjusting for inflation, than the government spent to build the entire interstate-highway system. "These are the largest single points of consumption of electricity in history," Jesse Jenkins, a climate modeler at Princeton, told me. Even conservative analyses forecast that the tech industry will drop the equivalent of roughly 40 Seattles onto America's grid within a decade; aggressive scenarios predict more than 60 in half that time. According to Siddharth Singh, an energy-investment analyst at the International Energy Agency, by 2030, U.S. data centers will consume more electricity than all of the country's heavy industries -- more than the cement, steel, chemical, car, and other industrial facilities put together. Roughly half of that demand will come from data centers equipped for the particular needs of generative AI -- programs, such as ChatGPT, that can produce text and images, solve complex math problems, and perhaps one day inform scientific discoveries. To power AI, energy and tech companies are turning to fossil fuels, which they regard as more reliable and readily available than wind, solar, or nuclear. Asked where the energy for data centers should come from, OpenAI CEO Sam Altman has repeatedly said, "Short-term: natural gas." (OpenAI and The Atlantic have a corporate partnership.) A Louisiana utility plans to build three natural-gas plants for a Meta data center that, upon completion, will be among the largest in this hemisphere. The lifespans of coal plants, too, are being extended to power new data centers. And the IEA estimates that data-center emissions could more than double by 2030 -- becoming one of the fastest-growing sources of greenhouse gases in the world. The optimist's case is that, by then, advanced nuclear reactors will have obviated many of the new fossil-fuel plants, and AI tools will have invented technologies that can solve the climate crisis. That may well happen. But today, "the market has converged on Add gas now, and then add nuclear later," Jenkins said. In other words, if natural-gas turbines seem to offer the most expedient path to an AI-enhanced future, then clean air may have to wait. A data center is a planet of contradictions: heat without motion, shelter without bodies, light without sky. "The lifeblood of the internet is essentially flowing through these sites," Jon Lin, the chief business officer at Equinix, one of the world's largest data-center companies, told me in an Equinix facility in Loudoun County, Virginia. Behind Lin, someone in a green hoodie fiddled with computer chips shelved in a row of humming, refrigerator-size cabinets on the data-center floor. There were no windows, to keep the facility secure and to guard against the sun's heat. As we walked along a corridor of cabinets, motion-activated lights illuminated the way. Farther ahead, only faint blue lights and blinking computer equipment pierced the darkness. Ever since the first data centers were built, in the mid-20th century, their purpose has remained constant: pack computer equipment close together to store and send information as efficiently as possible. But their scale has grown dramatically. The original data centers were simply large rooms housing mainframe computers. With the rise of the internet, in the 1990s, backroom computers gave way to entire buildings, such as the one Lin and I stood in -- facilities that enable us to stream movies, trade stocks, store medical records, manage supply chains, and make military decisions. Now the AI race is requiring vastly greater computing power, which has led to even bigger data centers, ones filled with computer chips that are much hungrier and run much hotter. Read: The lifeblood of the AI boom In a traditional data center, the cabinets are cooled by industrial fans -- as we walked through the Equinix facility, I felt a constant breeze on my cheek -- and rooftop cooling towers eventually expel the heat. The cabinets in a generative-AI data center use dozens of times more electricity. Lin showed me a row of AI-specialized cabinets used by Block, the firm that owns Square and Cash App, which radiated enough heat to make me break a sweat; to cool them, water runs into special metal plates that sit atop the chips inside the cabinets. AI data centers are filled with similar equipment, and cooling thousands of cabinets can require a lot of water. Public records from the Memphis water utility, for instance, show that the address for Colossus used more than 11 million gallons in September alone, as much as 150 homes use in an entire year. When a data center's cooling equipment malfunctions, spiraling heat combined with humid air has yielded that rarest of meteorological events: indoor rain. Placing servers in the same or neighboring buildings allows them to exchange information seamlessly and quickly, and Loudoun County has the highest concentration of data centers in the world, with 199 already operating and another 30 or so on the way. According to one report, 13 percent of global data-center capacity is squeezed into the county's 520 square miles. One particularly dense stretch is called "Data Center Alley." Northern Virginia offers a glimpse into what the AI rush may bring to the rest of the nation. Loudoun is running out of space, but new data-center hubs are popping up in Phoenix, Atlanta, and Dallas. Amazon and Meta are building AI data centers in Indiana and Louisiana, respectively, that will each require more than two gigawatts of electricity, dozens of times more than standard facilities. OpenAI has proposed that the U.S. establish "AI Economic Zones": little Loudouns everywhere. As I drove into Data Center Alley with Julie Bolthouse, the director of land use at the Piedmont Environmental Council, she explained how to distinguish data centers from warehouses: cooling towers on the roof, dozens of backup diesel generators to one side, no windows (or false ones, decorative glass panels backed by a wall of concrete). There didn't seem to be any warehouses, though, and I gave up counting data centers within minutes, unable to tell where one facility ended and the next one began. Bolthouse helps run a coalition aiming to slow data-center development throughout Virginia, but in Loudoun, it is too late. So many data centers are under construction just north of Dulles International Airport that hills of freshly dug dirt loom over roads and orange dust tints the air. Should Musk successfully colonize Mars, the early stages of terraforming might look like this. The architect of this labyrinth is Buddy Rizer, Loudoun's longtime executive director of economic development. Rizer has courted data centers with regulatory and state tax incentives, and when we met in his office, he told me that since 2009, at least one has been under construction at any given time. Data centers are typically operated by only a few dozen staff members, but building them has produced a steady source of employment. They also provide nearly 40 percent of the county's budget, helping to pay for police, schools, and parks for a population that has grown steadily since 2010. Within a 1.5-mile radius of us, Rizer said, were 12 substations: small jungles of metal poles and wiring that convert high-voltage electricity into a form you'd use to charge your iPhone or, in this case, run a data center. All around us were towering utility poles strung with high-voltage transmission lines that carry raw electricity from power plants to those substations; they hang over Loudoun like a canopy, or a cobweb. Follow any one cable far enough, and you're likely to reach a data center. For years to come, the AI race is projected to be the main force driving roughly 2 percent annual growth in U.S. electricity demand, which has been stagnant for nearly two decades. Nationally, this is not a crisis; regionally, it may be. Dominion Energy, the major electrical utility in Virginia, predicts growth of 5.5 percent each year, with overall electricity demand doubling by 2039. Aaron Ruby, a spokesperson for Dominion, told me that the company is preparing to meet that surge, though he was frank about the challenge: "We are experiencing the largest growth in power demand since the years following World War II." By the end of the decade, training the industry's most powerful AI model could require as much electricity as millions of American homes. In China, hundreds of data centers have been announced since 2023, and additional facilities are planned for beneath the ocean and in the desert. China's biggest advantage in the AI race is not the talent of its software engineers or the quantity of its data centers, but its abundance of energy: In 2024, the nation produced nearly as much electricity as the U.S., Europe, and India combined. President Trump has declared that the nation is in an "energy emergency," and been vocal about the need to build more power plants for the U.S. to win the AI race. A senior executive at OpenAI told me that the U.S. needs to activate every resource at its disposal -- solar panels, natural-gas turbines, nuclear reactors. And Anthropic, OpenAI's top rival, published a report arguing that the U.S. should streamline permitting for data centers and power plants in order to keep pace with China. But an internet-driven energy crisis has failed to materialize before: As fiber-optic cables were being laid in Loudoun in the 1990s, energy companies built more coal- and gas-fired plants. "Dig More Coal -- The PCs Are Coming," read a 1999 Forbes headline. When the demand didn't arrive, the nation was left with a glut of gas plants and multiple bankrupt energy companies. The generative-AI boom, too, could prove to be a bubble. The technology remains extraordinarily expensive, largely because of the cost of advanced computer chips, and no AI firm has presented a convincing business model. One path to profitability might be more efficient algorithms -- which would preclude the need for the new natural-gas plants. And if AI doesn't turn out to be as transformative a technology as experts predict, swaths of data centers could be left unused or unfinished -- ruins from a future that never came to pass. Either way, the rush to power data centers as fast as possible has already pushed the U.S. to expand its reliance on fossil fuels. Behind her one-story brick home in southwest Memphis, Sarah Gladney grows tomatoes, and when the vines wilted early last summer, she had a suspect in mind. "When the wind comes up early in the morning, I can smell it," Gladney told me, nodding in the direction of Colossus. One of her neighbors, Marilyn Gooch, told me the data center's turbines have made her uncertain about whether she should let her grandchildren visit. Their neighborhood, Boxtown, is named for the railway boxcars that formerly enslaved people used to build homes, and is still almost entirely Black. Virtually every heavy industry has set up nearby -- a wastewater facility, an oil refinery, a coal-fired power plant. Colossus itself, which is next to a steel mill and a trucking and rail yard, occupies the hull of an old oven factory. Life expectancy in and around Boxtown is more than five years below the national average, and the cancer risk in southwest Memphis is four times higher. What KeShaun Pearson and I smelled may not have been Colossus itself; xAI had chosen an area so besieged by heavy industry that any exhaust from the facility's turbines would mix in with a pervasive smog. Colossus was built so quickly that many Boxtown residents and elected officials didn't know what was happening until the project was well under way. Construction began in May 2024, and the project was announced the following month. Gladney, Pearson, and his younger brother Justin -- who represents the district in the Tennessee General Assembly -- found out about the project that day in June. By Labor Day weekend, less than three months after the press conference, Colossus was up and running. The company installed its own gas turbines because that was faster than waiting on the local grid, and argued that it did not need a permit to do so because the turbines would operate for less than a year, a claim that the Southern Environmental Law Center, representing the NAACP, contested in a letter threatening to sue the company. (xAI has since received a permit for 15 turbines, and is reportedly operating 12.) Meanwhile, residents report that they have had respiratory issues flare up since xAI moved in. Last June, when an analysis commissioned by the city of Memphis found "no dangerous levels" of pollutants in Boxtown and at two other test locations, the SELC criticized the study's methods. Using satellite data, researchers at the University of Tennessee at Knoxville found that levels of nitrogen dioxide -- which causes smog and is associated with asthma and other respiratory problems -- near Colossus have been substantially elevated since its public announcement. (xAI says on its website that it will install technology to reduce the pollution from its turbines. The company, the Shelby County Health Department, and the Memphis mayor's office did not respond to a list of questions about Colossus's environmental impacts and xAI's presence in Memphis; the Greater Memphis Chamber of Commerce declined to comment.) Fossil fuels have become the default for data centers around the country. OpenAI's first Stargate data center, in Texas, also has its own gas-fired power plant. Chevron and Exxon are angling to hook natural-gas facilities directly into data centers, and the world's three major manufacturers of natural-gas turbines all advertise their products as convenient energy sources for data centers. Michael Eugenis, the director of resource planning at Arizona Public Service, the state's largest utility, told me that because of the demand from data centers, the company is adding more fossil-fuel capacity than it otherwise would have; natural gas will help power Microsoft, Amazon, and Oracle data centers, too. In early 2025, a company affiliated with xAI purchased a former warehouse and nearly 200 acres south of Colossus to set up another data center, Colossus II. On a weekday afternoon, the road near the site was dense with traffic -- not dump trucks and forklifts, but sedans lining up outside the adjacent public school for pickup. An xAI affiliate bought a retired Duke Energy plant about a mile away in Mississippi that is likely to power this facility, and filed an application to operate 41 natural-gas turbines on the site. Those turbines could emit more carbon dioxide annually than the city of San Jose. On an island in the Susquehanna River, just south of Harrisburg, Pennsylvania, I saw another way to power the AI boom. Above me loomed four beige hourglass-shaped structures, each some 365 feet tall: the cooling towers for Three Mile Island, the site of the worst nuclear disaster in American history. On March 28, 1979, the facility was only a few years old, and nuclear-energy reactors were being built across the country. But a series of mechanical and human errors caused the core of one of the reactors, Unit Two, to rapidly overheat and leak radioactive material. The effects on human health and the environment were negligible, but together with the catastrophe at Chernobyl seven years later, the partial meltdown turned public sentiment strongly against nuclear power. Three Mile Island's Unit One went undamaged and continued operating, after a brief pause, until 2019. By then natural gas was too cheap, the regulatory environment was too unfriendly, and the losses -- hundreds of millions of dollars -- were too great for Constellation Energy, which owns Unit One, to keep the plant running. From the March 2023 issue: Jonathan Rauch on the real obstacle to nuclear power Nobody has ever resuscitated a fully shut-down U.S. nuclear-power plant, but in fall 2024, Constellation announced plans to do just that. Microsoft had agreed to purchase electricity from Unit One to power its data centers over the next two decades, a guarantee allowing Constellation to spend the $1.6 billion needed to restart the plant. It was the ultimate bellwether of the AI age: Experts have long argued that we need clean nuclear power to reduce the grid's existing carbon footprint. Instead, Three Mile Island will help offset a new source of emissions from a single company. Constellation is now reversing the steps it took to decommission the reactor: renewing its license, restoring equipment, retraining personnel. Dave Marcheskie, a community-relations manager, explained this to me in a conference room overlooking the nuclear core, which is housed in a building that resembles a large grain silo. Behind him, a clock counted down the time to launch: 650 days, zero hours, 42 minutes, and one second. As the need for carbon-free electricity grows more urgent, Americans are having to reckon with nuclear energy again, and the AI boom has provided the industry with wealthy backers and an army of tech cheerleaders. Meta and Amazon are buying electricity from large nuclear-power plants, and nearly every major data-center company is investing in experimental nuclear technologies -- especially small modular reactors, which in theory will make fission cheaper and easier to deploy. Read: A new reckoning for nuclear energy Nuclear energy has its downsides, of course. The waste is radioactive and must be stored almost indefinitely, and the meltdown at Japan's Fukushima plant in 2011 was a reminder of how spectacularly dangerous nuclear reactors can be. But the dangers posed by the burning of fossil fuels are far more imminent. At Three Mile Island, Marcheskie led me down a hall and into the actual power plant. Pipes, tubes, and hulking machines lined the floor and ceiling; a trefoil sign warned that a large tank potentially contained radioactive materials. The elevator was broken, so we walked a few stories up to the stadium-size room from which all of Three Mile Island's electricity will flow. Scaffolding and shipping containers were scattered around a row of pistachio-green semi-cylinders. Once the plant restarts, uranium atoms ripped apart in the adjacent core will generate immense amounts of heat, vaporizing water into steam that will spin blades inside those cylinders 1,800 times a minute, which will in turn produce hundreds of megawatts of electricity. This will be orchestrated from a nearby control room, where hundreds of lights and switches line muted-green walls. The shift manager, Bill Price, explained that one half of the main panel controls the nuclear core, while the other half controls the turbines. In the middle is the most important control of all: a red button that shuts down the reactor, and above it an identical button that serves as a backup. In the event of an emergency, Price said, you'd press both. I put a finger on each button and pushed. A small amount of the electricity generated here will support the plant itself. Microsoft is buying the remainder through a power-purchase agreement, a mechanism companies use to buy carbon-free electricity to match whatever their facilities draw from the grid. Power generated at Three Mile Island will help offset the energy used by data centers in Virginia and Illinois; Microsoft says it purchases enough clean energy to match all of its electricity consumption, as do Google, Amazon, and Meta. These companies are also investing in hydropower, geothermal plants, and solar panels; Google is exploring building a data center in space, to enable cloud-free access to the sun. Read: For now, there's only one good way to power AI Still, tech firms insist that nuclear and other clean technologies cannot be deployed quickly enough to meet their needs. President Trump has signed an executive order to accelerate permitting for natural-gas and coal-fired plants to support data centers. Yet China's energy advantage in the AI race comes from nuclear reactors and solar panels, not coal and oil; the country is building nearly two-thirds of the world's new solar and wind capacity. The U.S. could still catch up, thanks to private investments by the likes of Google and Microsoft. A majority of planned electricity generation in the U.S. will be carbon-free, and running data centers on renewables can be done, Jenkins, the Princeton climate modeler, told me. Meanwhile, natural-gas turbines are so far back-ordered that acquiring one in the next few years will be virtually impossible. For now, using existing power sources more wisely, rather than building new ones, may be all the AI industry needs. Electrical grids are designed for periods of peak demand -- cooling on summer afternoons, heating on winter mornings -- but mostly they run well below maximum capacity. Researchers at Duke University have shown that if data centers reduced their electricity consumption during some of those peaks, it would free up enough electricity to accommodate the country's planned data centers for years. Google and xAI have already entered agreements to do so. That strategy would allow tech companies to continue building more data centers without waiting for utilities to expand the grid. And time, not dollars or electrons, is the AI industry's primary currency. Google, Microsoft, and their competitors can afford to spend historic sums without near-term financial returns, but they cannot afford to slip behind one another. Time is also the biggest problem for Microsoft's deal with Three Mile Island, which is taking years to restart. As we left the facility, Marcheskie led me south, past the beige towers and through a fog that had settled over the river. At one point we passed a cluster of concrete barrels that had escaped my attention on the drive up. Marcheskie told me that they contained all of the nuclear waste from Unit One's 45 years of operation. Perhaps one day such casks will also line the perimeters of Colossus and Stargate. AI may well overhaul how humans think and work, but it's also pushing us toward another inflection point. We can unlock the promises of this technology by doubling down on the energy systems of the past, or we can seize the opportunity to push the grid into a carbon-free future. To get there, an industry that likes to move at warp speed will have to develop a quality it severely lacks: patience.
[4]
Why AI must shrink to reach its enterprise potential
From copilots and chatbots to advanced analytics and automation, AI systems are now embedded in how organizations operate and compete. Yet as adoption accelerates, a less visible issue is coming sharply into focus: energy. Data centers are already major consumers of electricity, and AI is pushing demand even higher. Power consumption linked to AI workloads is projected to grow by around 15% per year, far outpacing growth across other sectors. Training and running large language models (LLMs) requires enormous computational resources, and every additional layer of complexity translates directly into higher energy use. This trajectory raises a critical question for the future of AI: how long can innovation continue on a path that depends on ever-increasing power consumption? Power constraints are shaping AI's future The AI industry has spent the past decade chasing scale. Larger models, more parameters and bigger datasets have driven impressive gains in performance. At the same time, the cost of delivering those gains has risen sharply. Electricity prices, grid capacity and data center availability are no longer background considerations. They are becoming limiting factors. In many regions, access to sufficient power is now a strategic constraint, shaping where AI infrastructure can be built and who can afford to use it. For businesses, this creates growing tension. Advanced AI promises efficiency and competitive advantage, yet the operational costs of running large models can be prohibitive. For governments and regulators, the challenge is even broader: balancing AI-led economic growth with sustainability targets and grid resilience. Without changes in how AI systems are built and deployed, energy demand risks slowing progress at exactly the moment when momentum is strongest. Cost-effective AI is essential for wider adoption The conversation around democratizing AI often focuses on access to tools or models. In practice, affordability plays an equally important role. If advanced AI remains expensive to run, its benefits will concentrate in the hands of a few large organizations with the deepest pockets and the most robust infrastructure. Most companies do not need the largest possible model available. They need systems that deliver reliable results at a predictable cost. That applies just as much to public sector organizations, manufacturers and mid-sized enterprises as it does to startups. Energy-efficient AI lowers the barrier to entry. Reduced power requirements mean lower operational costs, simpler deployment and fewer infrastructure constraints. For data centers, this translates into more efficient use of existing capacity, reduced cooling demands and less need for constant expansion. Optimized models allow organizations to do more with the infrastructure they already have, easing pressure on energy supply while improving overall economics. Efficiency also enables new deployment models. Smaller, compressed AI systems can run locally on devices such as smartphones, laptops, vehicles and even home or industrial appliances. By bringing intelligence closer to where data is generated, organizations can reduce latency, improve reliability and limit dependence on centralized cloud infrastructure. For many use cases, this is a practical advantage as well as a sustainability win. Smaller models can still deliver strong results There is a widespread assumption that cutting down models inevitably means sacrificing accuracy. Advances in model optimization are challenging that idea. Techniques such as compression, pruning and optimization allow LLMs to be significantly reduced in size while preserving performance on real-world tasks. This allows organizations to deploy efficient AI models in environments where large-scale systems would be impractical or uneconomical, without sacrificing the performance required for enterprise applications. The impact is dramatic. Compressed models can be up to 95% smaller, requiring far less memory and compute. That reduction translates directly into lower energy consumption and faster inference, while maintaining the level of accuracy organizations expect. This approach shifts the emphasis from brute-force scaling to intelligent design. Rather than treating size as a proxy for quality, it prioritizes efficiency, precision and real-world applicability. Sustainability and competitiveness go hand in hand As AI becomes a core part of digital infrastructure, its environmental footprint will increasingly matter. Businesses are under pressure to meet ESG commitments, and customers are paying closer attention to how digital services are delivered. Governments, meanwhile, are assessing how AI fits into long-term energy planning. Energy-efficient AI aligns with all of these priorities. Lower power consumption reduces emissions, eases strain on grids and improves the economics of deployment. It also makes AI more resilient, less dependent on scarce resources and better suited to global scale. The shift toward efficiency does not require slowing innovation. On the contrary, it creates room for growth by removing one of the most significant constraints facing the industry. Building the next phase of AI The next chapter of AI will be shaped less by how large models can become and more by how effectively they can be deployed. Progress depends on systems that are powerful, practical and sustainable. Achieving that balance requires collaboration across the ecosystem - from researchers developing leaner architectures to organizations rethinking how and where AI is deployed. It also calls for a broader definition of innovation, one that values efficiency alongside raw performance. AI has the potential to transform industries, improve productivity and address complex global challenges. Ensuring that transformation remains accessible and sustainable will determine how widely those benefits are shared. Solving AI's energy challenge is part of that work. Done well, it opens the door to a future where advanced intelligence is not limited by power consumption, but enabled by smarter design. We've featured the best AI website builder. This article was produced as part of TechRadarPro's Expert Insights channel where we feature the best and brightest minds in the technology industry today. The views expressed here are those of the author and are not necessarily those of TechRadarPro or Future plc. If you are interested in contributing find out more here: https://www.techradar.com/news/submit-your-story-to-techradar-pro
Share
Share
Copy Link
The AI industry faces a critical bottleneck as data centers consume electricity at unprecedented rates, with projections showing demand doubling by 2030. Tech giants are investing $600 billion in AI infrastructure while researchers argue smaller, specialized models could deliver similar results using orders of magnitude less energy.
The AI industry has reached a turning point where computing power is no longer the primary constraint—electricity is. Data centers now face a fundamental challenge: securing enough reliable power to sustain AI's rapid expansion into everyday applications
2
. The International Energy Agency projects that data centers will consume more than twice as much electricity by the end of the decade, reaching levels comparable to major industrial economies2
.
Source: Live Science
This shift represents a stark departure from decades past when AI struggled due to slow, expensive computers. Today, specialized GPUs from companies like Nvidia and AMD can scale AI models in weeks rather than years
2
. Yet this progress comes at a steep cost. According to Siddharth Singh, an energy-investment analyst at the International Energy Agency, U.S. data centers will consume more electricity than all of the country's heavy industries combined by 2030—surpassing cement, steel, chemical, and car manufacturing facilities3
.The scale of AI infrastructure investments is staggering. Since ChatGPT's launch in November 2022, the capital expenditures of Amazon, Microsoft, Meta, and Google have exceeded $600 billion, with much of that spending directed toward data centers
3
. This exceeds, even after adjusting for inflation, what the government spent to build the entire interstate highway system.Elon Musk's xAI facility in Memphis exemplifies the extreme AI electricity demand. The Colossus data center, used to train the Grok model, would consume as much electricity as 200,000 American homes if run at full strength for a year
3
. When fully operational, Musk's three xAI facilities will require nearly two gigawatts of power—roughly twice the annual electricity consumption of Seattle3
. OpenAI has announced plans for facilities requiring more than 30 gigawatts of power in total, exceeding the largest recorded demand for all of New England3
.
Source: Bloomberg
The environmental consequences are mounting rapidly. The IEA estimates that data center emissions could more than double by 2030, becoming one of the fastest-growing sources of greenhouse gas emissions worldwide
3
. To meet this demand, energy and tech companies are increasingly turning to fossil fuels. OpenAI CEO Sam Altman has repeatedly stated that short-term energy needs should be met with natural gas3
. A Louisiana utility plans to build three natural-gas plants for a Meta data center, while coal plant lifespans are being extended to power new facilities3
.
Source: The Atlantic
The strain on power grids extends beyond energy supply to infrastructure capacity. Juan Arismendi-Zambrano from University College Dublin notes that the shortage is "less about an absolute global lack of electricity and more about local bottlenecks created by fast deployment of large data centres"
2
. These facilities scale faster than grid upgrades or government approvals can accommodate, creating physical constraints at specific grid nodes2
.Related Stories
A growing chorus of researchers and companies argues that the industry's bigger-is-better approach fundamentally misallocates resources. Daniela Rus, director of the MIT Computer Science and Artificial Intelligence Laboratory, told Bloomberg that decades of scaling have produced "huge models that have a very large energy costs and also very large water costs, and this translates into a big environmental footprint"
1
. This realization led Rus to co-found LiquidAI, which builds task-specific models using only 1,000 GPUs—a fraction of the resources employed by major AI labs1
. OpenAI, by contrast, expected to bring more than one million GPUs online by the end of last year1
.The operational costs of Large Language Models reveal striking inefficiencies. Simple courtesies like saying "please" and "thank you" to ChatGPT reportedly cost OpenAI tens of millions of dollars in energy and power costs
1
. Ramin Hasani, co-founder of LiquidAI, argues that smaller, specialized AI can match larger cloud-based counterparts on specific tasks while using "orders of magnitude less amounts of energy consumption"1
.Researchers from Nvidia and the Georgia Institute of Technology argued in a paper that insisting on large models for routine tasks "reflects a misallocation of computational resources" that is "economically inefficient and environmentally unsustainable at scale"
1
. They describe shifting workloads to smaller models as not merely a technical refinement, but a "moral" obligation .For businesses, AI sustainability directly impacts enterprise adoption viability. Power consumption linked to AI workloads is projected to grow by approximately 15% per year, far outpacing growth across other sectors
4
. Most companies don't need the largest possible model—they need systems that deliver reliable results at predictable operational costs4
.Advances in model optimization are challenging assumptions that smaller models sacrifice accuracy. Techniques such as compression, pruning, and optimization allow compressed models to be up to 95% smaller while maintaining performance on real-world tasks
4
. This reduction translates directly into lower energy consumption, faster inference, and decreased cooling systems requirements4
.The shift toward efficiency enables new deployment models beyond cloud computing. Smaller AI models can run locally on smartphones, laptops, and industrial appliances, reducing latency and limiting dependence on centralized AI infrastructure
4
. For many use cases, this represents both a practical advantage and a sustainability win that addresses growing pressure to meet ESG commitments4
.Yet inertia persists. The enormous capital already committed to existing centralized systems—with US tech giants collectively estimated to invest some $650 billion in AI infrastructure this year alone—makes it harder to question whether the underlying approach still makes sense
1
. Small models attract less marketing attention than the promise of superintelligence, even when better suited for business applications1
. The optimistic scenario involves advanced nuclear energy and renewable energy sources replacing fossil fuels, with AI tools inventing climate solutions. But as Princeton climate modeler Jesse Jenkins notes, "the market has converged on Add gas now, and then add nuclear later"3
.Summarized by
Navi
[1]
[3]
[4]
1
Technology

2
Science and Research

3
Startups
