How to Make AI Less of a Power Guzzler
The future of artificial intelligence may depend, in part, on whether providers can reduce their appetite for electricity and water
The future of artificial intelligence may depend, in part, on whether providers can reduce their appetite for electricity and water
Artificial intelligence is poised to transform both work and everyday life. But it has a dark underside: AI computer centres consume enormous amounts of electricity and water, to power their processing chips and cool the heat they emit.
Annual U.S. electricity use by data centres of all types will rise from 3% to 4% of the nation’s total today to between 11% and 12% in 2030, with AI being the main driver, according to projections from consulting firm McKinsey.
Meantime, AI’s demand for water globally in 2027 could account for more than the total annual amount withdrawn for use in Denmark or half of that in the U.K., according to researchers at the University of California, Riverside and University of Texas at Arlington.
All of that heavy use is causing logistical and public-image problems for the industry. Some utilities struggle to supply the needs of AI providers, and communities push back, fearing the added use will boost power prices and deplete water supplies.
The biggest AI providers, including Amazon , Alphabet Inc.’s Google, Meta and Microsoft , say they are working to be both carbon-neutral and replenish more water than they use—even as they continue to build massive data centres.
“It will be harder to build data centres, especially where energy already is at a premium or water might be scarce,” says Ed Anderson, research vice president at technology advisory firm Gartner. But, he adds, “the economic opportunity is rich enough that the providers will find a way.”
Below are some of the steps tech companies and researchers are hoping will reduce AI’s appetite for power and water.
One way of addressing power consumption is to make chips less power hungry. Nvidia , the largest maker of AI processors, says its newest ones, called Blackwell, will be about 25 times as energy efficient as its previous high-end version. Meanwhile, Amazon, Google, Meta and Microsoft are designing their own processing chips, in part to cut costs but also to make them use less power.
“Each generation has been significantly more efficient than the prior one,” says Google’s Partha Ranganathan , a vice president and engineering fellow, speaking of his company’s processing units.
Equipment used to cool data centres creates another issue: where to get the vast amount of water these systems consume. Google says its data centers globally used about 6.1 billion gallons of water in 2023, equivalent to the water used to irrigate and maintain 40 golf courses in the Southwest each year.
OpenAI’s GPT-3 model, meantime, consumes the equivalent of a 16.9-ounce bottle of water for every 10 to 50 responses it provides to users’ queries, according to the researchers at UC Riverside and UT Arlington. OpenAI declined to comment on the finding.
Data-centre water typically comes from municipal water systems. But in an era of water shortages, diverting drinking water for an industrial use has created tensions in some locales. That has sent AI companies searching for other sources, including rainwater, treated wastewater or water left over from factory processes.
Amazon, for example, uses recycled wastewater for cooling at its Santa Clara, Calif., data centers. The water comes from the city’s sewage-treatment system after it undergoes a three-step process that removes 99% of impurities.
Some researchers have experimented with carefully controlling what kind and how much information an AI model takes in during training. Usually, training a so-called large language model AI, such as OpenAI’s ChatGPT and Microsoft’s Copilot, involves ingesting hundreds of billions of words from the internet and elsewhere, then learning the relationships among them.
And that is energy and water intensive. Training an AI model called BLOOM over a 3½-month period consumed enough electricity to power the average U.S. home for 41 years, according to a Stanford University report.
As for water, training one of Google’s AI models, known as LaMDA, used about two million liters of it, both to produce the electricity used and keep the computers cool—enough to fill about 5,000 bathtubs, according to Shaolei Ren , a professor of electrical and computer engineering at the University of California, Riverside. Google declined to comment on the research, but said it is “committed to climate-conscious cooling of our data centres.”
One possible solution is to have AIs remove redundancy and low-quality data, instead of just vacuuming up the whole internet. The goal is a much smaller set of data that the AI system can more easily sift through when a user asks it a question.
This can lower electricity consumption, according to some researchers.
AI systems that limit the information they take in are also less likely to “hallucinate”—give false or misleading answers—and can respond in ways that are more on-point because of the higher quality of the data they contain, experts say. Microsoft found that one of its pared-down AIs exceeded that of vastly larger ones in measurements of common sense and logical reasoning .
Researchers at several universities have found that capping the amount of electricity used by AI computers has only a minor effect on the outcome, such as slightly more processing time.
Experts at the Massachusetts Institute of Technology and Northeastern University say that reducing the power to one of Meta’s AIs by 22% to 24% slowed the speed at which the AI responded to a query by only 5% to 8%. “These techniques can lead to significant reduction in energy consumption,” the researchers say. They add that the method also caused the processors to run at a lower temperature—which could trim the need for cooling.
Meta declined to comment on the research, but said it has had efforts to boost data-centre energy efficiency “since we started designing our first data center over a decade ago.”
Meantime, a team at the University of Michigan, University of Washington and University of California, San Diego devised an algorithm to modulate the use of power during training. The technique could cut power use by up to 30% , they say.
Some researchers believe companies should give users more context about the environmental impact of AI, to let them make more-informed decisions about the technology. Ren, of UC Riverside, proposes that AI providers disclose the approximate amount of electricity and water each query consumes—akin to how Google tells people searching for flights the amount of carbon emissions each trip would create.
Another proposal is to devise a rating system for the power efficiency of AI systems, akin to the government’s Energy Star ratings for home appliances and other products. Such a system could help people choose AI models for differing tasks based on their energy consumption, according to Sasha Luccioni , an AI researcher at Hugging Face, a company that makes machine-learning tools.
Academics and others have come up with other proposals to minimise AI’s environmental impact by tapping into green energy. For instance, companies might build more data centers in countries with abundant, low-emission power, such as hydropower in Norway or geothermal in Iceland. Or companies might do AI calculations at different locations at different times of the day, such as deploying computer centers with high use of solar power during the daytime or wind-powered ones when wind is more reliable at night.
Data-centre computers put out tremendous amounts of heat, and their temperature must be kept in a certain range, often 64 to 72 degrees, to prevent damaging the electronics. Traditionally, this has been done by high-power air conditioning. But air conditioning uses up to 40% of all the electricity consumed by a typical data centre, while devices called cooling towers that expel the heat to the outside air use a lot of water.
In response, the data-centre industry is moving to liquid cooling, which circulates a special liquid or cold water to “cold plates” that sit on top of the processor chips and keep them at a safe and efficient temperature range. The system, called direct-to-chip liquid cooling, uses less power than the traditional method—about 30% less, Nvidia says—because liquid is vastly better at removing heat from the electronics than blowing cold air over them.
Another method under development, called immersion cooling, involves placing the computers themselves inside big tanks of cooling liquid. While showing early promise, there are environmental concerns about the chemicals often used in the setup, says Mark Russinovich , chief technology officer of Microsoft’s Azure cloud-computing unit.
Some companies, meanwhile, are using computing gear that can withstand higher temperatures and doesn’t need as much cooling. Google says its data centres already are 1.8 times as energy efficient as the typical data centre, which it achieved in part by raising the inside temperature to 80 degrees. For every one-degree boost in their temperature, data centres can save 4% to 5% in energy costs, according to the Energy Star program.
A divide has opened in the tech job market between those with artificial-intelligence skills and everyone else.
A 30-metre masterpiece unveiled in Monaco brings Lamborghini’s supercar drama to the high seas, powered by 7,600 horsepower and unmistakable Italian design.
A divide has opened in the tech job market between those with artificial-intelligence skills and everyone else.
There has rarely, if ever, been so much tech talent available in the job market. Yet many tech companies say good help is hard to find.
What gives?
U.S. colleges more than doubled the number of computer-science degrees awarded from 2013 to 2022, according to federal data. Then came round after round of layoffs at Google, Meta, Amazon, and others.
The Bureau of Labor Statistics predicts businesses will employ 6% fewer computer programmers in 2034 than they did last year.
All of this should, in theory, mean there is an ample supply of eager, capable engineers ready for hire.
But in their feverish pursuit of artificial-intelligence supremacy, employers say there aren’t enough people with the most in-demand skills. The few perceived as AI savants can command multimillion-dollar pay packages. On a second tier of AI savvy, workers can rake in close to $1 million a year .
Landing a job is tough for most everyone else.
Frustrated job seekers contend businesses could expand the AI talent pipeline with a little imagination. The argument is companies should accept that relatively few people have AI-specific experience because the technology is so new. They ought to focus on identifying candidates with transferable skills and let those people learn on the job.
Often, though, companies seem to hold out for dream candidates with deep backgrounds in machine learning. Many AI-related roles go unfilled for weeks or months—or get taken off job boards only to be reposted soon after.
It is difficult to define what makes an AI all-star, but I’m sorry to report that it’s probably not whatever you’re doing.
Maybe you’re learning how to work more efficiently with the aid of ChatGPT and its robotic brethren. Perhaps you’re taking one of those innumerable AI certificate courses.
You might as well be playing pickup basketball at your local YMCA in hopes of being signed by the Los Angeles Lakers. The AI minds that companies truly covet are almost as rare as professional athletes.
“We’re talking about hundreds of people in the world, at the most,” says Cristóbal Valenzuela, chief executive of Runway, which makes AI image and video tools.
He describes it like this: Picture an AI model as a machine with 1,000 dials. The goal is to train the machine to detect patterns and predict outcomes. To do this, you have to feed it reams of data and know which dials to adjust—and by how much.
The universe of people with the right touch is confined to those with uncanny intuition, genius-level smarts or the foresight (possibly luck) to go into AI many years ago, before it was all the rage.
As a venture-backed startup with about 120 employees, Runway doesn’t necessarily vie with Silicon Valley giants for the AI job market’s version of LeBron James. But when I spoke with Valenzuela recently, his company was advertising base salaries of up to $440,000 for an engineering manager and $490,000 for a director of machine learning.
A job listing like one of these might attract 2,000 applicants in a week, Valenzuela says, and there is a decent chance he won’t pick any of them. A lot of people who claim to be AI literate merely produce “workslop”—generic, low-quality material. He spends a lot of time reading academic journals and browsing GitHub portfolios, and recruiting people whose work impresses him.
In addition to an uncommon skill set, companies trying to win in the hypercompetitive AI arena are scouting for commitment bordering on fanaticism .
Daniel Park is seeking three new members for his nine-person startup. He says he will wait a year or longer if that’s what it takes to fill roles with advertised base salaries of up to $500,000.
He’s looking for “prodigies” willing to work seven days a week. Much of the team lives together in a six-bedroom house in San Francisco.
If this sounds like a lonely existence, Park’s team members may be able to solve their own problem. His company, Pickle, aims to develop personalised AI companions akin to Tony Stark’s Jarvis in “Iron Man.”
James Strawn wasn’t an AI early adopter, and the father of two teenagers doesn’t want to sacrifice his personal life for a job. He is beginning to wonder whether there is still a place for people like him in the tech sector.
He was laid off over the summer after 25 years at Adobe , where he was a senior software quality-assurance engineer. Strawn, 55, started as a contractor and recalls his hiring as a leap of faith by the company.
He had been an artist and graphic designer. The managers who interviewed him figured he could use that background to help make Illustrator and other Adobe software more user-friendly.
Looking for work now, he doesn’t see the same willingness by companies to take a chance on someone whose résumé isn’t a perfect match to the job description. He’s had one interview since his layoff.
“I always thought my years of experience at a high-profile company would at least be enough to get me interviews where I could explain how I could contribute,” says Strawn, who is taking foundational AI courses. “It’s just not like that.”
The trouble for people starting out in AI—whether recent grads or job switchers like Strawn—is that companies see them as a dime a dozen.
“There’s this AI arms race, and the fact of the matter is entry-level people aren’t going to help you win it,” says Matt Massucci, CEO of the tech recruiting firm Hirewell. “There’s this concept of the 10x engineer—the one engineer who can do the work of 10. That’s what companies are really leaning into and paying for.”
He adds that companies can automate some low-level engineering tasks, which frees up more money to throw at high-end talent.
It’s a dynamic that creates a few handsomely paid haves and a lot more have-nots.
On October 2, acclaimed chef Dan Arnold will host an exclusive evening, unveiling a Michelin-inspired menu in a rare masterclass of food, storytelling and flavour.
A divide has opened in the tech job market between those with artificial-intelligence skills and everyone else.