Is there a More Environmentally Friendly Way to Train Artificial Intelligence?

Print Friendly, PDF & Email

In this special guest feature, Omri Geller, Co-founder and CEO at Run:AI, takes a timely and interesting look at one of the most pressing issues facing the computing industry by an accomplished data scientist. Run:AI is a company providing virtualization software for AI infrastructure. Omri earned a masters degree in Electrical and Electronics Engineering at Tel Aviv University.

Machine learning is changing the world, and it’s changing it fast. In just the last few years, it’s brought us virtual assistants that understand language, autonomous vehicles, new drug discoveries, AI-based triage for medical scans, handwriting recognition and more. 

One thing that machine learning shouldn’t be changing is the climate. 

The issue relates to how machine learning is developed. In order for machine learning (and deep learning) to be able to accurately make decisions and predictions, it needs to be “trained.” 

Imagine an online marketplace for selling shoes, that’s been having a problem with people trying to sell other things on the site – bikes and cats and theater tickets. The marketplace owners decide that they want to limit the site to shoes only, by building an AI to recognize photos of shoes and refusing any listing without shoes in the picture. 

The company collects tens of thousands of photos of shoes, and a similar number of photos with no shoes. It hires data scientists to design a complicated mathematical model and convert it into code. And then they start training their shoe-detecting machine learning model. 

This is the vital part: the computer model looks at all the pictures of shoes and tries to work out what makes them “shoey.” What do they have that the non-shoe pictures don’t? Without getting too bogged down in technical details, this process takes a lot of computing resources and time. Training accurate machine learning models means running multiple chips like GPUs, at full power, 24 hours a day for weeks or months as the models are trained, tweaked and refined. 

As well as time and expense, AI training uses a lot of energy. Modern computer chips use only minimal power when they are idle, but when they’re working at full capacity they can burn through electricity, generating masses of waste heat (which also needs to be pumped out using cooling systems that use, yup, more energy).

Any major energy use has implications for climate change, as most of our electricity is still generated from fossil fuels, producing carbon dioxide as they burn. One recent study from the University of Massachusetts claimed that training a single advanced language-processing AI produced 626,000lb of CO2, the same amount as five cars would produce over their lifetimes! 

In fact, a team from Canada’s Montreal Institute for Learning Algorithms (MILA) released the machine learning emissions calculator in December of last year to help researchers in the field of AI estimate how much carbon is produced in training their machine learning models.

This problem is getting worse, as data scientists and engineers solve more complicated AI problems by throwing more power at them, using bigger and more expensive computing to solve hard problems rather than focusing on efficiency. 

GPT-3, the AI-powered language model recently released by OpenAI, was trained on 45 terabytes of text data (the entirety of the English Wikipedia, spanning some 6 million articles, makes up only 0.6 percent of its training data) with the environmental costs of this hyper-powerful machine-learning technology still unknown.

To be fair, other computing processes are also on a worrying trajectory. A study by ICT specialist Anders Andrae found that according to his most optimistic projections, by 2030, the ICT industry, which delivers Internet, video, voice and other cloud services would be responsible for 8% of the total worldwide energy demand while his realistic projection put that number at 21% – with data centers using more than one-third of that. 

One of the key recommendations of the University of Massachusetts research for reducing the waste caused by AI training was “a concerted effort by industry and academia to promote research of more computationally efficient algorithms, as well as hardware that requires less energy”.  

Software can also be used to increase hardware efficiency, thereby reducing the computational power needed for AI models but perhaps the biggest impact will come from the use of renewable energy sources for the data centers themselves. Facebook’s Odense, Denmark data center is said to run completely on renewable energy sources. Google has its own energy-efficient data centers, like this one in Hamina, Finland. 

In the very long term, as the world’s industrial economies move away from fossil fuels, perhaps the link between computational load and CO2 production will be broken and perhaps all machine learning will be carbon-neutral. Even longer-term, deep learning on weather and climate patterns could help humanity gain a better understanding of how to combat and even reverse climate change. 

But until then, responsible businesses should consider the carbon impact of their new technologies including machine learning, and take steps to measure the carbon cost of their model development by upgrading efficiency in development, software and hardware.

Sign up for the free insideBIGDATA newsletter.

Join us on Twitter: @InsideBigData1 –

Speak Your Mind



  1. Our work becomes easy when you have some automated help to back your approach or your work techniques. The information that you have shared about machine learning and its different ways to train artificial intelligence was very helpful.
    Thank you for listing it together and sharing this useful information with us and letting us know more about it. Keep sharing more such information like this with us in the future also.