How sustainable are Generative AI models?

Nina Habicht • July 31, 2023

This blog post outlines why large language models, thus generative AI models need so much power to be trained, deployed and maintained. We also look why some solutions are more sustainable compared to others.


What impacts the carbon footprint of large language models?


There are three main values that impact the carbon footprint of llms like GPT-4:


1) The footprint of the training model

2) The footprint from inference. "Inference" in large language models is referring to the capability of llms predicting outcomes using new input data. The models essentially use past knowledge to make educated guesses about the meaning of new sentences or to predict what comes next in a conversation. It's like connecting the dots using what they've learned from reading lots of text.

3) The footprint needed to produce all the required hardware and capabilities of the cloud data center.


Most energy-intensive is the trainint part of such models. Importantly, larger models do use more energy during their deployment. However, this study shows that inference is also a major consumer of energy, with up to 90% of the ML workloads is due to inference processing.


How high are the energy-costs of a large language models used in generative AI?


We summarized some important facts for you to shed lights on the environmental impact of large language models:


  • The Megatron Turing model from NVIDIA needed hundreds of NVIDIA DGX A100 multi-GPU servers, each using up to ~ 6.5 kilowatts of power (1).
  • To train one BERT model (LLM by Google) is roughly the same amount of energy and carbon footprint as a trans-Atlantic flight (2).
  • Researchers from the paper "On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜" explain that training models such as OpenAI's GPT-4 or PaML from Google uses 300 tons of CO2 (= one average person uses 5 tons a year) as outlined by HBR.


What is a sustainable large language model strategy?


If possible, use existing, pre-trained models from llm providers such as Microsoft OpenAI, Google (bard, PaLM), Meta AI (LlaMA) and do not create your own generative models (sidemark: this also costs around $ 100 Mio. according to a recently published article on HuggingFace. You should - instead - still fine-tune a model for your use case.


What are Best-practices to build greener llms?


  • Besides pre-trained model usage, use small models (with less parameters) by removing unnecessary parameters (so-called "pruning") without jeopardizing the accuracy you need (e.g. your model should still answer with a good confidence)
  • Reduce training time by experimenting with a distilled version of BERT, DistilBERT. "Distilled models are also been shown to be more energy efficient" as outlined in sustainable AI in the cloud.
  • Fine-tune your model with own data for some period instead of training it from scratch
  • Use cloud-based environments which provide scalable infrastructure (Azure, AWS, Google)
  • There is specialized hardware that support training speed (GraphcoreHabana) and inference (Google TPUAWS Inferentia)
  • Other technical tricks: merge model layers (so-called "fusion"), store model parameters in smaller values (say, 8 bits instead of 32 bits, so-called "quantization")
  • Run ML models on small, low-powered edge devices without need to send the data to the server to process). Use for example  TinyML.
  • Monitor carbon footprint via tooling such as  CodeCarbonGreen algorithms, and ML CO2 Impact.
  • Encourage your data science team to set benchmarking standards and include sustainability in their model considerations
  • Finally, think if you really need Generative AI at all. If you need help we can support you by evaluating your use case.



Summary about sustainable generative AI


Please find a comparison between llms and their sizes in the previous blog. Additional material can be found here:



Responsible AI is a crucial foundation for all generative AI products. It is important that we consider carbon footprint with AI models to protect our earth and future. However, some studies also outline the difference between AI generated content vs. human generated content when it comes to emissions. This paper can be found here.


Need support with your Generative AI Strategy and Implementation?

🚀 AI Strategy, business and tech support 

🚀 ChatGPT, Generative AI & Conversational AI (Chatbot)

🚀 Support with AI product development

🚀 AI Tools and Automation

Get in touch
By Nina Habicht April 29, 2025
AI-powered chatbots, whether developed in-house or deployed through trusted platforms, are revolutionizing customer service, knowledge access, and internal communication. However, alongside these opportunities come new legal obligations: data protection , transparency , and EU AI Act compliance must be addressed carefully. This article covers: Where AI chatbots bring business value What compliance risks you must manage How to implement AI chatbots successfully and securely
By Nina Habicht March 25, 2025
What's RAG? The goal is for the language model is not to draw on its own knowledge (from the model), but for information to be enriched in the prompt. This is usually your own data you provide to the model (PDFs, systems).
Video Creation: The Ultimate Guide to Runway, Luma AI, Haiper.ai, and Hailuo AI
By Nina Habicht February 16, 2025
Video Creation: The Ultimate Guide to Runway, Luma AI, Haiper.ai, and Hailuo AI
What are the best AI powerpoint tools
By Nina Habicht February 16, 2025
What are the best AI powerpoint tools. Discover tools that create presenations with AI.
Image Creator Tools
By Nina Habicht November 24, 2024
Ultimative review of all relevant image creation tools
Optimizing your Website for AI: How to get found by ChatGPT
By Nina Habicht August 24, 2024
Optimizing your Website for AI: How to get found by ChatGPT. This article provides concrete Large Language Model Optimization strategies for SMEs and companies.
A Practical Guide for Midjourney Image Generation. Learn how to create professional images.
By Nina Habicht August 24, 2024
Since August 2024, users have been able to use the web version of the image creation tool Midjourney. This simplifies usage by providing a user-friendly interface to experiment with one of the top Generative AI image creation tools available. We tested it for you and are sharing helpful tips and tricks. How to prompt images with Midjourney? If you use Midjourney on discord, there is a clear prompt structure and prompt parameters to adhere to. Usually, it makes sense to stick to it: 1) To prompt use "/Imagine" 2) Then enter your subject (description and details) you want to see on the image and it's environment (see yellow highlighted below in the prompt example) 3) Then enter composition, lightning, colours (see green highlighted below in the prompt example) 4) Finally add technical parameters to adjust and finalize your image. Please find a useful parameter library here.
Agentic AI vs. Gen. AI vs. RPA
By Nina Habicht August 11, 2024
This article explains agentic AI and why it is so important when building generative AI and chatbot applications. Overview about Agentic AI vs. Gen. AI vs. RPA and all you need to know about these terms.
LLM Benchmarks: Finding the right LLM for your Needs
By Nina Habicht July 29, 2024
LLM Benchmarks: Finding the right LLM for your Needs
RAG vs. Finetuning. Open Source vs. Proprietary Models
By Nina Habicht June 25, 2024
RAG vs. Finetuning. Open Source vs. Proprietary Models. We explain what makes sense when.
Show More