I’m sure you’ve heard of ChatGPT, Copilot, or any of the other many (many, many, many) AI tools hitting the tech scene. But something a little more nuanced is coming to light–and that’s the environmental costs. I mean, ‘L’ in LLM doesn’t exactly stand for ’low-impact’.
So let’s dive into the environmental cost of that LinkedIn post you just generated, and the realities of training large language models.
Measuring large language model energy consumption
Diving into this subject, I found out that GPT-3, of which GPT-3.5 is a subclass, used around 1287 MWh to be trained (this is equivalent to the annual electricity consumption of an average household in the United States.) This was based on the understanding that they took 405 years (based on machine hours) to train in 2020 on V100 machines.