Solving Big AI’s Big Energy Problem

Solving Big AI’s Big Energy Problem


It seems that the more ground-breaking deep learning models are in AI, the more massive they get. This summer’s most buzzed-about model for natural language processing, GPT-3, is a perfect example. To reach the levels of accuracy and speed to write like a human, the model needed 175 billion parameters, 350 GB of memory and $12 million to train (think of training as the “learning” phase). But, beyond cost alone, big AI models like this have a big energy problem.  UMass Amherst researchers found that the computing power needed to train a large AI model can produce over 600,000 pounds…

This story continues at The Next Web
  1. No comment added yet!
  2. Leave Your Comment

    Your email address will not be published. Required fields are marked *

AtSign Innovations