ChatGPT may need 30,000 NVIDIA GPUs. Should PC gamers be worried?

NVIDIA A100 GPU
(Image credit: NVIDIA)

What you need to know

  • ChatGPT will require as many as 30,000 NVIDIA GPUs to operate, according to a report by research firm TrendForce.
  • Those calculations are based on the processing power of NVIDIA's A100, which costs between $10,000 and $15,000.
  • It's unclear if gamers will be affected by so many GPUs being dedicated to running ChatGPT.

ChatGPT stormed onto the tech scene near the end of 2022. By January of this year, the tool reportedly had 100 million monthly users. It would only get more users in the following month, as Microsoft announced a new version of Bing powered by ChatGPT. With the technology receiving so much interest and real-world use, it's not surprising that companies stand to profit from ChatGPT's popularity.

According to research firm TrendForce, NVIDIA could make as much as $300 million in revenue thanks to ChatGPT (via Tom's Hardware). That figure is based on TrendForce's analysis that ChatGPT needs 30,000 NVIDIA A100 GPUs to operate. The exact figure that NVIDIA stands to make will depend on how many graphics cards OpenAI needs and if NVIDIA gives the AI company a discount for having such a large order.

If NVIDIA prioritizes placing components within graphics card meant to run ChatGPT, it could affect the availability of other GPUs.

ChatGPT uses a different type of GPU than what people put into their gaming PCs. An NVIDIA A100 costs between $10,000 and $15,000 and is intended to handle the demand ChatGPT places on systems. So, it may seem like gamers would be unaffected by such a large number of A100 GPUs being used to operating ChatGPT but that's not entirely the case.

The global chip shortage was caused by a shortage of chips, not entire graphics cards. Because of that, we saw strange situations such as car companies and computer graphics card makers competing for the same components. Of course, most cars don't have gaming GPUs, Tesla being the exception, but chips are used in manufacturing, computing, and many other industries.

If NVIDIA prioritizes placing components within graphics card meant to run ChatGPT, it could affect the availability of other GPUs.

Another factor in all of this is Microsoft and OpenAI's recently extended partnership. While ChatGPT may require processing power equivalent to 30,000 NVIDIA A100 GPUs, that power could come in the form of Microsoft's own Azure systems. If that's the case, the number of physical GPUs needed to run ChatGPT would be drastically different.

Sean Endicott
News Writer and apps editor

Sean Endicott is a tech journalist at Windows Central, specializing in Windows, Microsoft software, AI, and PCs. He's covered major launches, from Windows 10 and 11 to the rise of AI tools like ChatGPT. Sean's journey began with the Lumia 740, leading to strong ties with app developers. Outside writing, he coaches American football, utilizing Microsoft services to manage his team. He studied broadcast journalism at Nottingham Trent University and is active on X @SeanEndicott_ and Threads @sean_endicott_.