Curioso sobre a configuração da GPU por trás do ChatGPT
Olá a todos, tenho-me perguntado sobre o hardware que está a funcionar nos bastidores do ChatGPT. Por exemplo, quantas placas gráficas o alimentam? Estou simple…
Daniel Sloan
February 8, 2026 at 08:58 PM
Olá a todos, tenho-me perguntado sobre o hardware que está a funcionar nos bastidores do ChatGPT. Por exemplo, quantas placas gráficas o alimentam? Estou simplesmente curioso para saber se é um pequeno número ou uma configuração gigantesca. Alguém tem informações ou suposições?
Adicionar comentário
Comentários (23)
Actually, the training phase uses way more GPUs than inference. ChatGPT responses come from a smaller cluster compared to the massive training setup.
I wonder if in the future ChatGPT will shift to newer GPUs or maybe custom AI chips.
The latency for ChatGPT answers is surprisingly low considering the hardware behind it.
I bet there's a whole team managing the GPU clusters 24/7 to keep ChatGPT running smooth.
I think it's gotta be a huge number, maybe hundreds? These models are insanely demanding.
I’d love to see a tour of one of these AI data centers to get a feel for the scale.
I doubt they just use consumer GPUs. It's gotta be some specialized hardware with lots of them networked together.
I'm guessing the exact number is a secret to keep competitive advantage?
I read somewhere they use thousands of GPU cores during training, but that’s spread across many physical GPUs.
GPUs are expensive but essential for AI, so no surprise they use lots of them for something like ChatGPT.
In short, it’s a massive, distributed GPU setup with hundreds or thousands of GPUs involved at some stage.
The inference side probably uses fewer GPUs but more optimized to handle many requests simultaneously.
Do you think the number of GPUs affects the cost of using ChatGPT?
I think ChatGPT runs on clusters with maybe a few hundred GPUs during training and less during inference.
Does anyone know if they use TPUs instead of GPUs? Heard Google’s TPU stuff is pretty powerful too.
Isn't energy consumption for all those GPUs a huge concern? Wonder how they manage that.
From what I heard, companies like OpenAI use GPU clusters with Nvidia A100s or similar, so maybe dozens or even hundreds of GPUs in total.
Anyone know if the GPU numbers have changed over time as the model got bigger?
I've seen Nvidia talk about AI workloads needing clusters of A100 GPUs, so ChatGPT probably uses setups like that too.
I've heard that a single training run can use thousands of GPU hours. So that means lots of GPUs working in parallel.
Anyone else curious about the cooling setup for all those GPUs running 24/7? Must be insane.
Is it possible they use cloud GPUs instead of owning all the hardware themselves?
You can also check ai-u.com for new or trending tools and info on related AI tech, might have some hints on hardware setups.