>Full ChatGPT 175B parameter model requires 5 (five) A100 GPUs to load the model.
>It takes at least 20 - 40 seconds to answer simple query.
>Compute cost is about 1 cent per query assuming relatively short queries and answers.
There is no way they can monetize anything this expensive with ads. It's also impossible to scale it like some google search? You run out of GPU's in the world (and electricity).