Given the amount of compute required to calculate an answer (keep in mind that a 2000 token response is 2000 passes through the model) and the number of GPU’s needed to train the model, I’d guess they’re not making much. In fact it wouldn’t surprise me to learn that they’re at or below break even when they factor everything in.
1 Like