- Selling those requests at less money than it cost to run the compute for those requests (because if you raise price clients go to openai)
The statements are not contradicting each other?
They keep subsidizing to try to grow customer base, but they can't serve the customer base they have, they're expecting customer base grows faster than it drops from people bothered with rate limits (it probably will, average user won't hit rate limits enough to change)
Probably expecting a breakthrough in efficiency for compute, or getting enough cash flow (IPO?) to get more compute before it all comes crashing down
Model inference compute over model lifetime is ~10x of model training compute now for major providers. Expected to climb as demand for AI inference rises.