Submitted by minimaxir t3_11fbccz in MachineLearning
farmingvillein t1_jajw0yj wrote
Reply to comment by badabummbadabing in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
> The training costs lie in the low millions (10M was the cited number for GPT3), which is a joke compared to the startup costs of many, many industries. So while this won't be something that anyone can train, I think it's more likely that there will be a few big players (rather than a single one) going forward.
Yeah, I think there are two big additional unknowns here:
-
How hard is it to optimize inference costs? If--for sake of argument--for $100M you can drop your inference unit costs by 10x, that could end up being a very large and very hidden barrier to entry.
-
How much will SOTA LLMs really cost to train in, say, 1-2-3 years? And how much will SOTA matter?
The current generation will, presumably, get cheaper and easier to train.
But if it turns out that, say, multimodal training at scale is critical to leveling up performance across all modes, that could jack up training costs really, really quickly--e.g., think the costs to suck down and train against a large subset of public video. Potentially layer in synthetic data from agents exploring worlds (basically, videogames...), as well.
Now, it could be that the incremental gains to, say, language are not that high--in which case the LLM (at least as these models exist right now) business probably heavily commoditizes over the next few years.
Viewing a single comment thread. View all comments