MrTacobeans t1_j7jwmn3 wrote
Reply to comment by el_chaquiste in What Large Language Models (LLMs) mean for the -near- future, from Search to Chatbots to personal Assistants. Some of my thoughts, predictions and hopes - and I would love to hear yours. by TFenrir
I donno stability although it seems like a well funded machine of a organization now, beat openAI incredibly fast at a time when their funding was no where near the level of openAI. All while producing a model that can throw strong punches against DALLE without using multiple industrial GPUs to inference each image.
Now stability has DeepFloyd which is a nebulous/ethereal model under lock&key atm that seems to be completely SOTA just from the base model.
I wouldn't discount the small players, especially the ones that plan on open source. People have done wild things with stable diffusion. The model I'm following right now for LLM, RWKV is creating pretty darn impressive results at 14B parameters. Compared to chatGPT it's ok but the big difference is you need 15k+ of hardware to even inference the chatGPT model. RWKV from it's base model is creating coherent results on consumer hardware. It hasn't even been tuned yet with RL training or q&a data.
Viewing a single comment thread. View all comments