Yardanico
Yardanico t1_jdls342 wrote
Reply to comment by wojtek15 in [D] Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
Yeah, I think there's a lot of overhyping going around "running ChatGPT-grade language models on consumer hardware". They can "follow" instructions they same way as ChatGPT, but obviously those models know far, far less than the ClosedAI models do, and of course they'll hallucinate much more.
Although it's not an entirely bad thing, at least the community will innovate more so we might get something interesting in the future from this "push" :)
Yardanico t1_jc49dlf wrote
Reply to comment by londons_explorer in [R] Stanford-Alpaca 7B model (an instruction tuned version of LLaMA) performs as well as text-davinci-003 by dojoteef
It does actually send a WebSocket request to join the queue and waits for it to complete.
Yardanico t1_jdstn3v wrote
Reply to [P] Using ChatGPT plugins with LLaMA by balthierwings
Have the author seen https://github.com/hwchase17/langchain? I think this is exactly the problem they're trying to solve.