What. The. ***k. [less than 1B parameter model outperforms GPT 3.5 in science multiple choice questions]
Submitted by Destiny_Knight t3_118svv7 in singularity
Reply to comment by turnip_burrito in What. The. ***k. [less than 1B parameter model outperforms GPT 3.5 in science multiple choice questions] by Destiny_Knight
What is the "catch" here? It sounds too good to be true
The catch is that it only outperforms large models in a narrow domain of study. It's not a general purpose tool like the really large models. That's still impressive though.
Can It be fine tuned ?
You can tune it to another data set and probably get good results, but you have to have a nice, high quality data set to work with.
I’m working on one that’s trained on JFK speeches and Bachlorette data to help people with conversation skills.
I can't tell if this is a joke or real
It’s real. Gonna launch after GME moons
Sounds like a viable AI implementation to me. I'll be your angel investor and throw some Doge your way or something.
I don't think that's true, but I do believe it was finetuned on the specific dataset to achieve the SOTA result they did.
It chooses the correct answer from multiple choices. it isn't actually comparable to chatGtp.
Viewing a single comment thread. View all comments