Jorycle t1_j86dx1u wrote
Reply to comment by PM_ME_GAY_STUF in Scientists Made a Mind-Bending Discovery About How AI Actually Works | "The concept is easier to understand if you imagine it as a Matryoshka-esque computer-inside-a-computer scenario." by Tao_Dragon
Yeah I work in ML and I don't get what the novel discovery is here based on the article. This all just sounds like... what we already know. Like this line:
>"We show that it is possible for these models to learn from examples on the fly without any parameter update we apply to the model."
That's so routine it's not even interesting.
I'm guessing the actual study goes into what was found, I'll have to read it when I have time.
MrChurro3164 t1_j87j8s7 wrote
Is this something we already know? I’m by no means an AI researcher but the model learning at run time without updating weights seems pretty novel no? What other ‘routine’ models do this?
SignificanceAlone203 t1_j87o8uo wrote
The weights that the AI updates and the "parameters we apply" are quite different. Weights are most definitely updated at run time during training. The fact that it learns without the researcher manually changing parameters is... kind of the whole point of AI.
MrChurro3164 t1_j87pn2y wrote
I think terms are being confused and it’s written poorly. From what I gather, the weights are not being updated, and this is not during training. This is someone chatting with the model and it learns new things “on the fly”.
From another article: > For instance, someone could feed the model several example sentences and their sentiments (positive or negative), then prompt it with a new sentence, and the model can give the correct sentiment. Typically, a machine-learning model like GPT-3 would need to be retrained with new data for this new task. During this training process, the model updates its parameters as it processes new information to learn the task. But with in-context learning, the model’s parameters aren’t updated, so it seems like the model learns a new task without learning anything at all.
jeffyoulose t1_j88xt1n wrote
How is it learning if no weights change? It's best simulating another training just for the session of input given at inference time.
professorDissociate t1_j89xizq wrote
Ah, so we’ve found d the novel discovery by the sound of this confusion then… yes?
Viewing a single comment thread. View all comments