Submitted by visarga t3_zzpe5h in singularity
So this time the question was - how can language models generalise better? Well, when you are in unfamiliar situations, you usually rely on analogies to make sense of the situations. How good is GPT-3 at analogies?
In order to test that it was necessary to create a task that was never seen before in its training, so it would be really unfamiliar.
> We found that GPT-3 displayed a surprisingly strong capacity for abstract pattern induction, matching or even surpassing human capabilities in most settings. Our results indicate that large language models such as GPT-3 have acquired an emergent ability to find zero-shot solutions to a broad range of analogy problems.
And these analogy problems are very dry. Just meaningless abstract patterns, nothing to relate to. It explains why GPT-3 is so good at coding and symbol manipulation.
[Emergent Analogical Reasoning in Large Language Models] (https://arxiv.org/abs/2212.09196v1)
wu_wey t1_j2d2z4z wrote
This is a gem. Thanks for sharing the paper.