massimosclaw2
massimosclaw2 t1_ix2w5bc wrote
Reply to comment by visarga in [D] Are researchers attempting to solve the ‘omnipotence’ requirement problem in LLMs? by [deleted]
Not quite. I think there’s value to this technique but it’s still constrained by probability of what GPT thinks an AI would say based on all the instances of similar texts in the data it consumed, which is not quite the same thing
massimosclaw2 t1_ishdjbw wrote
Reply to [R] UL2: Unifying Language Learning Paradigms - Google Research 2022 - 20B parameters outperforming 175B GTP-3 and tripling the performance of T5-XXl on one-shot summarization. Public checkpoints! by Singularian2501
I wonder how this will perform on out of distribution stuff + remembering obscure references like "Alfred Korzybski" (as GPT-3 does), and what they are related to or if 20B parameters is too small to memorize enough
massimosclaw2 t1_ir947iu wrote
Can't wait til this actually starts getting indistinguishable in terms of quality
massimosclaw2 t1_jdmvjlp wrote
Reply to comment by learn-deeply in [R] Reflexion: an autonomous agent with dynamic memory and self-reflection - Noah Shinn et al 2023 Northeastern University Boston - Outperforms GPT-4 on HumanEval accuracy (0.67 --> 0.88)! by Singularian2501
When you haven’t done much, best to obscure it in some complicated language /s