Submitted by RadioFreeAmerika t3_122ilav in singularity
FoniksMunkee t1_jdqs9x9 wrote
Reply to comment by RadioFreeAmerika in Why is maths so hard for LLMs? by RadioFreeAmerika
It's a limitation of LLM's as they currently stand. They can't plan ahead, and they can't backtrack.
So a human doing a problem like this would start, see where they get to, perhaps try something else. But LLM's can't. MS wrote a paper on the state of ChatGPT4 and they made this observation about why LLM's suck at math.
"Second, the limitation to try things and backtrack is inherent to the next-word-prediction paradigm that the model operates on. It only generates the next word, and it has no mechanism to revise or modify its previous
output, which makes it produce arguments “linearly”. "
They argue too that the model was probably not trained on as much mathematical data as code - and more training will help. But they also said the issue above "...constitutes a more profound limitation.".
Viewing a single comment thread. View all comments