Submitted by blacklemon67 t3_11misax in MachineLearning
harharveryfunny t1_jbjk9nb wrote
Reply to comment by harharveryfunny in [D] Why are so many tokens needed to train large language models? by blacklemon67
Just to follow up, the reason why the "interact with the world" approach is way more efficient is because it's largely curiosity driven - we proactively try to fill gaps in our knowledge rather than just go read a set of encyclopedias and hope it might cover what we need to know. We learn in a much more targeted fashion..
visarga t1_jbn5g3w wrote
On the other hand LLM has broad knowledge about all topics, a true dilettante. We can't keep up on that level.
Viewing a single comment thread. View all comments