Submitted by Vegetable-Skill-9700 t3_121agx4 in deeplearning
Vegetable-Skill-9700 OP t1_jdpqefg wrote
Reply to comment by FirstOrderCat in Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
But do we really need all that info in most of the practical use cases? Say, I am using LM to write Reddit posts, probably, it only needs to learn subjects I write about along my style of writing. A well-trained model on a highly refined dataset (which has high-quality examples of my posts) should perform better than GPT-4?
FirstOrderCat t1_jdpqp5l wrote
I do not know, it is hard to say if you will be able to create sufficient dataset for your case.
Viewing a single comment thread. View all comments