Submitted by starstruckmon t3_1027geh in MachineLearning
C0hentheBarbarian t1_j2sl0n3 wrote
Reply to comment by Purplekeyboard in [R] Massive Language Models Can Be Accurately Pruned in One-Shot by starstruckmon
What about BLOOMZ? Isn’t it fine tuned in a similar way to GPT-3? Instruction fine tuned?
yahma t1_j2ssc01 wrote
I wasn't very impressed with BLOOMZ. Responses seem short and optimized for Q/A style output. Perhaps Zero-Shot and single-shot worked better than Bloom, but Bloom seemed to produce better output for stories or writing in general.
I was only able to test the 6B models though, so not sure how the 176B models compare.
Viewing a single comment thread. View all comments