[R] SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot Submitted by Secure-Technology-78 t3_10mdhxb on January 27, 2023 at 5:48 AM in MachineLearning 29 comments 210
r2m2 t1_j64uah5 wrote on January 27, 2023 at 6:47 PM Isn’t this a (somewhat) well-known “free lunch” effect w/ naive one-shot magnitude pruning? I feel like this is a folklore fact for many models like ResNet/VGG (& a paper from a few years back validated the same for BERT) Permalink 2
Viewing a single comment thread. View all comments