r2m2 t1_j64uah5 wrote on January 27, 2023 at 6:47 PM Reply to [R] SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot by Secure-Technology-78 Isn’t this a (somewhat) well-known “free lunch” effect w/ naive one-shot magnitude pruning? I feel like this is a folklore fact for many models like ResNet/VGG (& a paper from a few years back validated the same for BERT) Permalink 2
r2m2 t1_j64uah5 wrote
Reply to [R] SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot by Secure-Technology-78
Isn’t this a (somewhat) well-known “free lunch” effect w/ naive one-shot magnitude pruning? I feel like this is a folklore fact for many models like ResNet/VGG (& a paper from a few years back validated the same for BERT)