Submitted by imgonnarelph t3_11wqmga in MachineLearning
mycall t1_jd0yi8i wrote
Reply to comment by currentscurrents in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
> if you're not shuffling the entire network weights across the memory bus every inference cycle
Isn't this common though?
Viewing a single comment thread. View all comments