[D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM Submitted by head_robotics t3_1172jrs on February 20, 2023 at 9:33 AM in MachineLearning 51 comments 220
marcus_hk t1_j9g5hns wrote on February 21, 2023 at 6:05 PM Seems it shouldn't be too difficult to run one stage or layer at a time and cache intermediate results. Permalink 1
Viewing a single comment thread. View all comments