Submitted by imgonnarelph t3_11wqmga in MachineLearning
keeplosingmypws t1_jd5xygm wrote
Reply to comment by KerfuffleV2 in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
I have the 16B parameter version of Alpaca.cpp (and a copy of the training data as well as the weights) installed locally on a machine with an Nvidia 3070 GPU. I know I can launch my terminal using the Discrete Graphics Card option, but I also believe this version was built for CPU use and I’m guessing that I’m not getting the most out of my graphics card
What’s the move here?
KerfuffleV2 t1_jd7sb4u wrote
llama.cpp
and alpaca.cpp
(and also related projects like llama-rs
) only use the CPU. So not only are you not getting the most out of your GPU, it's not getting used at all.
I have an old GPU with only 6GB so running larger models on GPU isn't practical for me. I haven't really looked at that aspect of it much. You could start here: https://rentry.org/llama-tard-v2
Keep in mind you will need to be pretty decent with technical stuff to be able to get it working based on those instructions even though they are detailed.
keeplosingmypws t1_jd9wpwm wrote
Thanks for leading me in the right direction! I’ll letcha know if I get it working
Viewing a single comment thread. View all comments