Submitted by imgonnarelph t3_11wqmga in MachineLearning
benfavre t1_jd2n1cg wrote
Reply to comment by cbsudux in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
1 epoch of finetuning the 30B model with llama-lora implementation, mini-batch-size=2, maxlen=384, is about 11 hours.
2muchnet42day t1_jd3pu0m wrote
Can you train with 24 gigs of vram ?
Viewing a single comment thread. View all comments