Submitted by Business-Lead2679 t3_12618zu in MachineLearning
gmork_13 t1_je7dho0 wrote
For a more stable compute, check out google cloud gpu.
Consider training a quantized model with LoRA. If you know enough, perhaps the model could be split between VRAM and DDR RAM to make it train on a smaller GPU.
edit: here, I found one: https://github.com/tloen/alpaca-lora
I think you could get this done for far less than your budget if need be.
Viewing a single comment thread. View all comments