Viewing a single comment thread. View all comments

blueSGL t1_jcjga2i wrote

Is it possible to split the model and do inference across multiple lower VRAM GPUs or does a single card have to have the minimum 16gig VRAM?

5