Submitted by NLP2829 t3_yu8nna in MachineLearning

(I only want to do inference, I don't need to finetune it.)

I want to use very-large language model (#parameters > 100B) to do some experiments, is that true the only very-large language model we can get access to is GPT3 API? Can we possibly get access to PaLM and Flan-PaLM 540B with no cost by chance?

I have searched over the internet but can't find a definite answer. As GPT-3 pricing for text-davinci-2 is not cheap, I am wondering if there's a chance to use other models.

Also, I can request up to 372GB VRAM, is there any large language model (#parameters > 100B) that I can actually download and run "locally"?

37

Comments

You must log in or register to comment.

ML4Bratwurst t1_iwaqxyc wrote

Do you really need such a big transformer? There a a couple of big transformers available in the Huggins face API.

2

csreid t1_iwlyxnz wrote

>Also, I can request up to 372GB VRAM, is there any large language model (#parameters > 100B) that I can actually download and run "locally"?

I've never done anything non-trivial with LLMs but even using 32 bit floats for 100B parameters should take 400 gigs of RAM, right?

1