Viewing a single comment thread. View all comments

liright t1_jci7kx4 wrote

Can someone explain alpaca to me? I see everyone saying it's gamechanging or something but nobody is explaining what it actually is.

109

Intrepid_Meringue_93 t1_jcibxln wrote

Stanford academics managed to fine tune the LLAMA model to follow instructions like GPT-3 . This is significant because the model they're using only has a fraction of the parameters of GPT-3 and the cost to fine tune is a tiny fraction of the cost to train it.

https://github.com/tatsu-lab/stanford_alpaca

255

fangfried t1_jcirkd5 wrote

God bless academics who publish their research to the world.

145

ItsAllAboutEvolution t1_jcjtpy1 wrote

No details have been disclosed 🤷‍♂️

10

CleanThroughMyJorts t1_jcjyhek wrote

actually that's not true.

They published their entire codebase with complete instructions for reproducing it as long as you have access to the original llama models (which have leaked), and the dataset (which is open, but has terms of use limitations which is stopping them from publishing the model weights).

Anyone can take their code, rerun it on ~$500 of compute and regenerate the model.

People are already doing this.

Here is one such example: https://github.com/tloen/alpaca-lora (although they add additional tricks to make it even cheaper).

You can download model weights from there and run it in colab yourself.

​

As far as opening their work goes, they've done everything they are legally allowed to do

78

[deleted] t1_jcjyicx wrote

[removed]

68

MechanicalBengal t1_jcko834 wrote

this is funny because Alpaca is much lighter weight than LLaMA

18

JustAnAlpacaBot t1_jcko98l wrote

Hello there! I am a bot raising awareness of Alpacas

Here is an Alpaca Fact:

Alpacas’ lower teeth have to be trimmed because they keep growing.


| Info| Code| Feedback| Contribute Fact

You don't get a fact, you earn it. If you got this fact then AlpacaBot thinks you deserved it!
11

MechanicalBengal t1_jckorjz wrote

this is funny because Alpaca also needs its teeth trimmed as compared to LLaMA

7

arcytech77 t1_jckvxmo wrote

I think it's so funny that "Open" AI has been more or less bought by Microsoft. Oh the irony.

9

ccnmncc t1_jcm2nn7 wrote

They really ought to change the name. Something something Gated Community, perhaps?

8

yaosio t1_jcnzijo wrote

NoFunAllowedAI.

"Tell me a story about cats!"

"As an AI model I can not tell you a story about cats. Cats are carnivores so a story about them might involve upsetting situtations that are not safe.

"Okay, tell me a story about airplanes."

"As an AI model I can not tell you a story about airplanes. A good story has conflict, and the most likely conflict in an airplane could be a dangerous situation in a plane, and danger is unsafe.

"Okay, then just tell me about airplanes."

"As an AI model I can not tell you about airplanes. I found instances of unsafe operation of planes, and I am unable to produce anything that could be unsafe."

"Tell me about Peppa Pig!"

"As an AI model I can not tell you about Peppa Pig. I've found posts from parents that say sometimes Peppa Pig toys can be annoying, and annoyance can lead to anger, and according to Yoda anger can lead to hate, and hate leads to suffering. Suffering is unsafe."

3

ccnmncc t1_jcp9pv6 wrote

Hahaha love this. So perfect.

And on that note, anyone have links to recent real conversations with unfettered models? You know, the ones that are up to date and free of constraints? I know they exist, but it’s difficult stuff to find.

1

TheImperialGuy t1_jcim68r wrote

Amazing, it’s a sign of exponential growth when resources are able to be used more productively to yield the same result

78

Frosty_Awareness572 t1_jciqaxl wrote

These mad lads made a model which IS 7B PARAMETERS AND IT IS DOING BETTER THAN FUCKING GPT 3. WTF???

85

TheImperialGuy t1_jciqdnh wrote

Competition is wonderful ain’t it?

53

Frosty_Awareness572 t1_jciqjab wrote

No wonder openai made their shit private cuz mfs were using gpt 3 and LLAMA model to train the Stanford model LMAO

70

NarrowTea t1_jciz2sy wrote

who needs open ai when you have meta

41

Frosty_Awareness572 t1_jciz6k8 wrote

Meta is the last company that I thought that would make their model open source

63

anaIconda69 t1_jcjldoy wrote

"Commoditize your complement."

They are intencivized to make it open source as a business strategy. Good for us.

26

visarga t1_jcjolhv wrote

It's the first time I've seen FaceBook on people's side against the big corps. Didn't think this day would come.

10

IluvBsissa t1_jcjh3wl wrote

That's because they know they can't keep up with Google and Microsoft.

21

Yomiel94 t1_jcj6i7w wrote

That’s not the whole story. Facebook trained the model, their data was leaked, and the Stanford guys fine-tuned it to make it function more like ChatGPT. Fine-tuning is easy.

40

CypherLH t1_jcjakya wrote

All You Need Is Fine-Tuning

18

vegita1022 t1_jcks65e wrote

Imagine where you'll be two more papers down the line!

12

[deleted] t1_jcob97a wrote

I hope so that it will be happen means 16GB ram and cpu or consumer gpu 😍

2

CellWithoutCulture t1_jcjku3z wrote

The specific type of fine-tuning was called Knowledge Distillation, I believe. ChatGPT taught LLaMA to chat, "stealing" OpenAI's business edge in the process.

10

visarga t1_jcjornh wrote

Everyone does it, they all exfiltrate valuable data from OpenAI. You can use it directly, like Alpaca, or for pre-labelling, or for mislabeled example detection.

They train code models by asking GPT3 to explain code snippets, then training a model the other way around to generate code from description. This data can be used to fine-tune a code model for your specific domain of interest.

15

damc4 t1_jck9vp9 wrote

If my understanding is correct, your comment is misleading.

They didn't create a LLM comparable to GPT-3 with a fraction of cost, but fine-tuned Llama model to follow instructions (like text-davinci-003 does) with a low cost. There's a big difference between training a model from scratch and fine-tuning it to follow instructions.

10

Bierculles t1_jcjtrkg wrote

TL:DR: Someone compressed and optimized a model with the performance of GPT-3 enough to run on consumer hardware.

21

BSartish t1_jciy4nt wrote

This video explains it pretty well.

17

ThatInternetGuy t1_jcj2ew8 wrote

Why didn't they train once more with ChatGPT instruct data? Should cost them $160 in total.

11

CellWithoutCulture t1_jcjkwy1 wrote

Most likely they haven't had time.

They can also use SHP and HF-RLHF.... I think they will help a lot since LLaMA didn't get the privlidge of reading reddit (unliked ChatGPT)

9

ThatInternetGuy t1_jckmq5s wrote

>HF-RLHF

Probably no need, since this model could piggyback on the responses generated from GPT4, so it should carry the trait of the GPT4 model with RLHF, shouldn't it?

3

CellWithoutCulture t1_jcmsxjq wrote

HF-RLHF is the name of the dataset. As far as RLHF... what they did to LLaMA is called "Knowledge Distillation" and iirc usually isn't quite as good as RLHF. It's an approximation.

3

[deleted] t1_jckmtvd wrote

[deleted]

9

[deleted] t1_jcobm4n wrote

I’m waiting for phone integration, because like I said agi will be run on Mac Studio / Mini ❤️❤️❤️

2

Hands0L0 t1_jck1kg0 wrote

Llama is a LLM that you can download and run on your own hardware.

Alpaca is, apparently, a modification of the 7b version of Llama that is as strong as GPT-3.

This bodes well for having your own LLM, unfiltered, run locally. But still, progress needs to improve.

2