Submitted by Destiny_Knight t3_11tab5h in singularity
BSartish t1_jciy4nt wrote
Reply to comment by liright in Those who know... by Destiny_Knight
This video explains it pretty well.
ThatInternetGuy t1_jcj2ew8 wrote
Why didn't they train once more with ChatGPT instruct data? Should cost them $160 in total.
CellWithoutCulture t1_jcjkwy1 wrote
Most likely they haven't had time.
They can also use SHP and HF-RLHF.... I think they will help a lot since LLaMA didn't get the privlidge of reading reddit (unliked ChatGPT)
ThatInternetGuy t1_jckmq5s wrote
>HF-RLHF
Probably no need, since this model could piggyback on the responses generated from GPT4, so it should carry the trait of the GPT4 model with RLHF, shouldn't it?
CellWithoutCulture t1_jcmsxjq wrote
HF-RLHF is the name of the dataset. As far as RLHF... what they did to LLaMA is called "Knowledge Distillation" and iirc usually isn't quite as good as RLHF. It's an approximation.
cartmanOne t1_jcof3eq wrote
That’s for their next paper…
CellWithoutCulture t1_jcjkycz wrote
decent video
Viewing a single comment thread. View all comments