2muchnet42day
2muchnet42day t1_jdgzocw wrote
Reply to comment by bias_guy412 in [D] What is the best open source chatbot AI to do transfer learning on? by to4life4
Neither was LLaMA before Alpaca.
2muchnet42day t1_jdgy2ht wrote
Reply to comment by immune_star in [P] CodeAlpaca Code and Data release by immune_star
Are they public now?
2muchnet42day t1_jda64te wrote
Reply to comment by immune_star in [P] CodeAlpaca Code and Data release by immune_star
!RemindMe 1 day
2muchnet42day t1_jd8u97a wrote
Reply to comment by Nice_Cod7781 in [P] CodeAlpaca Code and Data release by immune_star
We need to start iterating on the same weights and not start from scratch everytime
2muchnet42day t1_jd8fnje wrote
Reply to comment by immune_star in [P] CodeAlpaca Code and Data release by immune_star
Would you consider doing a LoRA version of CodeAlpaca and compare the ouputs of the two models?
2muchnet42day t1_jd7upsm wrote
Reply to [P] CodeAlpaca Code and Data release by immune_star
It's awesome. Thank you for your work.
I'd like to know why you didn't take the LoRA approach to finetuning LLaMA? Is a full finetuning better?
2muchnet42day t1_jd3pu0m wrote
Reply to comment by benfavre in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
Can you train with 24 gigs of vram ?
2muchnet42day t1_jczooi6 wrote
Reply to comment by UnusualClimberBear in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
Yeah, I wouldn't buy AMD either. It's a shame that NVIDIA is basically a monopoly in a AI, but it is what it is.
2muchnet42day t1_jczj8da wrote
Reply to comment by currentscurrents in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
>I'm gonna end up buying a bunch of 24GB 3090s at this rate.
Better hurry up...
2muchnet42day t1_jckjy9i wrote
Reply to comment by farmingvillein in [D] What is the best way to fine tune a LLM with your own data and build a custom text classifier? by pgalgali
Thank you
2muchnet42day t1_jckb1es wrote
Reply to comment by abstract000 in [D] Neat project that would "fit" onto a 4090? by lifesthateasy
Why not LLama/Alpaca ?
2muchnet42day t1_jcjsy5s wrote
Reply to comment by farmingvillein in [D] What is the best way to fine tune a LLM with your own data and build a custom text classifier? by pgalgali
Why do you suggest Roberta and not something like LLAMA or Standford Alpaca?
2muchnet42day t1_j9j5wl3 wrote
Reply to comment by pyonsu2 in [D] Large Language Models feasible to run on 32GB RAM / 8 GB VRAM / 24GB VRAM by head_robotics
And the hardware.
2muchnet42day t1_jee275g wrote
Reply to comment by gahblahblah in [D][N] LAION Launches Petition to Establish an International Publicly Funded Supercomputing Facility for Open Source Large-scale AI Research and its Safety by stringShuffle
It's kinda ironic there already is an OpenAI and it's exactly the opposite.
Let us hope these initiatives get all the necessary support and that they stay loyal to their foundational concepts.