light24bulbs
light24bulbs t1_jeeb4ag wrote
Reply to comment by ASlowDanceWithDeath in [P] Introducing Vicuna: An open-source language model based on LLaMA 13B by Business-Lead2679
I think people don't know if they can legally do this under the llama license.
It's part of why the Lora approach to fine tuning is so nice, you don't have to share the original weights.
light24bulbs t1_jeeb0rs wrote
Reply to comment by Alarming_Turnover578 in [P] Introducing Vicuna: An open-source language model based on LLaMA 13B by Business-Lead2679
No, it's a non-commercial license focused on research cases.
light24bulbs t1_jeeaqrs wrote
Reply to comment by AlmightySnoo in [P] Introducing Vicuna: An open-source language model based on LLaMA 13B by Business-Lead2679
Agreed. At least this work on top of llama is apache 2.0
light24bulbs t1_jee9lvm wrote
Reply to comment by Tostino in [D][N] LAION Launches Petition to Establish an International Publicly Funded Supercomputing Facility for Open Source Large-scale AI Research and its Safety by stringShuffle
I agree with you. Looking at papers like ToolFormer and so on, we are very close.
We are only a couple years away from AGI, which is what I've been saying for YEARS and getting yelled at here. The WaitButWhy article in 2016 was dead right
light24bulbs t1_jee9eey wrote
Reply to comment by gahblahblah in [D][N] LAION Launches Petition to Establish an International Publicly Funded Supercomputing Facility for Open Source Large-scale AI Research and its Safety by stringShuffle
So you went through and signed it?
light24bulbs t1_jee9dou wrote
Reply to comment by glichez in [D][N] LAION Launches Petition to Establish an International Publicly Funded Supercomputing Facility for Open Source Large-scale AI Research and its Safety by stringShuffle
Nice. You signed it then?
light24bulbs t1_jee9d2e wrote
light24bulbs t1_jee9clk wrote
Reply to [D][N] LAION Launches Petition to Establish an International Publicly Funded Supercomputing Facility for Open Source Large-scale AI Research and its Safety by stringShuffle
Please everyone, go through to the link and sign it
light24bulbs t1_je8d6bh wrote
Reply to comment by LetGoAndBeReal in [D] The best way to train an LLM on company data by jaxolingo
Continuous retraining is something else.
I'll be training llama soon, I'll get back to you with how it goes.
light24bulbs t1_je863pu wrote
Reply to comment by WokeAssBaller in [D] The best way to train an LLM on company data by jaxolingo
Yeah, he doesn't get it. That's ok though, but to be wrong and be sure about it is a bummer
light24bulbs t1_je7sn5e wrote
Reply to comment by LetGoAndBeReal in [D] The best way to train an LLM on company data by jaxolingo
The fact that the fine tuning can improve instruction following is EXACTLY that. There's no distinction between predicting the next word, following instructions, or deep knowledge. They are all the same thing as far as an LLM.
light24bulbs t1_je7pnxa wrote
Reply to comment by LetGoAndBeReal in [D] The best way to train an LLM on company data by jaxolingo
This IS training. That's what it is. This is how "knowledge" got into the model in the first place
light24bulbs t1_je7ob17 wrote
Reply to comment by LetGoAndBeReal in [D] The best way to train an LLM on company data by jaxolingo
Okay, here's my friend turning the alpaca instructions into training data
https://github.com/lxe/llama-peft-tuner/blob/main/convert_alpaca_to_text.py
See how it's just turning it into a fat string?
light24bulbs t1_je7mr9p wrote
Reply to comment by LetGoAndBeReal in [D] The best way to train an LLM on company data by jaxolingo
False, they all do. The process of fine-tuning is identical to the initial pre-training, though perhaps with different settings. They're mostly setup to take q&a data for getting llama to take instructions better, but actually that's just text wrapped in some context and passed in straight up.
I was very confused by this as well but no, you can train new stuff.
light24bulbs t1_je7ilvq wrote
Reply to comment by LetGoAndBeReal in [D] The best way to train an LLM on company data by jaxolingo
I disagree that it's not viable to train a model. The problem is that the best public model (llama) is non-commercial.
That said, it's extremely possible to train things into it. There's a ton of new scripts floating around online. The Lora training is especially good.
The trouble with vectors is they are so limited. They're fine if you need to look up one distinct thing (and the vector gets the match right) but they're utterly useless if you'd like the model to learn about something in general.
light24bulbs t1_jdzzeh4 wrote
Reply to comment by nemorocksharder in [R] Hello Dolly: Democratizing the magic of ChatGPT with open models by austintackaberry
Yes, I'm into it now. Code like this can be adapted to load bulk data instead of q&a.
I suspect some of the training parameters need to be adjusted a bit to prevent over fitting and obviously the data loading and templating needs to be removed.
https://github.com/lxe/llama-tune Or for a cooler approach where you make a Lora layer https://github.com/serp-ai/LLaMA-8bit-LoRA
light24bulbs t1_jdybyi9 wrote
Reply to [D] FOMO on the rapid pace of LLMs by 00001746
I'm just going in deep. These are FUN
light24bulbs t1_jduwgqt wrote
Reply to comment by was_der_Fall_ist in [D]GPT-4 might be able to tell you if it hallucinated by Cool_Abbreviations_9
Yeah, like it's actually using a huge amount of brain power to figure out what the next word is. Just because that's how it works doesn't mean it's not intelligent.
If you want to be really good at figuring out what the next word is you have to be really smart
light24bulbs t1_jduuuep wrote
Reply to comment by alexmin93 in [P] Using ChatGPT plugins with LLaMA by balthierwings
I do, still struggling with it
light24bulbs t1_jdtiq9w wrote
Reply to comment by endless_sea_of_stars in [P] Using ChatGPT plugins with LLaMA by balthierwings
I'm aware of that part. The wording of the test that's injected is not public. If it was, if use it in my langchain scripts.
Again i really expect there's fine-tuning, we will see eventually maybe.
light24bulbs t1_jdtgrjb wrote
Reply to comment by endless_sea_of_stars in [P] Using ChatGPT plugins with LLaMA by balthierwings
Based on how much langchain struggles to use tools and gets confused on them, I'd bet on fine tuning. I asked a contact to reveal what they're injecting into the prompt but it's not public information yet so i couldn't get it
light24bulbs t1_jdsulyn wrote
Reply to comment by endless_sea_of_stars in [P] Using ChatGPT plugins with LLaMA by balthierwings
By "in context learning" i take it you mean zero shot.
Yes, you can hot swap. Id be unsurprised if what Open-AI did is fine tune on how to use plugins in general by giving some examples combined with a little bit of zero-shot primer.
Something trained with ToolFormers technique and then told it can use a new, but similar, plugin is IMO going to generalize way better than something that's never used a plugin before.
light24bulbs t1_jdsapie wrote
Reply to comment by frequenttimetraveler in [D] Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
My mistake, correct you are:
https://www.beren.io/2022-08-06-The-scale-of-the-brain-vs-machine-learning/
This post recons we using 800b parameters for language processing.
light24bulbs t1_jds3mdl wrote
Reply to [P] Using ChatGPT plugins with LLaMA by balthierwings
What's the underlying approach here? Just prompt engineering right?
I really really want to apply the ToolFormer paper to llama. They're both Facebook systems, you can get they've done it.
ToolFormer just seems like SUCH a good and thorough approach. There are quite a few gaps between the paper and building a working example, IMO, but it's clearly doable.
The way Facebook licensed the weights is frustrating me. We should all be passing around Alpaca trained, GPTQ quantized, SparseGpt optimized Llama derived models by now. Is there some telegram group i need to be in or something?
light24bulbs t1_jeeb9cx wrote
Reply to comment by gliptic in [P] Introducing Vicuna: An open-source language model based on LLaMA 13B by Business-Lead2679
Nice way to get around the license problem.
Is Lora really associated with a quality loss? I thought it worked pretty well.