ghostfaceschiller
ghostfaceschiller t1_je8habj wrote
Reply to comment by EquipmentStandard892 in [R] LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention by floppy_llama
Could you extrapolate what you mean here? I'm not sure I'm following
ghostfaceschiller t1_je84v6j wrote
Reply to [Discussion] IsItBS: asking GPT to reflect x times will create a feedback loop that causes it to scrutinize itself x times? by RedditPolluter
You are talking about two different things here.
-
Reflexion/ReAct uses another system (like LangChain) to allow the bot to genuinely loop back over previous results to try and improve them. This indeed ends up getting you better results or outcomes in the end
-
You can also simply tell the bot, in ur prompt, something like "before you respond, review your first draft for errors, and only output your second draft". Now, this is not what the bot will actually do, but regardless, this will often result in higher quality output, presumably bc in the training data that kind of phrase is typically associated with a certain type of answer (IE: better answers)
ghostfaceschiller t1_je44ke8 wrote
Reply to comment by joeiyoma in [N] OpenAI may have benchmarked GPT-4’s coding ability on it’s own training data by Balance-
What?
ghostfaceschiller t1_je3abdo wrote
Reply to comment by -xXpurplypunkXx- in [N] OpenAI may have benchmarked GPT-4’s coding ability on it’s own training data by Balance-
Really? I def had that some with 3.5 but 4 has been v good. Not perfect obviously
ghostfaceschiller t1_jdzduen wrote
Reply to comment by Riboflavius in [N] OpenAI may have benchmarked GPT-4’s coding ability on it’s own training data by Balance-
Looolllll
ghostfaceschiller t1_jdz6vzn wrote
Reply to [N] OpenAI may have benchmarked GPT-4’s coding ability on it’s own training data by Balance-
I think this was shown awhile ago (like a week ago, which just feels like ten years)
While I do think this is important for several reasons, personally I don't see it as all that impactful for what I consider AI capable of going forward.
That's bc pretty much all my assumptions for the next couple years are based on the idea of systems that can loop and reflect on their own actions, re-edit code based on error messages, etc. Which they are very good at
ghostfaceschiller t1_jdyerkp wrote
Reply to comment by nxqv in [D] FOMO on the rapid pace of LLMs by 00001746
> Yan LeCun
That dude is becoming straight-up unhinged on Twitter
ghostfaceschiller t1_jdsnev1 wrote
Reply to comment by blose1 in [D] GPT4 and coding problems by enryu42
This line of thinking sounds sillier and sillier every week. Its like talking to someone who has had their eyes shut and fingers in their ears for the last two months.
EDIT: and tbc, i'm not trying to argue that it isn't statistics-based/trained on the internet/etc. I'm saying that it turns out that kind of system is powerful & capable than we ever would have intuitively thought it would be
ghostfaceschiller t1_jds855j wrote
Reply to comment by BeautifulLazy5257 in [D] GPT4 and coding problems by enryu42
I would just look up ReAct, CoT(chain of thought), and LangChain Agents. Its pretty simple to implement
ghostfaceschiller t1_jds202e wrote
Reply to comment by enryu42 in [D] GPT4 and coding problems by enryu42
Yeah it's essentially that at an automated level. Tbh it is powerful enough based on results so far that would actually be really surprised if it did not yield very significant gains in these tests.
I'm sure there will be a paper out doing it in like the next few days, so we'll see
ghostfaceschiller t1_jds0zez wrote
Reply to comment by Cool_Abbreviations_9 in [D] GPT4 and coding problems by enryu42
Basically just giving the model the ability to observe the results of its previous action and decide if it wants to try something different based on the feedback
ghostfaceschiller t1_jdrl62e wrote
Reply to [D] GPT4 and coding problems by enryu42
Ok. but what is the performance when you give GPT-4 a ReAct/Reflexion loop?
ghostfaceschiller t1_jdljq4g wrote
Reply to comment by MrEloi in [R] Artificial muses: Generative Artificial Intelligence Chatbots Have Risen to Human-Level Creativity by blabboy
tell me about the one with the microprocessor
ghostfaceschiller t1_jdljnsy wrote
Reply to comment by MysteryInc152 in [R] Artificial muses: Generative Artificial Intelligence Chatbots Have Risen to Human-Level Creativity by blabboy
exactly - no one wants to be the first one to say it for some reason. If it were me I'd be grabbing that place in history with both hands
ghostfaceschiller t1_jdgrba9 wrote
Reply to comment by willer in [N] ChatGPT plugins by Singularian2501
Fr??? Wow what an insane oversight
Or I guess maybe they don’t wanna rack up all the extra embeddings calls, bc I assume like 100% if users would turn that feature on
ghostfaceschiller t1_jdfc5uj wrote
Reply to [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
Leaving out the best part: a commented out line reveals that the original/alternate title of the paper was “First Contact With An AGI System”
ghostfaceschiller t1_jdenoo2 wrote
Reply to comment by BigDoooer in [N] ChatGPT plugins by Singularian2501
Here's a standalone product which is a chatbot with a memory. But look at LangChain for several ways to implement the same thing.
The basic idea is: periodically feed your conversation history to the embeddings API and save the embeddings to a local vectorstore, which is the "long-term memory". Then, any time you send a message or question to the bot, first send that message to embeddings API (super cheap and fast), run a local comparison, and prepend any relevant contextual info ("memories") to your prompt as it gets sent to the bot.
ghostfaceschiller t1_jdekpke wrote
Reply to comment by psdwizzard in [N] ChatGPT plugins by Singularian2501
Trivially easy to build using the embeddings api, already a bunch of 3rd party tools that give you this. I’d be surprised if it doesn’t exist as one of the default tools within a week of the initial rollout.
EDIT: OK yeah it does already exist a part of the initial rollout - https://github.com/openai/chatgpt-retrieval-plugin#memory-feature
ghostfaceschiller t1_jdekaf3 wrote
Reply to comment by drunk-en-monk-ey in [N] ChatGPT plugins by Singularian2501
People really need to update their priors on what kind of things are straightforwardly possible or not. Like if you majorly updated your expectations last week, you are way behind and need to update them again.
ghostfaceschiller t1_jazpo5t wrote
Reply to comment by mindfu in On Facebook, Visual Misinfo Widespread. In the runup to the 2020 U.S. Presidential election visual misinformation was widespread across the platform, and that it was highly asymmetric across party lines, with right-leaning images five to eight times more likely to be misleading. by Wagamaga
The issue is that the dude never managed “The Data Science Team” at Meta, he’s just lying
ghostfaceschiller t1_jax0gd8 wrote
Reply to comment by Duende555 in On Facebook, Visual Misinfo Widespread. In the runup to the 2020 U.S. Presidential election visual misinformation was widespread across the platform, and that it was highly asymmetric across party lines, with right-leaning images five to eight times more likely to be misleading. by Wagamaga
Genuinely amazing to me that ppl can look at the content and conclusions of this study and come back with “yep just like the New York Times”
ghostfaceschiller t1_j7h1xwd wrote
Reply to comment by DrTonyTiger in Analysis showed that 65.6% of women who took extra Vitamin D gave birth naturally. The study analysed results from the MAVIDOS trial which involved 965 women being randomly allocated an extra 1,000 International Units (IU) per day of vitamin D during their pregnancy or a placebo. by Wagamaga
It’s more than a 13% increase, that seems pretty substantial to me
ghostfaceschiller t1_j6ej1pw wrote
Reply to comment by kilopeter in Transition probabilities (shown as percentages) between successive letters in the names of girls born in 2021 in the USA [OC] by kilopeter
I recently put together a repo of character frequency analyses, bc they can be really useful when designing keyboard layouts. So I have an eye out rn for interesting ways to look at and visualize the data. I think this particular instance is probably too limited to be useful for keyboard layouts, but if you do anything more please let me know! It’s one of the more interesting visualizations I’ve seen so I’d love to include/link it
ghostfaceschiller t1_j6bz394 wrote
Reply to comment by kilopeter in Transition probabilities (shown as percentages) between successive letters in the names of girls born in 2021 in the USA [OC] by kilopeter
Have you done this analysis on any other corpora? I’d be super interested to see it on something like the Google Corpus (or a reasonable subset of it). Or BNC, etc
ghostfaceschiller t1_jeagl0y wrote
Reply to comment by Nhabls in [N] OpenAI may have benchmarked GPT-4’s coding ability on it’s own training data by Balance-
What?