Submitted by EducationalCicada t3_10vgrff in MachineLearning
st8ic t1_j7hf8yv wrote
given the volume of false information that chatGPT generates, I'm surprised that Google is jumping right in with a Google-branded product. They must be really scared of what chatGPT might do to search.
new_name_who_dis_ t1_j7hh479 wrote
Well obviously. Search is a tool for information retrieval (mostly). If you have an oracle, it's much more convenient than digging through the source material and doing the research yourself, even when it is presented to you in most relevant first order, which is the most convenient order and what made google successful in the first place.
But yes, anyone reading please don't use ChatGPT instead of google search unless you don't care about the responses being made up.
Zyansheep t1_j7hjddh wrote
Google search responses may be made up as well, its just a matter of there being more than one source to go through which makes it easier to spot potential discrepancies in any one source ;)
new_name_who_dis_ t1_j7hpz3q wrote
Well if you see variety in the top results in google that might give you pause. But you're not getting that from ChatGPT
JackandFred t1_j7hkrwq wrote
I like to tell people Gpt is more like writing an essay for English class or the sat than a research paper for a history class. It cares about grammatical correctness, readability is a better way to put that, that’s how you’re graded in English. It’s not graded on accuracy or truth. For the sat they used to say you can make up quotes for the essay section because they’re grading the writing, not the content. (I realize that’s dated, I don’t think they do an essay anymore)
worriedshuffle t1_j7jvjlu wrote
For the GRE our teacher said one of the easiest ways to get a high score was to have a strong ideology. Just be a Nazi, he said.
I did not end up using that advice but maybe if I did I would’ve done even better.
impermissibility t1_j7jywnd wrote
Uh, I'm sorry the English classes wherever you went to school sucked!
red75prime t1_j7k7hh0 wrote
I've run it thru GPT for your reading pleasure: "I like to tell people that GPT-3 is more like writing an essay for English class (or the SAT) than a research paper for a history class. It cares about grammatical correctness -- in other words, readability -- rather than accuracy or truth. For the SAT, they used to say "you can make up quotes", because they're grading your writing, not your content."
ginger_beer_m t1_j7ignoi wrote
> But yes, anyone reading please don't use ChatGPT instead of google search unless you don't care about the responses being made up.
Most people honestly don't care. They just want to get an answer quick, whether it's made up or not. This is true whether in real life or online.
artsybashev t1_j7k04qr wrote
If Xi Jing Ping, Putin and Trump have taught you anything, being correct is absolutely useless. Just having some sort of a plan, coming up with a good story and some fact sounding arguments is a lot more valuable that what the average person thinks. Nothing more is required to be one of the the most influential person alive.
joexner t1_j7j17v2 wrote
Like this one
ddavidovic t1_j7jwwc1 wrote
I think there's a lot more work to be done on that front. I tried to use ChatGPT and perplexity.ai instead of Google Search. It works for common knowledge, but once you get into more complex and niche queries it just falls apart. They're both very happy to lie to you and make up stuff, which is a huge time waste when you're trying to get work done.
harharveryfunny t1_j7kjohr wrote
I tried perplexity.ai for first time yesterday, and was impressed by it. While it uses GPT 3.5 it's not exactly comparable to ChatGPT since it's really an integration of Bing search with GPT 3.5, as you can tell by asking it about current events (and also by asking it about itself!). I'm not sure exactly how they've done the integration, but the gist of it seems to be more that GPT/chat is being used as an interface to search, rather than ChatGPT where the content itself is being generated by GPT.
Microsoft seem to be following a similar approach per the Bing/Chat verson that popped up and disappeared a couple of days ago. It was able to cite sources, which isn't possible for GPT-generated content which has no source as such.
[deleted] t1_ja6zwva wrote
[removed]
crazymonezyy t1_j7ojv39 wrote
> But yes, anyone reading please don't use ChatGPT instead of google search unless you don't care about the responses being made up.
The general public is not reading this sub, and ChatGPT is being sold to them by marketing and sales hacks without this disclaimer. We're way past the point of PSAs.
Sirisian t1_j7hy8td wrote
Google already has a knowledge graph which can be used to guard against common mistakes ChatGPT makes with trivia and basic information. Using such a system it's possible to prevent faults in the model and potentially stop some hallucination that can occur.
I've been hoping to see one of these companies construct and reference a complete probabilistic temporal knowledge graph. The bigger topic is being able to go from entity relationships back to training data sources to examine potential faults. I digress, this is a large topic, but it's something I've been very interested in seeing, especially since information can have a complex history with a lot of relationships. (Not just for our real timeline either. Every book has its own timeline of changing information that such a system should be able to unravel).
-Rizhiy- t1_j7ilv1v wrote
I feel that they won't be trying to generate novel responses from the model, but rather take knowledge graph + relevant data from the first few responses and ask the model to summarise that/change into an answer which humans find appealing.
That way you don't have to rely on the model to remember stuff, it can access all required information through attention.
PM_ME_YOUR_PROFANITY t1_j7lajpy wrote
Have you seen the work which connects ChatGPT to WolframAlpha?
[deleted] t1_j7ia76m wrote
[deleted]
_ModeM t1_j7it414 wrote
Haha this is great, thanks for sharing.
datasciencepro t1_j7i6msl wrote
They already had this up their sleeve having basically driven research in LLMs and having the largest dataset in the world. It's not a haphazard jumping in, more of a "okay we're starting to see some activity and commercial application in this space, now it's time to show what we've been working on". As a monopoly in search it would not have made sense for Google to move first.
jlaw54 t1_j7ihoki wrote
There are indications there has been some scrambling at google over this. But that they weren’t armed and researched, but they didn’t see this coming the way it did.
chiaboy t1_j7ivw24 wrote
Most of these “indications” are poorly sourced commentary, out of context internal docs, and absolute (or convient) ignorance re the space, it’s history, and Google’s work therein.
Go back and look at the articles. Very little actual indications Google is “scrambling” they’ve been thinking deeply about this space for longer than most folks have heard about it.
Among many other related asides, there aren’t many global (or even US) comprehensive AI rules. However Google has issued white papers and has lobby heavily for thoughtful regulation. Google not recklessly following the current AI-hype train doesn’t read to me that they were caught flat footed. Anything but.
But the headlines are catchy
opticd t1_j7jnou3 wrote
This is probably the most thoughtful take I’ve read in this. People forget how tilted the mainstream media is against big tech.
jlaw54 t1_j7j1k33 wrote
I agree with threads of what you are saying here.
That said, I think they were “prepared” for this in a very theoretical and abstract sense. I don’t think they were running around like fools at google hq aimlessly.
But that doesn’t mean it didn’t inherently create a shock to their system in real terms. Both can have some truth. Humans trend towards black and white absolutes, when the ground truth is most often grey.
chiaboy t1_j7j2bwp wrote
I agree.
They weren’t shocked per se, however clearly OAI is on their radar.
Not entirely unlike during COVID when Xoom taught most Americans about web conferencing. Arguably good for the entire space, but the company in the public imagination probably didn’t deserve all the accolades.
So the question for Google and other responsible AI companies, is how to capitalize on the consumer awareness/adoption, but do it in a way that acknowledges the real constraints (that OAI are less concerned with). MSFT is all ready running into some of those constraints viz the partnership (interesting to see Sataya get over his skis a little. That’s not his usual MO).
Fit-Meet1359 t1_j7iaw8u wrote
Given that this was announced only minutes before Microsoft announced the event tomorrow where they're expected to unveil the new GPT-powered Bing, they are probably scared of that rather than ChatGPT. I know Bing is a joke right now, but if it suddenly becomes a far better information assistant than Google simply by virtue of its ability to chat about search results and keep the context, that poses a huge threat (if the new Bing goes viral like ChatGPT did).
But it doesn't sound like Bard is going to be linked to the Google search engine just yet. The article mentions separate AI search integrations coming soon, but from the screenshots it just seems to generate a paragraph or two about the search, without citations.
krzme t1_j7i0to8 wrote
Given the volume of false information that Google gives hints to…
stml t1_j7hlueg wrote
It's not like Google vets the websites that show up in Google searches all that well regardless.
starstruckmon t1_j7i34u8 wrote
Retrieval augmented models ( whether via architecture or prompt ) don't have that issue.
Even GPT3 API based services like perplexity.ai that retrieval augment using just the prompt don't spew wrong information all that much.
farmingvillein t1_j7i4iiu wrote
> Retrieval augmented models ( whether via architecture or prompt ) don't have that issue.
Err. Yes they do.
They are generally better, but this is far from a solved problem.
starstruckmon t1_j7i5qoc wrote
It's not just better, wrong information from these models is pretty rare, unless the source it is retrieving from is also false. The LM basically just acts as a summary tool.
I don't think it needs to be 100% resolved for it to be a viable replacement for a search engine.
farmingvillein t1_j7ibgcn wrote
> wrong information from these models is pretty rare
This is not born at out all by the literature. What are you basing this on?
There are still significant problems--everything from source material being ambiguous ("President Obama today said", "President Trump today said"--who is the U.S. President?) to problems that require chains of logic happily hallucinating due to one part of the logic chain breaking down.
Retrieval models are conceptually very cool, and seem very promising, but statements like "pretty rare" and "don't have that issue" are nonsense--at least on the basis of published SOTA.
Statements like
> I don't think it needs to be 100% resolved for it to be a viable replacement for a search engine.
are fine--but this is a qualitative value judgment, not something grounded in current published SOTA.
Obviously, if you are sitting at Google Brain and privy to next-gen unpublished solutions, of course my hat is off to you.
starstruckmon t1_j7ie3ad wrote
Fair enough. I was speaking from a practical perspective, considering the types of questions that people typically ask search engines, not benchmarks.
RobbinDeBank t1_j7ky0ju wrote
Nice try. What are you hiding at Google Brain?
[deleted] t1_j7kl3ux wrote
[removed]
ktpr t1_j7ho1vo wrote
They don’t care that much about what ChatGPT will do search. They care about the advertising users of ChatGPT won’t be seeing.
wood_orange443 t1_j7jjnfc wrote
How exactly do you think chatgpt is going to get funded?
harharveryfunny t1_j7kmbzr wrote
OpenAI just got a second round $10B investment from Microsoft, so that goes a ways ... They are selling API access to GPT for other companies to use however they like, and Microsoft has integrated Copilot (also GPT-based, fine-tuned for code generation) into their dev tools, and MIcrosoft is also integrating OpenAI's LLM tech into Bing. While OpenAI are also selling access to ChatGPT to end users, I doubt that's going to really be a focus for them or major source of revenue.
marvinv1 t1_j7kpqoj wrote
Yup, OpenAI expects to generate $200 million in revenue for 2023 and $1 billion for next year.
astrange t1_j7jtrfh wrote
ChatGPT's a website and any website can show you ads. Of course, it has the same issue as Gmail where users aren't going to like ads being targeted based on what they say to it.
user4517proton t1_j7jda78 wrote
I'm not surprised. Honestly, Google is caught with their pants down on AI integration. They have focused on backend systems to make their ad revenue more profitable. What Microsoft is doing is adding value to the end user. That is a major shift in people's focus on what AI means to everyone, not just Google.
Microsoft is taking a very visible lead in AI for the masses by integrating ChatGPT with Bing, Microsoft 365, development tools, etc. If ChatGPT provides anything near the level of benefit that Co-Pilot does for developers Google has a very valid concern.
I think Microsoft's approach, focusing on the end user value, will make this event be pivotable for how AI is used. Also keep in mind Microsoft is also releasing the biochat GPT, and I suspect there will be a number of targeted releases in the next weeks or months.
A brave new world...
bartturner t1_j7k8fnb wrote
Geeze. What a bunch of nonsense. ChatGPT would NOT even be possible without Google.
Google has made most of the major AI fundemental breakthroughs in the last decade+. Google leads in every layer of the AI stack without exception.
A big one is silicon. They started 8 years ago and now on their fifth generation. Their fourth was settting all kinds of records.
https://blog.bitvore.com/googles-tpu-pods-are-breaking-benchmark-records
harharveryfunny t1_j7knqfa wrote
OpenAI trained GPT on Microsoft Azure - it has zero to do with Google's TPU. While the "Attention Is All You Need" paper did come out of Google, it just built on models//concepts that came before. OpenAI have proven themselves plenty capable of innovating.
bartturner t1_j7l64gq wrote
> OpenAI trained GPT on Microsoft Azure - it has zero to do with Google's TPU.
Geeze. ChatGPT would NOT exist if not for Google because the underlying tech was invented by Google.
OpenAI uses other people's stuff instead of inventing things themselves like Google.
Many of the big AI breakthroughs from the last decade+ have come from Google. GANs is another perfect example.
https://en.wikipedia.org/wiki/Transformer_(machine_learning_model)
The TPUs are key in being able to bring a large language model to market at scale. Not training but the inference aspect.
harharveryfunny t1_j7lu67f wrote
What underlying are you talking about? Are you even familiar with the "Attention" paper and it's relevance here? Maybe you think OpenAI use Google's Tensorflow? They don't.
GANs were invented by Ian Goodfellow while he was a student at. U.Montreal, before he ever joined Google.
No - TPUs are not key to deploying at scale unless you are targeting Google cloud. Google is a distant 3rd in cloud marketshare behind Microsoft and Amazon. OpenAI of course deploy on Microsoft Azure, not Google.
bartturner t1_j7lugv5 wrote
Geeze. Who do you think invented Transformers?
https://en.wikipedia.org/wiki/Transformer_(machine_learning_model)
NO!!! GANs were invented by Ian while he was working at Google. It is a pretty interesting story.
The vast majority of the major AI breakthroughs from the last decade+ came from Google.
OpenAI really does NOT do R&D. THey more use the R&D from others and mostly Google.
harharveryfunny t1_j7lvevz wrote
https://proceedings.neurips.cc/paper/2014/file/5ca3e9b122f61f8f06494c97b1afccf3-Paper.pdf
See page 1 footnote : "Goodfellow did this work as a UdeM student".
bartturner t1_j7lwvdt wrote
Ha! Go listen to Lex's podcast. Ian explains it all and it was ALL while working at Google.
harharveryfunny t1_j7ly2jz wrote
And then he travelled back in time to go write that paper at U.Montreal ?
Anyways, Schmidhuber was the real inventor ;-)
bartturner t1_j7lza3y wrote
Go listen to the podcast and Ian explains it all. Plus no Schmidhuber was NOT the inventor. It was Ian.
Go listen to the podcast and get back to me.
The key AI R&D from the last decade plus has all come from Google. Not from OpenAI and most definitely not from Microsoft.
Ill-Poet-3298 t1_j7kap8n wrote
Google is afraid to kill their ad business, so they're letting others pass them by. Classic business mistake. There are apparently a lot of Google stans going around telling everyone how Google invented AI, etc, but it really looks like they got caught flat footed on this one.
[deleted] t1_j7hh25u wrote
[deleted]
melodyze t1_j7j6h6t wrote
The Lamda paper has some interesting sidelines at the end about training the model to dynamically query a knowledge graph for context at inference time and stitch the result back in, to retrieve ground truth, which may also allow the state change at runtime without requiring constant retraining.
They are better positioned to deal with that problem than chatgpt, as they already maintain what is almost certainly the world's most complete and well maintained knowledge graph.
But yeah, while I doubt they have the confidence they would really want there, I would be pretty shocked if their tool wasn't considerably better at not being wrong on factual claims.
backafterdeleting t1_j7ju972 wrote
The problem with ChatGPT right now is that it has no way of expressing its confidence level with regard to its own output. So if its unsure about a possible response, it still has to write it as if its 100% undeniable fact.
---AI--- t1_j7ki2wb wrote
Eh, so like humans
marr75 t1_j7ksi6o wrote
They should be. I think LLMs will totally upset how content is indexed and accessed. It's one of the easiest and lowest stakes use cases for them, really.
Unfortunately, Google has such a huge incumbent advantage that they could produce the 5th or 6th best search specialized LLM and still be the #1 search provider.
yaosio t1_j7lnkh9 wrote
If you look at what you.com does they cite the claims their bot makes by linking to the pages the data come from, but only sometimes. When it doesn't cite something you can be sure that it's just making it up. In the supposed Bing leak it was doing the same thing, citing it's sources.
If they can force it to always provide a source, and if it can't then it won't say it, that could fix it. However, there's still the problem that the model doesn't know what's true and what's false. Just because it can cite a source doesn't mean the source is correct. This is not something that the model can learn by being told. To learn by being told assumes that it's data is correct, which can't be assumed. A researcher could tell the model, "all cats are ugly", which is obviously not true, but the model will say all cats are ugly because it was taught that. Models will need to have a way to determine on their own what is true and what isn't true, and explain it's reasoning.
emerging-tech-reader t1_j7kh681 wrote
> given the volume of false information that chatGPT generates
It actually generates mostly accurate information. The longer you have the conversation the more it starts to hallucinate, but it is considerably more accurate than most people.
yeluapyeroc t1_j7hlb5v wrote
Its a trivial configuration option to prevent OpenAI models from hallucinating answers and have them respond with an "I don't know" equivalent. I'm sure Google sees way beyond the novelty of the current publicly accessible ChatGPT model.
st8ic t1_j7hquaz wrote
> Its a trivial configuration option to prevent OpenAI models from hallucinating answers and have them respond with an "I don't know" equivalent.
How?
mettle t1_j7hrck0 wrote
is it though? how would you even do that? i think if you have that actually figured out, it's easily a $1b idea.
farmingvillein t1_j7i4ed7 wrote
> how would you even do that?
r/yeluapyeroc just reviews each post, np
mettle t1_j7i5ign wrote
the true human in the loop.
Viewing a single comment thread. View all comments