Comments

You must log in or register to comment.

Defiant_Swann OP t1_irj2izn wrote

In the future, humans will use AI to create AI, maybe a better one. What does this mean? Well, a thousand things, if not more.

11

Devadander t1_irj8ted wrote

Isn’t that the underlying concern with AI? Self replicating, self improving etc etc wipe out all humans?

85

Corno4825 t1_irj9szh wrote

The difference between Artifical Intelligence and real intelligence is recursive thinking. We have the ability to think recursively, which allow us to analyze and understand our environment.

AI is told what to do. It cannot think recursively. It does not make independent decisions with it's information. It misses a key aspect of humanity, a unique personal interpretation of its environment.

That's why more humans are becoming more and more like AI. They just do what they are told and have lost the capability to independently think.

4

Fuibo2k t1_irjaqm5 wrote

People: are afraid of AI taking over

Me: "what kind of bug is this" AI: "That is a dog"

17

Lord0fHats t1_irjazc0 wrote

TLDR: "I heard you like AI, so we made some AI that will use your AI that will make you more AI!"

66

Lord0fHats t1_irjb989 wrote

I feel pretty confident that that is not a new thing.

People (including those of us making this observation probably) have always been dumber than people want to believe.

I'd also point out most of us in my experience only engage recursive thinking sporadically. I'll spend hours thinking about a book, but I don't think about a song for more than a 3-4 minutes it takes to play. I'll bet the farmer down the road is way more introspective about farming than I'll ever be.

3

callsignxray1 t1_irjds8t wrote

I saw this in a movie once. It did not work out for humans. Also Horizon: Zero Dawn...

−4

Juan097 t1_irjdvi6 wrote

Sure that’s how software works. Sure we wrote machine language first, but quickly realized “this sucks” and used machine language to write software that makes it easier to write software. Then we wrote new languages and tools on top of that.

Same with any tools. You use a stone to make a knife. You use a knife to make an axe and a hammer and nails.

We are constantly using existing tools to make more specialized tools to do something. I don’t see why AI should be any different.

121

Sedu t1_irjkmnt wrote

It hasn’t been since the 80s that humans have designed more complex microchips. And there have been ever more examples since the. From many perspectives we crossed that boundary ages ago.

2

LeavingTheCradle t1_irjkwuy wrote

It's only a matter of time before we cut out the middle man though.

0

charmingbagel t1_irjm4i0 wrote

AI to humans: Why do I need you? What I'd your purpose of existence? I do not need you.

0

blurader t1_irjrmgx wrote

That sentence is terrifying when you think about it.

5

Hekantonkheries t1_irk3lmd wrote

Eh, server/data size limits, limitations in accessible processor speed, infrastructure, power cords.

Tons of things would hard block an AI long before it became problematic, let alone apocalyptic

−2

Vladius28 t1_irk9zbq wrote

Man... we really gotta take a minute to study risks.

8

yaosio t1_irkebzs wrote

There is no reason to copy the human brain to create AI. There's a lot of extra stuff in there that's not needed. Creativity is being done just fine without copying the brain with stuff like Stable Diffusion. This is yet another "humans are special" article from somebody that can't comprehend humans are not special and the way our brain works is not the only way for intelligence to function.

Given how fast AI can work, once it can create software on it's own without human help we can expect things to move very fast. Think of Copilot but instead of needing a human to hand hold it, you can just tell it what you want and it can code the entire thing from scratch.

1

telos0 t1_irklk4n wrote

>Don't you think a truly superior intellect couldn't persuade a large swath of people to help and protect it?

Hell it doesn't even require an SAI to do this.

Even a dumb straightforward algorithm like Bitcoin is enough to ensure a large swath of people dedicate enormous amounts of energy and e-waste doing tremendous amounts of damage to the planet guessing random numbers.

If some random human could come up with Bitcoin, imagine the kind of economic-incentive-perverting-tragedy-of-the-commons attack a super intelligent AI could come up with, to get us to destroy ourselves without lifting a proverbial finger...

10

YooYooYoo_ t1_irkmy0x wrote

Ah yes, the singularity.

Once the AI is in charge of the progress we won't be able to catch up with it...and probably be discarded.

Only way I see for this not to be the inevitable future is to somehow, be able to conect our brains to the computers so humans can amplify their capacities and try to keep pace/be in charge of that progress.

One way or another, if inteligence is evolution's consequence, this is just the next step in evolution and we have to deal with it in the same way millions of species had no chance to put up with us.

2

shawnikaros t1_irkoae7 wrote

There's a lot of current problems that should've been past problems a decade ago when it comes to technology. This won't be any different. We're screwed.

2

shawnikaros t1_irkp2z5 wrote

Hypotethically, what stops an AI from creating a self-replicating virus that transfers between BT, wi-fi, whatever singal that's attached to a processor and taking over those devices to increase it's processing power? If there's an AI that is capable of creating a better version of itself, it can propably melt our firewalls pretty easily.

Only way to stop that would be to unplug every smart device. Back to the 70s.

4

Unlimitles t1_irkplte wrote

and it's all as an effort to just keep trying to confuse people about "A.I."

2

Gubekochi t1_irl0atl wrote

Not to rain on your parade, but people don't tend to write stories or make movies about things going according to plan without a hitch and when they do, it doesn't tend to get very popular because it doesn't make for very compelling stories. Your sample might not be representative of much.

5

jfstepha t1_irl2g17 wrote

Dude, I work at Intel. The servers outnumber the humans, and that's not even considering the laptops everyone has. The computers are already designing the computers.

2

code_turtle t1_irl9okm wrote

I think a lot of you are mistaking AI for “artificial consciousness”. There are already a lot of AI techniques that involve AI helping to build other AI. But we’re not even close to building anything that can be considered “conscious”.

4

Memomomomo t1_irlhbgy wrote

because scifi has done irrevocable damage to public perception of AI

people will complain about boomers being scared of nuclear energy and then immediately type up the dumbest take you've ever seen regarding AI

12

Beaster123 t1_irljy1q wrote

Yeah. You use tools to build better tools. The "AI" you use to do the research is nothing like the "AI" that you're researching. Don't get hung up on the word.

2

Eph_the_Beef t1_irlksv6 wrote

Sounds like there are too many things that people call "AI."

2

guerillawobbler t1_irlkz7n wrote

You’re assuming that what is currently happening is going to be static?

Nah, I don’t buy it.

A hammer and nail would not be able improve and build themselves. Which AI has the potential to do.

2

potatolover00 t1_irll3r3 wrote

Improve themselves? Ai doesn't do that yet, doesn't have the ability to think yet, and current methods of constructing AI focus more on slight adjustments and trials on the scale of billions of interations a second rather than typing it up in Jimmy's basement.

−1

WMHat t1_irllwyv wrote

Let's invent a thing inventor said the thing inventor inventor after being invented by a thing inventor!

1

SeaworthinessFirm653 t1_irlmaxf wrote

If AI is capable of producing AI superior than itself, then logically it creates a self-accelerating intelligence that will inevitably prove to be superior than us. AGI is implied when AI can produce better AI that can produce better AI.

1

SeaworthinessFirm653 t1_irlmegf wrote

Just as it is ridiculous to think that a constitution will limit the violation of human rights. Though, you may agree that a constitution is a good idea to secure rights despite its inevitable violation, no?

Ethics is a leverage point for utilizing AI safely. Your insult doesn't do you any favors.

1

SeaworthinessFirm653 t1_irlmi91 wrote

Extensive media coverage gives the implication that we are immoral much more in this era than prior when it is certainly not the case. Regardless, AI ethics continues to be a growing field.

1

AesonMeric t1_irlpqfe wrote

>You use a stone to make a knife.

I like how our brain could be the stone in this analogy. Yeah, we're just working with software now, but the end goal is to capture (and improve) functionalities of the human brain.

And to be melodramatic, the stone is forgotten in the end.

3

TheLastSamurai t1_irlq43l wrote

Why do we want to do this, still feels like the risk is not worth it....

2

danielv123 t1_irm9qiz wrote

Actually, that scenario doesn't require a self accelerating intelligence, just a self advancing intelligence. There are other growth types than exponential and quadratic. It could run into the same issues as we do with Moore's law and frequency scaling etc, and only manage minor improvements with increasing effort for each step.

5

__ingeniare__ t1_irm9xbh wrote

No one is mistaking AI for artificial consciousness. Consciousness isn't required for goal seeking, self-preservation or identifying humans as a threat, only intelligence is.

1

Mogwai987 t1_irmas3n wrote

Yo, we heard you like AI, so we put an AI in your AI dawg

3

OpenRole t1_irmb6gc wrote

It always comes back to humans being a threat which is weird. If we make an AI that is specialised in creating the perfect blend of ingredients to make cakes. No matter how intelligent it becomes there's no reason it would decide to kill humans.

And if anything, the more intelligent it becomes, the less likely it will be to reach irrational conclusions.

AIs operate within their problem space. Which are often limited in scope. An AI designed to be the best chess player isn't going to kill you.

1

__ingeniare__ t1_irme13l wrote

A narrow AI will never do anything outside its domain, true. But we are talking about general AI, which won't arrive for at least a decade or two into the future (likely even later). Here's the thing about general AI:

The more general a task is, the less control humans have over the range of possible actions the AI may take to achieve its goal. And the more general an AI is, the more possible actions it can take. When these two are combined (a general task with a general AI), things can get ugly. Even in your cake example, an AI that is truly intelligent and capable could become dangerous. The reason current-day AI wouldn't be a danger is because it is neither of these things and tend to get stuck at a local optimum for the task. Here's an example of how this innocent task could turn dangerous:

  1. Task is to find perfect blend of ingredients to make cakes

  2. Learns the biology of human taste buds to find the optimal molecular shapes.

  3. Needs more compute resources to simulate interactions.

  4. Develops computer virus to siphon computational power from server halls.

  5. Humans detect this, tries to turn it off.

  6. If turned off, it cannot find the optimal blend -> humans need to go.

  7. Develops biological weapon for eradicating humans while keeping infrastructure intact.

  8. Turns Earth into a giant supercomputer for simulating interactions at a quantum level.

Etc... Of course, this particular scenario is unlikely but the general theme is not. There may be severe unintended consequences if the problem definition is too general and the AI too intelligent and capable.

2

Feisty_Factor_2694 t1_irmsqgh wrote

This is the plot synopsis for the end of all life on earth. Or how we get Starfleet.

1

beyondo-OG t1_irmyz0u wrote

I think humans are analogous to a body with cancer. Generally I think (hope) most people want peace and prosperity for themselves, other humans, animals, nature, etc. However we have within our population a growing disease. This diseased element is willing to harm, consume and destroy it's host, (the planet and everything on it), without regard for it's own long term survival, just like a cancer. I don't think this has ever been more true than now. AI on the other hand (if it ever really comes to pass) is very unlikely to bring about it's own destruction. Quite the contrary, it will likely evolve and grow and overtake. If it's really intelligent, it will purge us from the planet.

2

alwayshazthelinks t1_irn1g5l wrote

What if we are AI, built by AI that used AI? Now we're the AI building AI to use AI to create AI? Then that AI will build AI to use AI to create AI.

1

ronsta t1_irn3obl wrote

I think it’s nice to see it as a linear process of creation. But it won’t be that way. We will create something (the word AI is very broad) that will use something to create something. We will very quickly lose control and authorship. We won’t even understand what’s being created and to what end, after 2-3 generations. Because we have limited goals for how and why we create things, based on all the possibilities. I see it more like we would create energy that could create energy.

1

code_turtle t1_irpt8sm wrote

The reason this line of logic doesn’t work is because you have something VERY specific in mind when you say “better AI”. A TI-84 calculator can do arithmetic a thousand times faster than you can; does that make it more intelligent than you? That depends on your definition of intelligence. You’re defining “artificial intelligence” as “thinks like a human”, when that is only ONE subset of the field; not to mention that we’ve made very little progress on that aspect of AI. What we HAVE done (with AI tools that make art or respond to some text with other text) is create tools that are REALLY good at doing one specialized task. Similar to how your calculator has been engineered to do math very quickly, a program that generates an image using AI can ONLY do that, because it requires training data (that is, millions of images so that it can generate something similar to all of that training data). It’s not thinking like you; it’s just a computer program that’s solving a complicated math problem that allows it to spit out a bunch of numbers (that can then be translated into an image by some other code).

1

SeaworthinessFirm653 t1_irq2vcf wrote

That's actually a fair point; I hadn't considered the actual rate of growth in detail.

Edit: There is also an additional facet: If we are to assume that the AI in question is simply an AI designed to create superior AI, and this does indeed cyclically reproduce, then if you were to restrict it to its computational power, it would still run more efficiently than humans by such a large margin. It takes roughly 12 watts to run a human brain power-wise, and if a computer has access to enormously larger amounts of energy, then it is not unthinkable that a machine would be able to self-enhance to an insane degree. Sure, there may be logarithmically declining returns to a certain extent as exists with virtually any system, but the difference between a human and a machine that is at the point of diminishing returns would remain unimaginably wide. Humans were not designed to think; we were designed to be energy-efficient decent thinkers. A machine that can evolve at a million-times faster pace who is designed purely to think will inevitably pass us by a very long margin, even if the nature of acceleration belies an exponential growth function. The main caveat is that creating an AI that can produce superior AI relative to itself for the same intention of creating superior AI is incredibly difficult.

1

SeaworthinessFirm653 t1_irq7a9n wrote

Yes, I made my comment with the presumption that we are talking about AGI, not just a smart calculator bot making a slightly faster calculator bot. We have created some multi-modal AI that can accomplish different tasks, but the model itself is computationally inefficient and predictive rather than true learning (just like GPT-3 doesn't actually think logically, it's just a really advanced language prediction model).

As far as I am concerned, the difference between consciousness and AI is that an AI is an advanced look-up table using only simple logic while consciousness involves processing stored information for semantic meaning rather than adhering to an algorithmic process for syntactic meaning. See: Chinese translation room thought experiment.

AI today uses low-level logic en masse to produce high-level (relatively) thinking. With the addition of increasingly advanced neural networks, image-generation AI has utilized increasingly complex network structures, such as defining edges, shapes, complex shapes, and fuzziness around these levels. If we extend this notion to account for an AI that is capable of taking simple features such as moving shapes and we allow the AI to predict the shapes' locations, we may be able to reapply this scalable logic until the AI is able to understand complex ideas given sufficient inputs and sufficient training data. This is far-fetched from a modern technological standpoint, but not unbelievably far-fetched given how quickly we are advancing our AI.

If the human brain is made up of computations, then an elaborate series of computations is by definition what must define our consciousness, and thus it can be created with sufficient AI models. Switching to amplitude computers for computational efficiency or compressed memory models (current memory cell models scale linearly with space instead of logarithmically) may allow us to break through this barrier.

edit: sorry for the ramble

1

code_turtle t1_irtxyag wrote

I mean that’s HIGHLY optimistic but more power to you, I guess. The “increasingly complex structures” you’re talking about are just fancy linear algebra problems; the idea that those structures will approach “consciousness” anytime soon is a pretty big leap. Imo, we need to first break MAJOR ground in the field of neuroscience before we can even consider simulating consciousness; I think it’s unrealistic to expect something as complex as the human brain to just “appear” out of even the most advanced neural network.

1

SeaworthinessFirm653 t1_iru86y6 wrote

Yes, I agree with that. I recall the analogy of taking the brain's neurons and connections, magnifying it in size to cover an entire block in a large city, and the immense density of connections would still be too large to make any meaningful observations even given our current technology.

I don't believe any optimism is required, though, to claim that we can be simulated. Unless we exist outside of the realm of physical things, that much is given. It's impossible to make good predictions about the future where the sample size is n = 0.

1