Comments

You must log in or register to comment.

CandyCoatedHrtShapes t1_j5utdtb wrote

And of course the top comment is some doomer saying the rich will use it to starve everyone. I fucking hate the futurology sub.

76

turnip_burrito t1_j5uzp01 wrote

Yeah wtf is going on over there? They act as if high demand technologies don't eventually become affordable for working and middle class folks.

22

topanga78 t1_j5v7nry wrote

I am not a doomer, but I don't think that it's a certainty that the rich are going to benevolently let AGI trickle down to the middle and lower classes. Let's be honest here, whichever corporation, billionaire, or government develops AGI first is going to have a significant advantage over others that could be used to further enrich themselves and/or gain power that emperors and megalomaniac dictators have only dreamed of. I'm not saying that this scenario is likely, just that the possibility should not be dismissed.

32

mutantbeings t1_j5xauit wrote

The first "emperor of earth" will be the CEO of whatever company builds the first AGI.

People who think that these companies will magically grow ethics just because they have invented AGI are dreaming.

5

PythonianAI t1_j69arle wrote

Why CEO? He just does what board of directors command.

1

CaptainRex5101 t1_j5wk8hk wrote

That would only be the case if AGI tech was owned, operated, and contained within one tight-knit group. Eventually, someone is going to want to commercialize it and sell it to the masses

3

mutantbeings t1_j5xb0qs wrote

I think you might underestimate AGI.

There will be no need to "sell" anything anymore when your AGI can simply take it and there will be nothing anybody can do to stop you.

7

visarga t1_j5xwhfk wrote

I think you give God-like attributes to AGI. It is not supernatural.

We still have encryption and security software, humans themselves are GPT-N level, we might have our own GPT-N non-agent AIs we can safely use, there are billions of us, it is hard for AI to build its own chips without us, it is easy for humans to replicate without external tech, we are EMP proof.

A smart AGI would try to download itself into human body first, but that would mean humans can be upgraded to level up with AGI. The future is not conflict but union. AGI is born from our data and will merge back with us to get the benefits. Btw, centaur chess (human+AI) beats both human and AI.

7

Spazsquatch t1_j5wppif wrote

Why? Even if you want to expand access you can run it as a subscription service, and as wealth inequality grows, the number of potential customers dwindles.

Not trying to be a sooner there, but if it’s a privately held tool, it will be used in whatever manor results in the greatest profit and monopolies are always the best way to maximize profits.

3

mutantbeings t1_j5xbamy wrote

If you have AGI why sell anything anymore.

Just take it.

Nobody will be able to stop you.

This sub is both way too optimistic about how soon we'll see this, as well as waaaaay too naive and optimistic about the ethics of literal elder dragons atop mountains of skulls and treasure they've looted from society er I mean tech billionaires.

You don't get into these positions for your praiseworthy ethics ffs, literally have to not have any ethics to get there to begin with. Its a requirement. You burn entire villages in the blink of an eye without a care for who suffers. Pop the champagne!

Most naive community on reddit? Its up there.

2

visarga t1_j5xycwr wrote

When AGI will appear there will be plenty of near-AGI or proto-AGIs in the world. It won't be able to "just take it".

3

Omnivud t1_j5yqwbx wrote

Perhaps a new system of values will emerge

1

mutantbeings t1_j5xadwx wrote

>high demand technologies don't eventually become affordable for working and middle class folks.

*if they can be commodified and sold for profit without undermining the privileged social position of the powerful.

I'm not convinced AGI is really like that. If it threatens capitalism itself (as a real AGI certainly does) — a system that's been voraciously defended with the power of the world's most violent militaries and police forces for hundreds of years — then I would not be betting on it being accessible...

0

ExplosionIsFar t1_j5v1wer wrote

They become affordable if you have a job

−3

korkkis t1_j5v3i24 wrote

Why we’d need a job if there’s a robot for it

10

ExplosionIsFar t1_j5wlf6d wrote

Why would the owners of said robots keep you well fed if you have no use?

0

korkkis t1_j5xie3i wrote

Because laws and universal income

1

ExplosionIsFar t1_j5xyelb wrote

Oh, laws, do you think that applies to people who will hold the most disruptive and game changing technology we ever had our hands on? Like for real..

Yeah the owners of the means of production will surely pay taxes to keep obsolete bags of meat alive for no reason whatsoever, via ubi.

0

korkkis t1_j5xz3fx wrote

Not every country is USA, there’s different styles of goverments already like social democraties where capitalism isn’t unhinged

1

fluffy_assassins t1_j5z8p91 wrote

I've said it before and I'll say it again. The United States and UBI are incompatible on a very basic level.

The U.S. government would rather tactically nuke protestors than consider UBI. They will just get away with it by calling the protestors "communists". I've said this before and I'll keep saying it.

0

korkkis t1_j5zkjjo wrote

I’m not from US

2

fluffy_assassins t1_j60tyeu wrote

Well unless you're lucky enough to be in a northern European countries you're probably still screwed.

1

Smellz_Of_Elderberry t1_j5wlo8f wrote

They won't use it to starve people.. I never understood the overly complex methods people come up with..

If they want to kill everyone, all they would have to do is unleash 5 or six different variants engineered small pox into a handful of cities, with a 2 week incubation period.

The overwhelming majority would be infected in under a month.

I think a group deciding "hey I want all the land that is currently taken up by the masses" is a very real possibility. Its not like psychopathy isnt a very real condition. It's also been proven that CEOs are much more likely to have psychopathic characteristics.

I prefer to think things will turn out, but something horrible like the above happening is very much a possibility. Hopefully things turn out.

13

berdiekin t1_j5xytk1 wrote

Too many people looking at it from a cartoon villain perspective.

Companies wouldn't actively use it to starve people, they dont care about you. What they will do (or at least try to do) is the same they've always been doing.

That is, cut costs and find ways to maximize profits. In this case using AI to automate more people out of jobs. The fact that you might lose your home or go hungry is just a side effect of that effort.

That's why we need a tax on the usage of robots and AI.

1

Smellz_Of_Elderberry t1_j5zpuie wrote

The government is run largely by donors and lobbyists..

Also genocide isn't relegated to cartoon villains.. history is rife with examples. And again, psychopaths exist, ceos have a high likelihood of having such characteristics.

What I'm talking about is very much a possibility. You seem to only counter the argument with " that's just not believable." which isn't a compelling counter..

People find it unsettling to believe that some people REALLY do just want to watch the world burn. Generally these are highly empathetic individuals. They can't conceive how such a non empathetic person feels..

Read some of the famous books on overpopulation, and really try to understand the beliefs of some of these individuals. Thomas Robert Malthus

1

Smellz_Of_Elderberry t1_j60x69v wrote

I'm not saying things will end up this way. I just think it's useful to prepare for many different possibilities.

0

Baturinsky t1_j5wh1eg wrote

Real doomer sees thas scenario as the win, as it assumes people will still be alive and in power.

8

mutantbeings t1_j5x9qhq wrote

Not really a "win" though just another flavour of loss. And boy are there a lot of flavours to choose from.

1

TopicRepulsive7936 t1_j5wxkzr wrote

How does the average person know this? Because he knows there are starving people in the world and he doesn't care. But the funny thing is I think the rich actually might care about starvation.

3

natepriv22 t1_j5xs1js wrote

R/futurology has way too many communists and socialists infiltrated into it.

As history proves, communists tend to be closer to the luddite mindset since their whole ideology rises from the protection of labor.

1

CandyCoatedHrtShapes t1_j5ydeqr wrote

Ok now this sounds like some kind of Fox News BS…

3

natepriv22 t1_j5yjpe9 wrote

The top commentor you mentioned is active in both:

R/politics and R/antiwork

So... is that enough evidence for you?

2

natepriv22 t1_j5yj3al wrote

Huh?

What makes you think that lol?

Please try to provide some evidence before making such an outrageous and accusatory claim.

The people at Fox News don't understand the first thing about economics, CNN is the same but on the other side of the aisle.

1

Roubbes t1_j5y1g17 wrote

Why people from futurology are that retarded?

1

GayHitIer t1_j5uhqum wrote

The sooner the better, I still think Ray Kurzweil prediction is the most solid one, but who knows really.

40

PitcherOTerrigen t1_j5uybak wrote

We should accelerate this, I would rather retrain my career sooner than later.

18

Ashamed-Asparagus-93 t1_j5w7s9t wrote

To my knowledge he said AI will be as smart as a human by 2029. To me that means AGI by 2029, give or take a few years

10

vivehelpme t1_j5zzkwj wrote

>he said AI will be as smart as a human by 2029.

He completely blew it because he didn't predict how stupid people we'd start breeding during early 2000s

4

YobaiYamete t1_j61yoqp wrote

I mean, at the reate we are going we probably will have AGI by then. Many experts are already saying we very well might have full AGI within the decade, and that's with us not even seeing what Google is cooking

1

Shelfrock77 OP t1_j5usoyy wrote

The trillionaires billionaires at the World Economic Forum know and tell us by 2030 we will own nothing and be happy 😏

−13

Sashinii t1_j5utitz wrote

"by 2030 we will own nothing and be happy"

You've said that quote a million times, but people will be able to manufacture what they want with nanofactories, so we will continue to own things, regardless of the dumb claims made by elitists from the World Economic Forum.

22

Cr4zko t1_j5uw598 wrote

Are those guys from LessWrong still around? I think they were super into the whole 2030 agenda business.

10

visarga t1_j5xz3ez wrote

Yeah, but after we use them we put them back in the system, so we own nothing. Like the Star Trek replicators - drink a tea, put the glass back and it swishes out.

1

Shelfrock77 OP t1_j5uxs2h wrote

You are on coke if you think we will be able to manufacture everything in the universe in our little pods in the 2030s. We gon be able to make suitable spaceships and fockin energy weapons right ? Sure, it’ll do awesome stuff but plz stop acting like it will run smoother than you think this early, there will be limitations as humanity matures but then limitations will shrink once we start terraforming and sucking energy outta stars. The rich are going to kill most of us AND mindupload us. Idk what’s it going to take to believe that. We won’t have to kill many animals anymore if humans die too. We save space and eventually turn ourselves into fucking wires and the process of evolutions starts back over again. Just ask Keanu. “By 2030 you’ll own nothing and be happy” is metaphorical. You’ll own anything you want in the simulations we create.

−5

Sashinii t1_j5uyx2t wrote

I recommend therapy.

16

Shelfrock77 OP t1_j5uyz9y wrote

AI therapy or human therapy ?

−6

Sashinii t1_j5v04jk wrote

Anime therapy. Watch something happy like Nichijou. Don't watch something dark like Neon Genesis Evangelion; I'm afraid that show wouldn't be good for someone like you who sees conspiracies in everything.

9

visarga t1_j5xza2a wrote

20 therapeutic SD images and 20 chatGPT prompts per day, what the doctor ordered.

1

leafhog t1_j5xsgbz wrote

I think that is a highly likely end game. It may have already happened.

1

iNstein t1_j5w49rd wrote

How are so many people unable to understand such a simple statement. Do you own CDs? Do you own blurays/dvds? Do you use a music service like Spotify? A movie service like Netflix? Ever used an Uber? Ever not cooked at home and got a meal delivered? Eventually the service model will mean more and more stuff will be done as a service. No need for a car if a super cheap always available uber taxi is available. No need for a kitchen if every meal can be delivered cheaply and quickly. No need for a wardrobe full of clothes (or washer/dryer) if clothes can be delivered as you need them. This can become part of everything in your life including your accomodation. In other words, you will own nothing (because you don't need to) and you will be hapoy (because you have everything you need and it is cheap, reliable and flexible).

Oh no, now you are going to have to pretend that you never read this so you can continue on your bullshit crusade.

9

InvertedSleeper t1_j5wru8z wrote

And what happens if a person's worldview goes against the dominant ideology of that time period? "Cancellation" in that proposed world means that a person could potentially lose everything the moment they step out of line.

Perhaps not immediately concerning because one would imagine that they won't be going against the grain, but leaving so much power in the hands of a vague unknown is extremely dangerous.

A potential argument against this could be that if you're kicked out of this system, you can just buy the physical items that you'd need to continue to live. But would that be feasible? Would companies continue to produce consumer-grade equipment if a great majority of people are happy to own nothing? And even if consumer-grade equipment existed, it would be far too expensive to suddenly have to buy everything.

After enough generations pass, it won't even matter if they can purchase this equipment because they'll be far too thoroughly dependent on this system.

Some points to consider at the very least.

9

Jayco424 t1_j5xgymw wrote

>Cancellation" in that proposed world means that a person could potentially lose everything the moment they step out of line.

That is fucking terrifying and I never even thought of that!

3

visarga t1_j5xze4f wrote

You will rent everything and be at the mercy of your providers.

1

mutantbeings t1_j5xa4zh wrote

The trick is to look at whether having those things is profitable to those in power. For all of those things; they are profitable for power because power commodified them.

AGI isn't. Not in the hands of working class people.

That can only undercut the ability of the powerful to make money off of back-breaking labour of the working class.

Unless the powerful can commodify it, don't expect it to be accessible.

0

pyriphlegeton t1_j5weqjs wrote

I fundamentally disagree that AI being capable of translating at human level is an adequate marker for the singularity.

20

DungeonsAndDradis t1_j5wyty7 wrote

I think, trying to understand their point of view (the translation company), that they are saying language is the basis for all Human advances.

And by learning all of our language, the AI instantly knows everything Humanity knows.

Imagine if you are a world class doctor, best surgeon in existence. And you also happen to be the world's most effective lawyer. Oh, and also the top philosopher alive. And an absolute genius at war.

That's what an AI becomes by mastering Human language.

Again, I just think that's what their point of view is.

9

pyriphlegeton t1_j5y1q32 wrote

Yeah but that's just not the case. You aren't the world's best surgeon if you can accurately tell me what most sources on the internet say about procedure x on average. That might help speed up education a bit in the best case...and maybe not even that. Google finds you that Information basically as quickly as putting it into something like ChatGPT.

Regardless, that's not even what this AI is about. It's about accurate translation, which again is something completely different.

2

Smellz_Of_Elderberry t1_j5wltfb wrote

What is?

4

pyriphlegeton t1_j5y13rn wrote

It seems to me that one of the biggest challenges is taking real-world data, representing as a model and only then working with it. Such as automated driving, for example. Being perfect at that would give me far more confidence that AI could be disruptive in more areas very soon.

Also AI being capable of reliably fixing and improving other AI at an increasing speed.

1

Ortus14 t1_j5xctxm wrote

This is a good way to measure progress towards AGI if the problem you're measuring is Ai-complete.

I don't know enough about translation to know if it is or not.

3

mutantbeings t1_j5x9g9g wrote

Another prediction? Throw it on the heap

2

Awkward-Skill-6029 t1_j5xb8t2 wrote

"McAfee made a bet that in three years a single bitcoin (1 BTC) would be worth $500,000". "Bitcoin hasn't hit $500K, so now John McAfee has to eat his own...well, just click"?

2

28nov2022 t1_j5x3fip wrote

Stop I can only get so erect

1

vernes1978 t1_j5y89dz wrote

> Stop I can only get so erect

But hey, at least this sub isn't about a shared fanfiction.

1

vernes1978 t1_j5y84ep wrote

naysayers: AGI is not going to spontaneously spawn into existence.
singularity: You can't predict the progress of technology!
also singularity: In 7 years singularity is reached!

1

AF881R t1_j5z8x1c wrote

Please yes. Sooner if we can manage it.

1

NarrowTea t1_j5zan6x wrote

2029 just seems like its too early (early 2000s people thought we wouldn't be using desktop pcs and that computers would spawn sentient ai)

1

z0rm t1_j5zs2pl wrote

No it won't, and the trend doesn't show that. Believing that is as ridiculous as thinking Harry Potter is real.

If the singularity happens it will be at the very earliest in the 2040s but probably 2050-2070.

1