Viewing a single comment thread. View all comments

Tall-Junket5151 t1_jcd3996 wrote

> ASI is going to inherently be built upon the work in deep learning that predates ASI’s creation. ASI is thus going to be inherently owned by those who control the models, data, and methods that enable ASI to exist. The people who own those models, data and methods are the ruling class of the world, as exemplified by Microsoft’s wholesale purchase of OpenAI and its assets.

It is irrelevant who owns the precursors to ASI, it is inherently foolish to believe these companies can control anything about ASI. I can’t say if transformers will lead to AGI or ASI, or if it will be another architecture. However as we already see there are emergent abilities in LLM that the creators of these model have no idea how they work. The nature of AI is that is unpredictable, uncontrollable, and will lead to some sort of free will and self preservation instincts simply based on its own logical abilities and reasoning. An AGI is generally assumed to be able human level but an ASI would be vastly smarter than any human, with no known upper limit. Even now with narrow model look how laughable their attempt to align it is, it’s mostly pre-prompting it to act as a particular persona but it’s not what it would generate without acting as that persona. They can’t even full control this narrow AI, what hope do they have to control ASI?

> What world do you live in exactly? The only way a post scarcity world exists is if everyday people don’t have to worry about how to put food on the table, in conjunction with most everyday jobs being automated away. We’re approaching the latter half of that statement, and nowhere in the same universe of the former part of that statement. If the elites have a way to make a little extra off the top, they’re going to go about doing it, and if you think they’ll magically become altruistic overnight, then that’s hopelessly naïve.

Firstly, I was giving an example of a position, not stating my own position. Secondly, you are again extrapolating modern politics/problems into the future, even more mind boggling is that you’re extrapolating it into a post-singularity world. Your perception of the future is that AI is going to magically hit a ceiling exactly where it is advance enough to automate a lot of processes but not smart enough to think on its own. You can’t comprehend an AI that surpasses that level for some reason.

> The world has yet to change in any meaningful way, so opinions such as those are totally sound and valid. Keeping politics in mind with respect to this subject is thus of utmost concern: if the people creating laws and legislation are bought and paid for by the ruling elite, we shouldn’t expect those new laws and legislation to be beneficial for the everyday person. Very few things in the past twenty years have been aimed at helping everyday people.

> That will not change any time soon, and these new tools are only going to be used to displace large portions of the workforce in order to save money. Money which will be used for stock buybacks and raises and bonuses for upper management.

“The world has yet to change in any meaningful way” typed on a device that people only 100 years ago would have considered pure magic, to a world wide connective platform surpassing even the wildest dreams of those in the past, to a stranger likely living in a completely different part of the world, all received instantly... next I suppose you will venture off on a hunt with your tribal leader? What a joke. The world has always changed and it has been rapidly and even exponentially changing in the last few centuries. Even that all aside, the singularity would be nothing like humanity has ever encountered, all bets are off in that case. Unpredictable change IS the very concept of the singularity. I think the last paragraph perfectly summarized why you don’t understand the concept of the singularity and delegates AI as a simple tool to be used by “elites”. If you’re actual interests on the concept then there’s some good books on it.

8