Submitted by Malachiian t3_12348jj in Futurology
4354574 t1_jdu8nat wrote
Reply to comment by KnightOfNothing in Microsoft Suggests OpenAI and GPT-4 are early signs of AGI. by Malachiian
We're conscious. Subjective experience is magical. The experience of emotions is magical. Being aware of experience is magical. If that isn't magical to you, then...sucks to be you. What is even the point of existing? You might as well just go through the motions until you die.
There is no evidence at all that AI is conscious.
Surur t1_jdubb31 wrote
How do you know you are not the only one who is conscious?
4354574 t1_jdunko1 wrote
I don't. It's the classic "problem of other minds". This is not an issue for Buddhism and the Yogic tradition, however, and ultimately at the highest level all of the mystical traditions, whether Sufism, Christian mysticism (St. John of the Cross and others), shamanism, the Kabbalah etc. What's important to these traditions is what your own individual experience of being conscious is like. More precisely, from a subjective POV, there are no "other minds" - it's all the same mind experiencing itself as what it thinks are separate minds.
If your experience of being conscious is innately freeing, and infinite, and unified, and fearless, and joyous, as they all, cross-culturally and across time, claim the state of being called 'enlightenment' is, then whether there are other minds or not is academic. You help other people to walk the path to enlightenment because they perceive *themselves* to be isolated, fearful, angry, grieving individual minds, that still perceive the idea that there are "other minds" to be a problem.
In Buddhism, the classic answer to people troubled by unanswerable questions is that the question does not go away, but the 'questioner' does. You don't care about the answer anymore, because you've seen through the illusion that there was anyone who wanted an answer in the first place.
Surur t1_jdur5b3 wrote
Sure, but my point is that while you may be conscious, you can not really objectively measure it in others, you can only believe when they say it or not.
So when the AI says it's conscious....
audioen t1_jdw2frs wrote
The trivial counterargument is that I can write a python program that says it is conscious, while being nothing such, as it is literally just a program that always prints these words.
It is too much of a stretch to regard a language model as conscious. It is deterministic -- it always predicts same probabilities for next token (word) if it sees the same input. It has no memory except words already in its context buffer. It has no ability to process more or less as task needs different amount of effort, but rather data flows from input to output token probabilities with the exact same amount of work each time. (With the exception that as input grows, its processing does take longer because the context matrix which holds the input becomes bigger. Still, it is computation flowing through the same steps, accumulating to the same matrices, but it does get applied to progressively more words/tokens that sit in the input buffer.)
However, we can probably design machine consciousness from the building blocks we have. We can give language models a scratch buffer they can use to store data and to plan their replies in stages. We can give them access to external memory so they don't have to memorize contents of wikipedia, they can just learn language and use something like Google Search just like the rest of us.
Language models can be simpler, but systems built from them can display planning, learning from experience via self-reflection of prior performance, long-term memory and other properties like that which at least sound like there might be something approximating a consciousness involved.
I'm just going to go out and say this: something like GPT-4 is probably like 200 IQ human when it comes to understanding language. The way we test it shows that it struggles to perform tasks, but this is mostly because of the architecture of directly going prompt to answer in a single step. The research right now is adding the ability to plan, edit and refine the replies from the AI, sort of like how a human makes multiple passes over their emails, or realizes after writing for a bit that they said something stupid or wrong and go back and erase the mistake. These are properties we do not currently grant our language models. Once we do, their performance will go through the roof, most likely.
4354574 t1_jdwkos3 wrote
Well, I don’t believe consciousness is computational. I think Roger Penrose’s quantum brain theory is more likely to be accurate. So if an AI told me it was conscious, I wouldn’t believe it. If consciousness arose from complexity alone, we should have signs of it in all sorts of complex systems, but we don’t, and not even the slightest hint of it in AI. The AI people hate his theory because it means literal consciousness is very far out.
Surur t1_jdwqof7 wrote
> If consciousness arose from complexity alone, we should have signs of it in all sorts of complex systems
So do you believe animals are conscious, and if so, which is the most primitive animal you think is conscious, and do you think they are equally conscious as you?
4354574 t1_jdx1c88 wrote
If you want to know more about what I think is going on, research Orchestrated Objective Reduction, developed by Penrose and anaesthesiologist Stuart Hameroff.
It is the most testable and therefore the most scientific theory of consciousness. It has made 14 predictions, which is 14 more than any other theory. Six of these predictions have been verified, and none falsified.
Anything else would just be me rehashing the argument of the people who actually came up with the theory, and I’m not interested in doing that.
Viewing a single comment thread. View all comments