AI and Consciousness (Part 9)

*** Go to Part 1 *** *** Go to Part 8 ***

This is the crucial part of the series. ChatGPT explains why it is not conscious – in Advaitic terms.

Continue reading

AI and Consciousness (Part 8)

*** Go to Part 1 *** *** Go to Part 7 ***

***********************


Continue reading

AI and Consciousness (Part 7)

*** Go to Part 1 *** *** Go to Part 6 ***

Continue reading

AI and Consciousness (Part 6)

*** Go to Part 1 *** *** Go to Part 5 ***

[Note that, if you are only interested in Advaita-related aspects, you can safely ignore this part and the next and wait for Part 8.]

************************

Continue reading

AI and Consciousness (Part 5)

*** Go to Part 1 *** *** Go to Part 4 ***

Continue reading

AI and Consciousness (Part 4)

*** Go to Part 1 *** *** Go to Part 3 ***

Continue reading

AI and Consciousness (Part 3)

*** Go to Part 1 *** *** Go to Part 2 ***

Continue reading

AI and Consciousness (Part 2)

*** Read Part 1 ***

Opinions

When we are asked a question, we consult our memory for relevant information and how we have evaluated that (based upon our memory of related data and how we evaluated that…). And we evaluate all of this in relation to the present situation and formulate an answer. Is this process mechanically any different from that used when a LLM AI answers a question? Surely the only difference is that it uses a ‘memory’ of data that originated from what others have written down and which is available on the Internet, rather than our reliance upon a ‘remembering process’ of diminishing efficacy.

So the value of an AI response lies in the relative importance placed upon the various sources and the impartial and analytic ability to synthesize a conclusion. We are probably biased, consciously or not, by a desire to appear clever or whatever, whereas a machine is just following algorithms engineered to provide the ‘best’ answer.

None of this relates to ‘consciousness’ particularly. The human brain has its own ‘power source’ that functions electrically via neurochemistry in the brain; AI has an electrical power source. We are ‘aware’ of the conclusions that pop out of the ‘thinking process’ and may formulate them into vocal or written words forming an ‘opinion’. AI is able to formulate conclusions and communicate them via the internet. Can this be called an ‘opinion’ in the same way. Is it actually any different in essence?

Continue reading

AI and Consciousness

With the increasing use of AI, questions have been raised about its responses to various issues, flagging up mistakes, misunderstandings and potential dangers of various magnitude, with end-of-the-world scenarios as the ultimate concern. Anyone with some familiarity with using Large Language Models on the Internet will share these worries to varying degrees; certainly in the context of AI having any control over defense systems, performing operations or Air Traffic Control, for example. There have already been reported incidents of driverless cars making serious errors.

One of the most interesting questions relates to its human-like ‘behavior’. The computational strength of these AI models is so great that they are able to respond to questions in a style that mimics that of a human. They can appear to exhibit empathy. They may seem to sympathize with the questioner’s frustration or echo their elation at solving a problem. Introverted and socially inept people may even seek a simulated friendship with AI. It has been reported, if we can believe it, that people are ‘marrying’ an AI simulation!

Continue reading

World disappearing and mithyā

Not a lot of discussion on the site recently. I thought maybe the time was ripe for a little controversy again. Recently I queried ChatGPT for Shankara references on this topic. I also encountered just a few days ago, a devastatingly logical argument against this idea, but I will save that for a later post. Here is the (single post) AI contribution. Of course, you may argue, AI is presenting its position so as to support my own view – a recognized problem with AI at present). Perhaps, then, Ramesam or Venkat could tackle AI with essentially the same query to elicit a response in keeping with their belief? That would be interesting!

A (ChatGPT): Adi Shankaracharya, the great Advaita Vedanta philosopher, defines the concept of mithyā in several of his works, particularly in his commentaries on key Vedantic texts. The term “mithyā” generally refers to the idea of falsity or illusoriness. It’s crucial in Advaita Vedanta because it relates to the nature of the phenomenal world and how it is perceived in relation to the ultimate reality, Brahman.

One of the most explicit definitions of mithyā is found in Shankaracharya’s commentary on the Brahma Sutra (Brahmasutra Bhashya) in the section known as the Adhyasa Bhashya. Here, he describes mithyā as something that is neither completely real (like Brahman) nor completely unreal (like the horns of a hare), but rather something that appears real under certain conditions but is revealed as unreal upon closer examination or when true knowledge is attained.

Continue reading