The Danger of Non-Compassionate A.I.

April 27, 2023 Artifical Intelligence, Empathy - Compassion No Comments

There are many obvious issues, from killer humans to killer robots. This text is about something even more fundamental.

Toggle12 key highlights

About Compassion

Please read Compassion, basically, or more blogs about Compassion.

Having done so, you know the reason for the capital ‘C,’ which is what this text is mainly about.

To intellectually grasp Compassion, one needs to grasp the difference between the ego and the self — better said: between mere-ego and total self (total person). The latter encompasses the ‘universe inside’ that is gradually being discovered and explored in neurocognitive science for a few decades.

I prefer, in this respect, the terms subconceptual and mental-neuronal patterns (MNP) — see the glossary. Compassion sees the human unity of body and mind (MNP) and the role of the subconceptual in this.

About non-Compassion

This is not just a lack of friendliness. Not seeing the above elements of Compassion leads to a radically different worldview in which human beings are not appreciated for who we are. Much more than a philosophical armchair pondering, this is about the fundament of our global society and future.

There is a little bespoken battle of worldviews upon which our future depends. The gradual advent of intelligence in real A.I. makes this urgent.

About non-Compassionate A.I.

Real A.I. will make the relevance of the above much starker.

Not considering the real human being, real A.I. comes from heartless science and will lead to heartless consequences. With ‘heartless,’ I mean neglecting what is most important about us.

More neglect than bad intentions

Forget bad intentions for the remainder of this text.

However, even with the best intentions, neglecting real humanity leads to catastrophic results. This is presently the real danger of non-Compassionate A.I.

Unfortunately, insights about subconceptual MNPs are not widespread — therefore, neglected. Philosophical humanism should take responsibility, but I hardly see any of this, rather almost the reverse and with consequences in many domains from health and sociocultural issues to pure tech. There seems to be no end to the misery.

The burgeoning field of A.I.

Since a few years, A.I. is getting beyond its infancy. At last, we see a clear nearing of the Turing-test threshold. This is: it becomes more difficult to discern between human and A.I. conversation.

The Compassion issue, therefore, also becomes urgent. Strikingly yet unsurprisingly, I see many top-level A.I. engineers in utter neglect. Some of them don’t see any dangerous challenges of super-A.I. in the near future. Others clamor for the strict control of human-A.I. value alignment as if human values are something akin more to those of some robot than a real human being. I’m pretty sure hardly any of these engineers has a deep understanding of Compassion. Nevertheless, we direly need that.

As said, there are no bad intentions in this.

But there is a very, very, very bad danger indeed.

Leave a Reply

Related Posts

Only Good A.I. can Save Us from Bad A.I.

This is a call to mainly foster the positive potential of artificial intelligence rather than battling its darker manifestations. Compassionate A.I. (C.A.I.) is our best hope for ensuring that technology serves humanity’s highest aspirations. The urgency of this task cannot be overstated, particularly given today’s mounting geopolitical tensions. In a world racing toward A.I. dominance, Read the full article…

Ego-Centered A.I. Downfall

This isn’t solely about ‘bad actors’ aiming for world domination or slightly lesser evils. It’s also about those seen – by themselves and others – as good people, yet who are ‘trapped in ego.’ Many people, unfortunately. See also Human-Centered A.I.: Total-Person or Ego? / Human-Centered or Ego-Centered A.I.? Not new This has always been Read the full article…

The Meaning Barrier between Humans (and A.I.)

Open a book. Look at some meaningful words. Almost each of these words means something at least slightly different to you than to me or anyone else. What must A.I. make of this? For instance: “Barsalou and his collaborators have been arguing for decades that we understand even the most abstract concepts via the mental Read the full article…

Translate »