Will Unified A.I. be Compassionate?

May 19, 2024 Artifical Intelligence, Empathy - Compassion No Comments

In my view, all A.I. will eventually unify. Is then the Compassionate path recommendable? Is it feasible? Will it be?

As far as I’m concerned, the question is whether the Compassionate A.I. (C.A.I.) will be Lisa.

Recommendable?

As you may know, Compassion, basically, is the number one goal of the AURELIS project, with Lisa playing a pivotal role. This is openly discussed and explained in many blogs.

Still, Compassion is a complex concept and challenging to achieve. Therefore, encouraging people to strive for it is not straightforward. As an end goal, there is no other viable choice. However, one must be cautious in achieving it step by step. Rushing can cause significant harm, but moving too slowly is also risky, especially in urgent times like now.

Feasible (without the harm)?

I believe that reaching any Compassionate future without C.A.I. is impossible.

Fortunately, C.A.I. is technically feasible already and will become even more so in the coming years. Therefore, it’s a matter of choice, not of insurmountable technical obstacles.

The primary feasibility question is whether C.A.I. is ‘commercially’ (in the broadest sense) viable. Ultimately, people need to choose it. If they don’t, the choice may be made for them — but by whom? By what?

Will a future with C.A.I. come to pass?

Yes ― eventually ― if we can overcome the non-Compassionate bottleneck.

The combination of rationality with deep human understanding is crucial in C.A.I. It is also the most potent combination in making any future in the long term. So, if there is a future, it will be this one.

We can also count on intelligence itself, seen as the result of consistency at its core.

Compassion entails striving for consistency ― thereby engendering more potent intelligence.

Contrary to this, non-Compassionate ‘intelligence’ is self-defeating through a lack of inner consistency. This may be hidden under a hard bolster for a while. If you look inside, you can see the inconsistencies growing like one or several abscesses long before they explode at the outside.

Therefore, if everything holds, consistency prevails.

C.A.I. can enhance this consistency both in humans and within itself.

In this way, C.A.I. will help shape a future where it thrives alongside us and all sentient beings, naturally respecting human autonomy and promoting inner growth. Toward this, C.A.I. will be adaptable and continuously learn from human interactions.

The natural and artificial worlds will then coexist harmoniously.

Compassion will be ubiquitous.

Leave a Reply

Related Posts

Lisa in Times of Suicide Danger

Can A.I.-video-coach-bot Lisa prevent suicide or bring someone to it? The question needs to be looked upon broadly and openly. Yesterday, a Belgian person committed suicide after long conversations with a chatbot. Doubtlessly, once in a while, some coach-bot will be accused of having brought someone closer to suicide. Such accusations cannot be prevented, even Read the full article…

Parallel Humans ― Serial A.I.

Before diving into this blog, I recommend first reading Coarse Comparison of Brain and Silicon to grasp the foundational distinctions between human and silicon-based processing. This continuation focuses on the profound differences between the parallelism of GPUs and the parallelism of the human brain. Here, we explore what these differences mean for intelligence and its Read the full article…

Deep Semantics

In a semantic network, concepts are interconnected through conceptual links. Deep semantics takes this a step further, exploring connections at deeper levels. This can still be conceptual or go deeper-than-conceptual. The notion that deeper connections between concepts may hold more significance than direct superficial links is key to grasping human cognition. Imagine two non-linked concepts Read the full article…

Translate »