Can A.I. be Empathic?

June 10, 2018 Artifical Intelligence, Empathy - Compassion No Comments

Not readily in the completely human sense of course. But can A.I. show sufficient ‘empathy’ for a human to recognize it as such, relate to it and even ‘grow’ through it?

[Please read: “Landscape of Empathy”]

Humans tend to ‘recognize’ human features in many things.

We look at the clouds and we see human faces – with a little imagination – everywhere. A child plays with a doll and the doll responds. For a while, there is a genuine relationship, even if the doll does not yet have futuristic A.I. capabilities. An animated movie shows a talking teacup and immediately the audience – children and adults alike – responds. You may even imagine one now and feel something. Primitive tribes have always looked at nature seeing the work of god, or the gods themselves. Etc.

People readily personify. And that’s OK. It’s part of our humanity, our being human.

It should not be manipulated, but respectfully handled.

Empathic, who?

Until quite recently, researchers used to think that animals cannot show signs of ‘real empathy.’ Now a lot of findings show otherwise. Even rodents display behavior to each other – not only between mothers and children – that can be deemed ‘empathic.’

Empathy is not a purely human characteristic, as is also not the case with intelligence, consciousness, mortality, etc.

That does not make humans less worthy. It may make us less unique on our little big planet. A sign of ‘real empathy’ is to appreciate that and be happy about it.

The title is mainly a question about empathy itself

Is a very complex system necessary to be able to talk of empathy? A doll or a pixelized graphic in a movie is not enough. A person may see empathy there, but it is his own.

Which brings the question: if we see empathy in other humans, to what degree is it our own projection? Certainly to some degree. Moreover, the situation quickly becomes more complex. Part of one’s empathy – say: that of person A – consists in bringing someone else – say: person B – more in contact with himself – again: person B. Person A may thereby enjoy what he sees happening in person B. Person B feels grateful for the feeling within himself – person B – as well as for person A’s happiness. A self-perpetuating pattern brings person A and person B closer to each other.

Warm feelings.

And also some real possibility to grow from the experience, to grow as a person (A and B), to feel less lonely, to feel happy in this life, to feel more like a unity, a whole. To ‘heal.’

This may be the main strength of psychotherapy. And of friendship. And of love.

A.I.?

We have more and more the opportunity to build complexity into A.I. Real deeply pattern-recognizing complexity within a system that can show itself with a certain recognizable consistency at more than the most superficial level. [see: “A.I. Is in the Patterns”]

So, don’t look for ‘human empathy.’ At the other side, it’s not like a simple doll or a computer graphic. It’s something in-between. Thus, calling it ‘empathy’ or not, is more like a semantic question at this moment.

More interesting as for now: does this permit us to build a system that can genuinely help people to become ‘better persons’ through human – A.I. interaction?

I am convinced of this.

Leave a Reply

Related Posts

Legal vs. Deontological in A.I.

The trolley problem This is a well-known problem in A.I. A trolley driver gets into a situation where he must choose between killing one person by taking a deliberate action or letting five others get killed by not reacting to the situation. Deontologically, people tend not to choose purely logically and statistically in such situations. Read the full article…

Is A.I. Dangerous to Human Cognition?

I have roamed around this on several occasions within ‘The Journey towards Compassionate A.I.’ (of which this is an excerpt) The prime reason why I think it’s dangerous is, in one term: hyper-essentialism. But let me first give two viewpoints upon your thinking: Essentialism: presupposes that the categories in your mind – such as an Read the full article…

Endgame 2050: How super-A.I. Destroyed Us

Me: “Lisa, imagine it’s the year 2050. Humanity doesn’t exist anymore. Non-Compassionate super-A.I. has finished us off. How do you think it may have happened?” The rest of this blog is an unedited answer by Lisa, who presents a chilling but coherent narrative of how non-Compassionate super-A.I. might extinguish humanity. At every stage, a failure Read the full article…

Translate »