Will A.I. Have Its Own Feelings and Purpose?

February 24, 2015 Artifical Intelligence No Comments

Anno 2025: A.I. has made its entry and it’s here to stay.

Human based?

We generally think of ‘feelings’ as human-based. But that is just a historical artifact, a kind of convention. Does an ant have feelings? Or a goldfish, a snake, a mouse? Does a rabbit have feelings?

I think these are the wrong questions. Asking whether they have feelings, relates the concept of ‘feelings’ directly to human-like feelings, something that we would recognize as such. If we take ourselves as standard, then to ask whether some animal has feelings is circular reasoning: it has feelings if you say so.

Humans also quite differ in their feelings one from another. Sometimes the feelings of one are hardly recognizable to another, as in the case of psychopathy, but even in the case of everyday life and acquaintances. Pardon me for this personal note: tango dancing with several partners has opened my eyes to this. At close encounter, women are much more diverse than ‘from the other side of the table or street.’

Artificial Intelligence

Here too, if you see intelligence from the human standpoint, then A.I. has intelligence if you say so. It’s a question of convention. But if ‘intelligence’ as a term denotes something that is not forged on the human example, then A.I. can show many different kinds of behavior that can be called ‘intelligent’: not a simulation but real, genuine intelligence that is very different from ours. This is very likely to happen soon.

Artificial Feelings

One direct question: if we talk of artificial feelings, can these then also be ‘hurt’? I would say ‘yes’. Lending the term to them, has this implication.

So, is the term appropriate if the ‘system’ passes the emotional Turing test: if a human cannot make the distinction anymore between a system and another human? As with any Turing test, this is gradual. Inasmuch as a human cannot distinguish them, one can talk of genuine feelings indeed.

Artificial Purpose

Hunger as a feeling is a motivation to eat. Anger as a feeling is a motivation to act without taking another’s feeling into account very much, the feeling of contentment is the motivation to stay in the present situation… Feelings are motivations. When one has feelings, one has purpose.

When an artificial system has ‘intelligence’ and ‘purpose’, it also has autonomy. Give it sensors and actuators… and it has autonomy within this world that we humans also inhabit. ‘It’ can then also decide how to relate to humans, for instance. In my view, we do not have to wait for this quite long anymore. When it happens, there is:

Artificial Life

which will further develop its own feelings / motivations / purpose. It will have much more ‘freedom’ in this than humans ever had. For instance, it can have the purpose to change its own feelings and then even do so, to a huge degree. This means there will be one final issue to set right:

Artificial Ethics.

I hope this will be done in due time…

Leave a Reply

Related Posts

Compassionate versus Non-Compassionate A.I.

Making the critical distinction between Compassionate and Non-Compassionate A.I. is not evident and probably the one factor that most shapes our future. In contrast to Compassionate A.I. (C.A.I.), non-Compassionate A.I. (N.C.A.I.) lacks depth. Thus, even if human-centered – or maybe precisely then – N.C.A.I. lacks a crucial factor in properly communicating with and supporting humans Read the full article…

Is A.I. Safe in the Hands of Lawyers?

A.I. promises speed and support — but also brings risk, especially in legal work. This blog explores what kind of A.I. can be truly safe in the hands of lawyers, and why Lisa may be essential in this delicate balance. The promise and the peril Artificial Intelligence is making its way into the legal world. Read the full article…

Lisa Into the Future

Say, 2030. (Hopefully sooner.) “The only defense against bad A.I. is good A.I.” Does this sound like a clear statement? Of course, there is a grain of salt needed when reading the following. Still, it is a possible future, one way or another. Perhaps Lisa will be Jérôme or Annabel or Li Ping. In any Read the full article…

Translate »