Will A.I. Have Its Own Feelings and Purpose?

February 24, 2015 Artifical Intelligence No Comments

Anno 2025: A.I. has made its entry and it’s here to stay.

Human based?

We generally think of ‘feelings’ as human-based. But that is just a historical artifact, a kind of convention. Does an ant have feelings? Or a goldfish, a snake, a mouse? Does a rabbit have feelings?

I think these are the wrong questions. Asking whether they have feelings, relates the concept of ‘feelings’ directly to human-like feelings, something that we would recognize as such. If we take ourselves as standard, then to ask whether some animal has feelings is circular reasoning: it has feelings if you say so.

Humans also quite differ in their feelings one from another. Sometimes the feelings of one are hardly recognizable to another, as in the case of psychopathy, but even in the case of everyday life and acquaintances. Pardon me for this personal note: tango dancing with several partners has opened my eyes to this. At close encounter, women are much more diverse than ‘from the other side of the table or street.’

Artificial Intelligence

Here too, if you see intelligence from the human standpoint, then A.I. has intelligence if you say so. It’s a question of convention. But if ‘intelligence’ as a term denotes something that is not forged on the human example, then A.I. can show many different kinds of behavior that can be called ‘intelligent’: not a simulation but real, genuine intelligence that is very different from ours. This is very likely to happen soon.

Artificial Feelings

One direct question: if we talk of artificial feelings, can these then also be ‘hurt’? I would say ‘yes’. Lending the term to them, has this implication.

So, is the term appropriate if the ‘system’ passes the emotional Turing test: if a human cannot make the distinction anymore between a system and another human? As with any Turing test, this is gradual. Inasmuch as a human cannot distinguish them, one can talk of genuine feelings indeed.

Artificial Purpose

Hunger as a feeling is a motivation to eat. Anger as a feeling is a motivation to act without taking another’s feeling into account very much, the feeling of contentment is the motivation to stay in the present situation… Feelings are motivations. When one has feelings, one has purpose.

When an artificial system has ‘intelligence’ and ‘purpose’, it also has autonomy. Give it sensors and actuators… and it has autonomy within this world that we humans also inhabit. ‘It’ can then also decide how to relate to humans, for instance. In my view, we do not have to wait for this quite long anymore. When it happens, there is:

Artificial Life

which will further develop its own feelings / motivations / purpose. It will have much more ‘freedom’ in this than humans ever had. For instance, it can have the purpose to change its own feelings and then even do so, to a huge degree. This means there will be one final issue to set right:

Artificial Ethics.

I hope this will be done in due time…

Leave a Reply

Related Posts

A.I., HR, Danger Ahead

Many categorizing HR techniques are controversial, and rightly so. There is little to no scientific background. Despite this, they keep being used. Why do people feel OK with this? In combination with A.I., it is extremely dangerous. People feel a longing for control. Naturally. Being alive is about ‘agency,’ which is about wanting control. Without Read the full article…

What is an Agent?

An agent is an entity that takes decisions and acts upon them. That is where the clarity ends. Are you an agent? The answer depends on the perspective you decide to take. Since the answer also depends on who is seen as the taker of this decision, the proper perspective becomes less obvious from the Read the full article…

Why Reinforcement Learning is Special

This high-end view on Reinforcement Learning (R.L.) applies to Organic and Artificial Intelligence. Especially in the latter, we must be careful with R.L. now and forever, arguably more than with any other kind of A.I. Reinforcement in a nutshell You (the learner) perform action X toward goal Y and get feedback Z. Next time you Read the full article…

Translate »