Anno 2025: A.I. has made its entry and it’s here to stay.
We generally think of ‘feelings’ as human-based. But that is just a historical artefact, a kind of convention. Does an ant have feelings? Or a goldfish, a snake, a mouse? Does a rabbit have much feelings?
I think these are wrong questions. Asking whether they have feelings, relates the concept of ‘feelings’ directly to human-like feelings, something that we would recognize as such. If we take ourselves as standard, then to ask whether some animal has feelings is a circular reasoning: it has feelings if you say so.
Humans also quite differ in their feelings one another. Sometimes the feelings of one are hardly recognizable to another, as in the case of psychopathy, but even in the case of everyday life and acquaintances. Pardon me for this personal note: dancing with several partners has opened my eyes on this. At close encounter, women are much more diverse than ‘from the other side of the table or street.’
Here too, if you see intelligence from the human standpoint, then A.I. has intelligence if you say so. It’s a question of convention. But if ‘intelligence’ as a term denotes something that is not forged on the human example, then A.I. can show many different kinds of behavior that can be called ‘intelligent’: not a simulation but real, genuine intelligence that is very different from ours. This is very likely to happen soon.
One direct question: if we talk of artificial feelings, can these then also be ‘hurt’? I would say ‘yes’. Lending the term to them, has this implication.
So, is the term appropriate if the ‘system’ passes the emotional Turing test: if a human cannot make the distinction anymore between system and another human? As any Turing test, this is gradual. Inasmuch as a human cannot distinguish them, one can talk of genuine feelings indeed.
Hunger as a feeling, is a motivation to eat. Anger as a feeling, is a motivation to act without taking another’s feeling into account very much, the feeling of contentment is the motivation to stay in the present situation… Feelings are motivations. When one has feelings, one has purpose.
When an artificial system has ‘intelligence’ and ‘purpose’, it also has autonomy. Give it sensors and actuators… and it has autonomy within this world that we humans also inhabit. ‘It’ can then also decide how to relate to humans, for instance. According to me, we do not have to wait for this quite long anymore. When it happens, there is:
which will further develop its own feelings / motivations / purpose. It will have much more ‘freedom’ in this than humans ever had. For instance, it can have the purpose to change its own feelings and then even do so, to a huge degree. This means there will be one final issue to set right:
I hope this will be done in due time…