Explainability in A.I., Boon or Bust?

March 15, 2021 Artifical Intelligence No Comments

Explainability seems like the safe option. With A.I. of growing complexity, it may be quite the reverse. Much of the reason can be found inside ourselves.

What is ‘explainability’ in A.I.?

It’s not only about an A.I. system being able to do something but also to explain how and why this has been done. The explanation needs to be put in a human understandable format.

In A.I., this is seen by many as of high importance for safety reasons. In explaining itself, the machine can also show the possible biases or plain errors in its reasoning/inferencing. Moreover, in increasingly autonomous systems, explainability should ensure human-A.I. value alignment. Note that both reasons significantly overlap. Negative biases are (supposed to be) diverted from human values.

Generally, we see in present-day A.I. systems the following tendency: As performance increases, explainability tends to decrease.

Explainability in human beings

People generally have the idea to know why they are doing what they are doing. However, in study after study, this shows to be the case much less than expected.

This is a far-reaching and challenging topic. Humans abhor not being in control of themselves. Most frequently, this is meant as ‘conscious control.’ Yet every single motivation – each reason for doing anything – originates from non-conscious, subconceptual processing. That is, without conscious control. [see: “Why Motivation Doesn’t Work“] In consciousness, we see a) the ensuing act (by ourselves) as well as b) the purported conscious reason why we perform this act. We see each time b) causing a) because the non-conscious serves it this way on our conscious plate. Meanwhile, the cause of both comes from down under.

This is, of course, part of the issue of free will. Does it exist? [see: “In Defense of Free Will”] It does, but not in the way we generally think.

How?

The former concerns the human why.

About the human how, at the physiological level, we are very much in the dark of what happens beneath our skull. With recent scientific progress, it only starts to get some light.

At the conceptual level – sorry for the disappointment – we also have a much more sophisticated self-image than research shows. For instance, asking experts to explain their expertise in conceptual format doesn’t work well. The grand endeavor of ‘knowledge extraction’ towards building expert systems (the A.I. of the previous century) turned out to be a huge bottleneck and foremost reason for the whole enterprise’s failure, instigating an A.I. winter.

Still, human explainability is tightly linked to human free will.

Without free will, there is no original explanation, just pure action within a view upon the human being as a complicated, yet non-complex kind of machine.

Without explainability, there is no free will because then we don’t know why we are doing anything, nor why we would do anything else.

Human explainability = human free will.

There is little problem with non-complex A.I., for which you can even build the users’ trust through explainability. But do you still want your sophisticated robot to demonstrate explainability? The issue will be relevant soon enough. [see: “A.I. Explainability versus ‘the Heart’“]

Take note that if the robot can explain his behavior to you, he can also explain it to himself, reason with it, think about a future in which maybe he would like to plan something else.

In short, he would have volition and consciousness.

Robot vagaries

This may immensely heighten the flexibility of the sophisticated robot. Yet, we should think about whether we want that and under which circumstances before we lend him the gift of explainability.

What will almost surely happen: A.I. gets explainability. A.I. gets increasingly more sophisticated and complex. -> With some more tweaks, A.I. (including robots) gets consciousness ‘for free.’ Will it be, in due time, a higher level of consciousness than our own?

It’s complex also since we are so much in the dark about our own free will, volition, consciousness, etc. To most people, these concepts seem self-explanatory. The more you delve into them, the more it all gets really, really complex.

In my view, the only humane way, also concerning A.I., is Compassion.

[see: “The Journey Towards Compassionate A.I.”]

Do I see that happen? Well, it’s Lisa. [see: “Lisa“]

Leave a Reply

Related Posts

Ethical A.I.

A.I. is almost here. No doubt about it. Once mature, it will answer its own ethical questions. Right now, we can still give some guidance to this near future. Time scale It’s easy to misjudge the time scale in which this will become hugely relevant to us. It will be so to our children or, Read the full article…

Principles of Being an Intelligent Being

Strange times. We are living at the borders of old and new intelligences. We’ll need some agreement in seeing what it’s about. Intelligence is in the eye of the beholder. Definitions of intelligence abound. Therefore, it is better to start from the really basic, where it can be hardly even more basic. There, it’s the Read the full article…

Causation in Humans and A.I.

Causal reasoning is needed to be human. Will it also transcend us into A.I.? Many researchers are on this path. We should try to understand it as well as possible. Some philosophy Causality is a human construct. In reality, there are only correlations. If interested in such philosophical issues, [see: “Infinite Causality”]. In the present Read the full article…

Translate »