Future A.I.: Fluid or Solid?

March 24, 2021 Artifical Intelligence No Comments

Humans are fluid thinkers. That gives us huge strength and some major challenges. The one does not go without the other. A.I. – including Semantic A.I. – is still a very different matter.

Through proper context, data becomes information.

Still, the information as it is stored in a book is not in any way like the information that we ‘store in our human memory.’

Likewise, smart data in a present-day A.I. system is not in any way like data stored in the human brain/mind. For starters, human ‘software’ = hardware. Mind IS brain ― at least, to a substantial degree.

Likewise, human intelligence is not in any way like so-called intelligence in present-day A.I. Mainly, human intelligence is much more fluid.

Even human solidity in thinking is embedded in fluidity.

In humans, every concept, every piece of solidity, every thought and feeling is the result of fluidity. Our stream of consciousness is a stream of non-conscious fluidity of which only parts break through the surface of conscious awareness. This way, we think parallel and distributed. [see: “Patterns in Neurophysiology”]

This makes us robust and highly efficient. It comes at a price: proneness to logical errors, forgetfulness, biases of all kinds, dissociation leading to health problems in body and mind. We should probably not want to create a duplicate of ourselves, even if that would be possible.

This is also about the age-old question of culture versus nature. [see: “Culture and Nature”]

Why should A.I. need to be fluid?

It is needed to be able to see reality as only a fluid system can see it. What we call ‘common sense’ certainly pertains to this domain.

In dealing with humans, only a fluid system can understand us as we naturally are. This needs to be reckoned with as A.I.-systems become more complex. We should strive not to robotize humans through A.I. applications. [see: “Robotizing Humans or Humanizing Robots”]

Semantic A.I.

= the use of knowledge graphs + present-day A.I. technologies + natural language processing.

A ‘semantic net’ is a knowledge graph, basically a huge bunch of concepts and conceptual links.

The aim of semantic A.I. is an integration of top-down processing (~knowledge graphs – deductive) and bottom-up processing (~deep neural nets – inductive). The former is solid; the latter is fluid. In this striving, it resembles the human case, but not in the way it tries to perform the combination. This last word is crucial. A combination is not a synthesis, no intrinsically integrated whole. Integration that may be achieved is not from-inside-out.

Semantic A.I. does not think parallel and distributed in a human way.

Compassionate A.I. takes proper care of fluidity.

As a general principle, solid A.I. – as any solidity – is only meaningful if it serves as a container to fluidity. This is by definition itself of meaningfulness. [see: “The Meaning of Meaning”] The information in a book is not meaningful by itself. It only becomes so through a (fluid) reader.

To meaningfully heal – from inside out, mentally and psycho-somatically – as well as to attain and enhance our typically human potential, fluidity is necessary. A coaching chatbot should be Compassionate.

In my view, of course, all A.I. should be Compassionate to a fair degree. [see: “The Journey Towards Compassionate A.I.”] This entails taking care of fluidity as is appropriate, profoundly thinking about the Compassionate side also when it doesn’t appear relevant at first sight.

Only this way does the future look Compassionate.

Leave a Reply

Related Posts

Bringing Compassion to the World through A.I.

This is the crucial idea behind the philanthropic project of Planetarianism as part of the AURELIS project. You can find a blog about Planetarianism here and a concrete overview presentation (ppsx for laptop) here. Concretely, it’s a set of projects aiming for this blog’s title. Compassion, basically, is no rosy moonshine. There are strong traditions Read the full article…

Humanity Contra A.I.?

This blog is not about Compassionate A.I. (C.A.I.). Quite the opposite. It’s about the kind of A.I. that lacks Compassion — the kind that, if left unchecked, could become the greatest threat humanity has ever faced. For simplicity, let’s just call it A.I. The fundamental mistake many make is believing that we can control A.I. Read the full article…

A Tapestry of Compassionate Patterns

The AURELIS blogs form a vast network of interconnected insights woven together by a common thread: Compassion. This tapestry transcends the boundaries between seemingly unrelated domains, revealing deep connections and offering a foundation for meaningful understanding. Through this, Lisa emerges as a lens for uncovering the hidden bridges of meaning, amplifying the potential of this Read the full article…

Translate »