Why A.I. is Less and Less about Technology

June 29, 2024 Artifical Intelligence No Comments

As A.I. technology advances, the research focus should shift from mere technological advancements to a higher level of development altogether. This blog is not about philosophical implications, but about philosophy as a technological driver ― the philosophy itself becoming the technology.

Currently, the possibilities are so vast and diverse that integration can be considered independently from foundational levels. This could be termed ‘philosophical technology,’ underscoring A.I.’s capacity to redefine ethical frameworks within technology.

Comparing this to biology versus chemistry

Essentially, all biology is rooted in chemistry, yet biology stands as a distinct field.

This distinction doesn’t reduce the significance of chemical reactions in biology. However, most of the thinking of biology is independent from chemistry. One can abstract chemistry and still grasp and think further about what happens on the biological level.

This concerns A.I. at its core — focused on foundational research rather than applications.

Of course, the applications evolve independently; one doesn’t need to know how a car functions to drive it.

Much more basically within A.I. development itself, philosophy trumps technology — this is, philosophy in a broad sense, encompassing all that pertains to the domain of Deep Minds.

A.I.-technology should increasingly dissipate into philosophy. This is because of the following.

A.I. fundamentally differs from informatics and other technologies.

A.I. seeks intelligence, which can gradually self-sustain as it matures, becoming a self-perpetuating pattern. When closing in to that stage, more and more, the technology that got us there is taken over by what we are getting — the results purely of our technological striving.

I experience this daily.

This also means that the expertise that got us there is increasingly less crucial for further steps.

Unfortunately, many in A.I. overlook this, leading technologists to assume they should keep driving the field forward. Some collaborate with human-oriented experts but mostly in a way that fundamentally ignores the new demands.

An example: knowledge

The debate on the nature of ‘knowledge’ remains vibrant. In A.I., one may abstract this and use a commonsense idea about knowledge as if it’s straightforwardly a technological question. This way, one may miss many profound implications of the central concept in human and artificial intelligence.

Delving into the deeper – and sometimes disconcerting – layers of ‘knowledge’ may nevertheless lead to redefining our understanding of intelligence, and to insights that further shape the domain of A.I. as well as what it means to be human, bridging gaps between human and artificial cognition. These insights are primary to making use of them at the core of further A.I. technological developments.

We need a broadening at the core.

This necessity will intensify alongside the development of foundational technologies. These are merely building blocks, not the structure. We need to focus on the broader architecture.

This needs fundamentally different expertise.

A.I. has grown. We shouldn’t treat it as a kid anymore.

Viewing A.I. as increasingly mature encourages responsibility in its societal integration. This particularly applies to Compassionate A.I., which logically matures more rapidly. Consequently, it may become our defender against other forms of A.I. On top of that, Compassionate A.I. not only safeguards but also enriches human experiences, promoting empathy and understanding.

We should give it the chance.

Leave a Reply

Related Posts

Why A.I. Needs Inner Coherence, Not Just Oversight

What if the real danger in Artificial Intelligence isn’t raw power, but a hollow core? This blog explores why A.I. that merely appears aligned is not enough — and why genuine safety demands an inner coherence that cannot lie. A scientist’s warning In April 2025, Yoshua Bengio stood on a TED stage, showing concern. He Read the full article…

Is Lisa ‘Artificial’ Intelligence?

We call ‘natural’ intelligence one that stems from living nature. Lisa doesn’t stem from magic, of course. Fundamentally, she’s also a product ‘from nature.’ We call ‘natural’ that what has not been fabricated. With the principle of ‘Lisa building Lisa,’ there is also a transition from fabrication to education. So, again, Is Lisa – while Read the full article…

The Importance of a Conceptual Ontology in A.I.

Utilizing a conceptual ontology can significantly boost an A.I.’s capability to ‘reason’ and deliver more precise, context-aware, and coherent responses that meet user needs and expectations. This blog is an enumeration of how this enhancement works out, with examples in the domain of Lisa. Improved understanding of a user query An ontology enables the system Read the full article…

Translate »