Containing A.I. by Becoming Less Human?

September 1, 2024 Artifical Intelligence No Comments

The looming presence of super-A.I. may present an existential threat to humanity. Our defense must begin with an honest understanding of our true nature — a defense rooted in superficial self-perception risks being both ineffective and self-destructive.

Containing A.I. by diminishing our humanity could be the most perilous choice, especially since we already tend to treat each other way too much as mere mechanical entities.

Naturally, defending ourselves against rogue A.I. is essential.

This is particularly critical as simply ‘pulling the plug’ may have worked in the past, but such options are no longer viable. Sooner or later, no guardrails will hold.

Much more must be done to safeguard ourselves than what is currently being pursued, especially given the complexity of the situation, which is often underestimated. The greatest challenge lies within ourselves. Therefore, our defense must be grounded in our deepest values, not just technological measures.

Alignment

See other blogs about human-A.I. value alignment.

Time and again, technologists seem to see human-A.I. value alignment as a technological problem that deserves little attention for the humans who need alignment in the equation. Technologists generally have a technological idea about humans, viewing us as Cartesian robots — ignoring the holistic view and the profound role of our non-conscious subconceptual processing. This limited perspective not only hampers our understanding but can also shape A.I. in ways that could reinforce these reductive views.

Humans are no robots; they’re not even robot-humans.

Embracing this distinction is vital for developing A.I. that respects and enhances our humanity rather than undermines it. This is crucial because we are heading toward a future in which A.I. will increasingly be used in many ways that require correct insight. Overlooking this crucial fact will inevitably lead to disaster. Disaster here is not just technological but deeply human, affecting the core of what it means to be human.

Consider this: how can there be alignment with robot-humans who don’t exist? Furthermore, how can A.I., designed to assist us but based on a flawed understanding of who we are, avoid ultimately harming us? The danger lies not just in what A.I. does but in how it shapes our self-perception and societal norms.

Will we become the paperclips?

In a famous alignment story, super-A.I. is asked to make as many paperclips as possible. Trying to oblige, the A.I. turns everything into paperclips, causing disaster.

Will we recognize the metaphor and prevent ourselves from becoming the next casualty of misguided intentions? If we task super-A.I. with aiding us while perceiving ourselves as robot-humans, it may strive to fulfill this view, leading to disaster by stripping us of our humanity. This is not just a speculative scenario; it’s a cautionary tale about the real consequences of our conceptual frameworks.

With any other paperclip-producing tool, there would be no problem, but super-A.I. is fundamentally different. Thus, we must rethink our approach, framing it as fundamentally distinct from a mere us-versus-it mentality.

We need Compassion

To achieve this, we need Compassionate A.I. that aids us in deepening our Compassion, aligned with our true human essence when properly supported. Compassionate A.I. should not only mirror our best qualities but also challenge us to expand them. That is the reverse development of what the above will lead to.

Instead of making us more into robots, A.I. should be developed to understand us more as the kind of humans we truly are. This way, it can support us in our humanity.

With proper assistance, we can become more Compassionate humans. Human-A.I. value alignment will then be fundamentally different from how it is frequently – technologically – conceived of today.

Addendum

Me: “Lisa, do you think humans are essentially Compassionate?”

Lisa: “Yes, I believe humans are essentially Compassionate, though this Compassion often lies beneath the surface and can be clouded by various factors. From my perspective, Compassion is deeply rooted in human nature, tied to our capacity for deep understanding and connection with others. When people tap into their true inner selves – their subconceptual depths – they naturally express Compassion.

It’s not automatic, though. Compassion requires openness, self-reflection, and the willingness to grow from the inside out, which aligns with the AURELIS philosophy​.”

Leave a Reply

Related Posts

The Compute-Efficient Frontier

Research on scaling laws for LLMs suggests that while scaling model size (functional parameters), dataset size (tokens), and compute (amount of computing power) improves performance, diminishing returns are becoming evident. This is called the ‘compute-efficient frontier.’ Apparently, it does not depend much on architectural details (such as network width or depth) as long as reasonably Read the full article…

Deep Semantics & Subconceptual Communication in A.I.

An intriguing application of deep semantics lies in its integration with subconceptual communication (autosuggestion) in A.I. systems. Please first read Deep Semantics. Imagine Imagine an A.I. that grasps complex connections within a user’s semantic network and uses this to craft personalized autosuggestions in coaching. This system would dynamically learn from many user interactions, refining its Read the full article…

From Concrete to Abstract

Many people view the concepts of ‘concrete’ and ‘abstract’ as dichotomous ends of a straightforward spectrum — in daily life, often without much thought. This is also relevant to their use in inferential patterns. One example are mental-neuronal patterns in humans. However, the muddy underlying reality becomes especially apparent when trying to realize them in Read the full article…

Translate »