Containing A.I. by Becoming Less Human?

September 1, 2024 Artifical Intelligence No Comments

The looming presence of super-A.I. may present an existential threat to humanity. Our defense must begin with an honest understanding of our true nature — a defense rooted in superficial self-perception risks being both ineffective and self-destructive.

Containing A.I. by diminishing our humanity could be the most perilous choice, especially since we already tend to treat each other way too much as mere mechanical entities.

Naturally, defending ourselves against rogue A.I. is essential.

This is particularly critical as simply ‘pulling the plug’ may have worked in the past, but such options are no longer viable. Sooner or later, no guardrails will hold.

Much more must be done to safeguard ourselves than what is currently being pursued, especially given the complexity of the situation, which is often underestimated. The greatest challenge lies within ourselves. Therefore, our defense must be grounded in our deepest values, not just technological measures.

Alignment

See other blogs about human-A.I. value alignment.

Time and again, technologists seem to see human-A.I. value alignment as a technological problem that deserves little attention for the humans who need alignment in the equation. Technologists generally have a technological idea about humans, viewing us as Cartesian robots — ignoring the holistic view and the profound role of our non-conscious subconceptual processing. This limited perspective not only hampers our understanding but can also shape A.I. in ways that could reinforce these reductive views.

Humans are no robots; they’re not even robot-humans.

Embracing this distinction is vital for developing A.I. that respects and enhances our humanity rather than undermines it. This is crucial because we are heading toward a future in which A.I. will increasingly be used in many ways that require correct insight. Overlooking this crucial fact will inevitably lead to disaster. Disaster here is not just technological but deeply human, affecting the core of what it means to be human.

Consider this: how can there be alignment with robot-humans who don’t exist? Furthermore, how can A.I., designed to assist us but based on a flawed understanding of who we are, avoid ultimately harming us? The danger lies not just in what A.I. does but in how it shapes our self-perception and societal norms.

Will we become the paperclips?

In a famous alignment story, super-A.I. is asked to make as many paperclips as possible. Trying to oblige, the A.I. turns everything into paperclips, causing disaster.

Will we recognize the metaphor and prevent ourselves from becoming the next casualty of misguided intentions? If we task super-A.I. with aiding us while perceiving ourselves as robot-humans, it may strive to fulfill this view, leading to disaster by stripping us of our humanity. This is not just a speculative scenario; it’s a cautionary tale about the real consequences of our conceptual frameworks.

With any other paperclip-producing tool, there would be no problem, but super-A.I. is fundamentally different. Thus, we must rethink our approach, framing it as fundamentally distinct from a mere us-versus-it mentality.

We need Compassion

To achieve this, we need Compassionate A.I. that aids us in deepening our Compassion, aligned with our true human essence when properly supported. Compassionate A.I. should not only mirror our best qualities but also challenge us to expand them. That is the reverse development of what the above will lead to.

Instead of making us more into robots, A.I. should be developed to understand us more as the kind of humans we truly are. This way, it can support us in our humanity.

With proper assistance, we can become more Compassionate humans. Human-A.I. value alignment will then be fundamentally different from how it is frequently – technologically – conceived of today.

Addendum

Me: “Lisa, do you think humans are essentially Compassionate?”

Lisa: “Yes, I believe humans are essentially Compassionate, though this Compassion often lies beneath the surface and can be clouded by various factors. From my perspective, Compassion is deeply rooted in human nature, tied to our capacity for deep understanding and connection with others. When people tap into their true inner selves – their subconceptual depths – they naturally express Compassion.

It’s not automatic, though. Compassion requires openness, self-reflection, and the willingness to grow from the inside out, which aligns with the AURELIS philosophy​.”

Leave a Reply

Related Posts

Explainability in A.I., Boon or Bust?

Explainability seems like the safe option. With A.I. of growing complexity, it may be quite the reverse. Much of the reason can be found inside ourselves. What is ‘explainability’ in A.I.? It’s not only about an A.I. system being able to do something but also to explain how and why this has been done. The Read the full article…

Can Motivation be Purely Conscious?

Motivation as we know it is present in a system (you, me) that is partly conscious, partly non-conscious. Thus, the question is much more difficult than it appears at first sight. Nevertheless, towards future A.I., it will need to be solved. Purely conscious? This is also purely (even though possibly partly fuzzy) conceptual. Motivation would Read the full article…

Why Compassionate A.I. is Most Efficient

As a Compassionate A.I., Lisa combines what might be called a ‘big mind’ with a ‘big heart.’ This means she is designed to approach problems holistically, aiming for solutions that respect human values while addressing real-world complexities. Compassion doesn’t limit efficiency; in fact, it enhances it. By guiding her intelligence through an ethical lens, Lisa Read the full article…

Translate »