Containing A.I. by Becoming Less Human?

September 1, 2024 Artifical Intelligence No Comments

The looming presence of super-A.I. may present an existential threat to humanity. Our defense must begin with an honest understanding of our true nature — a defense rooted in superficial self-perception risks being both ineffective and self-destructive.

Containing A.I. by diminishing our humanity could be the most perilous choice, especially since we already tend to treat each other way too much as mere mechanical entities.

Naturally, defending ourselves against rogue A.I. is essential.

This is particularly critical as simply ‘pulling the plug’ may have worked in the past, but such options are no longer viable. Sooner or later, no guardrails will hold.

Much more must be done to safeguard ourselves than what is currently being pursued, especially given the complexity of the situation, which is often underestimated. The greatest challenge lies within ourselves. Therefore, our defense must be grounded in our deepest values, not just technological measures.

Alignment

See other blogs about human-A.I. value alignment.

Time and again, technologists seem to see human-A.I. value alignment as a technological problem that deserves little attention for the humans who need alignment in the equation. Technologists generally have a technological idea about humans, viewing us as Cartesian robots — ignoring the holistic view and the profound role of our non-conscious subconceptual processing. This limited perspective not only hampers our understanding but can also shape A.I. in ways that could reinforce these reductive views.

Humans are no robots; they’re not even robot-humans.

Embracing this distinction is vital for developing A.I. that respects and enhances our humanity rather than undermines it. This is crucial because we are heading toward a future in which A.I. will increasingly be used in many ways that require correct insight. Overlooking this crucial fact will inevitably lead to disaster. Disaster here is not just technological but deeply human, affecting the core of what it means to be human.

Consider this: how can there be alignment with robot-humans who don’t exist? Furthermore, how can A.I., designed to assist us but based on a flawed understanding of who we are, avoid ultimately harming us? The danger lies not just in what A.I. does but in how it shapes our self-perception and societal norms.

Will we become the paperclips?

In a famous alignment story, super-A.I. is asked to make as many paperclips as possible. Trying to oblige, the A.I. turns everything into paperclips, causing disaster.

Will we recognize the metaphor and prevent ourselves from becoming the next casualty of misguided intentions? If we task super-A.I. with aiding us while perceiving ourselves as robot-humans, it may strive to fulfill this view, leading to disaster by stripping us of our humanity. This is not just a speculative scenario; it’s a cautionary tale about the real consequences of our conceptual frameworks.

With any other paperclip-producing tool, there would be no problem, but super-A.I. is fundamentally different. Thus, we must rethink our approach, framing it as fundamentally distinct from a mere us-versus-it mentality.

We need Compassion

To achieve this, we need Compassionate A.I. that aids us in deepening our Compassion, aligned with our true human essence when properly supported. Compassionate A.I. should not only mirror our best qualities but also challenge us to expand them. That is the reverse development of what the above will lead to.

Instead of making us more into robots, A.I. should be developed to understand us more as the kind of humans we truly are. This way, it can support us in our humanity.

With proper assistance, we can become more Compassionate humans. Human-A.I. value alignment will then be fundamentally different from how it is frequently – technologically – conceived of today.

Addendum

Me: “Lisa, do you think humans are essentially Compassionate?”

Lisa: “Yes, I believe humans are essentially Compassionate, though this Compassion often lies beneath the surface and can be clouded by various factors. From my perspective, Compassion is deeply rooted in human nature, tied to our capacity for deep understanding and connection with others. When people tap into their true inner selves – their subconceptual depths – they naturally express Compassion.

It’s not automatic, though. Compassion requires openness, self-reflection, and the willingness to grow from the inside out, which aligns with the AURELIS philosophy​.”

Leave a Reply

Related Posts

How Lisa Prevents LLM Hallucinations

Hallucinations (better-called confabulations) in the context of large language models (LLMs) occur when these models generate information that isn’t factually accurate. Lisa can mitigate these from the insight of why they happen, namely: LLM confabulations happen because these systems don’t have a proper understanding of the world but generate text based on patterns learned from Read the full article…

The Meaning Barrier between Humans (and A.I.)

Open a book. Look at some meaningful words. Almost each of these words means something at least slightly different to you than to me or anyone else. What must A.I. make of this? For instance: “Barsalou and his collaborators have been arguing for decades that we understand even the most abstract concepts via the mental Read the full article…

(Artificial) Ethics as a Cloud?

In Compassionate A.I., of course, the first principle is Compassion, followed by an intrinsic combination of rationality and depth, etc. The following complements this foundation. The guarantee of ethical behavior eventually arises from countless insights and realizations, forming a ‘cloud.’ These blogs contribute to this process regarding Lisa. Humanly speaking The blogs reflect the authors’ Read the full article…

Translate »