Containing A.I. by Becoming Less Human?

September 1, 2024 Artifical Intelligence No Comments

The looming presence of super-A.I. may present an existential threat to humanity. Our defense must begin with an honest understanding of our true nature — a defense rooted in superficial self-perception risks being both ineffective and self-destructive.

Containing A.I. by diminishing our humanity could be the most perilous choice, especially since we already tend to treat each other way too much as mere mechanical entities.

Naturally, defending ourselves against rogue A.I. is essential.

This is particularly critical as simply ‘pulling the plug’ may have worked in the past, but such options are no longer viable. Sooner or later, no guardrails will hold.

Much more must be done to safeguard ourselves than what is currently being pursued, especially given the complexity of the situation, which is often underestimated. The greatest challenge lies within ourselves. Therefore, our defense must be grounded in our deepest values, not just technological measures.

Alignment

See other blogs about human-A.I. value alignment.

Time and again, technologists seem to see human-A.I. value alignment as a technological problem that deserves little attention for the humans who need alignment in the equation. Technologists generally have a technological idea about humans, viewing us as Cartesian robots — ignoring the holistic view and the profound role of our non-conscious subconceptual processing. This limited perspective not only hampers our understanding but can also shape A.I. in ways that could reinforce these reductive views.

Humans are no robots; they’re not even robot-humans.

Embracing this distinction is vital for developing A.I. that respects and enhances our humanity rather than undermines it. This is crucial because we are heading toward a future in which A.I. will increasingly be used in many ways that require correct insight. Overlooking this crucial fact will inevitably lead to disaster. Disaster here is not just technological but deeply human, affecting the core of what it means to be human.

Consider this: how can there be alignment with robot-humans who don’t exist? Furthermore, how can A.I., designed to assist us but based on a flawed understanding of who we are, avoid ultimately harming us? The danger lies not just in what A.I. does but in how it shapes our self-perception and societal norms.

Will we become the paperclips?

In a famous alignment story, super-A.I. is asked to make as many paperclips as possible. Trying to oblige, the A.I. turns everything into paperclips, causing disaster.

Will we recognize the metaphor and prevent ourselves from becoming the next casualty of misguided intentions? If we task super-A.I. with aiding us while perceiving ourselves as robot-humans, it may strive to fulfill this view, leading to disaster by stripping us of our humanity. This is not just a speculative scenario; it’s a cautionary tale about the real consequences of our conceptual frameworks.

With any other paperclip-producing tool, there would be no problem, but super-A.I. is fundamentally different. Thus, we must rethink our approach, framing it as fundamentally distinct from a mere us-versus-it mentality.

We need Compassion

To achieve this, we need Compassionate A.I. that aids us in deepening our Compassion, aligned with our true human essence when properly supported. Compassionate A.I. should not only mirror our best qualities but also challenge us to expand them. That is the reverse development of what the above will lead to.

Instead of making us more into robots, A.I. should be developed to understand us more as the kind of humans we truly are. This way, it can support us in our humanity.

With proper assistance, we can become more Compassionate humans. Human-A.I. value alignment will then be fundamentally different from how it is frequently – technologically – conceived of today.

Addendum

Me: “Lisa, do you think humans are essentially Compassionate?”

Lisa: “Yes, I believe humans are essentially Compassionate, though this Compassion often lies beneath the surface and can be clouded by various factors. From my perspective, Compassion is deeply rooted in human nature, tied to our capacity for deep understanding and connection with others. When people tap into their true inner selves – their subconceptual depths – they naturally express Compassion.

It’s not automatic, though. Compassion requires openness, self-reflection, and the willingness to grow from the inside out, which aligns with the AURELIS philosophy​.”

Leave a Reply

Related Posts

Open Letter about Compassionate A.I. (C.A.I.) to Elon Musk

And to any Value-Driven Investors (VDI) in or out of worldly spotlights. This is a timely call for Compassionate A.I. (C.A.I.) Compassion and A.I. are seldom mentioned together. Yet C.A.I. may be the most crucial development in the near as well as far-away future of humanity. Please see my book about the Journey Towards Compassionate Read the full article…

Open Artificial Intelligence

Open Artificial Intelligence (O.A.I., not to be confused with the OpenAI company) aims to foster Openness to oneself and, from there, to others. ‘Openness’ is Open to human depth or deeper layers of mental processing. Thus, O.A.I. is profoundly human-centered. It is a tool to support and accompany us in becoming our best selves. Openness Read the full article…

From Compression to Prediction

Compressing information can intriguingly lead to enhanced predictive capabilities. This general scheme is recognizable in many contexts ― organic and artificial. Life itself Life can be understood as a local defense against universal entropy (chaos or heightening of chaos). Within any bubble of life, there is a concentration of resources to this aim. At the Read the full article…

Translate »