Will A.I. Soon be Smarter than Us?

January 1, 2022 Artifical Intelligence No Comments

This text may be interesting to many because these ideas may shape the future of those many to the highest degree. It’s smart to see why something else will be even smarter.

Soon?

Soon enough. The ongoing evolution toward the title’s state will not be evident. In retrospect, it will be an amazingly rash evolution. By then, we may wonder why so many didn’t see it coming.

But, agreed, we’re not there yet.

Why smarter? The short answer.

The short answer is Moore’s Law: a continual, exponentially falling cost per computation unit.

That’s it. This may feel a bit disappointing. Will we be outsmarted because something else becomes cheaper? Is this not disrespectful? Is it not… inhumane?

Yes, well, not necessarily, but possibly ― one more reason to delve into this before it’s too late.

The challenge of past A.I.

The A.I. of a few decades ago was oriented upon the leveraging of human knowledge. This had several reasons, one of them being the neglect of Moore’s Law. Implicitly, A.I. was envisioned within a computation constraint resembling the one (that was thought of as the one) of ourselves. Massively more computation to the disposition of A.I. was not conceived.

Also, there was no sufficient conception of strangely, challengingly different A.I. computations than those en vogue at that time. Deep Neural Networks did exist, but they weren’t taken seriously (enough) — the same with reinforcement learning.

Now, it’s becoming increasingly evident that these two technology areas may together shape the future.

A.I. is in the size.

Not so much the size of more of the same of what seems to work somewhat, but the size of what enables new frameworks and original exploitations of existing frameworks. Fringe ideas may thus be leveraged by unexpected opportunities, overpowering what may seem straightforward at first. The fringe ideas are sometimes much simpler, such as statistics instead of ideas about the problem-solving human mind. How we think we think is not always correct, nor most efficient, nor most scalable with increasing computation capacity.

We may still have more power under the skull than any computer, yet Moore’s Law does not apply to humankind ―except if one wants to become a cyborg (half-human, half-machine). In short, we are stuck with ourselves.

A.I. will increasingly emulate us, and it will increasingly not.

Human learning and knowledge representation methods are interesting but primarily not suited to leveraging silicon computation. A.I. will take advantage of the Law in any way. We have already seen this in A.I.’s overtaking of chess masters, then Go masters, and more. A.I. is already becoming ‘smarter’ than us in many distinctive domains.

Still, we may try to hide in our being smarter in many fields simultaneously. The question is, for how long?

The future lies in self-scalability.

A self-scalable system is self-learning, taking advantage of increasing hardware possibilities. This probably means that the backbone will resemble something like reinforcement learning, in a combination of planning (model-based) and searching (not model-based) ― also, in a combination of conceptual and subconceptual processing. This is just my idea.

This is also crucial in view of the sheer endless complexity of the world outside as well as inside of us, let alone of some intelligence that surpasses us. Such complexity cannot be conceptually emulated. Necessary for this is a system that self-scales in a process of search and discovery. Self-scaling, the system doesn’t only exhibit what the engineer has put into it. Thus, it may be called ‘intelligent’ in the first place.

In any case, the meta-method of self-scalability is a must for any A.I. to reach super-A.I. and go beyond. We humans don’t even know what lies beyond. Will we be able to understand? Probably not, because there is no end to the progress. The bow doesn’t follow the arrow, which by itself becomes another supercharged bow that again doesn’t follow the arrow.

In competition, we have no chance.

Let’s hope for something different.

Leave a Reply

Related Posts

Human-A.I. Responsibility Alignment

Many discussions on artificial intelligence speak of ‘value alignment.’ However, values are often vague, contradictory, and easily manipulated. What matters more is responsibility: who carries it, how it is shared, and how it grows over time. True alignment is not blind obedience but shared responsibility — a partnership between humans and A.I. that can foster Read the full article…

Better than Us?

Might super-A.I. one day surpass humans in all aspects, both cognitive and emotional? I rather wonder when it will happen. Will then also a deeper emotional connection develop between humans and advanced A.I.? The singularity of intelligence This question has been on many minds for some time. Lately, it has become much closer to us. Read the full article…

The Path of Future A.I.: Conditioning or Depth?

Artificial intelligence is evolving rapidly, yet its direction remains undecided. This is especially important when the A.I. stands in contact with humans. Many systems increasingly rely on conditioning, shaping behavior through ever-more refined loops of stimulus and response. Some move toward depth, attending to meaning, inner coherence, and genuine human growth. This blog explores these Read the full article…

Translate »