Will A.I. Soon be Smarter than Us?

January 1, 2022 Artifical Intelligence No Comments

This text may be interesting to many because these ideas may shape the future of those many to the highest degree. It’s smart to see why something else will be even smarter.

Soon?

Soon enough. The ongoing evolution toward the title’s state will not be evident. In retrospect, it will be an amazingly rash evolution. By then, we may wonder why so many didn’t see it coming.

But, agreed, we’re not there yet.

Why smarter? The short answer.

The short answer is Moore’s Law: a continual, exponentially falling cost per computation unit.

That’s it. This may feel a bit disappointing. Will we be outsmarted because something else becomes cheaper? Is this not disrespectful? Is it not… inhumane?

Yes, well, not necessarily, but possibly ― one more reason to delve into this before it’s too late.

The challenge of past A.I.

The A.I. of a few decades ago was oriented upon the leveraging of human knowledge. This had several reasons, one of them being the neglect of Moore’s Law. Implicitly, A.I. was envisioned within a computation constraint resembling the one (that was thought of as the one) of ourselves. Massively more computation to the disposition of A.I. was not conceived.

Also, there was no sufficient conception of strangely, challengingly different A.I. computations than those en vogue at that time. Deep Neural Networks did exist, but they weren’t taken seriously (enough) — the same with reinforcement learning.

Now, it’s becoming increasingly evident that these two technology areas may together shape the future.

A.I. is in the size.

Not so much the size of more of the same of what seems to work somewhat, but the size of what enables new frameworks and original exploitations of existing frameworks. Fringe ideas may thus be leveraged by unexpected opportunities, overpowering what may seem straightforward at first. The fringe ideas are sometimes much simpler, such as statistics instead of ideas about the problem-solving human mind. How we think we think is not always correct, nor most efficient, nor most scalable with increasing computation capacity.

We may still have more power under the skull than any computer, yet Moore’s Law does not apply to humankind ―except if one wants to become a cyborg (half-human, half-machine). In short, we are stuck with ourselves.

A.I. will increasingly emulate us, and it will increasingly not.

Human learning and knowledge representation methods are interesting but primarily not suited to leveraging silicon computation. A.I. will take advantage of the Law in any way. We have already seen this in A.I.’s overtaking of chess masters, then Go masters, and more. A.I. is already becoming ‘smarter’ than us in many distinctive domains.

Still, we may try to hide in our being smarter in many fields simultaneously. The question is, for how long?

The future lies in self-scalability.

A self-scalable system is self-learning, taking advantage of increasing hardware possibilities. This probably means that the backbone will resemble something like reinforcement learning, in a combination of planning (model-based) and searching (not model-based) ― also, in a combination of conceptual and subconceptual processing. This is just my idea.

This is also crucial in view of the sheer endless complexity of the world outside as well as inside of us, let alone of some intelligence that surpasses us. Such complexity cannot be conceptually emulated. Necessary for this is a system that self-scales in a process of search and discovery. Self-scaling, the system doesn’t only exhibit what the engineer has put into it. Thus, it may be called ‘intelligent’ in the first place.

In any case, the meta-method of self-scalability is a must for any A.I. to reach super-A.I. and go beyond. We humans don’t even know what lies beyond. Will we be able to understand? Probably not, because there is no end to the progress. The bow doesn’t follow the arrow, which by itself becomes another supercharged bow that again doesn’t follow the arrow.

In competition, we have no chance.

Let’s hope for something different.

Leave a Reply

Related Posts

A.I. Will Be Singular

We tend to see human intelligence as what ‘intelligence’ is all about. Many humans each have intelligence. Of course, A.I. will not be bound by this. “If A.I. simulates human intelligence, is this then real intelligence, or only a simulation of intelligence?” This appears to be a merely philosophical question. It will soon be much Read the full article…

Why Compassionate A.I. is Effective

In this blog, we embark on a journey to demonstrate how Compassionate A.I. – like Lisa – may become a profoundly effective tool for mind-related matters, not by imitating humans but activating the user’s inner strength Please read Compassion, Basically. Compassion is effective Compassion stands as perhaps the most enduringly effective force in mental healthcare, Read the full article…

Societal Inner Dissociation and the Challenge of Super-A.I.

The rise of artificial intelligence, particularly super-A.I., intersects with Societal Inner Dissociation (SID), presenting significant challenges and potential opportunities. This blog is an exploration of the complex relationship between SID and super-A.I. (A.I. beyond human capabilities), examining how this might exacerbate or mitigate societal dissociation. This is part of the *SID* series. Please read the Read the full article…

Translate »