Patterns + Rewards in A.I.

May 29, 2023 Artifical Intelligence No Comments

Human-inspired Pattern Recognition and Completion (PRC) may significantly heighten the efficiency of Reinforcement Learning (RL) — also in A.I.

See for PRC: The Brain as a Predictor

See for RL: Why Reinforcement Learning is Special

Mutually reinforcing

PRC shows valid directions and tentatively also realizes them. RL consolidates/reinforces the best directions and attenuates the lesser ones.

Without RL, PRC may go forward pretty slowly, like a person who can learn only in small steps from what he already knows — crawling, never jumping.

Without PRC, RL may go forward pretty stochastically, like a person in a dark room searching for the light switch without any guidance. PRC then provides a glow through which the person may guess probable directions at least. He may find the switch after a limited search. Next time and in a different room, he can make use of the glow even better. To an inexperienced observer, he may seem to find the switch miraculously easily.

RL within PRC

This may take away some confusion from the reader who is already thinking beyond the above at this point.

Indeed, the recognized and completed pattern may incorporate the reward itself. It’s a semantic choice. However, this shows how things can overlap, and through their overlap may lead to new possibilities.

The lesson from humans

The combination enables humans to learn from few examples, as even children do spontaneously. It’s a significant part of the way the brain works.

It can be used for the same in A.I. Through PRC and rewards, the system can learn where to evolve toward in a way similar to humans. This makes the need for many rewards and smooth rewarding less stringent, as this is a substantial bottleneck in present-day A.I.

In humans, PRC is realized in a specifically human way that is intrinsically related to the human medium. Probably most crucial in this are our countless mental-neuronal patterns. These enable a flexible and performant kind of PRC, albeit a very fuzzy one.

Probably so fuzzy that we should never indulge this in a non-Compassionate vein of super-A.I. Yet I fear that we are precisely closing into that, unfortunately.

At the same time, a boon for Compassion

Broadly overlapping patterns are pointers to Compassion, basically. They enable intra-brain intuition and our natural kind of conceptual intelligence. They also enable our natural urge for social thinking. Within us, this goes together to a large degree, making us ‘social animals’ to whom ethics frequently means a big deal. Our intelligence has essentially been developed socially over a very long time.

The same provides hope for a Journey Towards Compassionate A.I. if that gets based on likewise principles even when from an utterly different background. An example of this is being realized in Lisa.

We may be very optimistic if we don’t blow it.

Leave a Reply

Related Posts

Depth: Hidden Key to A.I.’s True Potential

A.I. has brought us to the brink of extraordinary progress, yet its potential remains largely untapped. Why? Because people – and the systems they create – are ‘not ready for depth.’ This reluctance includes individuals, businesses, and entire regions, creating a fundamental barrier to fully leveraging the A.I. revolution. Dear reader, this probably includes you. Read the full article…

A.I. and In-Depth Sustainability

Soon enough, A.I. may become the biggest opportunity (and threat) to human-related sustainability. I hope that AURELIS/Lisa insights and tools can help counter the threat and realize the opportunity. This text is not an enumeration of what we may use present-day A.I. (or what carries that name) for to enhance sustainable solutions. It’s about Compassionate Read the full article…

Compassion as Basis for A.I. Regulations

To prevent A.I.-related mishaps or even disasters while going into a future of super-A.I., merely regulating A.I. is not sufficient ― presently nor in principle. Striving for Compassionate A.I. There will eventually be no security concerning A.I. if we don’t put Compassion into the core. The main reason is that super-A.I. will be much more Read the full article…

Translate »