Forward-Forward Neur(on)al Networks

April 19, 2023 Artifical Intelligence, Cognitive Insights No Comments

Rest assured, I don’t stuff technical details into this blog. Nevertheless, this new framework lies closer to how the brain works, which is interesting enough to go somewhat into it.

Backprop

In Artificial Neural Networks (ANN) – the subfield that sways a big scepter in A.I. nowadays – backpropagation (backprop) is one of the main buzzwords and is used ubiquitously. Present-day A.I. would be nowhere without it. However, that may change.

ANN schema with several layers from left to right

Shortly put, an explanation of backprop:

A multilayered ANN-in-training tries to accurately categorize, for instance, images of cats and dogs. But it makes many errors. After each pass forward through the network from entry layers to exit layers, an error estimation is fed into the system at the end side and, from there, backprops to earlier layers. So, signals go forward; error estimations go backward through the same network. ANNs work well this way for some problems, not for others.

In any case, the human brain hardly works this way.

Firstly, backprop involves a lot of mathematics, while no calculator exists in the brain. There are only neurons – and other cells – and groups of them. The brain may simulate a calculator, functionally approaching it to a small degree. But there is no way the brain could perform as many complex calculations as are needed for backprop.

Secondly, backprop would mean that, at the cellular level, there is a kind of memory that definitely isn’t there.

Thus, although low-level information in brainy neuronal networks can go in many directions, a backprop direction can hardly be one of them. Also, there is no evidence of it in reality. Backprop is especially (extremely) implausible with information that evolves over time.

The brain goes forward.

Neurons in the human brain are unidirectional. As a general flow, the information goes from one side (dendrites) to the cell body and from there to the axon connecting with other cells’ dendrites.

Neuron with flow from left to right

Thus, all in all, and taking into account many loops, forward processing is what the brain does, also when doing it backwards. For instance, much information travels from the brain to the eye, influencing how the eye itself reacts to what it ‘sees.’ This is no backprop but a forward feed in the backward (or ‘top-down’) direction.

A few months ago, Geoffrey Hinton proposed the Forward-Forward algorithm (FF) in neural networks.

Shortly put, if an error is made (cat, no dog). FF handles this network experience in a forward fashion ― erroneous and non-erroneous pass alike. The forward trails of both (thus ‘forward-forward’) are compared, and the system learns.

This is more aligned with how the brain works ― even to such a degree that both cases (natural and artificial) can be described pretty much in the same way. I just did so and will do some more.

Given good and bad examples, a neur(on)al network can proceed in two ways:

  • It can process each pretty much apart from each other. Hinton proposes this may be the main aim of why humans sleep ― a period in which the brain can generate its own erroneous ‘input.’
  • Alternatively, a system can compare good and bad examples as overlays and automatically find out where the overlays mainly differ. This way, the differences are found subconceptually. The conceptual interpretation results from emergence.

Are you following? For the sake of comprehension, imagine two transparent sheets. On one is drawn a cat; on the other a dog that overlaps as much as possible with the cat. Then you can look at this and immediately notice the differences without working anything out. It happens automatically.

The brain intrinsically likes automatic stuff.

Things that happen automatically don’t require much energy or other resources. This makes the brain very power-efficient and responsive, as it can also do with FF neural networks. One promise of the latter is that no massive amounts of memory or distinct processing are needed ― contrary to what is generally the case with backprop. In short, purely forward processing – using a suitable architecture – can do with very much less energy and memory.

Just as it is easier to drop an object on your toe than to calculate how much that may hurt.

Additionally interesting is that no magic is involved while we can explain more and more of the brain’s inner secrets in relatively simple terms that can be experimentally investigated in an artificial medium.

Society of FF-networks

This is a personal addition to FF. At least, I didn’t encounter it yet, but it’s logical. Such a society cannot be but close to the brain’s working as a network of networks (of networks).

One macro function (for instance, dancing the tango, or even one single move) is accomplished by many small brainy networks working together, each performing a part of the macro task. One network’s forward pass can relatively easily be combined with others in many parallel and serial ways ― which is what we see happening in the brain to a huge degree. Of course, nature’s playfield has been immense in this regard.

In such a society, many conglomerates of networks can work in different ways while accomplishing the same macro-functional result. In the human case, this is a well-investigated phenomenon. The same principle may be brought to the artificial world, creating many additional possibilities for efficiency and effectiveness.

(Only) two (of many possible) lessons

  1. We can learn much from the brain when developing genuinely artificial intelligence. The natural case (we) gives us much inspiration.
  2. We can learn from the above abstract thinking how humans are concocted ― including the positive and the flaws. Significantly, such insights may explain why our thinking does an excellent job while simultaneously being such a bunch of biases. It shows how hard it is to keep the intelligence while reducing bias. The biases are part of the intelligence. This explains why many people frequently act stupidly in so many interesting ways.

It also shows how we can ameliorate the situation.

In my view, we will direly need this more than ever soon enough.

Leave a Reply

Related Posts

The Quest for Abstract Patterns

This is about creation. The creative process Some see three levels of creativity: interpolation, extrapolation, and the invention of something new/out of the box. This progress in levels goes from the domain of the known toward the domain of the not-yet-known. One can also see these levels as the possible results of an increasing discernment Read the full article…

Transparency in A.I.

We should strive to the highest degree of transparency in A.I., but not at the detriment of ourselves. Information transparency In conceptual information processing systems (frequently called A.I.), transparency is the showing of all data (information, concepts) that are used in decision making of any kind. In system-human interaction, the human may ask the system Read the full article…

Let Us Go for Science with a Heart

Human-related science has, until now, not always taken the view from the heart. A.I.-based science may profoundly change this picture. If you are an avid reader of my blogs, this one has little new. Yet by putting things together from a different perspective, a new picture may emerge. Reproducibility Science is about many people being Read the full article…

Translate »