The Meaning of a Word

February 9, 2024 Artifical Intelligence, Cognitive Insights No Comments

“The meaning of a word is its use in the language.” — Quote by Wittgenstein, who talked about language games because ‘use in the language’ can be seen as having a very playful aspect.

A dictionary provides the meaning of a word ― in some way, but always as an approximation.

A word doesn’t have meaning like a book has pages.

Nor like I have two legs. The term ‘have’ is somewhat misleading in this. Meaning – either at surface or deep level – is not something to be had, nor sought or given away. It’s not something altogether.

‘Having meaning’ can, therefore, be very relational. It’s carried in relation to other words or to an environment, including people in that environment. Any context can determine or influence meaning ― again, either surface or deep, but mostly both. Since contexts are continually changing, so are the meanings of words.

What Wittgenstein referred to is that words are living beings, as are the ideas they convey. Everything flows.

Large Language Models (LLMs)

LLMs are based mathematically on Wittgenstein‘s mantra: words as used in language. An LLM derives its competence exclusively from this single principle. That’s enough to come to a sometimes pretty fantastic output.

It may be seen as a parrot, but an immensely sophisticated one, playing complex games without knowing the rules for itself.

An LLM’s ’intelligence’ is in the size. For each word, it uses an immense amount of context. It does so by using mathematical computing power that vastly outstrips ours. Noteworthy, this is also completely different from our way of mental processing. In this vein, ‘it’ will eventually (after adding conceptual processing) understand us better than we can ever understand ‘it’ ― or ourselves.

The step toward concepts.

In contrast to first-generation LLMs, we, humans, don’t use just words. We use them as pointers to concepts when these are positive in the cost-benefit analysis of efficient mental processing. Concepts are a step further in the formalization of communication.

However, the efficiency of formalization comes with diminishing complexity. Through formalization, we put the complexity into boxes, close the boxes, and are able to build with them as such, realizing many complicated – but not complex – things.

The meaning of a concept is its use in a conceptual context.

More so in relation to ontology. Using a strict ontology is a further step in formalization. Ontologization thus brings a huge processing advantage.

Crucially, with each step in formalization, we can use algorithms more efficiently just as we can use other tools. Once a tool – such as a hammer – is made, we can use it many times in slightly different contexts. Everyone knows this is a good thing — if the hammer is used correctly.

As with a hammer however, concepts, ontologies and algorithms can be used to destroy reality. Especially when applied to organic reality, their cautious use is mandatory.

Back to words

Words can carry meaning with a complexity beyond concepts. De-constructing concepts can partly draw this to the fore. The plain use of words in all playfulness – called daily life – brings more to the surface.

Therefore, it will be challenging to see LLMs being developed with the capacity to show a myriad of associations in human language. At the moment, this hasn’t been done yet (overtly). Still, it is a possibility and an occasion to know more about ourselves, how we use words, and what we are thinking when doing so even when we don’t know it ourselves.

From an immense amount of human words onwards, LLMs can ontologize the world as humans see it. Probably, AGI (Artificial General Intelligence) will be reached this way. It’s the perfect means for Compassion.

However!

Due to sheer processing power and if we don’t manage this well, it’s also the perfect means for manipulation.

We (humanity) wouldn’t survive.

Leave a Reply

Related Posts

Why Conscious A.I. is Near

Without pinning a date, it’s dangerous that many researchers/developers are making progress in many aspects of A.I. without deep insight into consciousness. Scary? ‘Near’ in the title is meant relative. The issue is the following. The ways are such, and the competition is such that I don’t see any other option than that we are Read the full article…

Introducing Lisa (Animated Video)

Without further delay, in this animated video, I bring you an introduction to Lisa. [Lisa animated video – 13:15′] If you want to cooperate, please contact us. If you have feedback, please let us know. This is a draft version. Here is the full written text. Hi, my name is Jean-Luc Mommaerts. I am a Read the full article…

Small Set Learning

This approach in A.I. differs significantly from big data learning. It may be the next revolution in town. Small set learning (SSL) is also called ‘few shot learning’ if done at run-time. This blog may interest those who want to know why we’re not at the end of a new A.I. upsurge but at the Read the full article…

Translate »