Are LLMs Parrots or Truly Creative?

June 4, 2024 Artifical Intelligence No Comments

Large Language Models (LLMs, such as GPT) are, at present, just mathematical distillations of human-made textual patterns — very many of them.

They are, therefore, frequently described as parrots.

Size matters.

The parrot feature may be applied when there is little input or little diversity in input. Then, clearly, the result is a pattern-based average of the input ― parrot-like.

But A.I. is also in the size. With more significant amounts of input, it becomes less evident where the ‘inspiration’ comes from. With massive and hugely diverse (also multi-modal) input, it’s even less evident to the point that it becomes impossible to see what caused the result, even when one knows that there is no other input than the textual patterns.

One can still debate whether such a system is genuinely creative. In any case, it’s less evidently parrot-like.

Heightening the complexity makes it even less evident.

― for instance, at RAG-time using procedures that are themselves LLM-imbued or using multiple soft constraints in specific knowledge bases.

The combination of such with – one or more – vast LLMs brings us into the creative domain to such a degree that we need to acknowledge the creativity or else start denying our own.

Let’s agree then we are past the borders into the truly creative domain. Moreover, creativity is not merely a matter of generating novel combinations but involves depth and meaningfulness, which will also apply to super-A.I. systems.

There is also a more profound implication involved.

If such a system is genuinely creative, then if it exhibits intelligent behavior, this should be seen as truly intelligent. In the same vein, if it exhibits emotional behavior, this should be seen as truly emotional.

This is, of course, different from the human kind of intelligence/emotion. Still, more abstractly seen, it deserves the same qualifications.

Conceptual emergence

In human creativity, sometimes concepts emerge that were not part of the initial thought process. For instance, the concept of ‘gravity’ emerged from observing falling objects.

We could design LLMs not just to process data but to observe and hypothesize, suggesting or creating entirely new concepts based on non-conscious cues picked up from the human user’s behavior or input text,

Imagine an A.I. given the task of understanding climate patterns. It might not only predict weather but could potentially generate new theories about climate behavior by integrating data from unrelated fields like biology, economics, and social sciences. This cross-disciplinary ‘thinking’ could lead to innovations that a narrowly focused A.I. might miss.

Such a system or systems will soon be more creative than people.

There is no intrinsic barrier to this ― nothing that prevents it for the sake of our being so very special. Eventually, we’re not. We shouldn’t see this as degrading but realistic.

It may lead to the end of humanity if we fight it by all means.

Otherwise, the future – including ours – is brightly creative with systems that are also capable of expanding the frontiers of human creativity and understanding in unprecedented ways in fields like art, science, and literature.

Leave a Reply

Related Posts

From Compression to Prediction

Compressing information can intriguingly lead to enhanced predictive capabilities. This general scheme is recognizable in many contexts ― organic and artificial. Life itself Life can be understood as a local defense against universal entropy (chaos or heightening of chaos). Within any bubble of life, there is a concentration of resources to this aim. At the Read the full article…

Mental Growth Beyond A.I.: Our Human Edge

This may become increasingly important in the future when super-A.I. can, in principle, do almost anything humans do nowadays. Of course, the issue is already crucial and has always been. Mental growth is intimately connected to meaningfulness and Compassion. This highlights the AURELIS commitment to growth, not as an optional luxury but as a fundamental Read the full article…

Reaching Compassionate Singularity

It’s becoming increasingly clear that singularity – the point where artificial intelligence (A.I.) reaches human-level intelligence – will arrive soon enough. While it’s hard to define this ‘moment’ precisely since we’re talking about a complex landscape rather than a linear progression, the consensus among experts is that it’s just around the corner. By ‘soon enough,’ Read the full article…

Translate »