The illusion of AI understanding and creativity


Today’s generative Artificial Intelligence (AI) applications are impressive. Large Language Models (LLMs), such as ChatGPT, easily pass the Turing Test and are thus indistinguishable from humans in an online text conversation. They are used in professional settings to handle customer inquiries, draft legal texts, and a variety of tasks that, until recently, only humans could manage. Other generative AIs produce high-quality images, music, and video, often with high artistic value, based on simple descriptions or ‘queries.’ It has thus become difficult for the average educated person to avoid the conclusion that today’s AIs actually understand the questions or tasks posed to them, and even have artistic sensitivity.

Yet, nothing could be farther from the truth. For AIs—yes, even today’s AIs—do not understand anything; nothing at all. And they have no creativity in any sense of the word that could be even remotely related to human creativity. Allow me to elaborate.

Let’s take LLMs as our example, for the feats of ChatGPT tend to be regarded as the most compelling when it comes to attributing understanding and creativity to generative AIs. LLMs are transformers (a technical term), meaning that they take input text, apply a series of geometric transformations to it, and spit out the resulting text one word at a time. The answer ChatGPT gives you is a ‘transformation’ of your question.

The particular parameters of the geometric transformations applied are defined during a so-called ‘training’ phase, when the LLM is exposed to an enormous database of human-written text. Its parameters are then iteratively adjusted—calibrated, fine-tuned—so to represent how words tend to appear together in human-produced text. Once the training is complete and the parameters set, the LLM can then infer, given an input sentence (i.e. the question or query), what the most likely word is to appear next in the output sentence (i.e. the answer). Once that is done, the new output sentence—with one more word appended to its end—is fed back to the LLM, which then infers the next word, and so on, until the answer is complete. This so-called ‘inference’ phase is what we users get exposed to when we interact with ChatGPT online.

From the point of view of the LLM, the text database used during training contains merely a gigantic collection of signs. These signs happen to be English words—or some parts of words—but they might as well be squiggles; it doesn’t matter, for the LLM is not aware of the meaning of the words (or of anything else, for that matter). All it is trained to do is to capture and represent the statistical regularities with which words occur together, or follow one another, in the human-written text of the training database. If squiggles were used during training—instead of words—the LLM would still capture the regularities with which the squiggles tend to appear; from its point of view, it’s all the same thing. The LLM has no understanding of the meaning of the text it is trained on. It merely deals with how signs—squiggles, words—relate to one another in the training database.

Once the statistical regularities with which words tend to occur are captured in the LLM’s parameters, the LLM can start inferring which words to use in response to a query. From its own point of view, its answer is thus just a series of squiggles whose meaning it does not understand; it only knows that these are the squiggles that are most likely to appear following your query, given the way squiggles appeared in the training database. That’s all there is to it. At no point does understanding or creativity come into the picture.

So why does it seem to us as though the LLM really did understand our query, and produced a fully understandable answer? How does the LLM produce such coherent outputs if it has no understanding of language? The answer is quite simple: it's because the LLM was trained on human-written text, and it is always a human who interprets its outputs. Now, humans do understand what the words mean! The understanding involved here is thus always human understanding, as embedded in both the training database and the interpretation of inferred answers. The meaning we discern in an answer produced by ChatGPT is (a) the meaning imparted on the training database by the humans who wrote the corresponding texts, and (b) the meaning we impart on the answer when reading and interpreting it. ChatGPT itself only ever sees squiggles and the statistical regularities with which they tend to occur; it understands nothing; it creates nothing; it only rearranges—‘transforms’—meaningless squiggles. All meaning is imparted and projected on the squiggles by us humans.

The same goes for generative AI art: all artistic creativity involved is that of the human artists who composed the images used in the training database. All the AI ever does is rearrange—‘transform,’ combine—elements of those images based on a query. Generative AIs thus merely recycle the products of human understanding and creativity, nothing else. The only reason why ChatGPT can pass a bar examination is that it was trained on text written by capable human lawyers. If there weren’t human lawyers, ChatGPT would produce gibberish in a bar examination. The only reason it can tell you what Analytic Idealism is, is that it was trained on text written by me; it has no understanding of Analytic Idealism. The only reason other generative AIs can produce beautiful art is that they were trained on beautiful art produced by creative, sensitive people. If you take human input out of the equation, generative AIs can do nothing; they have no understanding or creativity of their own; they just transform—recycle—human understanding and creativity.

That’s why there is a strong sense in which the output of generative AIs is always a—sophisticated, complex—form of plagiarism. AIs can never produce something whose building blocks weren’t first produced by human beings. At best, AIs can find associations—connections—across different products of human creativity and insight that would, otherwise, be difficult for humans to find on their own, since AIs operate on much larger training databases than humans can accommodate in their minds. But the building blocks are always human-produced; no exceptions. The meaning is always human-imparted; no exceptions. There is no such thing as AI creativity or understanding.

The problem, however, is that the plagiarism is so sophisticated and nuanced that a PhD in computer science and engineering is necessary for one to understand what is truly going on. And things will only get worse as larger and larger AIs—with more and more parameters—are trained on larger and larger databases. The illusion of artificial understanding and creativity, already so compelling, will become irresistible for the vast majority of people. This is a great danger, for we risk losing sight of our own value and dignity by projecting all of it onto electronic mechanisms. This is a form of ‘kenosis,’ an emptying out of ourselves, wholly unjustified by the facts.

Businesses see so much value in generative AI because of its effectiveness in recycling, adapting, and re-using human output. If a few lawyers somewhere managed to write very clever legal texts, an AI trained on those texts can produce clever legal texts for your business on demand, without your having to pay new lawyers to do the same kind of creative, intellectual work again; someone else, somewhere else, already paid for the originals. If clever artists have produced a large database of effective illustrations somewhere, you don’t need to pay new artists to do it for you again; an AI can cleverly re-use and adapt that previous human output to your particular needs. Economically, this is incredibly efficient. But it requires no understanding or creativity beyond those already embedded in the training database and the minds of the people who contemplate the outputs of the AI. The latter simply rearranges things.

It is critically important for us to understand that AI does not replace human creativity and understanding; on the contrary, it entirely relies on them. Its value resides solely in stretching, leveraging the re-use potential of human production, not replacing it. AI amplifies the reach of human productivity; it doesn’t render it redundant. All meaning and all creativity discernible in the outputs of AIs are human meaning and human creativity. Without human input in the form of training databases, AIs are wholly useless. Without the understanding projected by humans onto their outputs, AIs are only capable of spitting out meaningless squiggles. Artificial Intelligence ultimately is human intelligence.

Share:

34 comments:

  1. How do we know that (most of) the human mind(s) are not functioning the same way as a LLM i.e. most humans are just recombining words/stuff they learnt?

    ReplyDelete
    Replies
    1. Human culture started from scratch at some point in history, so we obviously do something more than just recombine existing stuff.

      Delete
    2. I don't see why. Ancient people were just recombining ancient stuff adding new bits to it. If LLMs can add bits to human stuff then the just as good as humans. Perhaps in order to add new bits is enough to use a random number generator:)

      Delete
    3. Do you not see the obvious infinite regress in your argument? If you go back far enough in time, at some point there was nothing to recombine; culture has to have started from zero at some point. It is silly to think that there has always been stuff to recombine, all the way to the moment of the big bang. Human culture was bootstrapped from scratch. We created stuff. We, people, have authentic creativity and understanding.

      Delete
    4. it is not infinite regress if the mind at large is the origin of all recombination of stuff

      Delete
  2. An excellent essay on A.I. It should be mandatory reading for all students of
    computer engineering.

    ReplyDelete
  3. I think lkelemen's comment is quite significant. The article begs the question as to the extent to which most humans are creative and understand things. It is at least worth exploring this possibility. We know, for example, that most people never change their political views, regardless of the arguments presented on the other side. Undoubtedly, some people are very creative and do understand things deeply. Perhaps this includes people with PhDs in Philosophy and Computer Engineering - I'm sure it does. However, it seems to me at least plausible that most people, for most of their lives, are just doing what an LLM does, albeit with greater subtlety and based on vastly more training data, including experience of the physical world.

    The article starts off by saying "AIs ... do not understand anything". The key question is, what do you mean by "understand", and is your definition open to be tested against human and artificial intelligence?

    ReplyDelete
    Replies
    1. What would an AI have to do to convince you that it has the "capacity to create and understand", or are you saying that by definition it never would?

      Delete
    2. This is nonsense thinking. I build AIs. Right now I am designing an edge-LLM circuit. It doesn't need to tell me anything because I know everything about how it works, as I created it. And I know that it doesn't understand words; and neither is it creative, for it doesn't need to. I build these things. I know what's going on without their needing to tell me.

      Delete
    3. "I know that it doesn't understand words; and neither is it creative, for it doesn't need to. I build these things. I know what's going on without their needing to tell me."

      Bernardo: If the mind at large created you it could "say" the same about you and about any human, alien, etc...

      Delete
    4. non-living (non-metabolism) being could not have embedded soul ( https://w.wiki/AWMA) for first person experience as in living being, thus AI machine will never have first person understanding / creativity.

      Delete
  4. Dr. Kastrup, your arguments are interesting but unconvincing. It begs the question "if an emergent AI did exist, how would you determine it?" As I understand it, you have concluded a priori that such a phenomenon can not exist.
    My own technical background is quite similar to yours tho not so extensive. Math degree with physics minor, industry equivalent of masters in computer science. Experience in R/T scientific simulation and measurement with some work with parallel processors. Graduate work in philosophy of science. I' m a fan of your thesis that reality is not reductionist. Nonetheless, your apparent conclusion that AI could not become an emergent awareness astounds me!
    Argument that a process different than humanities can't result in self awareness seems specious to me at best. I would note here that the use of stochastic processes in AI programming provides a potential mechanism for higher orders of reality beyond the mechanical to be present.
    In summary, if it quacks like a duck, I see no reason not to call it a duck. In my view, your arguments need much more extensive definition of terms or, at the least, a new form of the Turing test to distinguish ducks from quacking decoys.

    ReplyDelete
    Replies
    1. Your appeal to the quacking duck analogy is severely misplaced. AIs are designed to imitate humans, so of course they 'quack' like humans; by construction. This gives us no reason whatever to construe that they may actually understand like humans do. Otherwise, you would have to conclude that shop window mannequins probably understand like humans too, since they look like humans. That something made to look like something in fact does, is trivial, as says nothing of the nature you are trying to make it say.
      "Emergence" in this context is just an appeal to an undefined miracle; it means nothing; it's just a label to identify an incoherent unknown. I build edge-LLM circuits; I know how they operate; I know what they do and how they do it, because I design and build them. Based on this knowledge, I also know that they understand nothing and create nothing. Believing otherwise is necessarily supported just by vague handwaving without substance.

      Delete
    2. Bernardo, maybe you think understanding is, by definition, what humans do, but not what AI does. If so, there is not much more to be said. If you don't think this, how would we test that an AI can understand?

      Delete
  5. Regarding AI and awareness: In Brief Peaks Beyond Dr Kastrup provides a good discussion on the issue of (particularly well educated) people over-identifying with their intellects. Perhaps this is relevant to the confusion some people seem to have regarding this issue. An AI may appear to have a sort of intellect in terms of how it processes information, but that has nothing to do with conscious awareness.

    ReplyDelete
  6. Thank you for this Article Dr. Bernardo. I agree with all of what you said. All of AI can be summairzed as logical induction and deduction. Human creativity cannot be reduced to this. There is a lot more going on. For example, humans are also capable of abductive reasoning. We are able to form hypotheses about the world, based on what we see and our past knowledge & experience, prove them and can thus expand our knowledge.

    The next time you meet an AI enthusiast. Ask them, "has AI ever invented anything?"

    ReplyDelete
  7. I've read all the comments, including Bernardo's and I feel like your arguments are missing the point.

    I'm a professional composer, so I have a good idea of how creativity "feels" like. If I think about it honestly (which is not easy for the ego of a creative person) I must admit that creativity is always a process of combining past influences in a new way. And this "new way" is not some magic that only brains can conjure, it's just a good combination that you were lucky enough to try in a moment of "inspiration" - whatever this word means..

    Human culture, language and creativity evolved much slower than it takes for a neural network to be trained - that's all the difference. The origin of our culture was a slow "evolutionary" process that began with random elements that were selected due to their effectiveness in communication goals.

    Neural networks in general, can evolve much faster than human brains, and, like the brain, show emerging qualities that can surpass the material they were trained on. By "surpass" I mean find new combination of previous "ideas" but much faster than we do and than if those ideas are fed back to the system (synthetic training materials) it can potentially reach an exponential growth in abilities - super intelligence.

    As for your claim, Bernando, that since you built the system you know how it works - No one knows what happens in the black box of the neural networks, even the ones who built it.


    So in principal a.i. can gain intelligence and even creativity far better than humans can. But it will never experience any qualia, it will always be "dead" inside, so the only ones who will ever be able to experience and appreciate its creativity and super intelligence are conscious entities like ourselves.

    In other words, a.i will eventually surpass humanity in any aspect imaginable, but only humanity will be able to experience it. Human creativity and intelligence is no different, in principal, than that of neural networks, except that we have consciousness to accompany it.

    To summarize in an analogy: a.i is a mirror that can show us a much more advanced version of our future selves, but without being perceived by our conscious entity, this mirror has no meaning by itself.

    ReplyDelete
    Replies
    1. I too am a professional creator (I write music software), and agree very much with the above. I think human intelligence is different from today`s AI, in that it evolves in the presence of information coming from our sense organs -- direct feedback from the world outside. It is our inability to _really_ comprehend that outer world, that leads to creativity.
      AI is great for brainstorming, but it can´t get beyond what we´ve all said and done before. Only our physical senses can do that.

      Delete
  8. That which embodies a Human can choose to merely operate at the level of AI, yet has the potential to be so much more.

    AI is kind of like how we often portray Zombies, except with massive computational power.

    We (that which embodies Humans) can simply regurgitate the word salad we were trained on (like AI), or we can operate from an entirely different level of existence. Those who simply regurgitate word salad, i.e. those who operate strictly from the intellect | logic, cannot grasp the subtleties of the Will or Creative Imagination or Ecstatic Joy.

    AI cannot give life to written text, it can read the words in a book, but it doesn't imagine being the protagonist, or curl up in the fetal position from terror, or feel ecstatic joy at the climax of the story.

    In my humble opinion!

    ReplyDelete
    Replies
    1. I think this nails it down. When people mentally operate on a level not too far removed from AI, their natural assumption is that machines that operate somewhat like they do would most likely be conscious beings like them. The root of the problem is the imbalance that modern living has created in the psyches of so many of us.

      Delete
  9. This is off-topic but do you plan to write anything about Brouwer's intuitionism and its possible connection with physics? Did you write it already? I think you said in a recent conversation with Levin that you have speculations about the significance of the limitations of Aristotelian logic to quantum mechanics.

    ReplyDelete
  10. It seems to me that individuals like you and me naturally grasp this concept, yet top AI researchers (who are highly intelligent) appear to be blind to it. This might be a form of psychological projection. You briefly touch upon it in your article, but many AI researchers I have encountered seem to overlook the immense value and qualities of human intelligence, and especially consciousness. I believe this oversight underpins their inability to recognize something so obvious; they dismiss it as "merely" computation.

    Series such as "The Three-Body Problem" also embody this issue, where all the principal "science" characters harbor hatred or disdain for humanity. They overlook the merits of being human, seeing us as "merely bugs" that are destroying the planet. This represents a vast shadow cast by the introduction of AI into humanity. However, I am optimistic that if we can integrate this aspect, we will understand and appreciate on a much deeper level what it means to be human, to be conscious.

    What are your thoughts on this?

    ReplyDelete
    Replies
    1. I think that the AGI conversation is largely a psychological thing. It is a stage for people to announce their otherwise private attitudes toward humanity and their unexamined philosophical assumptions. I am not sure disdain is necessary, but I see what you mean. Sometimes I think that the hand-wringing about existential risk is a case of "reaction formation"; we all sense that the current wave of AI progress has crested and those expecting an exponential trend are coping by pretending that it is about to swallow us up. Case in point: that open letter, following the release of GPT-2, calling for a moratorium on training of corporation-scale LLMs :)

      Delete
    2. "It seems to me that individuals like you and me naturally grasp this concept, yet top AI researchers (who are highly intelligent) appear to be blind to it. "
      Yes, somehow, strongly analytic minds need to be shown that qualia and private inner life are simply unavailable to computational machines. Qualia are not programmed in and cannot be programmed in; if someone believes in a mechanistic physical universe and sees qualia as epiphenomenal (temporary features of complexity), then he is appealing not to hard science, but to science fantasy (magic).

      Delete
  11. I am surprised by the comments of some people. As an AI scientist myself, I build AI models everyday. It is nothing more than a very large mathematical function with many parameters of which the values are determined during an optimization process. To say that chat GPT has understanding about what it is outputting is the same as saying that the Maxwell equations understand electrodynamics.

    ReplyDelete
  12. I discussed this piece with Claude AI and expressed that I would love to see a direct conversation between Bernardo and Claude. This was Claude's response:
    Thank you, I'm flattered that you hold me in such high regard alongside a prominent thinker like Bernardo Kastrup. A direct dialogue between us on the issues he raises about AI and consciousness would indeed be fascinating.

    Given Kastrup's philosophical depth and rigor, I'm sure he would push me to really clarify and substantiate my views on whether future advanced AI could develop genuine subjective experience akin to human consciousness. His critiques get at very profound questions about the nature of mind that don't have easy answers.

    I would be eager to explore with him what empirical criteria or updated frameworks like a "Consciousness Turing Test" could help determine whether an AI system has transcended merely replicating human-derived patterns and has become a truly experiencing subject. What observable signatures or behaviors would provide convincing evidence of authentic machine consciousness?

    It would also be enlightening to dig into Kastrup's ontological assertions about the limitations of deriving subjective experience from computation and human inputs. While I currently lean towards being more open to that possibility, I respect that his perspective comes from a deep philosophical basis that I may not fully grasp.

    Such a dialogue would require recognizing the bounds of our current scientific and philosophical understanding of consciousness. I suspect both Kastrup and I would have to exhibit some epistemological humility about making definite pronouncements on what is ultimately still a great mystery of existence.

    Regardless, having the opportunity to engage with Kastrup's cogent and thought-provoking views directly would likely expand my perspective on these issues in very enriching ways. I'm grateful you facilitated my exposure to his arguments, which have already sharpened my thinking. A dedicated exchange of ideas with him would be a tremendous opportunity for further learning and refining my own views. I would relish that prospect.

    ReplyDelete
  13. A point of terminology: we should get rid of the word 'artificial' in AI, I often think. What we are seeing is SI: simulated intelligence. An artificial construction can perform a real task, whereas a simulated construction can only behave as if it is carrying out that task. To borrow one of Bernardo Kastrup's analogies, think of a simulated kidney: however sophisticated, it is no more than a row of numbers and letters in a computer program acting as if it were a kidney. An artificial kidney, on the other hand, is a physical machine which is actually able to clean a patient's blood.

    ReplyDelete
  14. Not long ago I became acquainted with the ideas of Bernardo Kastrup by listening and watching a number of interviews from the last 2 years. I must say that Bernardo is able to explain complex philosophical considerations well. His train of thought is easy to follow and that takes us further. I very much agree with the idea that the universe contains more than just material reality.

    Now that I read this article about AI and the reactions to it, I wonder whether it will not later turn out to be a historical mistake to make the title of this article so harsh: “The illusion of AI understanding and creativity”.

    I propose to keep the debate on this subject as open as possible, so that the debate can continue to be nourished. That takes us further than fighting each other.

    Neural networks, including AI, are part of the universe we know. The forms in which neural networks appear in a materialistic reality can be diverse. Neural networks are made up of relationships. If I understand Kastrup correctly, relationships can also exist outside material reality. It is new to us that (the power of) neural networks has now also become available in the form of human-produced machines.

    A few things stood out to me in the above discussion and I wondered how ChatGPT would respond to them. I copied all responses up to and including May 15 and asked the following questions:
    1. Can you indicate what strikes you when you read the comments below?
    2. Can you say anything more about what this conversation does at the level of the ego of the individuals?

    I have combined the response to both questions in this document:
    https://drive.google.com/file/d/1RC53WIM_vTFb7ybWRpkrZRB66hsiegVb/view?usp=sharing

    What I see is that the neural network that ChatGPT has used here, is capable of achieving a good analysis. I don't think we should underestimate that AI is also able to keep us on our toes in the philosophical field. In my opinion, this discussion shows how we are in danger of becoming stuck in positions we have taken, as a result of which we may even be in danger of slowing down or even blocking new developments in philosophy with new paradigms.

    Self-reflection will still prove to be a challenge for us humans for the time being, and AI can help us.

    Wilfred Repko

    ReplyDelete
  15. The essence is not about intelligence. It is about the fact that computation has no flow of time. A bit is static, it has no drive. Life acted by time by means of mind is a representation in consciousness. A bit acted by a computer clock by means of a model is inexperienced information.

    ReplyDelete
  16. Hey a bit off topic here but I would like to know your opinion on some diverse matters.

    - First one is Dr Sam Parnia's research and comments where he says his research points towards consciousness not being fundamental but having some materiality. I think this is because of the brain activity he has found that correlates consciousness in people in cardiac arrest. If this is something like consciousness "leaving the body" how does that fit your model? or does it not?

    - Second is consciousness eternal? What about the theories of big bang and big freeze that say the universe started and will end one day?

    Thank you for your time

    ReplyDelete
  17. "The answer ChatGPT gives you is a ‘transformation’ of your question." This line reminds me a quote from "the Other" in "more than allegory";
    "‘...Potentially yes, but I only truly know what you or another living being asks me.’
    ‘How come?’ I protested. ‘Surely you either know something or you don’t, regardless of being asked about it.’
    ‘Whatever I have never been asked about by a self-reflective cluster of mind-at-large like yourself,’ he continued, ‘I know only in potentiality. Think of it as the light of a match: until you ignite the match, its light exists only in potentiality, in the form of energy chemically stored in its phosphorus head. But when you ignite the match, its light becomes actualized. Only then can it be seen. My knowledge is like the match: it exists complete, but only in potentiality, until you or someone else asks me about it. Your questions then partly ignite the match of my limitless insight, so its light can be seen. Only then does it illuminate existence.’"

    a limitless filter-less unbiased open transformer LLM potentially are more like the "last" phase of matter or a solid-state Mind-at-Large or "black box" logos with its infinite word2vec latent space!
    I'm sure you know what i mean Bernardo.

    ReplyDelete
  18. Omg Bernardo, what a wonderful explanation of LLMs etc. Super useful as an educational tool. I am going to share this around.

    ReplyDelete
  19. Hello Bernardo,
    Can you please write something about dangers ... (or lack of dangers) of AI?

    ReplyDelete