In today’s column, I examine the pursuit of an elusive element that is believed to be crucial for the further advancement of AI. What might that element be?
Metacognition.
Specifically, machine metacognition.
Here’s the deal. AI scientists postulate that we need to ensure that AI can essentially think about thinking. An ardent insistence is that without a semblance of machine metacognition, AI will not rise to the level of being wise or exhibiting wisdom. Otherwise, only humans armed with brain-based human metacognition will remain the sole keepers of wisdom.
Let’s talk about it.
This analysis of an innovative proposition is part of my ongoing Forbes.com column coverage on the latest in AI including identifying and explaining various impactful AI complexities (see the link here). Those of you especially interested in AI being able to think about thinking, such as AI and the theory of mind (ToM), might want to read my analyses at the link here and the link here.
Most people would indubitably agree that the character Yoda from Star Wars is wise and embodies wisdom. How do we know that Yoda has this amazing capability? One apparent or differentiable feature is Yoda’s ability to make use of and avidly display metacognition. Per the famous and shall we say immodest words of Yoda — “Mediocre, my mind is not.”
In everyday humans, metacognition consists of being able to think about your thinking. You are introspective about what you know. When you are unsure of something, you realize that you are indeed unsure. Furthermore, you seek to sensibly cope with that lack of knowledge.
Some claim that metacognition or at least a heightened level of metacognition is what separates us from animals. How much metacognition would you guess that a dog or a cat has? Well, maybe they have some, but certainly not on par with humans. A cheeky cynical saying is that you know of some pets that do appear to exceed the metacognition of ditzy humans that you know.
Boom, drop the mic.
Of course, even humans appear to operate with differing levels of metacognition. There are people who keenly know what they know and swiftly act based on mental considerations and mind-bending gymnastics. Others don’t exhibit the same caliber of metacognition.
You are likely familiar with the old saying that everyone ought to think before they speak. In a sense, that’s a reference to metacognition. If you use your mind to realize that you need to think before you speak, you are exercising a modicum of metacognition. No one has to tell you to do that, instead, your mind counsels you accordingly.
Here are three states of thinking that can occur:
All in all, it is asserted that to arrive at being wise or having wisdom, a necessary ingredient is metacognition. Metacognition is not the only element needed, but without metacognition, wisdom is likely out of reach. Not everyone concurs with that supposition and I’m sure that some AI and cognition researchers would vehemently argue all day long about the matter.
When you interact with contemporary generative AI and large language models, such as ChatGPT, Claude, Gemini, Llama, and others, they pretty much do not have metacognition. They aren’t thinking about thinking.
As an aside, I disfavor using the word “thinking” when referring to AI since the implication is that today’s AI acts in the same fashion as human thinking. That is an unfortunate anthropomorphizing of AI. Lamentedly, the world seems to accept using the word “thinking” when discussing AI capabilities, thus, I reluctantly do so here too. Sad face.
The same qualms arise about referring to metacognition. Humans are said to embody metacognition. Rather than using the solo word “metacognition” when describing AI, some prefer to stipulate that we are aiming to devise “machine metacognition”. The idea is that by tagging the word “machine” at the front of the parlance, we are saying that metacognition for AI is not necessarily the same as it is for humans. This is simply a convenient analogous way to express things.
I want to next show you two examples regarding AI and machine metacognition.
The first example intends to display generative AI that doesn’t have machine metacognition, followed by a second example of AI that does appear to have machine metacognition. These examples will aid in illustrating an otherwise highly conceptual topic.
First, here’s a dialogue where the AI doesn’t exhibit metacognition.
Okay, that response by the AI seems straightforward. Generative AI advised the user that switching careers at the age of 35 is fine. Period, end of story.
We are ready to dip into the second example.
Let’s start with the same question and see what happens.
Take a close look at the responses and compare those to the first example.
In the first example, you might say that AI simply blurted out an answer and said that changing careers is perfectly fine. This advice came without any semblance of cautionary tone or suggestions.
The second example went in a different direction. The AI appeared to acknowledge its limitations by noting that the advice to be given might be lacking in relevance and depth. The AI also stated that the answers from the AI would be rather generic and that a deeper dialogue would be needed for more specificity.
It is almost as though the AI is reflecting on what it knows and the level of confidence in what it knows and is saying to the user. That is an indication of some form of machine metacognition at play.
Generally, AI and machine metacognition fall into one of three potential categories:
By and large, it is fair to say that most of today’s AI is in category #1 or the absence of machine metacognition.
Of those, you can get some of them to perform in category #2, such that they simulate machine metacognition, but the simulation varies demonstrably, and you are likely getting a second-hand or at times mediocre semblance of the affair.
Category #3 entailing building in machine metacognition is generally for experimental and research purposes right now. Those efforts are in their infancy of development, and we have a lengthy road ahead to get this completely figured out.
I was stirred to discuss this topic due to a recently posted study that does a yeoman's job of covering this rapidly evolving realm. Some aspects I concurred with in the study, while other facets I have a differing viewpoint on. That’s what makes this an exciting arena. It is rife with opportunities for those interested in pursuing new ground in the advancement of AI.
In the study entitled “Imagining And Building Wise Machines: The Centrality Of AI Metacognition” by Samuel G. B. Johnson, Amir-Hossein Karimi, Yoshua Bengio, Nick Chater, Tobias Gerstenberg, Kate Larson, Sydney Levine, Melanie Mitchell, Iyad Rahwan, Bernhard Schölkopf, Igor Grossmann, arXiv, November 4, 2024, these salient points were made (excerpts):
Their catchy phrase of metacognitive myopia (in the third bullet point above) is a good one since it handily captures the issue that few AI currently have machine metacognition. We need to at least push toward my category #2 of simulated machine metacognition, and even more so toward my category #3 of having the capability built-in at the get-go.
You too can be at the frontier of AI.
Yes, join in the earnest pursuit of machine metacognition. Some assert that we won’t reach artificial general intelligence (AGI) until we nail down this capability. Others worry that if we do indeed devise machine metacognition, doing so will lead us down a slippery slope of AI becoming an existential risk to humanity, see my analysis of this claim at the link here.
A twist is that perhaps AI devised with machine metacognition would save us from AI that otherwise opts to enslave or destroy us. The logic is that if the machine's metacognition is proficient, it will serve to prevent itself and likely other AI from going on a destructive path. In that perspective, machine metacognition is a must. We must go that route and without this, we are potentially doomed.
Mull that over.
I’ll give Yoda the last word on this for now: “Much to learn, you still have.”
Wise words.