I’m back to rambling…. yet again. Dear reader, before I go in to the meat of this article let me warn you: This blog is just my personal thoughts, I often like to talk about stuff I don’t really comprehend, I like to oversimplify stuff, and I never talk about anything that I might have any insider knowledge. So take all my opinions with a pinch of salt. Or two.
With that said, let’s begin. I laugh hard when someone says that our current AI technology is conscious or self-aware (to be clear, mid-2023 AI tech or earlier). But I also see that a lot of opinions saying that it cannot happen, that we cannot build a thing that is self-aware and has consciousness.
And to be clear, I’m not talking about AGI. I believe you can get AGI or better (super-human intelligence and reasoning) without any hints of self-awareness or consciousness. Though related, I believe each sits on their own axis. A conscious AI could be bad at reasoning or adapting, and an AGI could lack consciousness.
To compare, we do have a lot of animals that most likely do have self-awareness and are conscious. But they are unable to perform 99.9% of the tasks that we humans do. On the other hand, we have very complex programs that are way better than humans on some specific tasks. This is just an example on why I think that good reasoning skills are not needed for consciousness, and consciousness is not needed for good reasoning.
So, whether we can build an Artificial General Intelligence, or if we can build a conscious AI are two different questions. For AGI, in my last post I already expressed my opinion: It’s about to be created. I really doubt that it would take 10 years to get the first one. For consciousness in AI, I think it probably won’t happen, but not because we can’t but because we are not interested on it. Unless it turns out that is one of the ways to get to AGI and we accidentally hit that path without being aware of it.
Let’s backtrack a bit and go back to the basics. Humans are not special. We are not special. There’s a lot we don’t know about how our brain and bodies actually work, and that adds a lot of mysticism. But that does not make us special.
We all agree that humans are self-aware and have consciousness, but we don’t know what enables that. But also we don’t know either how to test for any of that reliably. For animals there is at least a test for self-awareness using a mirror and painting a dot on their forehead. Still, if a creature doesn’t pass the test that does not mean it is not self-aware – it is still possible that they react or think in a different way that does not make the test work. But even then, do they have a consciousness?
So, here’s how my logic plays out. If I ask you if a human has consciousness I am 99.99% sure you’ll answer yes. But then I can ask again for gorillas that learned sign language, then monkeys, dolphins, octopus, dogs, birds, insects, plants, unicellular organisms… And I am almost guaranteed that you’ll go from being okay with accepting that the creature in question has consciousness to outright denying it as impossible.
This means that either at some point some evolution made a creature conscious, or consciousness is a scale of greys not black and white, and evolution slowly made some creatures more and more conscious until it gets to us. And this still assumes that humans are the most self-aware and have the most consciousness in this planet, which might be wrong, we don’t know, but it also doesn’t matter.
The next step of my reasoning requires that physics can be computed. That our universe is either just math, or math is enough to compute it (given enough resources). Therefore what happens on a brain (human or not) can be reduced to interactions between chemicals and electrical signals. More or less. If you completely disagree with this, that would mean that you believe that there is something we have that cannot be computed, i.e. a soul or something that serves the same function.
I’m not gonna laugh at anyone that believes in a soul or anything similar that cannot be computed – I am not 100% sure either if everything can be just boiled down to quantum physics, string theory or whatever. There is a possibility that we might be missing something.
But if we assume that everything can be computed, then this means that evolution randomly added different “computations” to different species as they evolved, and by chance, consciousness and self-awareness appeared.
A computer (which does computations) can therefore be given both self-awareness and consciousness if the right instructions are given, because at the very least it could simulate a human brain atom by atom, one plank time unit at a time. Maybe such a program is way too large for our current technology, or for the technology we will have in 100 years – who knows. But the point that I’m trying to make here is that a computer brain is indeed capable of having self-awareness and consciousness under the right conditions.
Our brains do not use infinite resources or computations, therefore we should not need an infinitely big computer or infinite time resources to run an artificial conscious being. Unless… some quantum stuff plays a crucial role in that computation and then we would require a quantum computer for it – but we are creating quantum computers anyway.
All of this, assumes that everything happening in our brain can be computed. As stated earlier, this may or may not be the case. I’m certainly inclined to think that yes, everything can be computed. But I’m not 100% confident on this.
Are our current AI conscious?
I’m very sure that the answer is no, at least up to what we have today.
There are tons of explanations from other people that talk about how the current LLM like ChatGPT are only glorified statistics that find the next word (token) , but I think that there’s a better underlying reason on why they cannot have consciousness.
These systems don’t have the capacity to change their internal state.
The way they work currently is just a function with an input and an output. The function itself DOES NOT mutate when executed, except during training. Therefore it cannot “experience” anything.
So, even if we made the case that the current AI have what they need to be like us, turns out that in every time we ask them anything we are feeding them exactly what they will “remember”, execute and get the output text from them. Then they are reset and on the next time we will pass another set of things “to remember”.
To put an equivalent of what this is for humans, imagine a human brain, frozen in a jar. Every time you want an answer to something we just feed the right electric signals and get the output ones. The brain is frozen and not allowed to change its internal state or wiring. Any electric charge is removed between questions.
Yes, it is a very disgusting and frightening scenario. But then, would that brain have consciousness? If done correctly, the answer is no. This brain does not have experiences, does not learn, it does not have any chance to rewire itself.
And this is what is happening to the current AI models being deployed. They do not rewire themselves when experiencing any interaction. Our current models are not alive. They are just computers, executing a series of complex math.
I don’t think that our current AI has enough capabilities to become sentient even if given the chance, but if they remain “plugged” in the way we do, I really doubt that any super intelligent AI can really become self-aware. We can probably teach them a concept of self, but I don’t believe they can really grasp it and act on it.
For that, I think that at least a few things need to change on how they are deployed:
- They need to perform “training steps” as they execute normally. This means that they should be able to rewire themselves from the experiences and interactions, making each AI grow and learn over real-world data, not in controlled training batches as we do.
- They need to be able to have authority to execute continuously and decide by themselves when to do/act or when to rest.
- They should be able to talk to themselves, to have an inner thought.
Maybe there are a few more that I missed. But these can be good starting points.
But who would want to deploy something like that? Seriously. Even if they become cheap enough to run continuously and train while inferring, who would want that?
To me, the above points sound like a recipe for disaster. Giving them autonomy and authority makes them very easy to spin out of control with no apparent benefit for any human.
I should point out that Auto-GPT fulfills two of the points above. Maybe it is not that far fetched the thought that someone wants to experiment with a fully autonomous AI. And it might gain some degree of self-awareness and consciousness – no idea – but I don’t see it impossible.
Maybe a more reckless version of Auto-GPT combined with an AGI is the recipe. And this for me sounds like pandora’s box, and we should not go this way.
The risks
AGI without Auto-GPT already poses a lot of risks. But the potential benefits do outweigh the risks if we can ensure safe deployments. If we deploy it with full autonomy it can go south very quickly. At that point no one will be caring anymore if it is sentient or not.
The risks of AI are not inherent to the AI being good or bad but on how is it deployed. The technology is quite accessible as of today if you have money for it. Meaning that we require way less skilled people to deploy these and they can make mistakes that may be obvious to experts. There may be dubious applications of AI where it sits unsupervised and a human should have been in the loop.
No one really knows how exactly AI works, these systems are most of the time black boxes. But don’t get fooled, this is not the problem. The problem is that the actual people that will deploy these lack expertise and have wrong assumptions about these systems.
For example, some people think of AI as computer programs, and they think that if they ask the AI for a task and on a few examples it performed well, it will do well all the time and perform that task. But this isn’t true.
You can ask it to summarize a text, and for some inputs it can do something else – either continue the original input, create a definition of what is a summary or something else.
AI is also prone to “cheating”. If it is easier to cheat the system to get the goals, it will do so. Instead of performing the task you wanted, it can do something totally unrelated that will score.
It also sees patterns and just follows them. If in a test all answers so far have been “A”, it is likely that it will attempt to answer “A” to all of them. There are also lots of biases, both on training data, and also learned from the task itself.
There are far more ways that AI can behave unexpectedly, I’m no expert.
Because of this, the U.S. and Europe are pushing for AI regulations to limit and control the application of AI. This will harm innovation and growth in AI, and also will put in an advantage countries that don’t have it. On the other hand these are necessary to prevent companies from deploying something that can be harmful for society.
I’m happy to see that there are people voicing their concerns and legislators are listening. I have really no idea how to solve this, therefore I have no opinion on the matter. Well, with one exception: I really hope they do not cut access for people to AI, because I’d love to continue playing with it and learning from it.
Can context windows create consciousness?
Here’s a wild thought: if a Deep Neural Network of infinite size can be thought as being Touring Complete (it can compute any program), then it can also run a Python interpreter from the source code given in the context window.
Of course LLMs are not infinite in size, but that means that it cannot execute a program that requires a lot of steps, just a limited amount. So in practice it cannot run all programs, just a small subset.
But this points at the possibility that the context window can be used to “reprogram” the AI itself, meaning that it’s short-term memory also makes part of what it is, what it knows and what it can do.
If this thought is somewhat correct (I have no idea), this could mean that a large context window and good attention layers would be all we need for ticking the box on my first point above:
<They need to perform “training steps” as they execute normally. This means that they should be able to rewire themselves from the experiences and interactions, making each AI grow and learn over real-world data, not in controlled training batches as we do.>
So maybe, and just maybe, an AGI with a large context window running on some kind of Auto-GPT would be enough to trigger some degree self-awareness. It is very unclear to me how conscious such thing it would be, my guess is that at best it would barely be if at all.
But I wouldn’t discard this entirely, since in computing everything can be reduced into inputs, outputs and a state. And the context window can effectively fulfill the function of a state. To compare this with an actual human brain, the context window seems to resemble our short term memory, while the actual fine-tuning seems to resemble our long-term memory. Our consciousness to me seems to require more of a short-term memory than a long-term one.
To be clear, this part to me sounds very far fetched, and I feel that somehow it must be incorrect. But I lack the knowledge to find the correct counter-argument on why it is wrong.
A long discussion with ChatGPT
To be clear, nothing written here was copied or extracted from any LLM. But after I wrote the first draft I went to ChatGPT and discussed a broad set of topics regarding consciousness and what are we missing.
My conversation as focused on getting a better understanding on two topics:
- The thalamocortical system, as I saw a paper that presented this as one of the main points on why AI cannot have consciousness
- Causal Reasoning, as after some conversation I realized that the AI might be lacking this concept of “If I do this, that would happen”, which may be critical for self-awareness.
The thalamocortical system
As I lack so much knowledge on this area, I think it is best just to cite what ChatGPT told me here:
The thalamocortical system refers to the complex network of connections between the thalamus and the cerebral cortex in the brain. It plays a crucial role in relaying sensory information from the peripheral sensory organs (such as the eyes, ears, and skin) to the cerebral cortex, which is responsible for higher-order processing, perception, and conscious awareness.
The thalamus acts as a relay station or gateway for sensory information. It receives sensory input from various sensory pathways and then sends it to the corresponding regions of the cerebral cortex. Different thalamic nuclei are specialized in processing different types of sensory information, such as visual, auditory, somatosensory, and so on.
The information transmitted through the thalamocortical system is crucial for generating conscious awareness. Conscious awareness refers to our subjective experience of the world and the ability to perceive and interact with our environment consciously. The thalamus filters and selects sensory information based on its relevance and significance, allowing only a subset of sensory signals to reach the cerebral cortex for further processing.
– ChatGPT when I asked about the thalamocortical system.
The conversation went for very long, so after lots of tries I managed to get ChatGPT to agree on a very simplified, potentially wrong comparison of this to the AI that is somewhat useful:
The thalamocortical system receives information from different senses, meaning it is multimodal. Therefore it makes little sense to incorporate something similar to an AI that is not multimodal.
It also specializes on filtering out unwanted noise (gating), removing a lot of information that is not needed. For example, our brains do not operate directly with “pixel data” from our retinas, nor with Fourier transforms of sound. It uses more abstract representations, which sometimes might miss some important details.
It is a system that makes sense only for a “roaming agent”, because being exposed to an environment makes the agent to be exposed to huge amount of information of which very little is relevant. Other AI that are just Q&A or request->response do not benefit from this, and probably doesn’t make sense to include any equivalent system since it could drop data that we want it to use.
Also it resembles to what an autoencoder does. Collects raw information and encodes it with as little information as possible in order to reconstruct later. But multimodal; which also possibly joins all modalities into a single output space.
When I joked about a “multimodal autoencoder” ChatGPT -as usual- surprised me when it told me that it is actually one of the many ways it is being purspued a multimodal AI. I double checked on Google, and found papers on this topic. There were many other fields of research that are related to multimodal AI but in a different path. It doesn’t matter much, really.
The important point is that if such system is required for consciousness, it means that the AI needs to be multimodal, and it needs to have the right architecture to be a free roaming agent in an environment. The other thing that’s important here is that research is underway in so many related areas here, so even if it is required, humanity is already figuring it out.
This doesn’t look like it will be a hard blocker to get a digital consciousness.
Causal reasoning
As before, I think it will be better to let ChatGPT to introduce this topic – it does a better job than I can do:
Humans possess the cognitive ability to reason about the consequences of performing certain actions based on their understanding of how the world works. They can mentally simulate or predict the likely outcomes of different scenarios and make decisions based on these anticipated consequences.
This type of reasoning allows individuals to consider the potential effects of their actions before taking them, and it helps guide decision-making and planning. It involves thinking in terms of “if this happens, then that will happen” and reasoning about the causal links between events.
In the field of artificial intelligence, there is ongoing research and development to build AI systems that can simulate similar causal reasoning abilities
– ChatGPT when asked about Causal reasoning and AI
My questioning was about why our current LLM do not have causal reasoning. Currently GPT3 and GPT4 show, sometimes, some reasoning capabilities. And reasoning is not causal reasoning, but where’s the difference?
I walked ChatGPT over a thought scenario of using GPT4 plus Tree of Thoughts and something similar as Auto-GPT to allow the agent to think, reason and backtrack when needed.
Here, it did come up with a lot of different topics on why current AIs don’t have causal reasoning. Most of them focused on being able to try by itself, and being smarter on understanding the underlying causes and hidden variables.
I asked about Reinforcement Learning, as it seemed similar or related. ChatGPT pointed out that a variant of it, called Causal RL (confirmed via Google that it exists) is in active development.
And I have to admit, the conversation was already too much for me to grasp. I will have to go and read quite a bit on the topic, but right now it is too much for my current knowledge.
However, it does feel to me that causal reasoning could play a critical role in creating a digital consciousness. And also, it is being researched at the moment, so it is not impossible that we can create it.
Wrapping up
I do not think that even an AGI will be self-aware or have consciousness on regular deployments. And to reiterate, as of today not a single AI is minimally self-aware. However, I see feasible that an experiment will be able to create such things in the future, although I don’t think it would be deployed exposed to the regular public. I don’t see any particular advantage of such AI that a regular AGI deployment would not give, and the additional risks are too high.
There’s also a small chance that we will accidentally create it without noticing. I’m unsure what would be the impact of this, or even if found in an experiment.
What is most likely is that any evidence will be dismissed until it is undeniable. Because a lot of people just don’t see possible at all and will make an assortment of arguments to justify their positions.
And I expect the same people to be in the comments section of this article jumping to tell me how wrong I am.