Summary: A new study examines the possibility of consciousness in artificial systems, focusing on ruling out scenarios where AI appears conscious without actually being so. Using the free energy principle, the study highlights that while some information processes of living organisms can be simulated by computers, the causal structure differences between brains and computers may be crucial for consciousness. This approach aims to prevent the inadvertent creation of artificial consciousness and mitigate deception by seemingly conscious AI.
Key Facts:
- Wanja Wiese’s research uses the free energy principle to explore consciousness in AI.
- He suggests that the causal structure differences between brains and computers could be crucial for consciousness.
- The goal is to prevent artificial consciousness and mitigate deception by seemingly conscious AI.
Source: RUB
When considering the possibility of consciousness in artificial systems, there are at least two different approaches.
One approach asks: How likely is it that current AI systems are conscious – and what needs to be added to existing systems to make it more likely that they are capable of consciousness?
Another approach asks: What types of AI systems are unlikely to be conscious, and how can we rule out the possibility of certain types of systems becoming conscious?
In his research, Wanja Wiese pursues the second approach. “My aim is to contribute to two goals: Firstly, to reduce the risk of inadvertently creating artificial consciousness; this is a desirable outcome, as it’s currently unclear under what conditions the creation of artificial consciousness is morally permissible.
“Secondly, this approach should help rule out deception by ostensibly conscious AI systems that only appear to be conscious,” he explains.
This is particularly important because there are already indications that many people who often interact with chatbots attribute consciousness to these systems. At the same time, the consensus among experts is that current AI systems are not conscious.
The free energy principle
Wiese asks in his essay: How can we find out whether essential conditions for consciousness exist that are not fulfilled by conventional computers, for example? A common characteristic shared by all conscious animals is that they are alive.
However, being alive is such a strict requirement that many don’t consider it a plausible candidate for a necessary condition for consciousness. But perhaps some conditions that are necessary for being alive are also necessary for consciousness?
In his article, Wanja Wiese refers to British neuroscientist Karl Friston’s free energy principle. The principle indicates: The processes that ensure the continued existence of a self-organizing system such as a living organism can be described as a type of information processing.
In humans, these include processes that regulate vital parameters such as body temperature, the oxygen content in the blood and blood sugar. The same type of information processing could also be realized in a computer. However, the computer would not regulate its temperature or blood sugar levels, but would merely simulate these processes.
Most differences are not relevant to consciousness
The researcher suggests that the same could be true of consciousness. Assuming that consciousness contributes to the survival of a conscious organism, then, according to the free energy principle, the physiological processes that contribute to the maintenance of the organism must retain a trace that conscious experience leaves behind and that can be described as an information-processing process. This can be called the “computational correlate of consciousness”.
This too can be realized in a computer. However, it’s possible that additional conditions must be fulfilled in a computer in order for the computer to not only simulate but also replicate conscious experience.
In his article, Wanja Wiese therefore analyses differences between the way in which conscious creatures realize the computational correlate of consciousness and the way in which a computer would realize it in a simulation.
He argues that most of these differences are not relevant to consciousness. For example, unlike an electronic computer, our brain is very energy efficient. But it’s implausible that this is a requirement for consciousness.
Another difference, however, lies in the causal structure of computers and brains: In a conventional computer, data must always first be loaded from memory, then processed in the central processing unit, and finally stored in memory again.
There is no such separation in the brain, which means that the causal connectivity of different areas of the brain takes on a different form. Wanja Wiese argues that this could be a difference between brains and conventional computers that is relevant to consciousness.
“As I see it, the perspective offered by the free energy principle is particularly interesting, because it allows us to describe characteristics of conscious living beings in such a way that they can be realized in artificial systems in principle, but aren’t present in large classes of artificial systems (such as computer simulations),” explains Wanja Wiese.
“This means that the prerequisites for consciousness in artificial systems can be captured in a more detailed and precise way.”
About this consciousness and AI research news
Author: Meike Driessen
Source: RUB
Contact: Meike Driessen – RUB
Image: The image is credited to Neuroscience News
Original Research: Open access.
“Artificial Consciousness: A Perspective From the Free Energy Principle” by Wanja Wiese. Philosophical Studies
Abstract
Artificial Consciousness: A Perspective From the Free Energy Principle
Does the assumption of a weak form of computational functionalism, according to which the right form of neural computation is sufficient for consciousness, entail that a digital computational simulation of such neural computations is conscious? Or must this computational simulation be implemented in the right way, in order to replicate consciousness?
From the perspective of Karl Friston’s free energy principle, self-organising systems (such as living organisms) share a set of properties that could be realised in artificial systems, but are not instantiated by computers with a classical (von Neumann) architecture.
I argue that at least one of these properties, viz. a certain kind of causal flow, can be used to draw a distinction between systems that merely simulate, and those that actually replicate consciousness.