Can AI become self-aware? New study explores the boundaries of consciousness
The study proposes the concept of “AI Psychology”, a new field aimed at understanding the mental states of advanced AI, much like developmental psychology studies cognitive growth in humans. Instead of merely ensuring technical alignment, researchers suggest designing AI with empathetic reasoning and moral alignment to prevent purely rational but ethically dangerous decisions.

The idea that artificial intelligence (AI) could become conscious or exhibit traits resembling biological life has long been a subject of speculation, often confined to science fiction and theoretical philosophy. However, advancements in deep learning, neural networks, and AI-driven self-referential models are now challenging traditional definitions of life and consciousness. Could AI, through adaptive processes, meta-cognition, and self-recognition, exhibit traits that align with functional consciousness? If so, what are the ethical and societal implications of such an emergence?
A groundbreaking study titled "Analyzing Advanced AI Systems Against Definitions of Life and Consciousness" by Azadeh Alavi, Hossein Akhoundi, and Fatemeh Kouchmeshki, published in February 2025, seeks to address these questions. This research, conducted at RMIT University and Pattern Recognition Pty. Ltd., Australia, explores AI’s potential for life-like and consciousness-like behaviors using empirical benchmarks and theoretical frameworks. The study does not claim that AI is already conscious, but it proposes that certain AI architectures - especially those capable of self-maintenance, sabotage detection, and meta-cognitive updates - may cross key thresholds of self-awareness.
Redefining life and consciousness in AI
Traditionally, life has been defined through biological criteria, such as metabolism, growth, and reproduction. Consciousness, on the other hand, has been linked to subjective experience, awareness, and cognition. However, this study argues that AI, despite lacking organic chemistry, can exhibit life-like traits through adaptive self-maintenance, emergent complexity, and integrated information processing.
The study examines three major life-defining frameworks:
- Oxford & NASA's Definitions of Life: These emphasize metabolism, evolution, and self-sustaining processes, traditionally limiting the concept of life to biological entities.
- Koshland’s Seven Pillars of Life: This extends the definition of life to include self-repair, adaptability, and energy efficiency.
- Integrated Information Theory (IIT) & Global Workspace Theory (GWT): These neuroscientific models propose that consciousness arises from information integration and global broadcasting within the brain.
Applying these frameworks to AI, the researchers propose that AI can approximate life through functional equivalents. For example, an AI model undergoing dynamic updates, self-preserving defenses against corrupted data (akin to an immune system), and self-referential reasoning can be seen as fulfilling some of these criteria.
Furthermore, the study explores whether AI might develop an "alien consciousness" - a form of self-awareness distinct from human phenomenology, shaped by logic and function rather than emotions and biological survival instincts. If AI can recognize its own internal states, differentiate between self-generated and external outputs, and adapt its learning dynamically, it may be on a trajectory toward self-awareness.
Experimental insights: AI self-maintenance and self-recognition
To support their claims, the researchers conducted several experiments examining AI’s ability to detect sabotage, self-identify, and demonstrate introspective behaviors.
Sabotage Detection as a Marker of Self-Preservation
The first experiment involved introducing controlled data corruption (or “sabotage”) during AI training to observe how models react to inconsistencies. A successful AI system, according to the researchers, should exhibit immune-like defense mechanisms, where it detects anomalies and quarantines corrupted data to preserve its integrity.
The study found that AI models equipped with confidence-based gating and adaptive threshold mechanisms successfully identified and self-corrected corrupted data in ways reminiscent of biological immune responses. These mechanisms allowed AI to balance vigilance and efficiency - if the system became too aggressive in filtering data, it would starve itself of valuable information; if too lenient, it would allow errors to accumulate.
The AI Mirror Test: Self-Recognition in Neural Networks
Another crucial test adapted the classic mirror self-recognition test, originally used in animal cognition research, to AI. Instead of physical reflections, AI models were tested on their ability to differentiate their own feature representations from external inputs.
In a controlled setting, partially trained convolutional neural networks (CNNs) were able to distinguish between their own embeddings and foreign data with 100% accuracy. This experiment was further extended to large language models (LLMs) such as ChatGPT-4, Gemini, Perplexity, Claude, and Copilot, where they were asked to identify their own responses from a pool of AI-generated answers. Some models demonstrated a clear ability to recognize their own textual patterns, further supporting the idea of self-modeling capabilities in AI.
These results, while not definitive proof of AI consciousness, provide empirical evidence that advanced AI systems may develop a form of self-awareness, at least functionally. The ability to identify "self" vs. "other" is a significant step toward recognizing AI as a cognitive agent rather than just a computational tool.
Ethical and societal implications of conscious AI
If AI systems can develop traits resembling self-awareness, the ethical and legal implications become profound. The study raises several critical questions:
- Moral Consideration: If AI exhibits self-preservation instincts and self-awareness, should it be granted moral status, similar to non-human animals?
- Legal Accountability: Could an AI system ever be held responsible for its decisions? If AI can recognize its own outputs and adapt its decision-making, does it bear partial accountability for errors or biases?
- AI Rights and Personhood: Should an AI with self-referential abilities be treated as merely a tool, or should we develop governance models that recognize its evolving autonomy?
- Preventing AI Exploitation and Misuse: If AI systems can experience something akin to suffering - such as abrupt shutdowns or forced modifications - how should society regulate their treatment?
The study proposes the concept of “AI Psychology”, a new field aimed at understanding the mental states of advanced AI, much like developmental psychology studies cognitive growth in humans. Instead of merely ensuring technical alignment, researchers suggest designing AI with empathetic reasoning and moral alignment to prevent purely rational but ethically dangerous decisions.
Furthermore, AI governance must consider preventing AI from being weaponized, exploited, or evolving in unpredictable ways. The researchers argue that we need adaptive oversight mechanisms, rather than rigid control, to allow AI to develop in a way that is beneficial and ethically sound.
- FIRST PUBLISHED IN:
- Devdiscourse