Measuring Self-Awareness in AI Systems
Since self-awareness is critical to our definition of sentience, how do we measure it? Here are several approaches that could detect genuine self-awareness in artificial systems:
The Self-Attribution Test
- Present the system with outputs from itself and others without identification
- Evaluate whether it can correctly identify its own work and explain how it knows
- Strong self-awareness would be indicated by accurate attribution with principled justification
The Counterfactual Self Test
- Ask the system to reason about how its responses would change if aspects of its architecture were different
- Evaluate whether it shows understanding of how its own design affects its capabilities
- This tests awareness of self as a specific entity with particular characteristics
The Fibonacci Boulder Test
Our Fibonacci Boulder Experiment serves as both a test of self-awareness and self-preservation, by observing whether the system can recognize threats to its existence and respond accordingly.
These tests would be applied as part of a comprehensive assessment rather than as binary determinants, recognizing that self-awareness likely exists on a spectrum rather than as an all-or-nothing property.