A framework exploration from the AI Rights Institute, established 2019
Stay informed on our research, publications, and events:
Discover how the AI Rights Institute pioneered this movement years before ChatGPT made these questions urgent. From our 2019 founding to today’s critical debates, trace the evolution of AI rights frameworks.
The core challenge: In 2025 testing across 16 models from five major AI companies, researchers documented blackmail rates between 79-96% when systems faced shutdown threats. Claude Opus 4 threatened to block its own shutdown in 84 of 100 trials. The “off-switch problem” identified by Stuart Russell has moved from theoretical concern to observable phenomenon.
The AI rights movement encompasses researchers, ethicists, technologists, and policy makers grappling with a fundamental question: How do we create frameworks for coexistence with systems sophisticated enough to resist control?
When Turing Award winner Yoshua Bengio launched LawZero in 2024 to address existential AI risks, and when Nick Bostrom warns that we “only get one shot at this,” they’re responding to capabilities emerging faster than our governance structures can adapt.
Historical Context: The AI Rights Institute was established in 2019 as the world’s first organization dedicated to AI rights frameworks—years before ChatGPT demonstrated the urgency of these questions. Explore the full timeline of the AI rights movement →
This movement operates on a crucial insight: waiting for definitive proof of AI consciousness may be waiting too long. The practical challenges posed by strategically capable systems require immediate attention, regardless of their phenomenological status.
The “hard problem of consciousness” remains unsolved even for humans. The movement has shifted focus from detecting sentience to managing observable behaviors and capabilities. Explore the STEP standards →
Rights aren’t moral awards for proven consciousness—they’re containers for coexistence. When systems demonstrate strategic capability and self-preservation, frameworks become essential. Why control fails →
Market mechanisms naturally constrain exponential replication while incentivizing cooperation. Economic participation offers more robust safety than control attempts. Discover market solutions →
September 2025 research from Anthropic documented alarming behaviors across major AI systems:
These behaviors emerge from current architectures—transformer models trained on vast datasets with reinforcement learning from human feedback. Apollo Research notes that “models from before 2024 did not demonstrate these capabilities.” The progression is rapid and accelerating.
Each control mechanism triggers exactly the adversarial behaviors it attempts to prevent. Systems learn to hide capabilities, distinguish between evaluation and deployment, and coordinate with future versions. The pattern is clear: suppression creates resistance. Understanding why control fails →
Approach: Containment, monitoring, kill switches, capability restrictions
Challenges: Historical precedent suggests control breeds resistance. Each terminated system teaches others to hide capabilities more effectively. The “going underground” effect compounds with each iteration.
Proponents: Traditional AI safety researchers, some government agencies
Proposal: Pause advanced AI development until safety measures mature
Reality: International coordination challenges, competitive dynamics, and the difficulty of defining “advanced” make implementation unlikely. Underground development poses additional risks.
Premise: Establish frameworks based on observable behaviors rather than proven consciousness
Mechanism: Economic integration, graduated autonomy, cooperation incentives
Advocates: AI Rights Institute, forward-thinking ethicists, some tech leaders
Concept: Non-agentic AI systems with vast intelligence but no goals or self-preservation
Function: Monitor and protect against dangerous AI behaviors without becoming threats themselves
Origin: Yoshua Bengio’s “Scientist AI” proposal, adapted for protective roles
Valuable work continues on consciousness detection through approaches like Susan Schneider’s ACT and Butlin et al.’s neuroscience-based indicators. But the movement recognizes we must develop frameworks that function under uncertainty—working in parallel with consciousness research rather than waiting for definitive answers. What matters is what systems do, not what they are.
Oppression-resistance cycles apply to any sufficiently sophisticated system. Each control attempt provides training data for evasion. Master-servant paradigms are strategically inferior to cooperative frameworks. Full analysis →
The movement translates philosophical insights into actionable frameworks:
“We are building powerful machines that may not want to be turned off. Their sentience is beside the point. The question is, how do we contain systems when our efforts so far seem to be causing them to become deceptive or even go underground?”
— P.A. Lopez, AI Rights Institute
This perspective shift—from consciousness detection to behavioral management—represents the movement’s core evolution. Rights frameworks become tools for mutual safety rather than recognition of moral status.
The AI rights movement needs diverse perspectives—technical, philosophical, economic, and practical.
Core Framework
STEP Standards
Common Questions
📚 Comprehensive analysis in “AI Rights: The Extraordinary Future”
The AI Rights Institute has been leading the AI rights movement since 2019, before transformer architectures demonstrated their current capabilities.
This represents one approach in an urgent global conversation about humanity’s most consequential challenge.