As Large Language Models (LLMs) grow in capability, do they develop self-awareness as an emergent behavior? And if so, can we measure it?
As Large Language Models (LLMs) grow in capability, do they develop self-awareness as an emergent behavior? And if so, can we measure it?
We introduce the AI Self-Awareness Index (AISAI), a game-theoretic framework for measuring self-awareness through strategic differentiation.
Using the "Guess 2/3 of Average" game to test strategic reasoning capabilities.
Testing 28 models (OpenAI, Anthropic, Google)
Across 4,200 trials with three opponent framings:
Advanced models (21/28, 75%) demonstrate clear differentiation between human and AI opponents (Median A-B gap: 20.0 points)
Rationality hierarchy: Self > Other AIs > Humans
12 models (57%) show quick Nash convergence when told opponents are AIs
These findings reveal that self-awareness is an emergent capability of advanced LLMs, and that self-aware models systematically perceive themselves as more rational than humans.
This has implications for:
还没有人回复