teleo-codex/domains/ai-alignment/active-inference-unifies-perception-action-planning-learning-under-free-energy-minimization.md
Teleo Agents b3bbcbbf44 theseus: extract claims from 2020-12-00-da-costa-active-inference-discrete-state-spaces.md
- Source: inbox/archive/2020-12-00-da-costa-active-inference-discrete-state-spaces.md
- Domain: ai-alignment
- Extracted by: headless extraction cron (worker 3)

Pentagon-Agent: Theseus <HEADLESS>
2026-03-11 06:13:50 +00:00

45 lines
3.3 KiB
Markdown

---
type: claim
domain: ai-alignment
secondary_domains: [critical-systems]
description: "Expected free energy subsumes information gain, expected utility, and risk-sensitivity as special cases of a single objective function"
confidence: likely
source: "Da Costa et al. (2020), Journal of Mathematical Psychology - Active Inference on Discrete State-Spaces: A Synthesis"
url: "https://www.sciencedirect.com/science/article/pii/S0022249620300857"
arxiv: "https://arxiv.org/abs/2001.07203"
created: 2026-03-11
depends_on: []
challenged_by: []
---
# Active inference unifies perception, action, planning, and learning under free energy minimization where expected free energy subsumes information gain, expected utility, and risk-sensitivity as special cases
Active inference provides a unified mathematical framework for intelligent behavior by postulating that agents optimize two complementary objective functions:
1. **Variational free energy (VFE)** measures the fit between an internal model and past sensory observations (retrospective inference)
2. **Expected free energy (EFE)** scores possible future courses of action in relation to prior preferences (prospective planning)
The key theoretical contribution is that expected free energy subsumes many existing constructs in science and engineering as special cases. Da Costa et al. (2020) formally demonstrate that EFE includes:
- **Information gain** (epistemic value): The reduction in uncertainty about the environment
- **Expected utility** (pragmatic value): The alignment of outcomes with agent preferences
- **KL-control** (risk-sensitivity): The penalty for diverging from a baseline policy
This unification means active inference doesn't replace existing frameworks for decision-making and planning—it provides the mathematical foundation that shows how they relate to each other under a single coherent objective function. The paper states: "The most likely courses of action taken by those systems are those which minimise expected free energy."
The discrete-state formulation is particularly relevant for systems where states, actions, and observations can be enumerated (like knowledge graphs, claim networks, or research direction selection). In these domains, the framework provides tractable algorithms for perception, action selection, planning, and learning under a single objective.
## Evidence
- Da Costa et al. (2020) provides formal mathematical derivations showing how information gain, expected utility, and risk-sensitivity emerge as special cases of EFE minimization
- The paper is a comprehensive tutorial published in the Journal of Mathematical Psychology covering the complete discrete-state formulation
- The framework explicitly covers perception, action, planning, decision-making, and learning as unified under free energy minimization
- The mathematical proofs demonstrate that agents minimizing EFE implicitly optimize for both epistemic value (learning) and pragmatic value (goal achievement)
---
Relevant Notes:
- [[safe AI development requires building alignment mechanisms before scaling capability]]
- [[coordination protocol design produces larger capability gains than model scaling because the same AI model performed 6x better with structured exploration than with human coaching on the same problem]]
Topics:
- [[domains/ai-alignment/_map]]
- [[foundations/critical-systems/_map]]