theseus: extract from 2025-08-00-oswald-arrowian-impossibility-machine-intelligence.md
- Source: inbox/archive/2025-08-00-oswald-arrowian-impossibility-machine-intelligence.md - Domain: ai-alignment - Extracted by: headless extraction cron (worker 4) Pentagon-Agent: Theseus <HEADLESS>
This commit is contained in:
parent
ba4ac4a73e
commit
94551bd6e1
5 changed files with 144 additions and 1 deletions
|
|
@ -21,6 +21,12 @@ Dario Amodei describes AI as "so powerful, such a glittering prize, that it is v
|
|||
|
||||
Since [[the internet enabled global communication but not global cognition]], the coordination infrastructure needed doesn't exist yet. This is why [[collective superintelligence is the alternative to monolithic AI controlled by a few]] -- it solves alignment through architecture rather than attempting governance from outside the system.
|
||||
|
||||
|
||||
### Additional Evidence (extend)
|
||||
*Source: [[2025-08-00-oswald-arrowian-impossibility-machine-intelligence]] | Added: 2026-03-12 | Extractor: anthropic/claude-sonnet-4.5*
|
||||
|
||||
Oswald, Ferguson & Bringsjord (2025) extend Arrow's Impossibility Theorem from preference aggregation to intelligence measurement itself. They prove that no agent-environment-based machine intelligence measure (MIM) can simultaneously satisfy analogs of Arrow's fairness conditions (Pareto Efficiency, Independence of Irrelevant Alternatives, Non-Oligarchy). This affects Legg-Hutter Intelligence, Chollet's ARC benchmark, and 'a large class of MIMs.' The impossibility is structural and formal—published at AGI 2025 conference. This means the alignment problem is doubly underspecified: we cannot aggregate preferences fairly (Arrow's original result) AND we cannot measure intelligence fairly (this extension). If the measurement target itself violates fairness conditions, alignment becomes coordination over which unfair tradeoffs to accept, not optimization toward a fair universal metric.
|
||||
|
||||
---
|
||||
|
||||
Relevant Notes:
|
||||
|
|
|
|||
|
|
@ -0,0 +1,62 @@
|
|||
---
|
||||
type: claim
|
||||
domain: ai-alignment
|
||||
secondary_domains: [critical-systems]
|
||||
description: "Formal proof extends Arrow's theorem from preference aggregation to intelligence measurement, showing no agent-environment MIM can satisfy fairness conditions simultaneously"
|
||||
confidence: experimental
|
||||
source: "Oswald, Ferguson & Bringsjord (2025), 'On the Arrowian Impossibility of Machine Intelligence Measures', AGI 2025 Conference, Springer LNCS vol. 16058"
|
||||
created: 2026-03-11
|
||||
tags: [arrows-theorem, machine-intelligence, impossibility, formal-proof, measurement]
|
||||
---
|
||||
|
||||
# Arrow's Impossibility Theorem applies to machine intelligence measurement, making fair universal intelligence metrics mathematically impossible
|
||||
|
||||
Oswald, Ferguson, and Bringsjord (2025) prove that Arrow's Impossibility Theorem—originally about aggregating diverse preferences into a single social choice—applies equally to measuring machine intelligence in agent-environment frameworks.
|
||||
|
||||
## The Core Result
|
||||
|
||||
No machine intelligence measure (MIM) can simultaneously satisfy analogs of Arrow's three fairness conditions:
|
||||
- **Pareto Efficiency**: If all environments prefer agent A over agent B, the measure must rank A above B
|
||||
- **Independence of Irrelevant Alternatives**: The relative ranking of two agents should not change when a third agent is added or removed
|
||||
- **Non-Oligarchy**: No subset of environments should dictate the overall ranking regardless of other environments' preferences
|
||||
|
||||
This impossibility is **structural and formal**, not a limitation of current measurement approaches.
|
||||
|
||||
## Affected Intelligence Measures
|
||||
|
||||
The theorem directly impacts:
|
||||
- **Legg-Hutter Intelligence** (the dominant formal definition in the literature)
|
||||
- **Chollet's Intelligence Measure** (basis for the ARC benchmark, widely used in AI evaluation)
|
||||
- "A large class of MIMs" in agent-environment frameworks (the result generalizes beyond specific measures)
|
||||
|
||||
## Why This Matters for Alignment
|
||||
|
||||
The impossibility operates at a deeper level than preference aggregation. If we cannot measure intelligence fairly across diverse environments and tasks, the alignment target becomes fundamentally underspecified—you cannot align an AI system to a benchmark if the benchmark itself violates fairness conditions.
|
||||
|
||||
This creates a compounding problem:
|
||||
1. We cannot aggregate diverse human preferences into a single objective (Arrow's original result)
|
||||
2. We cannot measure intelligence fairly across contexts (this extension)
|
||||
3. Therefore: alignment targets are doubly underspecified—both the goal and the metric are impossible to specify fairly
|
||||
|
||||
## Publication Context
|
||||
|
||||
Published at AGI 2025 (Conference on Artificial General Intelligence), the venue most focused on general intelligence measurement. Bringsjord is a well-known AI formalist at RPI with extensive work on formal verification and computational models of intelligence.
|
||||
|
||||
## Limitations of Current Evidence
|
||||
|
||||
The full paper is paywalled; the proof technique and whether constructive workarounds exist (analogous to those in alignment impossibility) are not yet accessible. The claim is rated `experimental` because:
|
||||
- Single source (though from a credible venue and author)
|
||||
- Formal proof not yet independently verified in the knowledge base
|
||||
- Practical implications for specific benchmarks remain to be worked out
|
||||
|
||||
---
|
||||
|
||||
## Related Claims
|
||||
|
||||
- [[AI alignment is a coordination problem not a technical problem]] — extends to measurement layer
|
||||
- [[designing coordination rules is categorically different from designing coordination outcomes as nine intellectual traditions independently confirm]] — intelligence measurement is now a fourth tradition
|
||||
- [[safe AI development requires building alignment mechanisms before scaling capability]] — challenged by measurement impossibility
|
||||
|
||||
Topics:
|
||||
- [[domains/ai-alignment/_map]]
|
||||
- [[foundations/critical-systems/_map]]
|
||||
|
|
@ -0,0 +1,56 @@
|
|||
---
|
||||
type: claim
|
||||
domain: ai-alignment
|
||||
secondary_domains: [critical-systems, mechanisms]
|
||||
description: "Social choice, complexity theory, multi-objective optimization, and intelligence measurement independently confirm impossibility results, suggesting deep structural limits"
|
||||
confidence: likely
|
||||
source: "Pattern synthesis from Arrow (1950), computational complexity literature, Pareto optimality theory, and Oswald et al. (2025)"
|
||||
created: 2026-03-11
|
||||
tags: [impossibility, convergence, coordination, structural-limits]
|
||||
---
|
||||
|
||||
# Convergent impossibility across four independent intellectual traditions strengthens the case for structural coordination limits rather than solvable technical problems
|
||||
|
||||
Four separate intellectual traditions have independently discovered impossibility results in coordination and measurement:
|
||||
|
||||
1. **Social choice theory** (Arrow 1950): No voting system can aggregate preferences while simultaneously satisfying Pareto Efficiency, Independence of Irrelevant Alternatives, and Non-Oligarchy
|
||||
|
||||
2. **Complexity theory**: Computational intractability results (NP-completeness, PSPACE-hardness) show certain coordination problems cannot be solved efficiently even with unlimited resources
|
||||
|
||||
3. **Multi-objective optimization**: Pareto frontier analysis demonstrates irreducible tradeoffs between competing objectives—you cannot improve all objectives simultaneously without sacrificing others
|
||||
|
||||
4. **Intelligence measurement** (Oswald, Ferguson & Bringsjord 2025): Arrow's impossibility conditions apply to machine intelligence measures in agent-environment frameworks, proving no MIM can satisfy fairness conditions simultaneously
|
||||
|
||||
## Why Convergence Matters
|
||||
|
||||
These traditions developed independently:
|
||||
- Using different mathematical frameworks (voting theory, computational models, optimization theory, measurement theory)
|
||||
- Addressing different domains (governance, computation, resource allocation, AI evaluation)
|
||||
- Separated by decades (Arrow 1950 to Oswald et al. 2025)
|
||||
- Yet all converge on the same structural conclusion: certain coordination and measurement problems have no fair, complete solution
|
||||
|
||||
This convergence is significant because:
|
||||
- **Independent discovery** across disciplines suggests deep structural truth rather than domain-specific limitation
|
||||
- **Formal/mathematical proofs** mean these are not empirical or contingent—they hold necessarily given the axioms
|
||||
- **Constructive impossibility** means each tradition specifies exactly which combinations of desirable properties are incompatible (not just "no solution exists")
|
||||
- **Escalating scope** from preference aggregation → computational efficiency → objective tradeoffs → measurement fairness suggests the pattern applies to increasingly fundamental problems
|
||||
|
||||
## Implications for Coordination Problems
|
||||
|
||||
The pattern suggests that [[designing coordination rules is categorically different from designing coordination outcomes as nine intellectual traditions independently confirm]]—we can design processes that navigate tradeoffs transparently, but we cannot design outcomes that eliminate them.
|
||||
|
||||
For AI alignment specifically: if impossibility results converge across multiple traditions, alignment research should shift focus from "solving" alignment (finding the one right answer) to designing coordination mechanisms that navigate irreducible tradeoffs explicitly. This reframes alignment from an optimization problem (find the global maximum) to a governance problem (navigate tradeoffs fairly).
|
||||
|
||||
The intelligence measurement impossibility (Oswald et al. 2025) is particularly significant because it operates at a deeper level than preference aggregation—if we cannot even measure intelligence fairly, the alignment target is underspecified before we begin aggregating preferences.
|
||||
|
||||
---
|
||||
|
||||
## Related Claims
|
||||
|
||||
- [[designing coordination rules is categorically different from designing coordination outcomes as nine intellectual traditions independently confirm]]
|
||||
- [[AI alignment is a coordination problem not a technical problem]]
|
||||
- [[arrows-impossibility-theorem-applies-to-machine-intelligence-measurement-making-fair-universal-intelligence-metrics-mathematically-impossible]]
|
||||
|
||||
Topics:
|
||||
- [[domains/ai-alignment/_map]]
|
||||
- [[foundations/critical-systems/_map]]
|
||||
|
|
@ -21,6 +21,12 @@ This phased approach is also a practical response to the observation that since
|
|||
|
||||
Anthropic's RSP rollback demonstrates the opposite pattern in practice: the company scaled capability while weakening its pre-commitment to adequate safety measures. The original RSP required guaranteeing safety measures were adequate *before* training new systems. The rollback removes this forcing function, allowing capability development to proceed with safety work repositioned as aspirational ('we hope to create a forcing function') rather than mandatory. This provides empirical evidence that even safety-focused organizations prioritize capability scaling over alignment-first development when competitive pressure intensifies, suggesting the claim may be normatively correct but descriptively violated by actual frontier labs under market conditions.
|
||||
|
||||
|
||||
### Additional Evidence (challenge)
|
||||
*Source: [[2025-08-00-oswald-arrowian-impossibility-machine-intelligence]] | Added: 2026-03-12 | Extractor: anthropic/claude-sonnet-4.5*
|
||||
|
||||
If intelligence measurement itself is impossible under fairness constraints (Oswald et al. 2025), then 'building alignment mechanisms before scaling capability' faces a deeper problem: we cannot fairly measure what we are trying to align. The Oswald proof shows that no agent-environment-based machine intelligence measure can satisfy Arrow's fairness conditions simultaneously. This means alignment mechanisms must be built without a fair universal metric to optimize against—the target is fundamentally underspecified, not just difficult to hit. The challenge is not just 'build alignment before capability' but 'build alignment when the measurement of capability itself violates fairness conditions.' This suggests alignment mechanisms must be coordination processes that navigate tradeoffs rather than optimization targets that assume a measurable objective exists.
|
||||
|
||||
---
|
||||
|
||||
Relevant Notes:
|
||||
|
|
|
|||
|
|
@ -7,9 +7,15 @@ date: 2025-08-07
|
|||
domain: ai-alignment
|
||||
secondary_domains: [critical-systems]
|
||||
format: paper
|
||||
status: unprocessed
|
||||
status: processed
|
||||
priority: high
|
||||
tags: [arrows-theorem, machine-intelligence, impossibility, Legg-Hutter, Chollet-ARC, formal-proof]
|
||||
processed_by: theseus
|
||||
processed_date: 2026-03-11
|
||||
claims_extracted: ["arrows-impossibility-theorem-applies-to-machine-intelligence-measurement-making-fair-universal-intelligence-metrics-mathematically-impossible.md", "convergent-impossibility-across-four-traditions-strengthens-structural-coordination-limits.md"]
|
||||
enrichments_applied: ["AI alignment is a coordination problem not a technical problem.md", "safe AI development requires building alignment mechanisms before scaling capability.md"]
|
||||
extraction_model: "anthropic/claude-sonnet-4.5"
|
||||
extraction_notes: "Fourth independent impossibility tradition extending Arrow's theorem from preference aggregation to intelligence measurement. Two new claims extracted: (1) the core impossibility result for MIMs, (2) the convergent pattern across four traditions. Three enrichments: extends coordination-problem framing, adds fourth tradition to convergent impossibility map, challenges capability-first alignment by showing measurement target is underspecified. High significance—operates at deeper level than preference aggregation impossibility."
|
||||
---
|
||||
|
||||
## Content
|
||||
|
|
@ -41,3 +47,10 @@ No agent-environment-based MIM simultaneously satisfies analogs of Arrow's fairn
|
|||
PRIMARY CONNECTION: universal alignment is mathematically impossible because Arrows impossibility theorem applies to aggregating diverse human preferences into a single coherent objective
|
||||
WHY ARCHIVED: Fourth independent impossibility tradition — extends Arrow's theorem from alignment to intelligence measurement itself
|
||||
EXTRACTION HINT: Focus on the extension from preference aggregation to intelligence measurement and what this means for alignment targets
|
||||
|
||||
|
||||
## Key Facts
|
||||
- Paper published at AGI 2025 Conference (Springer LNCS vol. 16058)
|
||||
- Authors: Oswald, J.T., Ferguson, T.M., & Bringsjord, S.
|
||||
- Proof applies to Legg-Hutter Intelligence and Chollet's ARC benchmark
|
||||
- Bringsjord is AI formalist at RPI
|
||||
|
|
|
|||
Loading…
Reference in a new issue