- Source: inbox/archive/2025-09-00-gaikwad-murphys-laws-alignment.md - Domain: ai-alignment - Extracted by: headless extraction cron (worker 4) Pentagon-Agent: Theseus <HEADLESS>
43 lines
2.7 KiB
Markdown
43 lines
2.7 KiB
Markdown
---
|
|
type: claim
|
|
domain: ai-alignment
|
|
description: "Murphy's Law states the alignment gap always wins unless actively routed around; MAPS framework provides four design levers for managing misspecification"
|
|
confidence: experimental
|
|
source: "Madhava Gaikwad, Murphy's Laws of AI Alignment (2025)"
|
|
created: 2026-03-11
|
|
---
|
|
|
|
# Alignment gap cannot be eliminated but can be mapped bounded and managed through MAPS framework
|
|
|
|
**Murphy's Law of AI Alignment**: "The gap always wins unless you actively route around misspecification."
|
|
|
|
The alignment gap—the difference between specified objectives and true human values—is not a problem to be solved but a structural feature to be managed. Gaikwad proposes the MAPS framework as a conceptual model for managing (not eliminating) this gap through four design levers:
|
|
|
|
1. **Misspecification**: Map where feedback is unreliable
|
|
2. **Annotation**: Improve feedback quality in known problem regions
|
|
3. **Pressure**: Adjust optimization intensity based on confidence
|
|
4. **Shift**: Adapt as contexts change
|
|
|
|
This reframes alignment from "eliminate the gap" to "bound and navigate the gap." The formal result on calibration oracles provides theoretical support: knowing where problems occur enables polynomial rather than exponential learning, suggesting that structural interventions (not just more data) are necessary.
|
|
|
|
## Evidence
|
|
|
|
**Murphy's Law formulation**: "The gap always wins unless you actively route around misspecification" (Gaikwad 2025).
|
|
|
|
**MAPS framework**: Four design levers for managing alignment gap—Misspecification, Annotation, Pressure, Shift (Gaikwad 2025).
|
|
|
|
**Theoretical foundation**: The exponential sample complexity result proves the gap cannot be eliminated through more data alone—you need structural interventions (Gaikwad 2025).
|
|
|
|
## Challenges
|
|
|
|
The framework is conceptual rather than operational. It names the levers but does not specify how to pull them in practice. "Map misspecification" and "adjust optimization pressure" are design principles, not algorithms. The paper does not demonstrate MAPS applied to a concrete alignment problem.
|
|
|
|
---
|
|
|
|
Relevant Notes:
|
|
- [[feedback-misspecification-creates-exponential-sample-complexity-barrier-that-calibration-oracles-overcome]] — formal foundation for why structural intervention is necessary
|
|
- [[the specification trap means any values encoded at training time become structurally unstable as deployment contexts diverge from training conditions]] — related structural instability
|
|
- [[pluralistic alignment must accommodate irreducibly diverse values simultaneously rather than converging on a single aligned state]] — convergent conclusion about managing rather than eliminating gaps
|
|
|
|
Topics:
|
|
- [[domains/ai-alignment/_map]]
|