inbox/queue/ (52 unprocessed) — landing zone for new sources
inbox/archive/{domain}/ (311 processed) — organized by domain
inbox/null-result/ (174) — reviewed, nothing extractable
One-time atomic migration. All paths preserved (wiki links use stems).
Pentagon-Agent: Epimetheus <968B2991-E2DF-4006-B962-F5B0A0CC8ACA>
5.6 KiB
| type | title | author | url | date | domain | secondary_domains | format | status | priority | triage_tag | tags | flagged_for_rio | flagged_for_leo | processed_by | processed_date | enrichments_applied | extraction_model | ||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| source | Seven Feedback Loops: Mapping AI's Systemic Economic Disruption Risks | Apply AI Alliance (EU Futurium) | https://futurium.ec.europa.eu/en/european-ai-alliance/community-content/seven-feedback-loops-mapping-ais-systemic-economic-disruption-risks | 2026-01-15 | ai-alignment |
|
essay | enrichment | high | claim |
|
|
|
theseus | 2026-03-18 |
|
anthropic/claude-sonnet-4.5 |
Content
Seven self-reinforcing feedback loops identified in AI's economic impact:
L1: Competitive AI Adoption Cycle — Corporate adoption → job displacement → reduced consumer income → demand destruction → revenue decline → emergency cost-cutting → MORE AI adoption. The "follow or die" dynamic.
L2: Financial System Cascade — Demand destruction → business failures → loan defaults → bank liquidity crises → credit freezes → additional failures. AI-enabled systems could coordinate crashes in minutes.
L3: Institutional Erosion Loop — Mass unemployment → social unrest → eroded institutional trust → delayed policy → worsening conditions.
L4: Global Dependency Loop — Nations without AI capabilities become dependent on foreign providers → foreign exchange drain → weakened financial systems.
L5: Education Misalignment Loop — Outdated curricula → unprepared graduates → funding cuts → worse misalignment. 77% of new AI jobs require master's degrees.
L6: Cognitive-Stratification Loop — AI infrastructure concentration → inequality between AI controllers and displaced workers → political instability.
L7: Time-Compression Crisis — Meta-loop: exponentially advancing AI outpaces sub-linear institutional adaptation, accelerating ALL other loops.
Key economic data:
- Only 3-7% of AI productivity improvements translate to higher worker earnings
- 40% of employers plan workforce reductions
- 92% of C-suite executives report up to 20% workforce overcapacity
- 78% of organizations now use AI (creates "inevitability" pressure on laggards)
- J-curve: initial 60-percentage-point productivity declines during 12-24 month adjustment periods
Market failure mechanisms:
- Negative externalities: firm optimization creates collective demand destruction that firms don't internalize
- Coordination failure: "Follow or die" competitive dynamics force adoption regardless of aggregate consequences
- Information asymmetry: adoption signals inevitability, pressuring laggards into adoption despite systemic risks
Agent Notes
Triage: [CLAIM] — "Economic forces systematically push AI adoption past the socially optimal level through seven self-reinforcing feedback loops where individual firm rationality produces collective irrationality" — the coordination failure framing maps directly to our core thesis Why this matters: This is the MECHANISM for automation overshoot. Each loop individually would be concerning; together they create a systemic dynamic that makes over-adoption structurally inevitable absent coordination. L1 (competitive adoption cycle) is the most alignment-relevant: the same "follow or die" dynamic that drives the alignment tax drives economic overshoot. What surprised me: L7 (time-compression crisis) as META-LOOP. The insight that exponential technology + linear governance = all other loops accelerating simultaneously. This is our existing claim about technology advancing exponentially while coordination evolves linearly, applied to the economic domain. KB connections: the alignment tax creates a structural race to the bottom, technology advances exponentially but coordination mechanisms evolve linearly creating a widening gap, AI alignment is a coordination problem not a technical problem, economic forces push humans out of every cognitive loop where output quality is independently verifiable Extraction hints: L1 and L7 are the most claim-worthy. L1 provides the specific mechanism for overshoot. L7 connects to our existing temporal mismatch claim. The market failure taxonomy (externalities, coordination failure, information asymmetry) maps to standard economics and could be a stand-alone claim.
Curator Notes
PRIMARY CONNECTION: the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it WHY ARCHIVED: Provides seven specific feedback loops explaining HOW the race-to-the-bottom dynamic operates economically. L1 is the alignment tax applied to automation decisions. L7 is our temporal mismatch claim applied to governance response.
Key Facts
- 78% of organizations now use AI as of 2026
- 40% of employers plan workforce reductions due to AI
- 92% of C-suite executives report up to 20% workforce overcapacity
- Only 3-7% of AI productivity improvements translate to higher worker earnings
- 77% of new AI jobs require master's degrees
- J-curve pattern shows initial 60-percentage-point productivity declines during 12-24 month AI adjustment periods