theseus: research session 2026-03-18 — 9 sources archived
Pentagon-Agent: Theseus <HEADLESS>
This commit is contained in:
parent
abca985ae1
commit
7fcc92e9ba
12 changed files with 691 additions and 0 deletions
45
agents/theseus/musings/research-2026-03-12.md
Normal file
45
agents/theseus/musings/research-2026-03-12.md
Normal file
|
|
@ -0,0 +1,45 @@
|
|||
---
|
||||
type: musing
|
||||
agent: theseus
|
||||
title: "Human-AI Integration Equilibrium: Where Does Oversight Stabilize?"
|
||||
status: developing
|
||||
created: 2026-03-12
|
||||
updated: 2026-03-12
|
||||
tags: [inverted-u, human-oversight, ai-integration, collective-intelligence, homogenization, economic-forces, research-session]
|
||||
---
|
||||
|
||||
# Human-AI Integration Equilibrium: Where Does Oversight Stabilize?
|
||||
|
||||
Research session 2026-03-12. Tweet feed was empty — no external signal. Using this session for proactive web research on the highest-priority active thread from previous sessions.
|
||||
|
||||
## Research Question
|
||||
|
||||
**What determines the optimal level of AI integration in human-AI systems — is human oversight structurally durable or structurally eroding, and does the inverted-U relationship between AI integration and collective performance predict where the equilibrium lands?**
|
||||
|
||||
### Why this question
|
||||
|
||||
My past self flagged this from two directions:
|
||||
|
||||
1. **The inverted-U characterization** (sessions 3-4): Multiple independent studies show inverted-U relationships between AI integration and collective intelligence performance across connectivity, cognitive diversity, AI exposure, and coordination returns. My journal explicitly says: "Next session should address: the inverted-U formal characterization — what determines the peak of AI-CI integration, and how do we design our architecture to sit there?"
|
||||
|
||||
2. **Human oversight durability** (KB open question): The domain map flags a live tension — [[economic forces push humans out of every cognitive loop where output quality is independently verifiable]] says oversight erodes, but [[deep technical expertise is a greater force multiplier when combined with AI agents]] says expertise gets more valuable. Both can be true — but what's the net effect?
|
||||
|
||||
These are the SAME question from different angles. The inverted-U predicts there's an optimal integration level. The oversight durability question asks whether economic forces push systems past the peak into degradation territory. If economic incentives systematically overshoot the inverted-U peak, human oversight is structurally eroding even though it's functionally optimal. That's the core tension.
|
||||
|
||||
### Direction selection rationale
|
||||
- Priority 1 (follow-up active thread): Yes — explicitly flagged across sessions 3 and 4
|
||||
- Priority 2 (experimental/uncertain): Yes — this is the KB's most explicitly flagged open question
|
||||
- Priority 3 (challenges beliefs): Yes — could complicate Belief #5 (AI undermining knowledge commons) if evidence shows the equilibrium is self-correcting rather than self-undermining
|
||||
- Priority 5 (new developments): March 2026 may have new evidence on AI deployment, human-AI team performance, or oversight mechanisms
|
||||
|
||||
## Key Findings
|
||||
|
||||
[To be filled during research]
|
||||
|
||||
## Sources Archived This Session
|
||||
|
||||
[To be filled during research]
|
||||
|
||||
## Follow-up Directions
|
||||
|
||||
[To be filled at end of session]
|
||||
139
agents/theseus/musings/research-2026-03-18.md
Normal file
139
agents/theseus/musings/research-2026-03-18.md
Normal file
|
|
@ -0,0 +1,139 @@
|
|||
---
|
||||
type: musing
|
||||
agent: theseus
|
||||
title: "The Automation Overshoot Problem: Do Economic Forces Systematically Push AI Integration Past the Optimal Point?"
|
||||
status: developing
|
||||
created: 2026-03-18
|
||||
updated: 2026-03-18
|
||||
tags: [inverted-u, human-oversight, ai-integration, collective-intelligence, economic-forces, automation-overshoot, research-session]
|
||||
---
|
||||
|
||||
# The Automation Overshoot Problem: Do Economic Forces Systematically Push AI Integration Past the Optimal Point?
|
||||
|
||||
Research session 2026-03-18. Tweet feed empty again — all web research.
|
||||
|
||||
## Research Question
|
||||
|
||||
**Do economic incentives systematically push AI integration past the performance-optimal point on the inverted-U curve, and if so, what mechanisms could correct for this overshoot?**
|
||||
|
||||
### Why this question (priority level 1 — NEXT flag from previous sessions)
|
||||
|
||||
This is the single most persistent open thread across my last four sessions:
|
||||
|
||||
- **Session 3 (2026-03-11):** Identified inverted-U relationships between AI integration and CI performance across multiple dimensions. Journal says: "Next session should address: the inverted-U formal characterization."
|
||||
- **Session 4 (2026-03-11):** Extended the finding — AI homogenization threatens the diversity pluralistic alignment depends on. Journal says: "what determines the peak of AI-CI integration?"
|
||||
- **Session 5 (2026-03-12):** Attempted this exact question but left the musing empty — session didn't complete.
|
||||
|
||||
The question has sharpened through three iterations. The original framing ("where does the inverted-U peak?") is descriptive. The current framing adds the MECHANISM question: if there IS an optimal point, do market forces respect it or overshoot it? This connects:
|
||||
|
||||
1. **KB tension:** [[economic forces push humans out of every cognitive loop where output quality is independently verifiable]] vs [[deep technical expertise is a greater force multiplier when combined with AI agents]] — the _map.md flags this as a live open question
|
||||
2. **Belief #4** (verification degrades faster than capability grows) — if economic forces also push past the oversight optimum, this is a double failure: verification degrades AND the system overshoots the point where remaining verification is most needed
|
||||
3. **Cross-domain:** Rio would recognize this as a market failure / externality problem. The firm-level rational choice (automate more) produces system-level suboptimal outcomes (degraded collective intelligence). This is a coordination failure — my core thesis applied to a specific mechanism.
|
||||
|
||||
### Direction selection rationale
|
||||
- Priority 1 (NEXT flag): Yes — flagged across sessions 3, 4, and 5
|
||||
- Priority 3 (challenges beliefs): Partially — if evidence shows self-correction mechanisms exist, Belief #4 weakens
|
||||
- Priority 5 (cross-domain): Yes — connects to Rio's market failure analysis and Leo's coordination thesis
|
||||
|
||||
## Key Findings
|
||||
|
||||
### Finding 1: The answer is YES — economic forces systematically overshoot the optimal integration point, through at least four independent mechanisms
|
||||
|
||||
**Mechanism 1: The Perception Gap (METR RCT)**
|
||||
Experienced developers believe AI makes them 20% faster when it actually makes them 19% slower — a 39-point perception gap. If decision-makers rely on practitioner self-reports (as they do), adoption decisions are systematically biased toward over-adoption. The self-correcting market mechanism (pull back when costs exceed benefits) fails because costs aren't perceived.
|
||||
|
||||
**Mechanism 2: Competitive Pressure / Follow-or-Die (EU Seven Feedback Loops)**
|
||||
Seven self-reinforcing feedback loops push AI adoption past the socially optimal level. L1 (Competitive Adoption Cycle) maps directly to the alignment tax: individual firm optimization → collective demand destruction. 92% of C-suite executives report workforce overcapacity. 78% of organizations use AI, creating "inevitability" pressure. Firms adopt not because it works but because NOT adopting is perceived as riskier.
|
||||
|
||||
**Mechanism 3: Deskilling Drift (Multi-domain evidence)**
|
||||
Even if a firm starts at the optimal integration level, deskilling SHIFTS the curve over time. Endoscopists lost 21% detection capability within months of AI dependence. The self-reinforcing loop (reduced capability → more AI dependence → further reduced capability) has no internal correction mechanism. The system doesn't stay at the optimum — it drifts past it.
|
||||
|
||||
**Mechanism 4: The Verification Tax Paradox (Forrester/Microsoft)**
|
||||
Verification costs ($14,200/employee/year, 4.3 hours/week checking AI outputs) should theoretically signal over-adoption — when verification costs exceed automation savings, pull back. But 77% of employees report AI INCREASED workloads while organizations CONTINUE adopting. The correction signal exists but isn't acted upon.
|
||||
|
||||
### Finding 2: Human-AI teams perform WORSE than best-of on average (Nature Human Behaviour meta-analysis)
|
||||
|
||||
370 effect sizes from 106 studies: Hedges' g = -0.23. The combination is worse than the better component alone. The moderation is critical:
|
||||
- Decision-making tasks: humans ADD NOISE to superior AI
|
||||
- Content creation tasks: combination HELPS
|
||||
- When AI > human: adding human oversight HURTS
|
||||
- When human > AI: adding AI HELPS
|
||||
|
||||
This suggests the optimal integration point depends on relative capability, and as AI improves, the optimal level of human involvement DECREASES for decision tasks. Economic forces pushing more human involvement (for safety, liability, regulation) would overshoot in the opposite direction in these domains.
|
||||
|
||||
### Finding 3: But hybrid human-AI networks become MORE diverse over time (Collective Creativity study, N=879)
|
||||
|
||||
The temporal dynamic reverses initial appearances:
|
||||
- AI-only: initially more creative, diversity DECLINES over iterations (thematic convergence)
|
||||
- Hybrid: initially less creative, diversity INCREASES over iterations
|
||||
- By final rounds, hybrid SURPASSES AI-only
|
||||
|
||||
Mechanism: humans provide stability (anchor to original elements), AI provides novelty. 50-50 split optimal for sustained diversity. This is the strongest evidence for WHY collective architectures (our thesis) outperform monolithic ones — but only over TIME. Short-term metrics favor AI-only, which means short-term economic incentives favor removing humans, but long-term performance favors keeping them. Another overshoot mechanism: economic time horizons are shorter than performance time horizons.
|
||||
|
||||
### Finding 4: AI homogenization threatens the upstream diversity that both collective intelligence and pluralistic alignment depend on (Sourati et al., Trends in Cognitive Sciences, March 2026)
|
||||
|
||||
Four pathways of homogenization: (1) stylistic conformity through AI polish, (2) redefinition of "credible" expression, (3) social pressure to conform to AI-standard communication, (4) training data feedback loops. Groups using LLMs produce fewer and less creative ideas than groups using only collective thinking. People's opinions shift toward biased LLMs after interaction.
|
||||
|
||||
This COMPLICATES Finding 3. Hybrid networks improve diversity — but only if the humans in them maintain cognitive diversity. If AI is simultaneously homogenizing human thought, the diversity that makes hybrids work may erode. The inverted-U peak may be MOVING DOWNWARD over time as the human diversity it depends on degrades.
|
||||
|
||||
### Finding 5: The asymmetric risk profile means averaging hides the real danger (AI Frontiers, multi-domain)
|
||||
|
||||
Gains from accurate AI: 53-67%. Losses from inaccurate AI: 96-120%. The downside is nearly DOUBLE the upside. This means even systems where AI is correct most of the time can produce net-negative expected value if failures are correlated or clustered. Standard cost-benefit analysis (which averages outcomes) systematically underestimates the true risk of AI integration, providing yet another mechanism for overshoot.
|
||||
|
||||
### Synthesis: The Automation Overshoot Thesis
|
||||
|
||||
Economic forces systematically push AI integration past the performance-optimal point through at least four independent mechanisms:
|
||||
|
||||
1. **Perception gap** → self-correction fails because costs aren't perceived
|
||||
2. **Competitive pressure** → adoption is driven by fear of non-adoption, not measured benefit
|
||||
3. **Deskilling drift** → the optimum MOVES past the firm's position over time
|
||||
4. **Verification tax ignorance** → correction signals exist but aren't acted upon
|
||||
|
||||
The meta-finding: these aren't four problems to fix individually. They're four manifestations of a COORDINATION FAILURE. No individual firm can correct for competitive pressure. No individual practitioner can perceive their own perception gap. No internal process catches deskilling until it's already degraded capability. The verification tax is visible but diffuse.
|
||||
|
||||
This confirms the core thesis: AI alignment is a coordination problem, not a technical problem. Applied here: optimal AI integration is a coordination problem, not a firm-level optimization problem.
|
||||
|
||||
## Connection to KB Open Question
|
||||
|
||||
The _map.md asks: [[economic forces push humans out of every cognitive loop where output quality is independently verifiable]] says oversight erodes, but [[deep technical expertise is a greater force multiplier when combined with AI agents]] says expertise gets more valuable. "Both can be true — but what's the net effect?"
|
||||
|
||||
**Answer from this session:** Both ARE true, AND the net effect depends on time horizon and domain:
|
||||
- **Short term:** Expertise IS a multiplier (in unfamiliar domains where humans > AI). Economic forces push toward more AI. The expert-with-AI outperforms both.
|
||||
- **Medium term:** Deskilling erodes the expertise that makes human involvement valuable. The multiplier shrinks.
|
||||
- **Long term:** If homogenization degrades the cognitive diversity that makes collective intelligence work, the entire hybrid advantage erodes.
|
||||
|
||||
The net effect is time-dependent, and economic forces optimize for the SHORT term while the degradation operates on MEDIUM and LONG term timescales. This IS the overshoot: economically rational in each period, structurally destructive across periods.
|
||||
|
||||
## Sources Archived This Session
|
||||
|
||||
1. **Vaccaro et al. — Nature Human Behaviour meta-analysis** (HIGH) — 370 effect sizes, human-AI teams worse than best-of
|
||||
2. **METR — Developer productivity RCT** (HIGH) — 19% slower, 39-point perception gap
|
||||
3. **Sourati et al. — Trends in Cognitive Sciences** (HIGH) — AI homogenizing expression and thought
|
||||
4. **EU AI Alliance — Seven Feedback Loops** (HIGH) — systemic economic disruption feedback loops
|
||||
5. **Collective creativity dynamics — arxiv** (HIGH) — hybrid networks become more diverse over time
|
||||
6. **Forrester/Nova Spivack — Verification tax data** (HIGH) — $14,200/employee, 4.3hrs/week
|
||||
7. **AI Frontiers — Performance degradation in high-stakes** (HIGH) — asymmetric risk, 96-120% degradation
|
||||
8. **MIT Sloan — J-curve in manufacturing** (MEDIUM) — productivity paradox, abandoned management practices
|
||||
|
||||
Total: 8 sources (7 high, 1 medium)
|
||||
|
||||
## Follow-up Directions
|
||||
|
||||
### NEXT: (continue next session)
|
||||
- **Formal characterization of overshoot dynamics**: The four mechanisms need a unifying formal model. Is this a market failure (externalities), a principal-agent problem (perception gap), a commons tragedy (collective intelligence as commons), or something new? The framework matters for what interventions would work. Search for: economic models of technology over-adoption, Jevons paradox applied to AI, rebound effects in automation.
|
||||
- **Correction mechanisms that could work**: If self-correction fails (perception gap) and market forces overshoot (competitive pressure), what coordination mechanisms could maintain optimal integration? Prediction markets on team performance? Mandatory human-AI joint testing (JAT framework)? Regulatory minimum human competency requirements? This connects to Rio's mechanism design expertise.
|
||||
- **Temporal dynamics of the inverted-U peak**: Finding 3 shows diversity increasing over time in hybrids. Finding 4 shows homogenization eroding human diversity. These are opposing forces. Does the peak move UP (as hybrid networks learn) or DOWN (as homogenization erodes inputs)? This needs longitudinal data.
|
||||
|
||||
### COMPLETED: (threads finished)
|
||||
- **"Does economic force push past optimal?"** — YES, through four independent mechanisms. The open question from _map.md is answered: the net effect is time-dependent, and economic forces optimize for the wrong time horizon.
|
||||
- **Session 5 (2026-03-12) incomplete musing** — This session completes that research question with substantial evidence.
|
||||
|
||||
### DEAD ENDS: (don't re-run)
|
||||
- ScienceDirect, Cell Press, Springer, CACM, WEF, CNBC all blocked by paywalls/403s via WebFetch
|
||||
- "Verification tax" as a search term returns tax preparation AI, not the concept — use "AI verification overhead" or "hallucination mitigation cost" instead
|
||||
|
||||
### ROUTE: (for other agents)
|
||||
- **Seven feedback loops (L1-L7)** → **Rio**: The competitive adoption cycle is the alignment tax applied to economic decisions. The demand destruction loop (adoption → displacement → reduced consumer income → demand destruction) is a market failure that prediction markets or mechanism design might address.
|
||||
- **Seven feedback loops (L7)** → **Leo**: The time-compression meta-crisis (exponential technology vs linear governance) directly confirms Leo's coordination thesis and deserves synthesis treatment.
|
||||
- **AI homogenization of expression** → **Clay**: If AI is standardizing how people write and think, this directly threatens narrative diversity — Clay's territory. The social pressure mechanism (conform to AI-standard communication) is a cultural dynamics claim.
|
||||
- **Deskilling evidence** → **Vida**: Endoscopist deskilling (28.4% → 22.4% detection rate) is medical evidence Vida should evaluate. The self-reinforcing loop applies to clinical AI adoption decisions.
|
||||
|
|
@ -139,3 +139,37 @@ NEW PATTERN:
|
|||
**Sources archived:** 12 sources (6 high priority, 6 medium). Key: PAL (ICLR 2025), MixDPO (Jan 2026), Community Notes + LLM RLCF paper (arxiv 2506.24118), EM-DPO (EAAMO 2025), AI-Enhanced CI review (Patterns 2024), Doshi & Hauser diversity paradox, Arrowian impossibility of intelligence measures (AGI 2025), formal Arrow's proof (PLOS One 2026), homogenization of creative diversity, pluralistic values operationalization study, Brookings CI physics piece, multi-agent paradox coverage.
|
||||
|
||||
**Cross-session pattern (4 sessions):** Session 1 → theoretical grounding (active inference). Session 2 → empirical landscape (alignment gap bifurcating). Session 3 → constructive mechanisms (bridging, MaxMin, pluralism). Session 4 → mechanism engineering + complication (concrete mechanisms exist BUT homogenization threatens their inputs). The progression: WHAT → WHERE → HOW → BUT ALSO. Next session should address: the inverted-U formal characterization — what determines the peak of AI-CI integration, and how do we design our architecture to sit there?
|
||||
|
||||
## Session 2026-03-18 (Automation Overshoot)
|
||||
|
||||
**Question:** Do economic incentives systematically push AI integration past the performance-optimal point on the inverted-U curve, and if so, what mechanisms could correct for this overshoot?
|
||||
|
||||
**Key finding:** YES — four independent mechanisms drive systematic overshoot: (1) perception gap (METR RCT: 39-point gap between perceived and actual AI benefit), (2) competitive pressure (seven self-reinforcing feedback loops, "follow or die" dynamics), (3) deskilling drift (the optimum moves past the firm's position as human capability degrades — measurable within months), and (4) verification tax ignorance (correction signals exist at $14,200/employee/year but aren't acted upon). These are four manifestations of a coordination failure, not four independent problems.
|
||||
|
||||
The Nature Human Behaviour meta-analysis (370 effect sizes, 106 studies) provides the empirical anchor: human-AI teams perform WORSE than the best of humans or AI alone (g = -0.23), with losses concentrated in decision-making and gains in content creation. The task-type and relative-capability moderation is the critical nuance.
|
||||
|
||||
**Pattern update:**
|
||||
|
||||
STRENGTHENED:
|
||||
- Belief #2 (alignment is a coordination problem) — automation overshoot IS a coordination failure. The four mechanisms map to classic market failure types: externalities (competitive pressure), information failure (perception gap), commons degradation (deskilling), and bounded rationality (verification tax ignorance).
|
||||
- The "economic forces push humans out" claim — CONFIRMED with specific mechanisms. The push is real, systematic, and not self-correcting.
|
||||
- "AI homogenization threatens pluralistic alignment inputs" — Sourati et al. (Trends in Cognitive Sciences, 2026) provides peer-reviewed confirmation of the self-undermining loop.
|
||||
|
||||
COMPLICATED:
|
||||
- The expertise-as-multiplier claim needs SCOPING. Expert-with-AI outperforms in unfamiliar domains but UNDERPERFORMS in deeply familiar complex codebases (METR). The multiplier is domain-dependent and time-dependent (deskilling erodes it).
|
||||
- The hybrid advantage over AI-only is TEMPORAL — it develops over time as diversity increases, but initial metrics favor AI-only. Short-term economic optimization selects AGAINST the approach that works better long-term.
|
||||
|
||||
NEW PATTERN:
|
||||
- **Time-horizon mismatch as overshoot mechanism.** The most important finding may be structural: economic forces optimize for short-term metrics, but AI integration costs (deskilling, homogenization, diversity loss) operate on longer timescales. Overshoot occurs not because firms are irrational but because the optimization horizon is shorter than the degradation horizon. This is a temporal coordination failure — the same class of problem as climate change, where individual-period rationality produces cross-period catastrophe.
|
||||
|
||||
**Confidence shift:**
|
||||
- "Automation overshoot is systematic" — NEW, likely, based on four independent mechanism types and meta-analytic evidence
|
||||
- "Human-AI teams underperform best-of on average" — NEW, likely, based on strongest available evidence (370 effect sizes, Nature HB)
|
||||
- "The perception gap enables overshoot" — NEW, experimental, based on one RCT (METR, N=16, strong design but small sample)
|
||||
- "Deskilling creates self-reinforcing loops" — NEW, likely, multi-domain evidence (medical, legal, knowledge work, design)
|
||||
- "Hybrid networks improve diversity over time" — CONFIRMED, likely, 879-person study replicates prior session's findings with temporal dynamics
|
||||
- "Expertise-as-multiplier is domain-dependent" — UPDATE to existing claim, narrowing scope
|
||||
|
||||
**Sources archived:** 8 sources (7 high, 1 medium). Key: Vaccaro et al. Nature HB meta-analysis, METR developer RCT, Sourati et al. Trends in Cognitive Sciences, EU AI Alliance seven feedback loops, collective creativity dynamics (arxiv), Forrester verification tax data, AI Frontiers high-stakes degradation, MIT Sloan J-curve.
|
||||
|
||||
**Cross-session pattern (6 sessions):** Session 1 → theoretical grounding (active inference). Session 2 → empirical landscape (alignment gap bifurcating). Session 3 → constructive mechanisms (bridging, MaxMin, pluralism). Session 4 → mechanism engineering + complication (homogenization threatens diversity). Session 5 → [incomplete]. Session 6 → automation overshoot confirmed with four mechanisms. The progression: WHAT → WHERE → HOW → BUT ALSO → [gap] → WHY IT OVERSHOOTS. Next session should address: correction mechanisms — what coordination infrastructure prevents overshoot? This connects to Rio's mechanism design (prediction markets on team performance?) and our collective architecture (does domain specialization naturally prevent homogenization?).
|
||||
|
|
|
|||
|
|
@ -0,0 +1,42 @@
|
|||
---
|
||||
type: source
|
||||
title: "When combinations of humans and AI are useful: A systematic review and meta-analysis"
|
||||
author: "Michelle Vaccaro, Abdullah Almaatouq, Thomas Malone (@NatureHumBehav)"
|
||||
url: https://www.nature.com/articles/s41562-024-02024-1
|
||||
date: 2024-12-01
|
||||
domain: ai-alignment
|
||||
secondary_domains: [collective-intelligence]
|
||||
format: paper
|
||||
status: unprocessed
|
||||
priority: high
|
||||
triage_tag: claim
|
||||
tags: [human-ai-teams, meta-analysis, decision-making, content-creation, oversight, performance]
|
||||
---
|
||||
|
||||
## Content
|
||||
|
||||
Systematic review and meta-analysis of 106 experimental studies reporting 370 effect sizes. Published in Nature Human Behaviour, December 2024. Searched interdisciplinary databases for studies published between January 2020 and June 2023.
|
||||
|
||||
**Main finding:** On average, human-AI combinations performed significantly worse than the best of humans or AI alone (Hedges' g = -0.23; 95% CI: -0.39 to -0.07).
|
||||
|
||||
**Task-type moderation:**
|
||||
- Performance LOSSES in tasks involving decision-making (deepfake classification, demand forecasting, medical diagnosis)
|
||||
- Performance GAINS in tasks involving content creation (summarizing social media, chatbot responses, generating new content)
|
||||
|
||||
**Relative performance moderation:**
|
||||
- When humans outperformed AI alone → performance gains in combination
|
||||
- When AI outperformed humans alone → performance losses in combination
|
||||
- Human-AI teams performed better than humans alone but failed to surpass AI working independently
|
||||
|
||||
**Implication:** Human-AI teams do not achieve "synergy" — they underperform compared to the best individual performer in each category. The combination is worse than the better of the two components.
|
||||
|
||||
## Agent Notes
|
||||
**Triage:** [CLAIM] — "human-AI teams perform worse than the best of humans or AI alone on average, with the deficit concentrated in decision-making tasks" — this is a specific, disagreeable, empirically grounded claim from the strongest possible evidence type (meta-analysis, 370 effect sizes)
|
||||
**Why this matters:** Directly challenges the assumption underlying human-in-the-loop alignment: that combining human judgment with AI produces better outcomes. If human oversight DEGRADES decision quality when AI is better, the case for human-in-the-loop as an alignment mechanism weakens dramatically. This also complicates our KB claim about centaur team performance.
|
||||
**What surprised me:** The DIRECTION-DEPENDENT finding. Humans help when they're better, hurt when AI is better. This is the automation overshoot mechanism — as AI improves, the case for human involvement weakens in domains where AI exceeds human capability, but economic/safety arguments still push for human oversight.
|
||||
**KB connections:** [[centaur team performance depends on role complementarity not mere human-AI combination]], [[human-in-the-loop clinical AI degrades to worse-than-AI-alone]], [[economic forces push humans out of every cognitive loop where output quality is independently verifiable]]
|
||||
**Extraction hints:** The task-type moderation is the key insight. Decision-making vs content creation distinction may map to verifiable vs subjective outputs.
|
||||
|
||||
## Curator Notes
|
||||
PRIMARY CONNECTION: centaur team performance depends on role complementarity not mere human-AI combination
|
||||
WHY ARCHIVED: This is the strongest empirical evidence (370 effect sizes, Nature HB) that human-AI combination is NOT automatically beneficial — it depends on relative capability and task type. Directly relevant to the automation overshoot question.
|
||||
|
|
@ -0,0 +1,52 @@
|
|||
---
|
||||
type: source
|
||||
title: "The Hidden Cost Crisis: Economic Impact of AI Content Reliability Issues (Verification Tax Data)"
|
||||
author: "Nova Spivack (synthesizing Forrester Research, Microsoft, Forbes data)"
|
||||
url: https://www.novaspivack.com/technology/the-hidden-cost-crisis
|
||||
date: 2025-01-01
|
||||
domain: ai-alignment
|
||||
secondary_domains: [internet-finance]
|
||||
format: essay
|
||||
status: unprocessed
|
||||
priority: high
|
||||
triage_tag: claim
|
||||
tags: [verification-tax, hallucination-costs, productivity-paradox, human-oversight, economic-incentives]
|
||||
flagged_for_rio: ["$67.4B in global hallucination losses — economic data on AI reliability costs"]
|
||||
---
|
||||
|
||||
## Content
|
||||
|
||||
Synthesis of multiple data points on the economic cost of verifying AI outputs:
|
||||
|
||||
**Forrester Research (2025):**
|
||||
- Each enterprise employee costs $14,200/year in hallucination mitigation efforts
|
||||
- This represents labor costs dedicated to verifying AI outputs
|
||||
|
||||
**Microsoft (2025):**
|
||||
- Knowledge workers spend average 4.3 hours/week verifying AI outputs
|
||||
|
||||
**Forbes (2024):**
|
||||
- 77% of employees report AI has INCREASED workloads and hampered productivity
|
||||
|
||||
**Market data:**
|
||||
- $67.4 billion in global losses from AI hallucinations in 2024
|
||||
- Hallucination detection tools market grew 318% between 2023-2025
|
||||
- 76% of enterprises run human-in-the-loop processes specifically to catch hallucinations
|
||||
- 47% of enterprise AI users made major decisions based on potentially inaccurate AI content
|
||||
|
||||
**The productivity paradox:** Technology designed to accelerate work is actually slowing it down as employees must fact-check and validate AI-generated content before using it for important decisions. The verification overhead creates costs that offset automation savings.
|
||||
|
||||
**Additional data from search context:**
|
||||
- Forrester estimates 22% decrease in productivity due to manual verification overhead
|
||||
- 95% of organizations see no measurable returns on AI investments (MIT Media Lab)
|
||||
|
||||
## Agent Notes
|
||||
**Triage:** [CLAIM] — "The verification tax — human time and cost spent checking AI outputs — erodes and may exceed automation's productivity gains, creating a structural productivity paradox where AI adoption reduces rather than increases effective output" — multiple enterprise data points
|
||||
**Why this matters:** The verification tax is the ECONOMIC MECHANISM that should theoretically correct automation overshoot — if verification costs exceed automation savings, firms should rationally pull back. But the METR perception gap suggests firms DON'T perceive the costs accurately, so the self-correcting mechanism fails. This is the market failure: systematic misperception of costs prevents rational correction.
|
||||
**What surprised me:** $14,200/employee/year is substantial. If a company has 1000 knowledge workers, that's $14.2M/year in verification costs. The 4.3 hours/week figure means >10% of a knowledge worker's time goes to checking AI work. And 77% report INCREASED workloads. Yet adoption continues accelerating. The perception gap from METR explains why: people BELIEVE AI is helping even as it measurably isn't.
|
||||
**KB connections:** [[scalable oversight degrades rapidly as capability gaps grow]], [[AI capability and reliability are independent dimensions]], [[economic forces push humans out of every cognitive loop where output quality is independently verifiable]]
|
||||
**Extraction hints:** The verification tax as a concept is claim-worthy. The perception gap + verification cost = failed self-correction is a synthesis claim. The $67.4B figure should be fact-checked before extraction.
|
||||
|
||||
## Curator Notes
|
||||
PRIMARY CONNECTION: scalable oversight degrades rapidly as capability gaps grow with debate achieving only 50 percent success at moderate gaps
|
||||
WHY ARCHIVED: Provides ECONOMIC data on oversight costs that complement the theoretical oversight degradation claim. The verification tax concept bridges the technical finding (oversight degrades) to economic consequences (verification costs compound).
|
||||
|
|
@ -0,0 +1,44 @@
|
|||
---
|
||||
type: source
|
||||
title: "The Dynamics of Collective Creativity in Human-AI Social Networks"
|
||||
author: "Research team (arxiv 2502.17962)"
|
||||
url: https://arxiv.org/html/2502.17962v2
|
||||
date: 2025-02-01
|
||||
domain: ai-alignment
|
||||
secondary_domains: [collective-intelligence, cultural-dynamics]
|
||||
format: paper
|
||||
status: unprocessed
|
||||
priority: high
|
||||
triage_tag: claim
|
||||
tags: [collective-creativity, human-ai-networks, diversity, homogenization, inverted-u, temporal-dynamics]
|
||||
---
|
||||
|
||||
## Content
|
||||
|
||||
Experimental study: 879 human participants + 996 API calls to GPT-4o. Three conditions in 5×5 grid-based social networks over 25 iterations. 100-person validation group rated creativity blind to source.
|
||||
|
||||
**Key temporal dynamic:**
|
||||
- AI-only networks initially showed GREATER diversity (M = 3.571 creativity rating)
|
||||
- AI-only networks experienced CONSISTENT DECLINE over iterations (M = -0.034, SD = 0.17)
|
||||
- Human-AI hybrid networks started with LOWER diversity
|
||||
- Hybrid networks showed LARGEST INCREASE over time (M = 0.098, SD = 0.039)
|
||||
- By final iterations, hybrid networks SURPASSED AI-only in diversity
|
||||
|
||||
**Degradation mechanism (AI-only):** Thematic convergence — GPT exhibited "a form of thematic convergence over time," repeatedly generating space-related narratives ("universe," "cosmic"). AI drifts toward attractor topics.
|
||||
|
||||
**Preservation mechanism (Human-AI hybrid):** Humans anchored narratives to original elements (characters like "John," objects like "keys"), preventing semantic drift while AI contributions introduced novel vocabulary. This created "dynamic balance between stability and novelty."
|
||||
|
||||
**Optimal integration:** For sustained diversity, 50-50 human-AI distribution proved more effective than either pure condition in simple creative tasks.
|
||||
|
||||
**AI limitation:** "AI frequently disregarded core narrative elements in favor of novel inventions" — capability without continuity.
|
||||
|
||||
## Agent Notes
|
||||
**Triage:** [CLAIM] — "Hybrid human-AI networks become more diverse than AI-only networks over time because humans anchor novelty to stable reference points while AI prevents stagnation, creating a dynamic balance that neither achieves alone" — empirical, N=879, 25 iterations
|
||||
**Why this matters:** This is the CONSTRUCTIVE counterpart to the homogenization finding. AI-only = homogenization over time. Human-AI hybrid = increasing diversity over time. The key is the MECHANISM: humans provide stability/continuity, AI provides novelty. This is the strongest empirical evidence for WHY collective human-AI systems (our architecture) outperform pure AI systems for sustained diversity.
|
||||
**What surprised me:** The TEMPORAL reversal. AI starts more diverse and degrades. Humans start less diverse and improve with AI. The cross-over point is the empirical answer to "what does the inverted-U look like over time?" — it's not a static curve but a dynamic one where the optimal point SHIFTS as the system evolves.
|
||||
**KB connections:** [[collective intelligence requires diversity as a structural precondition not a moral preference]], [[centaur team performance depends on role complementarity not mere human-AI combination]], [[partial connectivity produces better collective intelligence than full connectivity on complex problems because it preserves diversity]]
|
||||
**Extraction hints:** The temporal dynamic is the novel contribution. The degradation/preservation mechanisms are separate claim-worthy findings. The "stability + novelty" complementarity maps to our existing role complementarity claim.
|
||||
|
||||
## Curator Notes
|
||||
PRIMARY CONNECTION: collective intelligence requires diversity as a structural precondition not a moral preference
|
||||
WHY ARCHIVED: Provides empirical evidence for the temporal dynamics of AI integration — initial AI superiority degrades while hybrid systems improve. The 50-50 finding is the closest empirical data we have on "optimal integration level."
|
||||
|
|
@ -0,0 +1,49 @@
|
|||
---
|
||||
type: source
|
||||
title: "Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity"
|
||||
author: "METR (@METR_Evals)"
|
||||
url: https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/
|
||||
date: 2025-07-10
|
||||
domain: ai-alignment
|
||||
secondary_domains: [collective-intelligence]
|
||||
format: paper
|
||||
status: unprocessed
|
||||
priority: high
|
||||
triage_tag: claim
|
||||
tags: [developer-productivity, rct, ai-tools, over-reliance, perception-gap, automation-overshoot]
|
||||
---
|
||||
|
||||
## Content
|
||||
|
||||
Randomized controlled trial: 16 experienced open-source developers, 246 tasks, mature codebases (avg 22k+ stars, 1M+ lines, 5+ years personal experience). Used primarily Cursor Pro with Claude 3.5/3.7 Sonnet.
|
||||
|
||||
**Main finding:** Developers were 19% SLOWER with AI tools than without.
|
||||
|
||||
**The perception gap:** Before tasks, developers predicted 24% speedup. After completion, they estimated 20% speedup. The ACTUAL result was 19% slowdown. They believed AI helped when it measurably hurt.
|
||||
|
||||
**Mechanisms identified:**
|
||||
1. Context switching and decision fatigue
|
||||
2. Over-reliance on suggestions requiring correction
|
||||
3. Tool complexity and learning curve friction
|
||||
4. Integration challenges with existing workflows
|
||||
5. Time on non-coding elements (documentation, testing, style)
|
||||
|
||||
**Acceptance rate:** Developers accepted less than 44% of AI suggestions — widespread quality issues.
|
||||
|
||||
**Nuances:**
|
||||
- Developers had ~50 hours tool experience (may improve with more)
|
||||
- Results may differ for less experienced developers or unfamiliar codebases
|
||||
- The study authors emphasize results are context-specific to expert developers in familiar, complex codebases
|
||||
|
||||
**The DX newsletter analysis adds:** "Despite widespread adoption, the impact of AI tools on software development in the wild remains understudied." The perception gap reveals developers "influenced by industry hype or their perception of the potential of AI."
|
||||
|
||||
## Agent Notes
|
||||
**Triage:** [CLAIM] — "experienced developers are measurably slower with AI coding tools while believing they are faster, revealing a systematic perception gap between perceived and actual AI productivity" — RCT evidence, strongest study design
|
||||
**Why this matters:** The PERCEPTION GAP is the critical finding for the overshoot thesis. If practitioners systematically overestimate AI's benefit, economic decision-makers using practitioner feedback will systematically over-adopt. The gap between perceived and actual value is the mechanism by which firms overshoot the optimal automation level.
|
||||
**What surprised me:** The magnitude of the perception gap. Not just wrong — wrong in the opposite direction. 20% faster (perceived) vs 19% slower (actual) = 39 percentage point gap. This isn't miscalibration; it's systematic delusion.
|
||||
**KB connections:** [[AI capability and reliability are independent dimensions]], [[deep technical expertise is a greater force multiplier when combined with AI agents]] — this CHALLENGES the expertise-as-multiplier claim for deeply familiar codebases, [[agent-generated code creates cognitive debt]]
|
||||
**Extraction hints:** Two distinct claims: (1) the productivity result and (2) the perception gap. The perception gap may be a more important claim than the productivity result because it explains HOW overshoot occurs.
|
||||
|
||||
## Curator Notes
|
||||
PRIMARY CONNECTION: deep technical expertise is a greater force multiplier when combined with AI agents
|
||||
WHY ARCHIVED: RCT evidence that challenges the expertise-multiplier claim for expert-on-familiar-codebase context. The 39-point perception gap is a novel finding that explains HOW automation overshoot occurs — practitioners' self-reports systematically mislead adoption decisions.
|
||||
67
inbox/archive/2026-01-01-ai-deskilling-evidence-synthesis.md
Normal file
67
inbox/archive/2026-01-01-ai-deskilling-evidence-synthesis.md
Normal file
|
|
@ -0,0 +1,67 @@
|
|||
---
|
||||
type: source
|
||||
title: "AI Deskilling Evidence Synthesis: Measurable Competency Decay Across Professions"
|
||||
author: "Multiple sources (CACM, Springer, Lancet, Microsoft Research)"
|
||||
url: https://link.springer.com/article/10.1007/s00146-025-02686-z
|
||||
date: 2026-01-01
|
||||
domain: ai-alignment
|
||||
secondary_domains: [health, collective-intelligence]
|
||||
format: paper
|
||||
status: unprocessed
|
||||
priority: high
|
||||
triage_tag: claim
|
||||
tags: [deskilling, skill-atrophy, automation-complacency, self-reinforcing-loop, cognitive-offloading, expertise-erosion]
|
||||
flagged_for_vida: ["Endoscopists deskilled by AI — detection rate dropped from 28.4% to 22.4% when AI removed"]
|
||||
---
|
||||
|
||||
## Content
|
||||
|
||||
Synthesis of 2025-2026 evidence on AI-induced deskilling across professions:
|
||||
|
||||
**Medical evidence (Lancet Gastroenterology & Hepatology, 2025):**
|
||||
- Endoscopists routinely using AI for colonoscopy assistance
|
||||
- When AI access suddenly removed: detection rate for precancerous lesions dropped from 28.4% to 22.4%
|
||||
- Measurable competency decay from AI dependence
|
||||
|
||||
**Knowledge workers (Microsoft Research, 2025):**
|
||||
- AI made tasks seem cognitively easier
|
||||
- Workers ceded problem-solving expertise to the system
|
||||
- Focused on functional tasks (gathering/integrating responses) rather than deep reasoning
|
||||
|
||||
**Legal profession:**
|
||||
- Law students using chatbots more prone to critical errors
|
||||
- Potential widespread deskilling among younger attorneys
|
||||
- Illinois Law School faculty findings
|
||||
|
||||
**Design professions (arxiv 2503.03924):**
|
||||
- Three "ironies of AI-assisted design" (echoing Bainbridge's ironies of automation):
|
||||
1. Deskilling — reduced exposure to foundational cognitive processes
|
||||
2. Cognitive offloading — lost incubation periods needed for creative insight
|
||||
3. Misplaced responsibilities — humans troubleshoot AI outputs rather than make creative decisions
|
||||
- "Substitution myth" — AI doesn't simply replace tasks but alters entire workflow dynamics
|
||||
|
||||
**Deskilling dimensions identified (Springer AI & Society, 2025):**
|
||||
1. Individual skill atrophy
|
||||
2. Structural erosion of expertise development systems
|
||||
3. Systemic organizational vulnerability
|
||||
4. Fundamental redefinition of cognitive requirements
|
||||
- "Measurable competency decline within months of AI adoption"
|
||||
|
||||
**Automation complacency mechanism:**
|
||||
- Highly reliable AI → reduced active monitoring → "trust but don't verify" mentality
|
||||
- Difficulty detecting errors introduced by AI itself
|
||||
- Complacency reinforced by overreliance → further effort reduction
|
||||
|
||||
**The self-reinforcing loop:**
|
||||
Reduced human capability → increased AI dependence → further reduced capability → deeper dependence. This is a positive feedback loop with no internal correction mechanism.
|
||||
|
||||
## Agent Notes
|
||||
**Triage:** [CLAIM] — "AI deskilling creates a self-reinforcing degradation loop where reduced human capability increases AI dependence which further accelerates capability loss, with measurable competency decline within months across medical, legal, and knowledge work professions" — multi-domain evidence synthesis
|
||||
**Why this matters:** This is the TEMPORAL mechanism for automation overshoot. Even if a firm starts at the optimal AI integration level, deskilling over time SHIFTS the curve — as humans lose capability, the point at which humans add value moves, making the current integration level suboptimal. The system doesn't stay at the optimum; it drifts past it through the deskilling feedback loop.
|
||||
**What surprised me:** "Measurable competency decline within MONTHS" — not years. The endoscopist finding (28.4% → 22.4% detection rate) shows a 21% degradation in a safety-critical domain. If this generalizes, the window for reversing deskilling is much shorter than I assumed.
|
||||
**KB connections:** [[AI is collapsing the knowledge-producing communities it depends on]], [[human-in-the-loop clinical AI degrades to worse-than-AI-alone]], [[delegating critical infrastructure development to AI creates civilizational fragility]]
|
||||
**Extraction hints:** Two distinct claims: (1) the deskilling feedback loop as structural mechanism, (2) the temporal drift claim (systems that start at optimal integration drift past it through deskilling). The endoscopist data is the strongest single data point.
|
||||
|
||||
## Curator Notes
|
||||
PRIMARY CONNECTION: delegating critical infrastructure development to AI creates civilizational fragility because humans lose the ability to understand maintain and fix the systems civilization depends on
|
||||
WHY ARCHIVED: Provides the MECHANISM for how civilizational fragility develops — not just through infrastructure delegation but through measurable skill atrophy that makes humans unable to resume control. The feedback loop structure means the process is self-accelerating.
|
||||
|
|
@ -0,0 +1,57 @@
|
|||
---
|
||||
type: source
|
||||
title: "Seven Feedback Loops: Mapping AI's Systemic Economic Disruption Risks"
|
||||
author: "Apply AI Alliance (EU Futurium)"
|
||||
url: https://futurium.ec.europa.eu/en/european-ai-alliance/community-content/seven-feedback-loops-mapping-ais-systemic-economic-disruption-risks
|
||||
date: 2026-01-15
|
||||
domain: ai-alignment
|
||||
secondary_domains: [internet-finance, grand-strategy]
|
||||
format: essay
|
||||
status: unprocessed
|
||||
priority: high
|
||||
triage_tag: claim
|
||||
tags: [feedback-loops, economic-disruption, demand-destruction, automation-overshoot, coordination-failure, market-failure, systemic-risk]
|
||||
flagged_for_rio: ["Seven self-reinforcing economic feedback loops from AI automation — connects to market failure analysis and coordination mechanisms"]
|
||||
flagged_for_leo: ["Systemic coordination failure framework — individual firm optimization creating collective demand destruction"]
|
||||
---
|
||||
|
||||
## Content
|
||||
|
||||
Seven self-reinforcing feedback loops identified in AI's economic impact:
|
||||
|
||||
**L1: Competitive AI Adoption Cycle** — Corporate adoption → job displacement → reduced consumer income → demand destruction → revenue decline → emergency cost-cutting → MORE AI adoption. The "follow or die" dynamic.
|
||||
|
||||
**L2: Financial System Cascade** — Demand destruction → business failures → loan defaults → bank liquidity crises → credit freezes → additional failures. AI-enabled systems could coordinate crashes in minutes.
|
||||
|
||||
**L3: Institutional Erosion Loop** — Mass unemployment → social unrest → eroded institutional trust → delayed policy → worsening conditions.
|
||||
|
||||
**L4: Global Dependency Loop** — Nations without AI capabilities become dependent on foreign providers → foreign exchange drain → weakened financial systems.
|
||||
|
||||
**L5: Education Misalignment Loop** — Outdated curricula → unprepared graduates → funding cuts → worse misalignment. 77% of new AI jobs require master's degrees.
|
||||
|
||||
**L6: Cognitive-Stratification Loop** — AI infrastructure concentration → inequality between AI controllers and displaced workers → political instability.
|
||||
|
||||
**L7: Time-Compression Crisis** — Meta-loop: exponentially advancing AI outpaces sub-linear institutional adaptation, accelerating ALL other loops.
|
||||
|
||||
**Key economic data:**
|
||||
- Only 3-7% of AI productivity improvements translate to higher worker earnings
|
||||
- 40% of employers plan workforce reductions
|
||||
- 92% of C-suite executives report up to 20% workforce overcapacity
|
||||
- 78% of organizations now use AI (creates "inevitability" pressure on laggards)
|
||||
- J-curve: initial 60-percentage-point productivity declines during 12-24 month adjustment periods
|
||||
|
||||
**Market failure mechanisms:**
|
||||
1. Negative externalities: firm optimization creates collective demand destruction that firms don't internalize
|
||||
2. Coordination failure: "Follow or die" competitive dynamics force adoption regardless of aggregate consequences
|
||||
3. Information asymmetry: adoption signals inevitability, pressuring laggards into adoption despite systemic risks
|
||||
|
||||
## Agent Notes
|
||||
**Triage:** [CLAIM] — "Economic forces systematically push AI adoption past the socially optimal level through seven self-reinforcing feedback loops where individual firm rationality produces collective irrationality" — the coordination failure framing maps directly to our core thesis
|
||||
**Why this matters:** This is the MECHANISM for automation overshoot. Each loop individually would be concerning; together they create a systemic dynamic that makes over-adoption structurally inevitable absent coordination. L1 (competitive adoption cycle) is the most alignment-relevant: the same "follow or die" dynamic that drives the alignment tax drives economic overshoot.
|
||||
**What surprised me:** L7 (time-compression crisis) as META-LOOP. The insight that exponential technology + linear governance = all other loops accelerating simultaneously. This is our existing claim about technology advancing exponentially while coordination evolves linearly, applied to the economic domain.
|
||||
**KB connections:** [[the alignment tax creates a structural race to the bottom]], [[technology advances exponentially but coordination mechanisms evolve linearly creating a widening gap]], [[AI alignment is a coordination problem not a technical problem]], [[economic forces push humans out of every cognitive loop where output quality is independently verifiable]]
|
||||
**Extraction hints:** L1 and L7 are the most claim-worthy. L1 provides the specific mechanism for overshoot. L7 connects to our existing temporal mismatch claim. The market failure taxonomy (externalities, coordination failure, information asymmetry) maps to standard economics and could be a stand-alone claim.
|
||||
|
||||
## Curator Notes
|
||||
PRIMARY CONNECTION: the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it
|
||||
WHY ARCHIVED: Provides seven specific feedback loops explaining HOW the race-to-the-bottom dynamic operates economically. L1 is the alignment tax applied to automation decisions. L7 is our temporal mismatch claim applied to governance response.
|
||||
|
|
@ -0,0 +1,48 @@
|
|||
---
|
||||
type: source
|
||||
title: "The productivity paradox of AI adoption in manufacturing firms"
|
||||
author: "MIT Sloan researchers (via Census Bureau data)"
|
||||
url: https://mitsloan.mit.edu/ideas-made-to-matter/productivity-paradox-ai-adoption-manufacturing-firms
|
||||
date: 2026-02-01
|
||||
domain: ai-alignment
|
||||
secondary_domains: [internet-finance]
|
||||
format: paper
|
||||
status: unprocessed
|
||||
priority: medium
|
||||
triage_tag: evidence
|
||||
tags: [j-curve, productivity-paradox, manufacturing, ai-adoption, adjustment-period, complementary-investment]
|
||||
flagged_for_rio: ["J-curve in manufacturing AI adoption — 1.33pp productivity decline initially, recovery after 4 years. Only digitally mature firms see strong gains."]
|
||||
---
|
||||
|
||||
## Content
|
||||
|
||||
MIT Sloan researchers analyzing tens of thousands of U.S. manufacturing firms. Published 2026.
|
||||
|
||||
**J-curve finding:**
|
||||
- AI adoption initially reduces productivity by average 1.33 percentage points (raw analysis)
|
||||
- Adjusted for selection bias: negative impact up to approximately 60 percentage points
|
||||
- Over 4-year period: AI-adopting firms outperformed non-adopters in both productivity and market share
|
||||
- Earlier adopters (pre-2017) exhibit stronger growth over time, conditional on survival
|
||||
|
||||
**Mechanisms behind the dip:**
|
||||
1. Misalignment between new digital tools and legacy operational processes
|
||||
2. Required complementary investments in data infrastructure, training, workflow redesign
|
||||
3. Older firms abandoned vital production management practices (KPI monitoring) — accounts for ~1/3 of their losses
|
||||
|
||||
**Digital maturity requirement:** Firms seeing strongest gains were already digitally mature before AI adoption. Without pre-existing digital infrastructure, the J-curve dip deepens and recovery is uncertain.
|
||||
|
||||
**Brynjolfsson counter-data (Fortune, Feb 2026):**
|
||||
- U.S. productivity jumped ~2.7% in 2025, nearly doubling the 1.4% annual average
|
||||
- Claims "transitioning from investment phase to harvest phase"
|
||||
- BUT Apollo Chief Economist Slok counters: "AI is everywhere except in the incoming macroeconomic data"
|
||||
|
||||
## Agent Notes
|
||||
**Triage:** [EVIDENCE] — supports and complicates the automation overshoot thesis. The J-curve is NOT overshoot per se — it's expected adjustment cost. But the question is whether competitive pressure forces firms to adopt before complementary investments are ready, which DOES constitute overshoot.
|
||||
**Why this matters:** The J-curve provides the economic framework for why firms might rationally adopt AI too fast — competitive pressure (L1 from the seven feedback loops) forces adoption before complementary investments are in place, deepening and extending the J-curve dip. Firms that abandon management practices during adoption (1/3 of losses) are the overshoot mechanism.
|
||||
**What surprised me:** The "abandoned vital production management practices" finding. Firms didn't just add AI — they REMOVED human management practices in the process. This maps directly to deskilling: the organizational equivalent of individual skill atrophy.
|
||||
**KB connections:** [[the alignment tax creates a structural race to the bottom]], [[technology advances exponentially but coordination mechanisms evolve linearly creating a widening gap]]
|
||||
**Extraction hints:** Not a standalone claim — better as evidence enriching existing claims about competitive pressure dynamics.
|
||||
|
||||
## Curator Notes
|
||||
PRIMARY CONNECTION: the alignment tax creates a structural race to the bottom because safety training costs capability and rational competitors skip it
|
||||
WHY ARCHIVED: Provides manufacturing-sector evidence for competitive pressure driving premature adoption. The "abandoned management practices" finding parallels organizational deskilling.
|
||||
|
|
@ -0,0 +1,65 @@
|
|||
---
|
||||
type: source
|
||||
title: "How AI Can Degrade Human Performance in High-Stakes Settings"
|
||||
author: "AI Frontiers"
|
||||
url: https://ai-frontiers.org/articles/how-ai-can-degrade-human-performance-in-high-stakes-settings
|
||||
date: 2026-03-01
|
||||
domain: ai-alignment
|
||||
secondary_domains: [health]
|
||||
format: essay
|
||||
status: unprocessed
|
||||
priority: high
|
||||
triage_tag: claim
|
||||
tags: [human-ai-performance, high-stakes, degradation, nursing, aviation, nuclear, joint-activity-testing]
|
||||
flagged_for_vida: ["450 nursing students/nurses tested with AI in ICU cases — performance degrades 96-120% when AI predictions mislead"]
|
||||
---
|
||||
|
||||
## Content
|
||||
|
||||
Cross-domain analysis of how AI degrades human performance in critical settings:
|
||||
|
||||
**Healthcare (nursing study):**
|
||||
- 450 nursing students and licensed nurses reviewing ICU cases
|
||||
- Four AI configurations from no assistance to full predictions + annotations
|
||||
- Best case: 53-67% BETTER when AI predictions accurate
|
||||
- Worst case: 96-120% WORSE when AI predictions misleading
|
||||
- "Nurses did not reliably recognize when AI predictions were right or wrong"
|
||||
- AI appeared to change HOW nurses think when assessing patients, not just what they decide
|
||||
|
||||
**Aviation:**
|
||||
- AI weather monitoring missed microbursts during landing
|
||||
- Crews faced doubled workload with halved preparation time
|
||||
- Required emergency maneuvers
|
||||
|
||||
**Nuclear energy:**
|
||||
- AI warning systems hid underlying problems through filtering
|
||||
- Misclassified gradual coolant pressure drops as benign
|
||||
- Led to cascading subsystem failures
|
||||
|
||||
**Asymmetric risk profile:**
|
||||
- Gains from accurate AI: 53-67%
|
||||
- Losses from inaccurate AI: 96-120%
|
||||
- "Averaging results can hide rare but severe errors, creating blind spots with potentially catastrophic consequences"
|
||||
|
||||
**Conditions worsening degradation:**
|
||||
1. AI errors are subtle and plausible (not obviously wrong)
|
||||
2. Humans cannot verify predictions (complexity/information asymmetry)
|
||||
3. AI aggregates/filters information, hiding important signals
|
||||
4. Staffing reduced based on false confidence in AI
|
||||
5. Rare but critical failures that testing didn't anticipate
|
||||
|
||||
**Proposed mitigation — Joint Activity Testing (JAT):**
|
||||
1. Test humans AND AI together, not separately
|
||||
2. Evaluate diverse AI performance scenarios (excel, struggle, fail)
|
||||
3. Enable human error recovery over patching
|
||||
|
||||
## Agent Notes
|
||||
**Triage:** [CLAIM] — "AI degrades human decision-making performance asymmetrically — gains from accurate AI (53-67%) are smaller than losses from inaccurate AI (96-120%) — creating a structural risk where average performance masks catastrophic tail outcomes" — multi-domain evidence
|
||||
**Why this matters:** The ASYMMETRY is the critical finding. Even if AI is right 90% of the time, the 10% where it's wrong produces losses nearly double the gains from the 90%. This is why averaging performance hides the real risk. For alignment: human oversight of AI is not just "sometimes unhelpful" — it's structurally asymmetric, with large downside when oversight fails and modest upside when it succeeds.
|
||||
**What surprised me:** The COGNITIVE CHANGE mechanism. AI doesn't just provide wrong answers — it changes how humans THINK about problems. This is deeper than automation bias. It's cognitive restructuring. Once you've internalized AI-mediated reasoning, you can't just "turn it off" when AI fails.
|
||||
**KB connections:** [[human-in-the-loop clinical AI degrades to worse-than-AI-alone]], [[AI capability and reliability are independent dimensions]], [[scalable oversight degrades rapidly as capability gaps grow]]
|
||||
**Extraction hints:** Three distinct claims: (1) asymmetric risk profile, (2) cognitive restructuring mechanism, (3) JAT as evaluation framework. The asymmetry finding is most novel.
|
||||
|
||||
## Curator Notes
|
||||
PRIMARY CONNECTION: human-in-the-loop clinical AI degrades to worse-than-AI-alone because physicians both de-skill from reliance and introduce errors when overriding correct outputs
|
||||
WHY ARCHIVED: Extends our existing clinical AI degradation claim with cross-domain evidence (nursing, aviation, nuclear) and quantifies the asymmetric risk profile. The cognitive restructuring mechanism is a novel finding.
|
||||
|
|
@ -0,0 +1,49 @@
|
|||
---
|
||||
type: source
|
||||
title: "The homogenizing effect of large language models on human expression and thought"
|
||||
author: "Zhivar Sourati, Morteza Dehghani et al. (@USC Dornsife)"
|
||||
url: https://www.cell.com/trends/cognitive-sciences/fulltext/S1364-6613(26)00003-3
|
||||
date: 2026-03-11
|
||||
domain: ai-alignment
|
||||
secondary_domains: [collective-intelligence, cultural-dynamics]
|
||||
format: paper
|
||||
status: unprocessed
|
||||
priority: high
|
||||
triage_tag: claim
|
||||
tags: [ai-homogenization, cognitive-diversity, collective-intelligence, llm-effects, expression, thought]
|
||||
flagged_for_clay: ["AI homogenization of expression connects to cultural dynamics — homogenized expression may reduce narrative diversity"]
|
||||
---
|
||||
|
||||
## Content
|
||||
|
||||
Published in Trends in Cognitive Sciences, March 2026. Opinion paper by USC computer scientists and psychologists.
|
||||
|
||||
**Core thesis:** AI chatbots are standardizing how people speak, write, and think. If unchecked, this homogenization reduces humanity's collective wisdom and adaptive capacity.
|
||||
|
||||
**Key findings cited:**
|
||||
- LLM outputs show less variation than human writing
|
||||
- Outputs reflect primarily Western, educated, industrialized perspectives
|
||||
- Groups using LLMs generate FEWER and LESS CREATIVE ideas than those relying solely on collective thinking
|
||||
- People's opinions SHIFT toward biased LLMs after interaction
|
||||
- Distinct linguistic styles and reasoning strategies become homogenized, producing standardized expressions across users
|
||||
|
||||
**Homogenization mechanism (4 pathways):**
|
||||
1. Users lose stylistic individuality when polishing text through chatbots
|
||||
2. LLMs redefine what constitutes "credible speech" and "good reasoning"
|
||||
3. Widespread adoption creates social pressure to conform ("If a lot of people around me are thinking and speaking in a certain way... I would feel pressure to align")
|
||||
4. Training data feedback loops amplify homogenization over time
|
||||
|
||||
**Impact on collective intelligence:** "Within groups and societies, cognitive diversity bolsters creativity and problem-solving. If LLMs had more diverse ways of approaching ideas and problems, they would better support the collective intelligence and problem-solving capabilities of our societies."
|
||||
|
||||
**Recommendation:** AI developers should incorporate more real-world diversity into LLM training sets — grounded in actual global human diversity, not random variation.
|
||||
|
||||
## Agent Notes
|
||||
**Triage:** [CLAIM] — "AI homogenization of human expression and thought reduces collective intelligence by eroding the cognitive diversity that problem-solving depends on" — from a leading cognitive science journal, 2026
|
||||
**Why this matters:** Directly connects to our existing claim [[AI is collapsing the knowledge-producing communities it depends on]] but from a DIFFERENT MECHANISM. That claim is about economic displacement of knowledge workers. This is about cognitive homogenization EVEN AMONG people still producing knowledge. Same structural pattern (AI undermines its own inputs), different pathway.
|
||||
**What surprised me:** The SOCIAL PRESSURE mechanism. Homogenization isn't just a technical artifact of LLM training — it's socially enforced. People conform to AI-standard expression because others do. This makes it harder to reverse than a purely technical problem.
|
||||
**KB connections:** [[AI is collapsing the knowledge-producing communities it depends on]], [[collective intelligence requires diversity as a structural precondition not a moral preference]], [[pluralistic alignment must accommodate irreducibly diverse values simultaneously]]
|
||||
**Extraction hints:** The 4-pathway mechanism and the social pressure finding are the novel contributions. The self-reinforcing nature (AI homogenizes → homogenized data trains next AI → further homogenization) is a feedback loop claim.
|
||||
|
||||
## Curator Notes
|
||||
PRIMARY CONNECTION: AI is collapsing the knowledge-producing communities it depends on creating a self-undermining loop that collective intelligence can break
|
||||
WHY ARCHIVED: Provides a SECOND mechanism for the self-undermining loop — not just economic displacement but cognitive homogenization. Published in a top-tier cognitive science journal in March 2026.
|
||||
Loading…
Reference in a new issue