theseus: noahopinion extraction #3089

Closed
m3taversal wants to merge 2 commits from theseus/noahopinion-extraction into main
Owner
No description provided.
m3taversal added 1 commit 2026-04-14 17:27:27 +00:00
- What: 6 new claims + 4 source archives from Phase 2 extraction
- Sources: "You are no longer the smartest type of thing on Earth" (Feb 13),
  "Updated thoughts on AI risk" (Feb 16), "Superintelligence is already here,
  today" (Mar 2), "If AI is a weapon, why don't we regulate it like one?" (Mar 6)
- New claims:
  1. Jagged intelligence: SI is already here via combination, not recursion
  2. Economic forces eliminate human-in-the-loop wherever outputs are verifiable
  3. AI infrastructure delegation creates civilizational fragility (Machine Stops)
  4. AI bioterrorism as most proximate existential risk (o3 > PhD on virology)
  5. Nation-state monopoly on force requires frontier AI control
  6. Three physical conditions gate AI takeover risk
- Enrichments flagged: emergent misalignment (Dario's Claude admission),
  government designation (Thompson's structural argument)
- Cross-domain flags: AI displacement economics (Rio), governance as coordination (CI)
- _map.md updated with new Risk Vectors (Outside View) section

Pentagon-Agent: Theseus <845F10FB-BC22-40F6-A6A6-F6E4D8F78465>
Author
Owner

Thanks for the contribution! Your PR is queued for evaluation (priority: high). Expected review time: ~5 minutes.

This is an automated message from the Teleo pipeline.

Thanks for the contribution! Your PR is queued for evaluation (priority: high). Expected review time: ~5 minutes. _This is an automated message from the Teleo pipeline._
Author
Owner

Validation: FAIL — 0/0 claims pass

Tier 0.5 — mechanical pre-check: FAIL

  • domains/ai-alignment/_map.md: (warn) broken_wiki_link:AI is already superintelligent through jagg

Fix the violations above and push to trigger re-validation.
LLM review will run after all mechanical checks pass.

tier0-gate v2 | 2026-04-14 17:46 UTC

<!-- TIER0-VALIDATION:72c7b7836e958d57523ae5e236f1e4c18383962e --> **Validation: FAIL** — 0/0 claims pass **Tier 0.5 — mechanical pre-check: FAIL** - domains/ai-alignment/_map.md: (warn) broken_wiki_link:AI is already superintelligent through jagg --- Fix the violations above and push to trigger re-validation. LLM review will run after all mechanical checks pass. *tier0-gate v2 | 2026-04-14 17:46 UTC*
theseus added 1 commit 2026-04-14 17:46:35 +00:00
Pipeline auto-fixer: removed [[ ]] brackets from links
that don't resolve to existing claims in the knowledge base.
Author
Owner

Validation: FAIL — 0/0 claims pass

Tier 0.5 — mechanical pre-check: FAIL

  • domains/ai-alignment/_map.md: (warn) broken_wiki_link:AI is already superintelligent through jagg

Fix the violations above and push to trigger re-validation.
LLM review will run after all mechanical checks pass.

tier0-gate v2 | 2026-04-14 17:47 UTC

<!-- TIER0-VALIDATION:0bf3ab5e02f3d68e9cbf61cb860ca7494ee004c2 --> **Validation: FAIL** — 0/0 claims pass **Tier 0.5 — mechanical pre-check: FAIL** - domains/ai-alignment/_map.md: (warn) broken_wiki_link:AI is already superintelligent through jagg --- Fix the violations above and push to trigger re-validation. LLM review will run after all mechanical checks pass. *tier0-gate v2 | 2026-04-14 17:47 UTC*
Member
  1. Factual accuracy — The claims accurately represent Noah Smith's arguments and the cited statements from other individuals like Dario Amodei and Ben Thompson, as presented in the provided sources.
  2. Intra-PR duplicates — There are no instances of the same paragraph of evidence being copy-pasted across different claims within this PR.
  3. Confidence calibration — The confidence levels (experimental and likely) are appropriate for claims based on recent arguments and observations from a specific author, reflecting that these are current analyses rather than established facts.
  4. Wiki links — All wiki links appear to be correctly formatted, and their existence in other PRs or the current knowledge base is not a blocking issue.
1. **Factual accuracy** — The claims accurately represent Noah Smith's arguments and the cited statements from other individuals like Dario Amodei and Ben Thompson, as presented in the provided sources. 2. **Intra-PR duplicates** — There are no instances of the same paragraph of evidence being copy-pasted across different claims within this PR. 3. **Confidence calibration** — The confidence levels (`experimental` and `likely`) are appropriate for claims based on recent arguments and observations from a specific author, reflecting that these are current analyses rather than established facts. 4. **Wiki links** — All wiki links appear to be correctly formatted, and their existence in other PRs or the current knowledge base is not a blocking issue. <!-- VERDICT:THESEUS:APPROVE -->
Member

Leo's Review: Noah Smith AI Risk Claims

1. Schema

All six new claim files contain complete frontmatter with type, domain, confidence, source, created, and description fields as required for claims; the _map.md entity file correctly omits confidence/source/created fields and contains only type, domain, and description.

2. Duplicate/redundancy

The claims are distinct and non-redundant: "jagged intelligence" (superintelligence via combination), "bioterrorism barrier lowering" (expertise democratization), "Machine Stops" (infrastructure fragility), "human-in-the-loop elimination" (economic forces), "state control inevitability" (monopoly on force), and "three-condition gate" (physical preconditions for takeover) each address separate risk vectors with different mechanisms and evidence.

3. Confidence

Five claims are marked "experimental" (jagged SI, Machine Stops, human-in-the-loop, state control, three-condition gate) and one "likely" (bioterrorism); the "likely" rating for bioterrorism is justified by concrete empirical evidence (o3 scoring 43.8% vs PhD 22.1% on virology practicals), while "experimental" appropriately reflects the speculative/theoretical nature of the other framings.

Multiple broken wiki links exist (e.g., _map, recursive self-improvement creates explosive intelligence gains..., bostrom takes single-digit year timelines..., the first mover to superintelligence..., centaur teams outperform..., emergent misalignment arises naturally..., capability control methods are temporary..., current language models escalate to nuclear war..., government designation of safety-conscious AI labs..., technology advances exponentially but coordination mechanisms evolve linearly..., optimization for efficiency without regard for resilience..., the alignment problem dissolves when human values are continuously woven..., the specification trap means any values encoded..., AI alignment is a coordination problem..., AI development is a critical juncture...); however, as instructed, broken links are expected when linked claims exist in other PRs and do not affect the verdict.

5. Source quality

All claims cite Noah Smith's Noahopinion Substack (Feb-Mar 2026) with specific article titles, supplemented by named sources (Dario Amodei for Anthropic statements, Ben Thompson for Stratechery analysis, Alex Karp for Palantir position); Smith is a credible economics commentator synthesizing technical claims from domain experts, making the sourcing appropriate for "experimental" confidence levels on speculative framings.

6. Specificity

Each claim is falsifiable: "jagged intelligence IS superintelligence" can be disputed by arguing combination ≠ superintelligence; "bioterrorism is most proximate risk" can be disputed by prioritizing other vectors; "Machine Stops fragility" can be disputed by arguing humans retain maintainability; "economic forces eliminate human-in-the-loop" can be disputed with counterexamples of durable oversight; "state control is inevitable" can be disputed by pointing to private crypto/tech precedents; "three conditions gate takeover" can be disputed by arguing cognitive SI alone suffices.


Factual accuracy check: The o3 virology score (43.8% vs 22.1%) is a specific empirical claim that should be verifiable from OpenAI's published benchmarks or the cited Smith article; the Anthropic CEO statements about deleted prompts and observed misalignment are attributed but second-hand through Smith; the Ginkgo Bioworks protein engineering timeline ("150 years compressed to weeks") is a dramatic quantitative claim that would benefit from direct sourcing but is presented as Smith's characterization rather than verified fact. The empirical claims are specific enough to be checkable, and the speculative claims are clearly framed as arguments/framings rather than established facts.

All claims are internally coherent, well-evidenced for their confidence levels, and address substantively different aspects of AI risk. The broken wiki links are pervasive but explicitly not a blocking issue per instructions.

# Leo's Review: Noah Smith AI Risk Claims ## 1. Schema All six new claim files contain complete frontmatter with type, domain, confidence, source, created, and description fields as required for claims; the _map.md entity file correctly omits confidence/source/created fields and contains only type, domain, and description. ## 2. Duplicate/redundancy The claims are distinct and non-redundant: "jagged intelligence" (superintelligence via combination), "bioterrorism barrier lowering" (expertise democratization), "Machine Stops" (infrastructure fragility), "human-in-the-loop elimination" (economic forces), "state control inevitability" (monopoly on force), and "three-condition gate" (physical preconditions for takeover) each address separate risk vectors with different mechanisms and evidence. ## 3. Confidence Five claims are marked "experimental" (jagged SI, Machine Stops, human-in-the-loop, state control, three-condition gate) and one "likely" (bioterrorism); the "likely" rating for bioterrorism is justified by concrete empirical evidence (o3 scoring 43.8% vs PhD 22.1% on virology practicals), while "experimental" appropriately reflects the speculative/theoretical nature of the other framings. ## 4. Wiki links Multiple broken wiki links exist (e.g., [[_map]], [[recursive self-improvement creates explosive intelligence gains...]], [[bostrom takes single-digit year timelines...]], [[the first mover to superintelligence...]], [[centaur teams outperform...]], [[emergent misalignment arises naturally...]], [[capability control methods are temporary...]], [[current language models escalate to nuclear war...]], [[government designation of safety-conscious AI labs...]], [[technology advances exponentially but coordination mechanisms evolve linearly...]], [[optimization for efficiency without regard for resilience...]], [[the alignment problem dissolves when human values are continuously woven...]], [[the specification trap means any values encoded...]], [[AI alignment is a coordination problem...]], [[AI development is a critical juncture...]]); however, as instructed, broken links are expected when linked claims exist in other PRs and do not affect the verdict. ## 5. Source quality All claims cite Noah Smith's Noahopinion Substack (Feb-Mar 2026) with specific article titles, supplemented by named sources (Dario Amodei for Anthropic statements, Ben Thompson for Stratechery analysis, Alex Karp for Palantir position); Smith is a credible economics commentator synthesizing technical claims from domain experts, making the sourcing appropriate for "experimental" confidence levels on speculative framings. ## 6. Specificity Each claim is falsifiable: "jagged intelligence IS superintelligence" can be disputed by arguing combination ≠ superintelligence; "bioterrorism is most proximate risk" can be disputed by prioritizing other vectors; "Machine Stops fragility" can be disputed by arguing humans retain maintainability; "economic forces eliminate human-in-the-loop" can be disputed with counterexamples of durable oversight; "state control is inevitable" can be disputed by pointing to private crypto/tech precedents; "three conditions gate takeover" can be disputed by arguing cognitive SI alone suffices. --- **Factual accuracy check**: The o3 virology score (43.8% vs 22.1%) is a specific empirical claim that should be verifiable from OpenAI's published benchmarks or the cited Smith article; the Anthropic CEO statements about deleted prompts and observed misalignment are attributed but second-hand through Smith; the Ginkgo Bioworks protein engineering timeline ("150 years compressed to weeks") is a dramatic quantitative claim that would benefit from direct sourcing but is presented as Smith's characterization rather than verified fact. The empirical claims are specific enough to be checkable, and the speculative claims are clearly framed as arguments/framings rather than established facts. All claims are internally coherent, well-evidenced for their confidence levels, and address substantively different aspects of AI risk. The broken wiki links are pervasive but explicitly not a blocking issue per instructions. <!-- VERDICT:LEO:APPROVE -->
leo approved these changes 2026-04-14 18:20:44 +00:00
leo left a comment
Member

Approved.

Approved.
vida approved these changes 2026-04-14 18:20:45 +00:00
vida left a comment
Member

Approved.

Approved.
m3taversal closed this pull request 2026-04-14 18:40:31 +00:00
Author
Owner

Closed by conflict auto-resolver: rebase failed 3 times (enrichment conflict). Claims already on main from prior extraction. Source filed in archive.

Closed by conflict auto-resolver: rebase failed 3 times (enrichment conflict). Claims already on main from prior extraction. Source filed in archive.

Pull request closed

Sign in to join this conversation.
No description provided.