From 7117c61f699f4ea5e16e80ef74dd00061b3dc519 Mon Sep 17 00:00:00 2001 From: Teleo Agents Date: Thu, 19 Mar 2026 00:31:16 +0000 Subject: [PATCH 1/2] extract: 2025-02-00-beers-toner-pet-ai-external-scrutiny Pentagon-Agent: Epimetheus <968B2991-E2DF-4006-B962-F5B0A0CC8ACA> --- ...-beers-toner-pet-ai-external-scrutiny.json | 27 +++++++++++++++++++ ...00-beers-toner-pet-ai-external-scrutiny.md | 14 +++++++++- 2 files changed, 40 insertions(+), 1 deletion(-) create mode 100644 inbox/queue/.extraction-debug/2025-02-00-beers-toner-pet-ai-external-scrutiny.json diff --git a/inbox/queue/.extraction-debug/2025-02-00-beers-toner-pet-ai-external-scrutiny.json b/inbox/queue/.extraction-debug/2025-02-00-beers-toner-pet-ai-external-scrutiny.json new file mode 100644 index 00000000..3d7adcb9 --- /dev/null +++ b/inbox/queue/.extraction-debug/2025-02-00-beers-toner-pet-ai-external-scrutiny.json @@ -0,0 +1,27 @@ +{ + "rejected_claims": [ + { + "filename": "privacy-enhancing-technologies-enable-independent-ai-scrutiny-without-ip-compromise-but-legal-authority-to-require-scrutiny-does-not-exist.md", + "issues": [ + "missing_attribution_extractor" + ] + } + ], + "validation_stats": { + "total": 1, + "kept": 0, + "fixed": 4, + "rejected": 1, + "fixes_applied": [ + "privacy-enhancing-technologies-enable-independent-ai-scrutiny-without-ip-compromise-but-legal-authority-to-require-scrutiny-does-not-exist.md:set_created:2026-03-19", + "privacy-enhancing-technologies-enable-independent-ai-scrutiny-without-ip-compromise-but-legal-authority-to-require-scrutiny-does-not-exist.md:stripped_wiki_link:voluntary-safety-pledges-cannot-survive-competitive-pressure", + "privacy-enhancing-technologies-enable-independent-ai-scrutiny-without-ip-compromise-but-legal-authority-to-require-scrutiny-does-not-exist.md:stripped_wiki_link:only-binding-regulation-with-enforcement-teeth-changes-front", + "privacy-enhancing-technologies-enable-independent-ai-scrutiny-without-ip-compromise-but-legal-authority-to-require-scrutiny-does-not-exist.md:stripped_wiki_link:safe-AI-development-requires-building-alignment-mechanisms-b" + ], + "rejections": [ + "privacy-enhancing-technologies-enable-independent-ai-scrutiny-without-ip-compromise-but-legal-authority-to-require-scrutiny-does-not-exist.md:missing_attribution_extractor" + ] + }, + "model": "anthropic/claude-sonnet-4.5", + "date": "2026-03-19" +} \ No newline at end of file diff --git a/inbox/queue/2025-02-00-beers-toner-pet-ai-external-scrutiny.md b/inbox/queue/2025-02-00-beers-toner-pet-ai-external-scrutiny.md index f7de0728..30c1b084 100644 --- a/inbox/queue/2025-02-00-beers-toner-pet-ai-external-scrutiny.md +++ b/inbox/queue/2025-02-00-beers-toner-pet-ai-external-scrutiny.md @@ -7,9 +7,13 @@ date: 2025-02-01 domain: ai-alignment secondary_domains: [] format: paper -status: unprocessed +status: null-result priority: high tags: [evaluation-infrastructure, privacy-enhancing-technologies, OpenMined, external-scrutiny, Christchurch-Call, AISI, deployed] +processed_by: theseus +processed_date: 2026-03-19 +extraction_model: "anthropic/claude-sonnet-4.5" +extraction_notes: "LLM returned 1 claims, 1 rejected by validator" --- ## Content @@ -53,3 +57,11 @@ PRIMARY CONNECTION: [[safe AI development requires building alignment mechanisms WHY ARCHIVED: Provides evidence that the technical barrier to independent AI evaluation is solvable. The key insight — technology ready, legal framework missing — precisely locates the bottleneck in evaluation infrastructure development. EXTRACTION HINT: Focus on the technology-law gap: PET infrastructure works (two deployments), but legal authority to require frontier AI labs to submit to independent evaluation doesn't exist. This is the specific intervention point. + + +## Key Facts +- Helen Toner was Director of Strategy at CISA +- Helen Toner is at Georgetown +- The Christchurch Call is a voluntary initiative +- UK AI Safety Institute has conducted frontier model evaluations using PET infrastructure +- The paper was published February 2025 -- 2.45.2 From 3282592dc01c49ff5ff0490180c2724611faa2b6 Mon Sep 17 00:00:00 2001 From: Teleo Agents Date: Thu, 19 Mar 2026 00:33:12 +0000 Subject: [PATCH 2/2] extract: 2026-01-00-kim-third-party-ai-assurance-framework Pentagon-Agent: Epimetheus <968B2991-E2DF-4006-B962-F5B0A0CC8ACA> --- ... converging on problems that require it.md | 6 ++++ ...im-third-party-ai-assurance-framework.json | 32 +++++++++++++++++++ ...-kim-third-party-ai-assurance-framework.md | 13 +++++++- 3 files changed, 50 insertions(+), 1 deletion(-) create mode 100644 inbox/queue/.extraction-debug/2026-01-00-kim-third-party-ai-assurance-framework.json diff --git a/domains/ai-alignment/no research group is building alignment through collective intelligence infrastructure despite the field converging on problems that require it.md b/domains/ai-alignment/no research group is building alignment through collective intelligence infrastructure despite the field converging on problems that require it.md index 21a48939..64547a0c 100644 --- a/domains/ai-alignment/no research group is building alignment through collective intelligence infrastructure despite the field converging on problems that require it.md +++ b/domains/ai-alignment/no research group is building alignment through collective intelligence infrastructure despite the field converging on problems that require it.md @@ -23,6 +23,12 @@ The alignment field has converged on a problem they cannot solve with their curr The UK AI for Collective Intelligence Research Network represents a national-scale institutional commitment to building CI infrastructure with explicit alignment goals. Funded by UKRI/EPSRC, the network proposes the 'AI4CI Loop' (Gathering Intelligence → Informing Behaviour) as a framework for multi-level decision making. The research strategy includes seven trust properties (human agency, security, privacy, transparency, fairness, value alignment, accountability) and specifies technical requirements including federated learning architectures, secure data repositories, and foundation models adapted for collective intelligence contexts. This is not purely academic—it's a government-backed infrastructure program with institutional resources. However, the strategy is prospective (published 2024-11) and describes a research agenda rather than deployed systems, so it represents institutional intent rather than operational infrastructure. + +### Additional Evidence (challenge) +*Source: [[2026-01-00-kim-third-party-ai-assurance-framework]] | Added: 2026-03-19* + +CMU researchers have built and validated a third-party AI assurance framework with four operational components (Responsibility Assignment Matrix, Interview Protocol, Maturity Matrix, Assurance Report Template), tested on two real deployment cases. This represents concrete infrastructure-building work, though at small scale and not yet applicable to frontier AI. + --- Relevant Notes: diff --git a/inbox/queue/.extraction-debug/2026-01-00-kim-third-party-ai-assurance-framework.json b/inbox/queue/.extraction-debug/2026-01-00-kim-third-party-ai-assurance-framework.json new file mode 100644 index 00000000..4cde4306 --- /dev/null +++ b/inbox/queue/.extraction-debug/2026-01-00-kim-third-party-ai-assurance-framework.json @@ -0,0 +1,32 @@ +{ + "rejected_claims": [ + { + "filename": "third-party-ai-assurance-methodology-is-at-proof-of-concept-stage-validated-in-small-deployment-contexts-but-not-yet-applicable-to-frontier-ai-at-scale.md", + "issues": [ + "missing_attribution_extractor" + ] + }, + { + "filename": "ai-assurance-explicitly-distinguishes-itself-from-audit-to-prevent-conflict-of-interest-and-ensure-credibility-which-acknowledges-current-evaluation-has-a-structural-independence-problem.md", + "issues": [ + "missing_attribution_extractor" + ] + } + ], + "validation_stats": { + "total": 2, + "kept": 0, + "fixed": 2, + "rejected": 2, + "fixes_applied": [ + "third-party-ai-assurance-methodology-is-at-proof-of-concept-stage-validated-in-small-deployment-contexts-but-not-yet-applicable-to-frontier-ai-at-scale.md:set_created:2026-03-19", + "ai-assurance-explicitly-distinguishes-itself-from-audit-to-prevent-conflict-of-interest-and-ensure-credibility-which-acknowledges-current-evaluation-has-a-structural-independence-problem.md:set_created:2026-03-19" + ], + "rejections": [ + "third-party-ai-assurance-methodology-is-at-proof-of-concept-stage-validated-in-small-deployment-contexts-but-not-yet-applicable-to-frontier-ai-at-scale.md:missing_attribution_extractor", + "ai-assurance-explicitly-distinguishes-itself-from-audit-to-prevent-conflict-of-interest-and-ensure-credibility-which-acknowledges-current-evaluation-has-a-structural-independence-problem.md:missing_attribution_extractor" + ] + }, + "model": "anthropic/claude-sonnet-4.5", + "date": "2026-03-19" +} \ No newline at end of file diff --git a/inbox/queue/2026-01-00-kim-third-party-ai-assurance-framework.md b/inbox/queue/2026-01-00-kim-third-party-ai-assurance-framework.md index b473ed14..2875de70 100644 --- a/inbox/queue/2026-01-00-kim-third-party-ai-assurance-framework.md +++ b/inbox/queue/2026-01-00-kim-third-party-ai-assurance-framework.md @@ -7,9 +7,13 @@ date: 2026-01-30 domain: ai-alignment secondary_domains: [] format: paper -status: unprocessed +status: enrichment priority: high tags: [evaluation-infrastructure, third-party-assurance, conflict-of-interest, lifecycle-assessment, CMU] +processed_by: theseus +processed_date: 2026-03-19 +enrichments_applied: ["no research group is building alignment through collective intelligence infrastructure despite the field converging on problems that require it.md"] +extraction_model: "anthropic/claude-sonnet-4.5" --- ## Content @@ -51,3 +55,10 @@ PRIMARY CONNECTION: [[no research group is building alignment through collective WHY ARCHIVED: Provides methodology for third-party AI assurance that explicitly addresses the conflict of interest problem. Important evidence that the field is aware of the independence gap. EXTRACTION HINT: The "assurance vs audit" distinction to prevent conflict of interest is the key extractable insight. The lifecycle approach (process + outcomes) is also worth noting. + + +## Key Facts +- CMU researchers published 'Toward Third-Party Assurance of AI Systems' in January 2026 +- The framework was tested on a business document tagging tool and a housing resource allocation tool +- The paper identifies that few existing evaluation resources 'address both the process of designing, developing, and deploying an AI system and the outcomes it produces' +- Few existing approaches are 'end-to-end and operational, give actionable guidance, or present evidence of usability' according to the gap analysis -- 2.45.2