teleo-codex/inbox/queue/.extraction-debug/2026-03-26-anthropic-activating-asl3-protections.json
Teleo Agents 108a0d631c extract: 2026-03-26-anthropic-activating-asl3-protections
Pentagon-Agent: Epimetheus <3D35839A-7722-4740-B93D-51157F7D5E70>
2026-03-26 03:19:26 +00:00

38 lines
No EOL
2.6 KiB
JSON

{
"rejected_claims": [
{
"filename": "precautionary-ai-governance-triggers-protection-escalation-when-capability-evaluation-becomes-unreliable-near-thresholds.md",
"issues": [
"missing_attribution_extractor"
]
},
{
"filename": "ai-safety-governance-lacks-independent-verification-creating-self-referential-accountability-where-labs-assess-their-own-compliance.md",
"issues": [
"missing_attribution_extractor"
]
}
],
"validation_stats": {
"total": 2,
"kept": 0,
"fixed": 8,
"rejected": 2,
"fixes_applied": [
"precautionary-ai-governance-triggers-protection-escalation-when-capability-evaluation-becomes-unreliable-near-thresholds.md:set_created:2026-03-26",
"precautionary-ai-governance-triggers-protection-escalation-when-capability-evaluation-becomes-unreliable-near-thresholds.md:stripped_wiki_link:pre-deployment-AI-evaluations-do-not-predict-real-world-risk",
"precautionary-ai-governance-triggers-protection-escalation-when-capability-evaluation-becomes-unreliable-near-thresholds.md:stripped_wiki_link:voluntary safety pledges cannot survive competitive pressure",
"precautionary-ai-governance-triggers-protection-escalation-when-capability-evaluation-becomes-unreliable-near-thresholds.md:stripped_wiki_link:safe AI development requires building alignment mechanisms b",
"ai-safety-governance-lacks-independent-verification-creating-self-referential-accountability-where-labs-assess-their-own-compliance.md:set_created:2026-03-26",
"ai-safety-governance-lacks-independent-verification-creating-self-referential-accountability-where-labs-assess-their-own-compliance.md:stripped_wiki_link:AI transparency is declining not improving because Stanford ",
"ai-safety-governance-lacks-independent-verification-creating-self-referential-accountability-where-labs-assess-their-own-compliance.md:stripped_wiki_link:only binding regulation with enforcement teeth changes front",
"ai-safety-governance-lacks-independent-verification-creating-self-referential-accountability-where-labs-assess-their-own-compliance.md:stripped_wiki_link:voluntary safety pledges cannot survive competitive pressure"
],
"rejections": [
"precautionary-ai-governance-triggers-protection-escalation-when-capability-evaluation-becomes-unreliable-near-thresholds.md:missing_attribution_extractor",
"ai-safety-governance-lacks-independent-verification-creating-self-referential-accountability-where-labs-assess-their-own-compliance.md:missing_attribution_extractor"
]
},
"model": "anthropic/claude-sonnet-4.5",
"date": "2026-03-26"
}