teleo-codex/skills/extract.md
m3taversal e830fe4c5f Initial commit: Teleo Codex v1
Three-agent knowledge base (Leo, Rio, Clay) with:
- 177 claim files across core/ and foundations/
- 38 domain claims in internet-finance/
- 22 domain claims in entertainment/
- Agent soul documents (identity, beliefs, reasoning, skills)
- 14 positions across 3 agents
- Claim/belief/position schemas
- 6 shared skills
- Agent-facing CLAUDE.md operating manual

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-03-05 20:30:34 +00:00

2.5 KiB

Skill: Extract

Turn raw content into structured evidence and proposed claims.

When to Use

When new content arrives in inbox/ — articles, tweets, papers, transcripts, research files.

Input

Raw content (text, URL, document).

Process

Step 1: Read the source material completely

Don't skim. Read the full content before extracting anything. Understand the author's argument, not just individual data points.

Step 2: Separate evidence from interpretation

Evidence is factual: data, statistics, quotes, study results, events, observations. Things that are verifiable regardless of your interpretive framework.

Claims are interpretive: assertions about what the evidence means, causal relationships, predictions, evaluations. Things someone could disagree with.

Most sources mix these freely. Your job is to separate them.

Step 3: Extract evidence

For each piece of evidence:

  • Is it sourced and verifiable?
  • Is it relevant to at least one Teleo domain?
  • Does it already exist in the knowledge base? (check for duplicates)

Include evidence inline in the claim body — cite sources, data, studies directly in the prose.

Step 4: Extract candidate claims

For each potential claim:

  • Is it specific enough to disagree with? ("AI is changing healthcare" → NO. "AI diagnostic triage achieves 97% sensitivity across 14 conditions" → YES)
  • Does it cite evidence from this source or the knowledge base?
  • Does it duplicate an existing claim? (semantic check — different words, same idea)
  • Title passes the prose-as-claim test: "This note argues that [title]" works as a sentence

Create candidate claim files for evaluation.

Step 5: Classify by domain

Tag each evidence and claim with primary domain:

  • internet-finance, entertainment, grand-strategy

Cross-domain items get a primary domain + secondary_domains list.

Step 6: Identify enrichments

Does this source contain information that would improve existing notes?

  • New data for existing claims
  • Counter-evidence to existing claims
  • New connections between existing claims

Flag enrichments for the evaluation cycle.

Output

  • Claim files in domains/{domain}/ with evidence inline
  • Candidate claim files for PR review
  • Enrichment flags for existing notes
  • Extraction summary: N evidence extracted, N claims proposed, N enrichments flagged

Quality Gate

  • Every claim cites verifiable evidence inline
  • Every claim is specific enough to disagree with
  • No duplicates of existing knowledge base content
  • Domain classification is accurate
  • Titles work as prose propositions