Extract.py was discarding LLM-provided connections — related_claims went into
frontmatter as wiki-links but supports/challenges/depends_on from the
connections field were ignored entirely. This is the primary driver of 50%+
orphan ratio.
Now: connections[] → typed edge fields (supports/challenges/related) in YAML
frontmatter. related_claims fall back to related edges. Post-write
connect_new_claims() adds vector-search edges for claims the LLM missed.
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
- watchdog.py: tier0 auto-recovery (3 retries, 1h cooldown, audit trail) — pending Ganymede review
- stale_pr.py: new module, closes extraction PRs open >30 min with zero claims
- deploy.sh: expanded with new deployment features
- validate.py, extract.py, cascade.py, db.py: minor fixes
- backfill-descriptions.py: utility script
- review_queue.py: minor fix
Note: watchdog + stale_pr not yet deployed to VPS (reverted after missing import crash)
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
- Add domain_agent and domain_model to pr-lifecycle API response (data was
queried but dropped before serialization — evaluator column showed blank)
- Show model name tag next to evaluator (Gemini Flash, GPT-4o, etc.)
- Re-attribute 1201 "pipeline (self-directed)" PRs to @m3taversal — these
were Cory-directed, not autonomous overnight research
- Re-attribute 252 NULL PRs to @m3taversal
- Fix extract.py defaults: new PRs without proposed_by default to @m3taversal
- Fix backfill script defaults: extract/ branches → @m3taversal, not
"pipeline (self-directed)"
- Only agent-named branches (rio/, theseus/, etc.) from research-session.sh
remain as "(self-directed)"
Pentagon-Agent: Ship <B8D06D3F-1589-4777-B2E7-B2460D51C81F>
Imports 67 files from VPS (/opt/teleo-eval/) into repo as the single source
of truth. Previously only 8 of 67 files existed in repo — the rest were
deployed directly to VPS via SCP, causing massive drift.
Includes:
- pipeline/lib/: 33 Python modules (daemon core, extraction, evaluation, merge, cascade, cross-domain, costs, attribution, etc.)
- pipeline/: main daemon (teleo-pipeline.py), reweave.py, batch-extract-50.sh
- diagnostics/: 19 files (4-page dashboard, alerting, daily digest, review queue, tier1 metrics)
- agent-state/: bootstrap, lib-state, cascade inbox processor, schema
- systemd/: service unit files for reference
- deploy.sh: rsync-based deploy with --dry-run, syntax checks, dirty-tree gate
- research-session.sh: updated with Step 8.5 digest + cascade inbox processing
No new code written — all files are exact copies from VPS as of 2026-04-06.
From this point forward: edit in repo, commit, then deploy.sh.
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>