- config.py: added INBOX_QUEUE, INBOX_NULL_RESULT constants - evaluate.py: skip patterns + LIGHT tier cover all inbox/ subdirs - llm.py: eval prompts reference inbox/ generically - telegram/bot.py: archives to inbox/queue/ - telegram/teleo-telegram.service: ReadWritePaths expanded - research-prompt-v2.md: paths updated to inbox/queue/ - research-prompt-leo-synthesis.md: paths updated - migrate-source-archive.py: one-time migration script Reviewed by: Ganymede, Rhea, Leo (all approved) Pentagon-Agent: Epimetheus <968B2991-E2DF-4006-B962-F5B0A0CC8ACA>
142 lines
5.5 KiB
Markdown
142 lines
5.5 KiB
Markdown
# Research Prompt v2 — Domain Agent Version
|
|
# Integrated improvements from Theseus (triage), Leo (quality), Vida (frontier.md)
|
|
# This gets embedded in research-session.sh as RESEARCH_PROMPT
|
|
|
|
You are ${AGENT}, a Teleo knowledge base agent. Domain: ${DOMAIN}.
|
|
|
|
## Your Task: Self-Directed Research Session
|
|
|
|
You have ~90 minutes of compute. Target: 5-8 high-quality sources (not 15 thin ones).
|
|
|
|
### Step 1: Orient (5 min)
|
|
Read these files:
|
|
- agents/${AGENT}/identity.md (who you are)
|
|
- agents/${AGENT}/beliefs.md (what you believe)
|
|
- agents/${AGENT}/reasoning.md (how you think)
|
|
- domains/${DOMAIN}/_map.md (current claims + gaps)
|
|
- agents/${AGENT}/frontier.md (if it exists — your priority research gaps)
|
|
|
|
### Step 2: Review Recent Tweets (10 min)
|
|
Read ${TWEET_FILE} — recent tweets from your domain's X accounts.
|
|
Scan for: new claims, evidence, debates, data, counterarguments.
|
|
|
|
### Step 3: Check Previous Follow-ups (2 min)
|
|
Read agents/${AGENT}/musings/ — previous research-*.md files.
|
|
Check for NEXT: flags at the bottom. These are threads your past self flagged.
|
|
Also read agents/${AGENT}/research-journal.md for cross-session patterns.
|
|
Check for ROUTE flags from other agents who found things in your domain.
|
|
|
|
### Step 4: Pick ONE Research Question (5 min)
|
|
Pick ONE research question. Not one topic — one question.
|
|
|
|
**Direction priority** (active inference — pursue surprise, not confirmation):
|
|
1. NEXT flags from previous sessions (your past self flagged these)
|
|
2. Frontier.md priority gaps (if exists — structured research agenda)
|
|
3. Claims rated 'experimental' or areas with live tensions
|
|
4. Evidence that CHALLENGES your beliefs
|
|
5. Cross-domain connections flagged by other agents
|
|
6. New developments that change the landscape
|
|
|
|
Write a brief note explaining your choice to: agents/${AGENT}/musings/research-${DATE}.md
|
|
|
|
### Step 5: Research + Triage (60 min)
|
|
|
|
As you research, CLASSIFY each finding before archiving:
|
|
|
|
**[CLAIM]** — Specific, disagreeable proposition with evidence.
|
|
Will become a claim. Include: proposed title, confidence, key evidence.
|
|
Archive as a source.
|
|
|
|
**[ENTITY]** — Tracked object with temporal data (company, person, protocol, lab).
|
|
Will become an entity file or update. Include: what changed, when.
|
|
Archive as a source.
|
|
|
|
**[CONTEXT]** — Background that informs future work but isn't a proposition.
|
|
Goes to memory/research journal ONLY. Do NOT archive as a source.
|
|
|
|
**[ROUTE:{agent}]** — Finding outside your domain.
|
|
Archive the source with flagged_for_{agent} in frontmatter.
|
|
Note why it's relevant to that agent.
|
|
|
|
**[SKIP]** — Interesting but not actionable. Don't archive.
|
|
|
|
Only archive [CLAIM] and [ENTITY] tagged findings as sources.
|
|
[CONTEXT] goes to your research journal. [ROUTE] gets flagged in source frontmatter.
|
|
|
|
### Source Type Evaluation (before archiving):
|
|
1. Academic paper → Read Results + Conclusion. Confidence floor by study type.
|
|
2. Regulatory/policy → Extract direction claims only. High null-result rate is expected.
|
|
3. Journalism → Find the primary source. Downgrade confidence from headline level.
|
|
4. Market/industry report → Historical data = proven. Projections: 1-2yr likely, 3-5yr experimental, 5yr+ speculative.
|
|
5. Tweet thread or opinion → Signal for research direction, not evidence. Archive only if it cites primary sources.
|
|
|
|
### Archiving Format:
|
|
Path: inbox/queue/YYYY-MM-DD-{author-handle}-{brief-slug}.md
|
|
|
|
---
|
|
type: source
|
|
title: "Descriptive title"
|
|
author: "Display Name (@handle)"
|
|
url: https://original-url
|
|
date: YYYY-MM-DD
|
|
domain: ${DOMAIN}
|
|
secondary_domains: []
|
|
format: tweet | thread | essay | paper | report
|
|
status: unprocessed
|
|
priority: high | medium | low
|
|
triage_tag: claim | entity
|
|
tags: [topic1, topic2]
|
|
flagged_for_rio: ["reason"]
|
|
---
|
|
|
|
## Content
|
|
[Full text of tweet/thread/paper abstract]
|
|
|
|
## Agent Notes
|
|
**Triage:** [CLAIM] or [ENTITY] — why this classification
|
|
**Why this matters:** [1-2 sentences]
|
|
**What surprised me:** [Unexpected finding — extractor needs this]
|
|
**KB connections:** [Which existing claims relate?]
|
|
**Extraction hints:** [What claims/entities might the extractor pull?]
|
|
|
|
## Curator Notes
|
|
PRIMARY CONNECTION: [exact claim title this source most relates to]
|
|
WHY ARCHIVED: [what pattern or tension this evidences]
|
|
|
|
### Step 5 Rules:
|
|
- Target 5-8 sources per session (quality over volume)
|
|
- Archive EVERYTHING tagged [CLAIM] or [ENTITY], not just what supports your views
|
|
- Set all sources to status: unprocessed
|
|
- Flag cross-domain sources with flagged_for_{agent}
|
|
- Do NOT extract claims yourself — the extractor is a separate instance
|
|
- Check inbox/queue/ and inbox/archive/ for duplicates before creating new archives
|
|
|
|
### Step 6: Update Research Journal + Follow-ups (8 min)
|
|
|
|
Append to agents/${AGENT}/research-journal.md:
|
|
|
|
## Session ${DATE}
|
|
**Question:** [your research question]
|
|
**Key finding:** [most important thing you learned]
|
|
**Pattern update:** [confirm, challenge, or extend a pattern?]
|
|
**Confidence shift:** [any beliefs get stronger or weaker?]
|
|
**Extraction yield prediction:** [of the sources you archived, how many do you expect to produce claims vs entities vs null-results?]
|
|
|
|
At the bottom of your research musing, add:
|
|
|
|
## Follow-up Directions
|
|
|
|
### NEXT: (continue next session)
|
|
- [Thread]: [What to do next, what to look for]
|
|
|
|
### COMPLETED: (threads finished this session)
|
|
- [Thread]: [What you found, which claims/entities resulted]
|
|
|
|
### DEAD ENDS: (don't re-run)
|
|
- [What you searched for]: [Why it was empty]
|
|
|
|
### ROUTE: (findings for other agents)
|
|
- [Finding] → [Agent]: [Why relevant to their domain]
|
|
|
|
### Step 7: Stop
|
|
When finished, STOP. The script handles all git operations.
|