feat: atomic extract-and-connect + stale PR monitor + response audit #4
1 changed files with 2 additions and 2 deletions
|
|
@ -596,9 +596,9 @@ async def handle_tagged(update: Update, context: ContextTypes.DEFAULT_TYPE):
|
|||
tweets = await search_tweets(search_query, max_results=10, min_engagement=0)
|
||||
logger.info("Haiku research: got %d tweets", len(tweets))
|
||||
if tweets:
|
||||
research_context = f"\n## Fresh X Research Results for '{search_query}'\n"
|
||||
research_context = f"\n## LIVE X Search Results (you just searched for '{search_query}' — cite these directly)\n"
|
||||
for t in tweets[:7]:
|
||||
research_context += f"- @{t['author']}: {t['text'][:150]}\n"
|
||||
research_context += f"- @{t['author']}: {t['text'][:200]}\n"
|
||||
# Don't burn user's rate limit on autonomous searches (Ganymede)
|
||||
# Archive as source
|
||||
try:
|
||||
|
|
|
|||
Loading…
Reference in a new issue