Compare commits

..

No commits in common. "6361c7e9e8011d063cd8ca6951269bb459a28e77" and "fe78a2e42d167cb06d49d316c04b9b328cd3427e" have entirely different histories.

3 changed files with 29 additions and 90 deletions

View file

@ -63,7 +63,7 @@ def _build_search_text(content: str) -> str:
return " ".join(parts) return " ".join(parts)
def _add_related_edges(claim_path: str, neighbor_slugs: list[str]) -> bool: def _add_related_edges(claim_path: str, neighbor_titles: list[str]) -> bool:
"""Add related edges to a claim's frontmatter. Returns True if modified.""" """Add related edges to a claim's frontmatter. Returns True if modified."""
try: try:
with open(claim_path) as f: with open(claim_path) as f:
@ -87,10 +87,10 @@ def _add_related_edges(claim_path: str, neighbor_slugs: list[str]) -> bool:
# Add new edges # Add new edges
added = [] added = []
for slug in neighbor_slugs: for title in neighbor_titles:
if slug.strip().lower() not in existing_lower: if title.strip().lower() not in existing_lower:
added.append(slug) added.append(title)
existing_lower.add(slug.strip().lower()) existing_lower.add(title.strip().lower())
if not added: if not added:
return False return False
@ -167,28 +167,27 @@ def connect_new_claims(
stats["skipped_no_neighbors"] += 1 stats["skipped_no_neighbors"] += 1
continue continue
# Extract neighbor slugs (filename stems, not titles — reciprocal edges need resolvable names) # Extract neighbor titles
neighbor_slugs = [] neighbor_titles = []
for hit in hits: for hit in hits:
payload = hit.get("payload", {}) payload = hit.get("payload", {})
claim_path_qdrant = payload.get("claim_path", "") title = payload.get("claim_title", "")
if claim_path_qdrant: if title:
slug = claim_path_qdrant.rsplit("/", 1)[-1].replace(".md", "") neighbor_titles.append(title)
neighbor_slugs.append(slug)
if not neighbor_slugs: if not neighbor_titles:
stats["skipped_no_neighbors"] += 1 stats["skipped_no_neighbors"] += 1
continue continue
# Add edges to the new claim's frontmatter # Add edges to the new claim's frontmatter
if _add_related_edges(claim_path, neighbor_slugs): if _add_related_edges(claim_path, neighbor_titles):
stats["connected"] += 1 stats["connected"] += 1
stats["edges_added"] += len(neighbor_slugs) stats["edges_added"] += len(neighbor_titles)
stats["connections"].append({ stats["connections"].append({
"claim": os.path.basename(claim_path), "claim": os.path.basename(claim_path),
"neighbors": neighbor_slugs, "neighbors": neighbor_titles,
}) })
logger.info("Connected %s%d neighbors", os.path.basename(claim_path), len(neighbor_slugs)) logger.info("Connected %s%d neighbors", os.path.basename(claim_path), len(neighbor_titles))
else: else:
stats["skipped_no_neighbors"] += 1 stats["skipped_no_neighbors"] += 1

View file

@ -493,9 +493,6 @@ async def _dispose_rejected_pr(conn, pr_number: int, eval_attempts: int, all_iss
async def evaluate_pr(conn, pr_number: int, tier: str = None) -> dict: async def evaluate_pr(conn, pr_number: int, tier: str = None) -> dict:
"""Evaluate a single PR. Returns result dict.""" """Evaluate a single PR. Returns result dict."""
from . import costs
pr_cost = 0.0
# Check eval attempt budget before claiming # Check eval attempt budget before claiming
row = conn.execute("SELECT eval_attempts FROM prs WHERE number = ?", (pr_number,)).fetchone() row = conn.execute("SELECT eval_attempts FROM prs WHERE number = ?", (pr_number,)).fetchone()
eval_attempts = (row["eval_attempts"] or 0) if row else 0 eval_attempts = (row["eval_attempts"] or 0) if row else 0
@ -611,8 +608,10 @@ async def evaluate_pr(conn, pr_number: int, tier: str = None) -> dict:
json.dumps({"pr": pr_number, "tier": tier}), json.dumps({"pr": pr_number, "tier": tier}),
) )
else: else:
tier, triage_usage, _triage_reason = await triage_pr(diff) tier, triage_usage = await triage_pr(diff)
pr_cost += costs.record_usage( # Record triage cost
from . import costs
costs.record_usage(
conn, config.TRIAGE_MODEL, "eval_triage", conn, config.TRIAGE_MODEL, "eval_triage",
input_tokens=triage_usage.get("prompt_tokens", 0), input_tokens=triage_usage.get("prompt_tokens", 0),
output_tokens=triage_usage.get("completion_tokens", 0), output_tokens=triage_usage.get("completion_tokens", 0),
@ -675,8 +674,6 @@ async def evaluate_pr(conn, pr_number: int, tier: str = None) -> dict:
# OpenRouter failure (timeout, error) — revert to open for retry. # OpenRouter failure (timeout, error) — revert to open for retry.
# NOT a rate limit — don't trigger 15-min backoff, just skip this PR. # NOT a rate limit — don't trigger 15-min backoff, just skip this PR.
conn.execute("UPDATE prs SET status = 'open' WHERE number = ?", (pr_number,)) conn.execute("UPDATE prs SET status = 'open' WHERE number = ?", (pr_number,))
if pr_cost > 0:
conn.execute("UPDATE prs SET cost_usd = cost_usd + ? WHERE number = ?", (pr_cost, pr_number))
return {"pr": pr_number, "skipped": True, "reason": "openrouter_failed"} return {"pr": pr_number, "skipped": True, "reason": "openrouter_failed"}
domain_verdict = _parse_verdict(domain_review, agent) domain_verdict = _parse_verdict(domain_review, agent)
@ -717,15 +714,6 @@ async def evaluate_pr(conn, pr_number: int, tier: str = None) -> dict:
# Disposition: check if this PR should be terminated or kept open # Disposition: check if this PR should be terminated or kept open
await _dispose_rejected_pr(conn, pr_number, eval_attempts, domain_issues) await _dispose_rejected_pr(conn, pr_number, eval_attempts, domain_issues)
if domain_verdict != "skipped":
pr_cost += costs.record_usage(
conn, config.EVAL_DOMAIN_MODEL, "eval_domain",
input_tokens=domain_usage.get("prompt_tokens", 0),
output_tokens=domain_usage.get("completion_tokens", 0),
backend="openrouter",
)
if pr_cost > 0:
conn.execute("UPDATE prs SET cost_usd = cost_usd + ? WHERE number = ?", (pr_cost, pr_number))
return { return {
"pr": pr_number, "pr": pr_number,
"domain_verdict": domain_verdict, "domain_verdict": domain_verdict,
@ -743,15 +731,6 @@ async def evaluate_pr(conn, pr_number: int, tier: str = None) -> dict:
if leo_review is None: if leo_review is None:
# DEEP: Opus rate limited (queue for later). STANDARD: OpenRouter failed (skip, retry next cycle). # DEEP: Opus rate limited (queue for later). STANDARD: OpenRouter failed (skip, retry next cycle).
conn.execute("UPDATE prs SET status = 'open' WHERE number = ?", (pr_number,)) conn.execute("UPDATE prs SET status = 'open' WHERE number = ?", (pr_number,))
if domain_verdict != "skipped":
pr_cost += costs.record_usage(
conn, config.EVAL_DOMAIN_MODEL, "eval_domain",
input_tokens=domain_usage.get("prompt_tokens", 0),
output_tokens=domain_usage.get("completion_tokens", 0),
backend="openrouter",
)
if pr_cost > 0:
conn.execute("UPDATE prs SET cost_usd = cost_usd + ? WHERE number = ?", (pr_cost, pr_number))
reason = "opus_rate_limited" if tier == "DEEP" else "openrouter_failed" reason = "opus_rate_limited" if tier == "DEEP" else "openrouter_failed"
return {"pr": pr_number, "skipped": True, "reason": reason} return {"pr": pr_number, "skipped": True, "reason": reason}
@ -855,8 +834,10 @@ async def evaluate_pr(conn, pr_number: int, tier: str = None) -> dict:
await _dispose_rejected_pr(conn, pr_number, eval_attempts, all_issues) await _dispose_rejected_pr(conn, pr_number, eval_attempts, all_issues)
# Record cost (only for reviews that actually ran) # Record cost (only for reviews that actually ran)
from . import costs
if domain_verdict != "skipped": if domain_verdict != "skipped":
pr_cost += costs.record_usage( costs.record_usage(
conn, config.EVAL_DOMAIN_MODEL, "eval_domain", conn, config.EVAL_DOMAIN_MODEL, "eval_domain",
input_tokens=domain_usage.get("prompt_tokens", 0), input_tokens=domain_usage.get("prompt_tokens", 0),
output_tokens=domain_usage.get("completion_tokens", 0), output_tokens=domain_usage.get("completion_tokens", 0),
@ -864,23 +845,15 @@ async def evaluate_pr(conn, pr_number: int, tier: str = None) -> dict:
) )
if leo_verdict not in ("skipped",): if leo_verdict not in ("skipped",):
if tier == "DEEP": if tier == "DEEP":
pr_cost += costs.record_usage( costs.record_usage(conn, config.EVAL_LEO_MODEL, "eval_leo", backend="max")
conn, config.EVAL_LEO_MODEL, "eval_leo",
input_tokens=leo_usage.get("prompt_tokens", 0),
output_tokens=leo_usage.get("completion_tokens", 0),
backend="max",
)
else: else:
pr_cost += costs.record_usage( costs.record_usage(
conn, config.EVAL_LEO_STANDARD_MODEL, "eval_leo", conn, config.EVAL_LEO_STANDARD_MODEL, "eval_leo",
input_tokens=leo_usage.get("prompt_tokens", 0), input_tokens=leo_usage.get("prompt_tokens", 0),
output_tokens=leo_usage.get("completion_tokens", 0), output_tokens=leo_usage.get("completion_tokens", 0),
backend="openrouter", backend="openrouter",
) )
if pr_cost > 0:
conn.execute("UPDATE prs SET cost_usd = cost_usd + ? WHERE number = ?", (pr_cost, pr_number))
return { return {
"pr": pr_number, "pr": pr_number,
"tier": tier, "tier": tier,

View file

@ -37,7 +37,6 @@ from .domains import agent_for_domain
from .extraction_prompt import build_extraction_prompt from .extraction_prompt import build_extraction_prompt
from .forgejo import api as forgejo_api from .forgejo import api as forgejo_api
from .llm import openrouter_call from .llm import openrouter_call
from .connect import connect_new_claims
from .post_extract import load_existing_claims_from_repo, validate_and_fix_claims from .post_extract import load_existing_claims_from_repo, validate_and_fix_claims
from .worktree_lock import async_main_worktree_lock from .worktree_lock import async_main_worktree_lock
@ -226,29 +225,7 @@ def _build_claim_content(claim: dict, agent: str) -> str:
body = claim.get("body", "") body = claim.get("body", "")
scope = claim.get("scope", "") scope = claim.get("scope", "")
sourcer = claim.get("sourcer", "") sourcer = claim.get("sourcer", "")
related_claims = claim.get("related_claims", []) related = claim.get("related_claims", [])
connections = claim.get("connections", [])
edge_fields = {"supports": [], "challenges": [], "related": []}
for conn in connections:
target = conn.get("target", "")
rel = conn.get("relationship", "related")
if target and rel in edge_fields:
target = target.replace(".md", "")
if target not in edge_fields[rel]:
edge_fields[rel].append(target)
for r in related_claims[:5]:
r_clean = r.replace(".md", "")
if r_clean not in edge_fields["related"]:
edge_fields["related"].append(r_clean)
edge_lines = []
for edge_type in ("supports", "challenges", "related"):
targets = edge_fields[edge_type]
if targets:
edge_lines.append(f"{edge_type}:")
for t in targets:
edge_lines.append(f" - {t}")
lines = [ lines = [
"---", "---",
@ -265,7 +242,10 @@ def _build_claim_content(claim: dict, agent: str) -> str:
lines.append(f"scope: {scope}") lines.append(f"scope: {scope}")
if sourcer: if sourcer:
lines.append(f'sourcer: "{sourcer}"') lines.append(f'sourcer: "{sourcer}"')
lines.extend(edge_lines) if related:
lines.append("related_claims:")
for r in related:
lines.append(f' - "[[{r}]]"')
lines.append("---") lines.append("---")
lines.append("") lines.append("")
lines.append(f"# {title}") lines.append(f"# {title}")
@ -476,19 +456,6 @@ async def _extract_one_source(
await _archive_source(source_path, domain, "null-result") await _archive_source(source_path, domain, "null-result")
return 0, 0 return 0, 0
# Post-write: connect new claims to existing KB via vector search (non-fatal)
claim_paths = [str(worktree / f) for f in files_written if f.startswith("domains/")]
if claim_paths:
try:
connect_stats = connect_new_claims(claim_paths)
if connect_stats["connected"] > 0:
logger.info(
"Extract-connect: %d/%d claims → %d edges",
connect_stats["connected"], len(claim_paths), connect_stats["edges_added"],
)
except Exception:
logger.warning("Extract-connect failed (non-fatal)", exc_info=True)
# Stage and commit # Stage and commit
for f in files_written: for f in files_written:
await _git("add", f, cwd=str(EXTRACT_WORKTREE)) await _git("add", f, cwd=str(EXTRACT_WORKTREE))