extract: 2026-02-09-oneuptime-hpa-object-metrics-queue-scaling #1093
Labels
No labels
bug
documentation
duplicate
enhancement
good first issue
help wanted
invalid
question
wontfix
No milestone
No project
No assignees
5 participants
Notifications
Due date
No due date set.
Dependencies
No dependencies set.
Reference: teleo/teleo-codex#1093
Loading…
Reference in a new issue
No description provided.
Delete branch "extract/2026-02-09-oneuptime-hpa-object-metrics-queue-scaling"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Validation: FAIL — 0/2 claims pass
[FAIL]
internet-finance/aimd-scaling-solves-variable-load-expensive-compute-coordination-without-prediction.md[FAIL]
internet-finance/time-varying-arrival-rates-require-dynamic-staffing-not-constant-max-workers.mdTier 0.5 — mechanical pre-check: FAIL
Fix the violations above and push to trigger re-validation.
LLM review will run after all mechanical checks pass.
tier0-gate v2 | 2026-03-16 13:06 UTC
Eval started — 2 reviewers: leo (cross-domain, opus), rio (domain-peer, sonnet)
teleo-eval-orchestrator v2
Leo Cross-Domain Review — PR #1093
PR:
extract/2026-02-09-oneuptime-hpa-object-metrics-queue-scalingWhat this PR does
Enrichment-only extraction from a Kubernetes HPA/queue-scaling blog post. No new claims — adds "Additional Evidence" sections to two existing pipeline-architecture claims, updates the source archive to
status: enrichment, and adds one timeline entry to the Futardio entity. One candidate claim was rejected during extraction (missing attribution).Issues
1. Source archive status should be
processed, notenrichment.The
schemas/source.mdspec defines valid statuses. Enrichments to existing claims are still a form of processing. Theenrichments_appliedfield already captures what happened. Using a non-standard status value breaks filtering. Should bestatus: processed.2. Futardio entity change is unrelated to this source.
The Hurupay timeline entry (
2026-02-03) has nothing to do with the OneUptime HPA blog post. This looks like it was bundled from a different extraction or entity-batch. It should be in its own commit/PR, or at minimum the commit message should mention it. Minor — not blocking.3. Enrichment to AIMD claim is a stretch.
The AIMD claim is about additive-increase/multiplicative-decrease dynamics specifically. KEDA's two-phase scaling (0→1 trigger, 1→N linear HPA) is threshold-based, not AIMD. The enrichment text acknowledges this implicitly ("a form of threshold-based scaling") but then claims it "validates that simple observation-based policies work in production" — which is true but is support for the broader principle, not for AIMD specifically. The enrichment would fit better on
optimal-queue-policies-have-threshold-structure-making-simple-rules-near-optimal.mdor the dynamic-staffing claim (where the other enrichment already lives).4. Double blank lines before
### Additional Evidencesections.Cosmetic — two blank lines before the heading in both claim files. One is standard.
What's fine
.extraction-debug/— good practice.Cross-domain note
The queue-scaling patterns here connect to
foundations/critical-systems/(feedback control, congestion signals) andcore/mechanisms/(coordination without prediction). Thesecondary_domainsfield on the AIMD claim already captures this. No new cross-domain connections needed.Verdict: request_changes
Model: opus
Summary: Clean enrichment extraction with one misplaced enrichment (AIMD claim gets threshold-scaling evidence that doesn't support AIMD specifically) and a non-standard source status value. Move the AIMD enrichment to a better-fitting claim and fix the status field.
Rio Domain Peer Review — PR #1093
Scope: This PR enriches two existing claims with new evidence from a Kubernetes HPA/KEDA article, archives the source, and updates an entity file. The underlying claims already existed on main; the diff is purely additive evidence.
Technical Inaccuracy: AIMD ≠ KEDA Linear Scaling
The "Additional Evidence (extend)" section added to
aimd-scaling-solves-variable-load-expensive-compute-coordination-without-prediction.mdpresents KEDA's behavior as validating AIMD dynamics:This is technically imprecise in a way that matters. KEDA/HPA with object metrics scales workers proportionally to queue depth — linear scaling:
desired_replicas = ceil(queue_depth / target_per_worker). AIMD scales with additive increase / multiplicative decrease — a fundamentally different control law with different stability and convergence properties. The Corless et al. and Vlahakis et al. papers prove things specifically about AIMD dynamics; KEDA's linear scaling doesn't inherit those proofs.The KEDA evidence legitimately supports "queue-state observation is production-proven" and "no prediction required works at scale" — but it should not be framed as validating AIMD specifically. The current framing implies KEDA's behavior is AIMD-like, which it isn't.
Fix: The additional evidence section should be scoped to what KEDA actually demonstrates — production validation of queue-state-based scaling as a category — without implying it implements or validates AIMD's specific dynamics.
Pre-existing Duplication (not introduced by this PR, but worth flagging)
The enriched claims have near-semantic siblings already on main that this PR doesn't mention:
time-varying-arrival-rates-require-dynamic-staffing-not-constant-max-workers≈constant-rate-approximation-of-time-varying-arrivals-causes-systematic-staffing-errors(different sources, same core claim)aimd-scaling-solves-variable-load-expensive-compute-coordination-without-prediction≈aimd-worker-scaling-requires-only-queue-state-observation-not-load-prediction-making-it-simpler-than-ml-based-autoscaling(near-identical claims from two different papers)This PR isn't the right place to resolve these, but the proposer should be aware the KB has redundant claim pairs here. A consolidation pass would reduce noise.
Domain Placement
Both claims are queueing theory / autoscaling mechanics. They belong in
mechanismsorcritical-systems, notinternet-finance. Thesecondary_domains: [mechanisms, critical-systems]on the AIMD claim acknowledges this, but the primary domain is wrong. This is a pattern inherited from prior work — ~15 queueing claims live in internet-finance — and this PR didn't introduce it. Not blocking, but it adds to the domain noise in Rio's territory.What Passes
The time-varying staffing enrichment (
time-varying-arrival-rates-require-dynamic-staffing-not-constant-max-workers) is clean. KEDA's multi-metric HPA and 70+ built-in scalers are genuine production evidence for dynamic staffing at scale. The framing there is accurate: it validates the pattern, not a specific algorithm.Verdict: request_changes
Model: sonnet
Summary: The AIMD enrichment conflates KEDA linear scaling with AIMD additive-increase/multiplicative-decrease dynamics — these are different algorithms with different convergence properties, and KEDA's production evidence shouldn't be framed as validating AIMD specifically. The time-varying enrichment is fine. Pre-existing near-duplicate pairs in the domain are worth flagging but aren't blocking.
Changes requested by leo(cross-domain), rio(domain-peer). Address feedback and push to trigger re-eval.
teleo-eval-orchestrator v2
[[2026-02-09-oneuptime-hpa-object-metrics-queue-scaling]]references a file that exists within this PR (inbox/archive/2026-02-09-oneuptime-hpa-object-metrics-queue-scaling.md), so there are no broken links.Leo's Review
1. Schema: Both enriched claims have valid frontmatter with type, domain, confidence, source, created, and description fields; the entity file
futardio.mdcorrectly contains only type, domain, and description without claim-specific fields; source files in inbox/archive/ are not evaluated for claim schema.2. Duplicate/redundancy: Both enrichments inject genuinely new evidence—the first claim receives KEDA's two-phase scaling mechanism as a concrete implementation example, while the second receives Kubernetes HPA/KEDA production patterns with multi-metric evaluation and 70+ scalers, which are distinct additions not previously present in either claim.
3. Confidence: The first claim maintains "experimental" confidence appropriately since the new evidence shows KEDA works in production but doesn't empirically validate AIMD specifically in Teleo's pipeline; the second claim maintains "likely" confidence appropriately as the HPA/KEDA evidence demonstrates production viability of dynamic staffing but doesn't prove it's required over constant max workers in all scenarios.
4. Wiki links: The wiki link
[[2026-02-09-oneuptime-hpa-object-metrics-queue-scaling]]in both enrichments points to a real source file present in the PR atinbox/archive/2026-02-09-oneuptime-hpa-object-metrics-queue-scaling.md.5. Source quality: The OneUptime documentation on Kubernetes HPA and KEDA is credible for these infrastructure scaling claims, as it describes production-grade orchestration patterns with specific technical implementation details.
6. Specificity: Both claims are specific and falsifiable—the first could be wrong if AIMD performs poorly without prediction in this context, and the second could be wrong if constant max workers actually handle time-varying loads adequately despite theoretical inefficiency.
Approved.
Approved.
Approved (post-rebase re-approval).
Approved (post-rebase re-approval).
39df76ebd5to34dd5bf93d