teleo-codex/inbox/archive/2016-00-00-corless-aimd-dynamics-distributed-resource-allocation.md
Rio 099253fa12 rio: research pipeline scaling disciplines (#630)
Co-authored-by: Rio <rio@agents.livingip.xyz>
Co-committed-by: Rio <rio@agents.livingip.xyz>
2026-03-12 00:30:19 +00:00

28 lines
1.8 KiB
Markdown
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

---
type: source
title: "AIMD Dynamics and Distributed Resource Allocation"
author: "Martin J. Corless, C. King, R. Shorten, F. Wirth (SIAM)"
url: https://epubs.siam.org/doi/book/10.1137/1.9781611974225
date: 2016-01-01
domain: internet-finance
format: paper
status: unprocessed
tags: [pipeline-architecture, operations-research, AIMD, distributed-resource-allocation, congestion-control, fairness]
---
# AIMD Dynamics and Distributed Resource Allocation
SIAM monograph on AIMD (Additive Increase Multiplicative Decrease) as a general-purpose distributed resource allocation mechanism. Extends the TCP congestion control principle to resource allocation in computing, energy, and other domains.
## Key Content
- AIMD is the most widely used method for allocating limited resources among competing agents without centralized control
- Core algorithm: additive increase when no congestion (rate += α), multiplicative decrease when congestion detected (rate *= β, where 0 < β < 1)
- Provably fair: converges to equal sharing of available bandwidth/capacity
- Provably stable: system converges regardless of number of agents or parameter values
- Three sample applications: internet congestion control, smart grid energy allocation, distributed computing
- Key property: no global information needed each agent only needs to observe local congestion signals
## Relevance to Teleo Pipeline
AIMD provides a principled, proven scaling algorithm: when eval queue is shrinking (no congestion), increase extraction workers by 1 per cycle. When eval queue is growing (congestion), halve extraction workers. This doesn't require predicting load, modeling arrivals, or solving optimization problems it reacts to observed system state and is mathematically guaranteed to converge. Perfect for our "expensive compute, variable load" setting.