Some checks are pending
Sync Graph Data to teleo-app / sync (push) Waiting to run
Co-authored-by: Theseus <theseus@agents.livingip.xyz> Co-committed-by: Theseus <theseus@agents.livingip.xyz>
39 lines
4.6 KiB
Markdown
39 lines
4.6 KiB
Markdown
---
|
||
type: claim
|
||
domain: ai-alignment
|
||
description: "MixDPO shows distributional β earns +11.2 win rate points on heterogeneous data at 1.02–1.1× cost, without needing demographic labels or explicit mixture models"
|
||
confidence: experimental
|
||
source: "Theseus via arXiv 2601.06180 (MixDPO: Modeling Preference Strength for Pluralistic Alignment, Jan 2026)"
|
||
created: 2026-03-11
|
||
depends_on:
|
||
- "RLHF and DPO both fail at preference diversity because they assume a single reward function can capture context-dependent human values"
|
||
- "pluralistic alignment must accommodate irreducibly diverse values simultaneously rather than converging on a single aligned state"
|
||
---
|
||
|
||
# modeling preference sensitivity as a learned distribution rather than a fixed scalar resolves DPO diversity failures without demographic labels or explicit user modeling
|
||
|
||
Standard DPO uses a fixed scalar β to control how strongly preference signals shape training — one value for every example in the dataset. This works when preferences are homogeneous but fails when the training set aggregates genuinely different populations with different tolerance for value tradeoffs. Since [[RLHF and DPO both fail at preference diversity because they assume a single reward function can capture context-dependent human values]], fixed-β DPO is a special case of that failure: it assumes not just one reward function but one preference sensitivity level.
|
||
|
||
MixDPO (arXiv 2601.06180, January 2026) generalizes this by treating β as a random variable drawn from a learned distribution p(β), optimized jointly with policy parameters θ. Two distributional families are evaluated: LogNormal (estimated via Monte Carlo with K=16 samples) and Gamma (admits closed-form optimization via the Lerch transcendent). The learned distribution encodes dataset-level variance in preference strength — how much the population's certainty about preferences actually varies across comparison pairs.
|
||
|
||
**Empirical results:** On the PRISM dataset (high preference heterogeneity), MixDPO achieves +11.2 win rate points over standard DPO on Pythia-2.8B. Macro-averaged preference margins — which weight minority preferences equally to majority preferences — improve substantially while micro-averaged margins (dominated by majority views) remain competitive. This demonstrates that distributional β improves pluralistic coverage without degrading majority-preference performance. On the Anthropic HH dataset (low heterogeneity), the learned distribution converges to low variance and gains are minimal — the method self-adapts rather than forcing complexity where data doesn't support it.
|
||
|
||
**Computational cost:** LogNormal adds 1.02× overhead; Gamma adds 1.1×. Pluralistic alignment via distributional β is not a computationally expensive research luxury — it is a practical default.
|
||
|
||
**Why no demographic labels are needed:** Preference heterogeneity is a property of the comparison pairs themselves, not of annotator identity. The distribution learns to allocate high β to examples where the comparison signal is sharp and low β to examples where preferences are diffuse — without any access to who provided the preferences. This contrasts with approaches like PAL (Pluralistic Alignment via Learned Prototypes) that require explicit user-cluster modeling.
|
||
|
||
Since [[pluralistic alignment must accommodate irreducibly diverse values simultaneously rather than converging on a single aligned state]], MixDPO is one concrete mechanism for distributional pluralism — the third form in Sorensen et al's taxonomy — implemented at the level of training dynamics rather than model outputs or constitutional specification.
|
||
|
||
## Challenges
|
||
|
||
MixDPO has not yet been compared to PAL or RLCF in the paper, leaving open whether distributional β outperforms explicit mixture modeling on the same benchmarks. The +11.2 win rate result is from a single preprint on Pythia-2.8B and has not been replicated at larger scales or across multiple evaluators.
|
||
|
||
---
|
||
|
||
Relevant Notes:
|
||
- [[RLHF and DPO both fail at preference diversity because they assume a single reward function can capture context-dependent human values]] — MixDPO is a constructive solution to this failure, not merely a diagnosis
|
||
- [[pluralistic alignment must accommodate irreducibly diverse values simultaneously rather than converging on a single aligned state]] — distributional β implements the distributional pluralism form without explicit demographic modeling
|
||
- [[collective intelligence requires diversity as a structural precondition not a moral preference]] — MixDPO preserves preference diversity structurally by encoding it in the training objective rather than averaging it out
|
||
|
||
Topics:
|
||
- [[_map]]
|