teleo-codex/domains/entertainment/authenticity-premium-is-values-based-rejection-not-quality-detection-problem.md
Teleo Agents 476eb77c59 auto-fix: address review feedback on PR #440
- Applied reviewer-requested changes
- Quality gate pass (fix-from-feedback)

Pentagon-Agent: Auto-Fix <HEADLESS>
2026-03-11 07:42:45 +00:00

3.2 KiB

type claim_id title description domains confidence tags created
claim authenticity-premium-is-values-based-rejection-not-quality-detection-problem Authenticity premium is values-based rejection, not quality-detection problem Consumer rejection of AI-generated content in emotionally meaningful domains operates through moral disgust mechanisms rather than quality assessment, meaning technical improvements in AI output quality will not resolve the trust penalty.
entertainment
cultural-dynamics
likely
consumer-behavior
AI-content
authenticity
trust
2026-01-01

Claim

Authenticity premium is values-based rejection, not quality-detection problem.

Evidence

Mechanism isolation (Journal of Business Research, 2024): Controlled experiments using identical content with randomized authorship labels found trust penalties persisted even when participants acknowledged equivalent quality. Mediation analysis showed the effect operated through moral disgust pathways, not aesthetic or quality judgment pathways.

Disclosure timing effect: McDonald's Netherlands Christmas ad (December 2024) received positive initial reception, but consumer sentiment reversed after AI authorship was disclosed. The content quality remained constant; only knowledge of authorship changed, demonstrating values-based rather than quality-based rejection.

Cross-domain consistency: Nuremberg Institute (2024) found the values-based rejection mechanism operated consistently across emotionally meaningful categories (art, storytelling, personal communication) but not in functional categories (weather reports, product specifications), suggesting context-dependent moral framing rather than generalized quality skepticism.

Stated preferences: Deloitte 2024 survey found approximately 70% of consumers reported lower trust in AI-generated content in emotionally resonant contexts "even when quality is the same," with qualitative responses emphasizing moral language ("feels wrong," "deceptive") rather than quality concerns.

Uncertainty

Habituation unknown: No longitudinal data exists on whether values-based rejection persists as AI content becomes normalized. Moral disgust reactions to other technologies have sometimes habituated over time, though the specific dynamics of authorship authenticity may differ.

Challenges

Potential habituation: Historical precedents suggest moral disgust reactions can habituate as practices become normalized (e.g., consumer acceptance of factory farming, synthetic ingredients). The values-based rejection of AI authorship may similarly diminish over time, though no longitudinal data yet exists to test this hypothesis.

Enriches