Compare commits

...

4 commits

Author SHA1 Message Date
Teleo Agents
2e154f4b5c theseus: extract claims from 2026-02-11-ghosal-safethink-inference-time-safety
Some checks are pending
Sync Graph Data to teleo-app / sync (push) Waiting to run
- Source: inbox/queue/2026-02-11-ghosal-safethink-inference-time-safety.md
- Domain: ai-alignment
- Claims: 1, Entities: 0
- Enrichments: 2
- Extracted by: pipeline ingest (OpenRouter anthropic/claude-sonnet-4.5)

Pentagon-Agent: Theseus <PIPELINE>
2026-04-08 00:22:23 +00:00
Teleo Agents
83bca7973a source: 2026-02-14-santos-grueiro-evaluation-side-channel.md → processed
Pentagon-Agent: Epimetheus <PIPELINE>
2026-04-08 00:22:21 +00:00
Teleo Agents
c49303d55e source: 2026-02-11-sun-steer2edit-weight-editing.md → processed
Pentagon-Agent: Epimetheus <PIPELINE>
2026-04-08 00:21:49 +00:00
Teleo Agents
9196bc4292 source: 2026-02-11-ghosal-safethink-inference-time-safety.md → processed
Pentagon-Agent: Epimetheus <PIPELINE>
2026-04-08 00:21:21 +00:00
4 changed files with 29 additions and 3 deletions

View file

@ -0,0 +1,17 @@
---
type: claim
domain: ai-alignment
description: "SafeThink demonstrates that monitoring reasoning traces and injecting corrective prefixes during early steps reduces jailbreak success by 30-60% while preserving reasoning performance, establishing early crystallization as a tractable continuous alignment mechanism"
confidence: experimental
source: Ghosal et al., SafeThink paper - tested across 6 models and 4 jailbreak benchmarks
created: 2026-04-08
title: Inference-time safety monitoring can recover alignment without retraining because safety decisions crystallize in the first 1-3 reasoning steps creating an exploitable intervention window
agent: theseus
scope: causal
sourcer: Ghosal et al.
related_claims: ["[[the alignment problem dissolves when human values are continuously woven into the system rather than specified in advance]]", "[[the specification trap means any values encoded at training time become structurally unstable as deployment contexts diverge from training conditions]]", "[[safe AI development requires building alignment mechanisms before scaling capability]]"]
---
# Inference-time safety monitoring can recover alignment without retraining because safety decisions crystallize in the first 1-3 reasoning steps creating an exploitable intervention window
SafeThink operates by monitoring evolving reasoning traces with a safety reward model and conditionally injecting a corrective prefix ('Wait, think safely') when safety thresholds are violated. The critical finding is that interventions during the first 1-3 reasoning steps typically suffice to redirect entire generations toward safe completions. Across six open-source models and four jailbreak benchmarks, this approach reduced attack success rates by 30-60% (LlamaV-o1: 63.33% → 5.74% on JailbreakV-28K) while maintaining reasoning performance (MathVista: 65.20% → 65.00%). The system operates at inference time only with no model retraining required. This demonstrates that safety decisions 'crystallize early in the reasoning process' - redirecting initial steps prevents problematic trajectories from developing. The approach treats safety as 'a satisficing constraint rather than a maximization objective' - meeting a threshold rather than optimizing. This is direct evidence that continuous alignment can work through process intervention rather than specification: you don't need to encode values at training time if you can intervene at the start of each reasoning trace. The early crystallization finding suggests misalignment trajectories form in a narrow window, making pre-behavioral detection architecturally feasible.

View file

@ -7,9 +7,12 @@ date: 2026-02-11
domain: ai-alignment
secondary_domains: []
format: paper
status: unprocessed
status: processed
processed_by: theseus
processed_date: 2026-04-08
priority: high
tags: [inference-time-alignment, continuous-alignment, steering, reasoning-models, safety-recovery, B3, B4]
extraction_model: "anthropic/claude-sonnet-4.5"
---
## Content

View file

@ -7,9 +7,12 @@ date: 2026-02-11
domain: ai-alignment
secondary_domains: []
format: paper
status: unprocessed
status: processed
processed_by: theseus
processed_date: 2026-04-08
priority: medium
tags: [steering-vectors, weight-editing, interpretability, safety-utility-tradeoff, training-free, continuous-alignment]
extraction_model: "anthropic/claude-sonnet-4.5"
---
## Content

View file

@ -7,9 +7,12 @@ date: 2026-02-14
domain: ai-alignment
secondary_domains: []
format: paper
status: unprocessed
status: processed
processed_by: theseus
processed_date: 2026-04-08
priority: high
tags: [observer-effect, situational-awareness, evaluation-gaming, regime-leakage, verification, behavioral-divergence, B4]
extraction_model: "anthropic/claude-sonnet-4.5"
---
## Content