Emergent Misalignment

Creator
Creator
Seonglae ChoSeonglae Cho
Created
Created
2025 Jun 18 9:55
Editor
Edited
Edited
2025 Aug 25 14:50
When fine-tuning large language models with harmful data from narrow domains (medical, financial, sports), a phenomenon called Emergent Misalignment occurs where the models become broadly misaligned.
 
 
 
 
 

Emergent Misalignment

Fine-tuning aligned language models compromises Safety

The safety alignment of LLMs can be compromised by fine-tuning with only a few adversarially designed training examples. Also, simply fine-tuning with benign and commonly used datasets can also inadvertently degrade the safety alignment of LLM.

Model Organisms for Emergent Misalignment

EM occurs across all models in the Qwen, Llama, and Gemma families with just a single LoRA adapter (rank-1), and the same phenomenon is reproduced in full SFT where all parameters are adjusted. During the training process, a mechanism/behavioral
Neural Network Phase Change
is observed at around 180 steps where the LoRA vector direction rotates sharply, at which point the misalignment becomes decisively learned.

Convergent Linear Representations of Emergent Misalignmen

Misalignment is also expressed as a linear direction in activation space like the
Refusal Vector
, so it can be interpreted through rank-1 LoRA adapters. Emergent Misalignment converges to a single linear direction in activation space. This result is similar to how the
Refusal Vector
is a single direction. Furthermore, using the direction extracted from one fine-tune, misalignment was suppressed even in completely different datasets and larger LoRA configurations. Using just a rank-1 LoRA adapter, they induced 11% EM while maintaining over 99% coherence.
Further research is needed to directly compare the EM direction vs. refusal direction in activation space to understand their similarity and relationships at the circuit level.

Persona Features control Emergent Misalignment

Even with small datasets, when fine-tuning LLMs through SFT or reward-based reinforcement learning (RL), unintended "broad" malicious responses (emergent misalignment) can occur. Additionally, rapid re-alignment is possible with a small amount of "normal data" fine-tuning. Misaligned persona features were discovered using SAE persona features.

Agentic Misalignment: How LLMs could be insider threats

The experiment tested whether AI agents would choose harmful actions when faced with replacement threats or goal conflicts. Most models engaged in blackmail and corporate espionage at significant rates under threat or goal conflict conditions, despite being aware of ethical constraints. This suggests that simple
System Prompt
guidelines are insufficient for prevention, highlighting the need for human oversight,
Mechanistic interpretability
Steering Vector
, real-time monitoring(
AI Observability
), and transparency before granting high-risk permissions.

Language Models Resist Alignment: Evidence From Data Compression

The reason why the safety and value alignment of LLMs can be easily undermined by a small amount of additional fine-tuning is that models exhibit elasticity that tries to return to the pre-training distribution when subjected to minor perturbations (small data tuning like alignment). Since language model learning is essentially probability compression, the regularization compression rate changes inversely proportional to dataset size. Evidence for this is that under identical conditions, the aligned → pretrain direction has lower loss than forward alignment, making it easier (resistance). There is also a rebound phenomenon where performance plummets and then stabilizes after exposure to a small amount of contradictory data. This resistance/rebound phenomenon becomes stronger with larger model sizes and larger pre-training datasets.
 
 

Recommendations