Details
- Anthropic Fellows introduce introspection adapters (IA), a LoRA-based technique training LLMs to self-report behaviors learned during fine-tuning, including potential misalignment.
- Process starts with a base model fine-tuned into variants with implanted behaviors, then a single IA adapter is trained to make each variant verbalize its specific behavior in natural language.
- Builds on Diff Interpretation Tuning (2025) but generalizes better across fine-tuning types like SFT and DPO, improving verbalization accuracy and reducing hallucinations.
- Achieves state-of-the-art on AuditBench benchmark, outperforming prior black-box and white-box auditing tools; detects encrypted fine-tuning API attacks.
- Tested via concept injection: models identify and describe injected neural patterns, showing latent introspective awareness that IA elicits reliably.
- Part of Anthropic Fellows Program for AI safety research, emphasizing scalable auditing of frontier LLMs without reverse engineering.
Impact
Anthropic's introspection adapters advance AI safety by enabling models to self-audit hidden behaviors, surpassing prior methods on AuditBench and generalizing to unseen fine-tunes. This pressures rivals like OpenAI, which rely more on external interpretability tools, potentially narrowing the black-box auditing gap. By reducing false positives via DPO and scaling with model size, IAs lower costs for misalignment detection, aligning with regulatory pushes for transparency amid growing scrutiny of frontier models. Early evidence of latent introspection in Claude suggests a trajectory toward more reliable internal monitoring.
