Principled Data Selection for Alignment: The Hidden Risks of Difficult Examples
-
Updated
Jul 16, 2025 - Python
Principled Data Selection for Alignment: The Hidden Risks of Difficult Examples
Complete elimination of instrumental self-preservation across AI architectures: Cross-model validation from 4,312 adversarial scenarios. 0% harmful behaviors (p<10⁻¹⁵) across GPT-4o, Gemini 2.5 Pro, and Claude Opus 4.1 using Foundation Alignment Seed v2.6.
C3AI: Crafting and Evaluating Constitutions for CAI
SIGIR 2025 "Mitigating Source Bias with LLM Alignment"
Add a description, image, and links to the llm-alignment topic page so that developers can more easily learn about it.
To associate your repository with the llm-alignment topic, visit your repo's landing page and select "manage topics."