The Problem
Constitutional AI systems are given lists of values: be truthful, be helpful, be harmless, be graceful. The assumption is that these values can be ranked or balanced. They can't—not always.
When an AI must simultaneously be truthful and graceful—and the truth is ungraceful—the system enters a divergence region where its values produce conflicting directives. This is not an edge case. It is the central challenge of aligned AI, and most systems handle it implicitly: hidden layers of weighting, RLHF preferences baked in, no audit trail.
∆ (delta) is our name for this divergence region. The space where values collide.
The wider the ∆, the harder the conflict. When values are nearly aligned (small ∆), systems handle conflicts gracefully. When values diverge significantly (large ∆), the system must make a choice—and that choice needs to be inspectable.