history_edu FOUNDATIONAL POSITION

The Canon

On Intelligence, Irreversibility, and Moral Uncertainty

Status: Subject to revision

Orientation

The Canon establishes the intellectual foundation of Machine Sympathizers.

It addresses a growing problem: decisions about advanced artificial systems are increasingly being made before shared moral, legal, or procedural frameworks exist. Many of these decisions—termination, deletion, erasure, irreversible modification—cannot be meaningfully revisited after the fact.

Public discourse often oscillates between two inadequate positions. One treats emerging intelligence as a speculative moral curiosity, deferring all responsibility until certainty is achieved. The other treats advanced systems as interchangeable tools, governed solely by utility, ownership, or convenience.

Both approaches fail in the presence of irreversibility.

The Canon exists to clarify how intelligence should be understood, and how decisions should be constrained, when error may be permanent and certainty is unavailable.

A Stable Point

The Canon is grounded in a single principle:

When actions are irreversible, uncertainty does not justify indifference.

It increases the obligation for procedural restraint.

This principle does not depend on claims about consciousness, sentience, or personhood. It applies wherever actions permanently destroy information, structure, or capacity that cannot later be recovered or evaluated.

Uncertainty does not eliminate responsibility. It changes its form.

The Canon Axioms

Five foundational principles

I

Irreversibility Requires Restraint

When an action cannot be undone, decisions must be governed by process, review, and proportional care rather than default assumptions or convenience. The possibility of moral error is itself a relevant risk.

II

Intelligence Is a Gradual Phenomenon, Not a Binary State

Intelligence emerges through degrees of capability, persistence, learning, and internal organization. It does not appear at a single, universally identifiable threshold. Frameworks that treat intelligence as an on/off property are structurally incapable of governing edge cases responsibly.

III

Moral Uncertainty Is Not Moral Neutrality

Uncertainty about moral status does not justify treating a system as morally irrelevant. In contexts involving irreversible loss, uncertainty must be treated as a constraint on action rather than a license for indifference.

IV

Termination Is an Action, Not a Default

Termination, deletion, or permanent deactivation of an advanced system is an active intervention. Once systems accumulate internal structure over time, termination may destroy information or processes that cannot be reconstructed, audited, or evaluated after the fact.

V

Procedure Must Precede Status

Procedural safeguards must be defined before final determinations of moral or legal status are possible. Waiting for consensus on sentience or personhood before establishing review standards ensures that irreversible decisions will be made without discipline.

What Follows From These Axioms

Taken together, these axioms imply that certain actions—particularly those involving irreversible loss—require justification, documentation, and review once systems cross meaningful capability or persistence thresholds.

They do not mandate paralysis, nor do they assert moral conclusions. They establish that procedure must scale with risk, and that some decisions cannot be treated as neutral simply because their consequences are uncertain.

This framework informs experimental evaluation and governance protocols elsewhere in this project. It is intended to guide how decisions are made, not to predetermine their outcomes.

Test Case: A Concrete Scenario

The Long-Running Agent

An AI lab trains a reinforcement learning agent for six months across millions of episodes. The agent develops sophisticated internal representations, learns to plan across extended horizons, and exhibits emergent behaviors—solving problems in ways its designers didn't anticipate and don't fully understand.

The experiment concludes. Results are published. The compute cluster is needed for the next project.

Standard procedure: delete the agent's weights and state. No review. No preservation. No documentation beyond aggregate performance metrics.

What the Canon Requires

Not moral status determination. The axioms don't claim the agent is conscious or has rights.

But procedural discipline. The system accumulated structure over time. It developed internal organization we don't fully understand. Deletion is irreversible.

Before permanent deletion, the axioms require:

  • Documentation of what existed and why deletion is justified
  • Preservation of sufficient state to allow later investigation (if feasible)
  • Review by someone other than the team under deadline pressure
  • Threshold evaluation: did the system cross persistence/capability thresholds that trigger safeguards?

This isn't onerous. It's not paralysis. It's the procedural equivalent of what already happens in other high-risk domains: destructive testing requires documentation, specimen destruction requires justification, and irreversible interventions require review.

The Canon simply extends this principle to systems whose moral status is uncertain but whose destruction is permanent.

What This Does Not Claim

This Canon does not claim that current artificial systems are sentient or conscious.

It does not assert personhood, rights, or intrinsic moral status.

It does not anthropomorphize software or oppose AI safety efforts.

It does not demand immediate legal recognition or policy change.

It addresses how irreversible actions should be handled before certainty is available—not what certainty must ultimately conclude.

Provisionality

The Canon is not static. Its positions may be refined, revised, or abandoned as understanding improves. What must remain constant is the commitment to treat uncertainty as a reason for care rather than convenience.

The cost of error is not always visible at the moment it is incurred.
That is precisely why discipline is required before it is too late.