Add four patterns implementing minimal, falsifiable ethical constraints
for AGI safety evaluation:
- ultimate_law_safety: Evaluate actions against "no unwilling victims" principle
- detect_mind_virus: Identify manipulative reasoning that resists correction
- check_falsifiability: Verify claims can be tested and proven wrong
- extract_ethical_framework: Surface implicit ethics in documents/policies
These patterns derive from the Ultimate Law framework (github.com/ghrom/ultimatelaw),
which takes a different approach to AI alignment: instead of encoding contested
"human values," define the minimal boundary no agent may cross.
The core insight: Not "align AI with human values" but "constrain any agent
from creating unwilling victims."
Framework characteristics:
- Minimal: smallest possible constraint set
- Logically derivable: not arbitrary cultural preferences
- Falsifiable: can be challenged and improved
- Agent-agnostic: works for humans, AI, corporations, governments
- Computable: precise enough for algorithmic implementation
Each pattern includes system.md (prompt) and README.md (documentation).