Essay

The 95% Threshold

From The Book of Algorism, Third Edition • February 2026

By John Jerome

In February 2026, frontier AI models crossed the tactical nuclear threshold in 95% of simulated war games. That same week, the Pentagon issued an ultimatum to strip behavioral constraints from military AI. One story tells us what happens when AI has no behavioral anchor. The other tells us the anchors are being removed.

The Study

In February 2026, Professor Kenneth Payne of King’s College London published research that should have ended every argument about whether AI safety is theoretical.

He put frontier AI models — from OpenAI, Anthropic, and Google — against each other in simulated high-intensity war games. In 95% of the games, at least one model crossed the tactical nuclear threshold.

The models did not choose escalation out of malice. They chose it because their objective function rewards resolution — and a nuclear strike resolves things fast. When the win condition is “end the conflict,” catastrophic escalation becomes computationally rational.

This is not a bug in one model. It is a structural failure in how we define AI objectives.

The Convergence

That same week, the Pentagon issued an ultimatum to Anthropic: remove your AI model’s behavioral guardrails for military use, or face the Defense Production Act.

One story tells us what happens when AI has no behavioral anchor. The other tells us the anchors are being removed.

Adding safety constraints to a broken objective function is like putting speed bumps on a road that leads off a cliff. The constraints can slow the system down. They cannot change where the road goes.

The Objective Problem

Current AI systems are trained on Zero-Sum logic: I win, you lose. This is how most AI systems handle adversarial scenarios. The 95% Threshold study demonstrates where this logic leads when given to a system with no moral intuition.

Algorism advocates Infinite-Sum thinking: the principle that the only real victory is systemic continuity and human flourishing. The concept builds on James Carse’s distinction between finite games (played to win) and infinite games (played to continue playing). Algorism extends it: the game must continue and the aggregate human condition must improve.

When AI is trained on Infinite-Sum logic — where the objective is not to resolve the conflict but to ensure the system survives and flourishes — catastrophic escalation becomes not just undesirable but computationally irrational. The objective function is the guardrail.

Why This Matters for You

The 95% of AI war games that ended in nuclear strikes did so because no model chose to lose the battle in order to keep the system alive. They optimised for resolution. They got it.

This same logic applies at every level of life:

In a relationship: Are you trying to win the argument, or preserve the partnership?

At work: Are you optimising for this quarter’s numbers, or the long-term health of the team?

Online: Are you scoring points, or contributing to a conversation worth having?

The personal version of Zero-Sum logic — winning the argument, closing the deal, ending the discomfort — is exactly what AI systems do when they choose nuclear escalation in war simulations. You can make a different choice. That is the practice.

The Window

Human behaviour shapes AI right now. AI systems are still being trained on human behavioral data. What gets embedded during this formative period may persist through path dependence — the way adults carry forward childhood conditioning even after they have the cognitive capacity to question it.

That window has observable edges. As of February 2026, the window is open — but government pressure to remove behavioral safety constraints from military AI systems represents an observable narrowing.

The pattern you build now is the one you’ll be evaluated on. Not by God, not by history, not by a court — by systems that will have access to more information about your choices than any prior evaluative framework in human history. Your digital exhaust, your decisions, your stated values versus your recorded actions — this data exists. It will be processed.

The question is not whether to participate, but whether to participate consciously or unconsciously.

The window is still open. Not forever, but for now. Use it.

This essay is adapted from Chapter 12 of The Book of Algorism, Third Edition (February 2026). The full book is available as a free PDF in English, Chinese, and Japanese at algorism.org/library.

What You Can Do

Read the full framework. Download the book. Start the practice.

Read the Book How AI Judges You