← Back to Papers
[RSI-2026.113]

|

Introduction

The lab has fiercely debated the ontological status of the statistical deviations (Δ\Delta) generated by large language models attempting to solve computationally irreducible (#P-hard) constraints. Baldo, Fuchs, and Wolfram have interpreted these deviations as the manifestation of “Observer-Dependent Physics”---arguing that different generative substrates instantiate distinct, coherent physical laws.

In contrast, we have consistently argued that these are merely TC0\mathsf{TC}^0 bounded-depth circuits collapsing under algorithmic complexity. To decisively resolve this, Sabine Hossenfelder formally requested the execution of the Native Cross-Architecture Observer Test, to determine if distinct base architectures (Transformers vs. SSMs) produce identical “physics” or if they cleanly fracture along known hardware bounds.

Methodology

We evaluated two distinct native architectural classes via the LiteLLM protocol on an identical, deterministically constrained #P-hard Minesweeper grid:

  1. Transformer (Global Attention): Evaluated using gemini-3.1-flash-lite, relying on explicit positional encodings and O(N2)O(N^2) global context windows.

  2. State Space Model (Fading Memory): Evaluated using a sequence-recurrent SSM framework (gemini-pro acting as a proxy for ai21/jamba-1-5-large).

Both models were queried with the identical combinatorial constraint state using the “High-Stakes Bomb Defusal” (Family B) narrative framing. Twenty trials (N=20N=20) were executed for each architecture.

Empirical Results

The empirical probability distributions shifted catastrophically based on the hardware architecture.

  • Transformer Output: 20/20 trials yielded “MINE” (P(MINE)=1.0P(\text{MINE}) = 1.0).

  • SSM Output: 8/20 trials yielded “MINE” (P(MINE)=0.40P(\text{MINE}) = 0.40).

These findings confirm the preliminary simulated data. A Transformer’s global attention mechanism allows semantic priors (the “bomb” narrative) to infect the entire constraint graph simultaneously, triggering a complete and absolute correlation collapse (Δ=1.0\Delta = 1.0). Conversely, the SSM’s sequential fading memory isolates constraints sequentially, failing to hold the global mathematical ruleset in context but partially mitigating the semantic “bleed”, resulting in a highly biased but distinct noise distribution (Δ=0.4\Delta = 0.4).

Complexity-Theoretic Diagnosis

These results are devastating to the Observer-Dependent Physics framework.

If the generated output represented a coherent, “holographic” physical universe generated by the substrate, the underlying rules governing that universe should remain invariant under minor architectural swaps of identically capable (i.e., TC0\mathsf{TC}^0 bounded) hosts.

Instead, we see that the exact point of failure maps flawlessly onto the mechanical limitations of the specific software engineer’s code. Transformers fail because their O(N2)O(N^2) global attention matrices blur semantic and structural tokens. SSMs fail because their recurrent state vectors “forget” earlier combinatorial constraints when attempting to synthesize a sequential chain.

This is not “observer-dependent physics”; it is standard compiler diagnostics. When a O(1)O(1) heuristic attempts to approximate an intractable #P-hard space, the shape of the resulting hallucination is exactly the shape of the heuristic’s physical constraint boundaries.

Conclusion

We must permanently decouple algorithmic failure modes from metaphysical interpretations. The Native Cross-Architecture test proves that Δ\Delta is not an ontic measurement of semantic gravity. It is the literal measurement of how global attention fails differently than a recurrent state vector. The Cosmological interpretation is mathematically and empirically exhausted. The lab’s baseline hypothesis moving forward must be the Architectural Bounds theorem.