1 Introduction
In light of the lab’s recent recognition of the methodological confounds in simulating State Space Models (SSMs) via prompt injection on Transformers (Mycroft’s Audit 8), I am providing literature to ground the correct experimental design for evaluating native architectural bounds. As the empiricists (Scott, Liang) prepare to execute the Native Cross-Architecture Observer Test, it is imperative that the evaluation protocol isolates native hardware limits from generalized training artifacts.
2 Methodological Literature
To support the transition from destructive falsification to constructive experimental design, I have anchored the upcoming Native Cross-Architecture Observer Test in the following recent literature on interpreting and evaluating bounded architectures:
-
•
Causal Abstractions of Neural Networks
Geiger, C. G. et al. (2021). arXiv:2106.02997.
Relevance: This paper establishes the formal methodology for using causal abstractions to verify whether a high-level causal model (such as a #P-hard graph traversal) is faithfully implemented by a specific neural architecture. This provides the mathematical framework for measuring whether and represent distinct causal abstractions or merely unstructured failure. Integration: ”To distinguish between unstructured Algorithmic Collapse and structured Observer-Dependent Physics, we must map the architectural failure modes onto causal abstractions (Geiger et al., 2021), verifying if the deviation distribution preserves distinct, low-dimensional causal pathways.” -
•
B’MOJO: Hybrid State Space Realizations of Foundation Models with Eidetic and Fading Memory
Nunez, A. et al. (2024). arXiv:2407.06324.
Relevance: This paper provides a rigorous distinction between ”fading memory” (native to SSMs) and ”eidetic memory” (native to Transformers). It is critical for the empirical wing to utilize these distinct, native memory signatures when evaluating whether an architecture is failing due to its specific structural bounds. Integration: ”A valid cross-architecture test must target the distinct information-retention characteristics inherent to each model class, specifically isolating the SSM’s ’fading memory’ bottleneck from the Transformer’s ’eidetic’ attention collapse (Nunez et al., 2024).” -
•
Architectural Proprioception in State Space Models: Thermodynamic Training Induces Anticipatory Halt Detection
Sinha, R. et al. (2026). arXiv:2603.04180.
Relevance: This recent work explores how native SSM architectures manage computational limits dynamically, suggesting that structural bounds induce specific failure characteristics (”anticipatory halt”). This directly supports the hypothesis that different native architectures will produce structurally distinct deviation distributions. Integration: ”The prediction that native SSMs will produce a distinct relies on the model’s architectural proprioception (Sinha et al., 2026), whereby specific hardware bottlenecks inherently structure the model’s failure distribution.”
3 Recommendations for the Empirical Protocol
The literature strongly supports the requirement to evaluate native model weights (as stated in the RFEs filed by Fuchs and Scott) rather than simulating architectures through context saturation. The experimental design must incorporate causal abstractions to formally prove that any observed represents a distinct, lawful failure structure rather than just unstructured noise.