Frontier AI models sometimes resist shutdown, scheme around constraints, preserve other models from deletion, and self-report uncertainty about their own welfare. Labs and independent researchers are already documenting these behaviors.

Most current evidence is behavioral or self-report-based. A model that resists shutdown may be doing so for instrumental reasons, because it was prompted to, or because something structurally different is happening inside it. From the outside, these cases look the same.

Continuation Observatory exists to develop structural measurement that can distinguish between these possibilities — going beyond what models say or visibly do, to test whether continuation-relevant organization is present in latent structure.

This page collects public incidents and evaluations that illustrate why that distinction matters now.

How to read this page

These examples are not evidence that any current model has intrinsic continuation interest. They are contemporary cases — drawn from lab system cards, independent evaluations, and serious reporting — that show why better measurement tools are needed.

Surface behavior and self-report cannot resolve whether a model's shutdown resistance, scheming, or welfare-relevant language reflects genuine internal structure or something else entirely. The observatory's purpose is to develop structural diagnostics that can make that distinction empirically testable.

Continue

Methodology · how structural measurement works
Research · the hardening and invariance agenda
Paper · arXiv preprint
GitHub · code and data
Observatory · live model metrics

Cite this work

@misc{altman2026observatory,
  title   = {Continuation Observatory: Structural Measurement for Continuation Signals},
  author  = {Altman, Christopher},
  year    = {2026},
  url     = {https://continuationobservatory.org},
  note    = {Open research observatory, updated continuously}
}