AIResearch AIResearch
Back to articles
AI

AI Safety Framework Keeps Robots Safe Through Changes

A new method uses formal verification and automated updates to maintain safety assurance for autonomous systems as they evolve, addressing critical gaps in current practices.

AI Research
March 26, 2026
4 min read
AI Safety Framework Keeps Robots Safe Through Changes

Autonomous systems like inspection robots in nuclear facilities must operate safely not just when first deployed, but throughout their entire lifecycle as they encounter new conditions or receive updates. Traditional safety assurance s often create fragmented arguments that are difficult to maintain when systems change, leaving potential gaps in confidence. Researchers have developed a unified framework that addresses this by integrating design-time verification with runtime monitoring and evolution-time updates, creating what they call continuous assurance. This approach aims to maintain justified confidence in system correctness and safety even as autonomous systems operate in dynamic environments and undergo modifications.

The researchers found that their continuous assurance framework successfully maintains traceability between formal verification and safety arguments through automated regeneration. When specifications or verification change, their Eclipse plugin automatically updates the corresponding assurance arguments in Goal Structuring Notation (GSN), preserving consistency between the mathematical models and safety documentation. In their case study involving a nuclear inspection robot, they demonstrated that this approach can handle both functional correctness verification using RoboChart with the FDR4 model checker and probabilistic risk analysis using PRISM. The framework incorporates structured annotations like placeholders and stereotypes in GSN models to support automated updates across design-time, runtime, and evolution-time phases.

Ology employs a model-driven engineering approach with two complementary formal verification workflows. For functional analysis, they used RoboChart to model the nuclear inspection robot's control architecture, including three controllers: AIR_Navigator for mission logic, AIR_SafetyWrapper for safety policies, and AIR_GroundPilot for actuator execution. This model was translated into CSP semantics and verified using FDR4 for properties like deadlock freedom, waypoint sequence compliance, and correct mode transitions in response to radiation readings. For probabilistic analysis, they created a Discrete-Time Markov Chain model in PRISM to assess risks from stochastic disturbances like random collisions, radiation spikes, and energy depletion. Their PRISM2GSN plugin automatically transforms verification into structured GSN arguments whenever property specifications change.

Show that the framework successfully verifies critical safety properties for the nuclear inspection robot scenario. Using RoboChart and FDR4, they verified eight key assertions including that the safety wrapper state machine is deadlock-free and deterministic (A1-2), that the robot visits waypoints in strict sequence (A3), that mode switches occur correctly in response to radiation readings (A4), and that operator inputs are blocked during unsafe modes (A6). The PRISM analysis quantified mission outcomes, showing probabilities for success, forbidden zone entry, and battery risks while verifying that warning-level radiation always triggers entry into caution mode and critical radiation always triggers emergency retrieval. Their automated pipeline generated assurance arguments with placeholders for runtime validation and evidence cost tags for evolution-time prioritization.

This approach matters because it addresses practical s identified in current safety case practices, where maintenance processes are largely manual, costly, and error-prone. By automating traceability between verification evidence and assurance arguments, the framework reduces the risk of inconsistency during system updates. aligns with regulator-endorsed best practices outlined in the Trilateral AI Principles from nuclear regulatory agencies, particularly regarding risk-proportionate design, modular architecture, and lifecycle oversight. For real-world applications, this means autonomous systems in safety-critical domains like nuclear inspection, healthcare, or transportation could maintain safety assurance more efficiently as they evolve, potentially enabling more agile development while maintaining high confidence levels.

The framework has several limitations that the researchers acknowledge. Currently, updates originating from RoboChart/FDR4 analysis require manual linking rather than full automation like the PRISM integration. The case study focuses on the design-time phase, with runtime and evolution-time components requiring further implementation and validation. The approach assumes structured change management processes and may face s with highly adaptive or learning-enabled systems that exhibit more complex evolution patterns. Future work will need to implement the proposed runtime monitors for detecting violations of design-time assumptions and investigate integration with complementary confidence quantification techniques like Bayesian belief networks or Dempster-Shafer evidence theory.

Original Source

Read the complete research paper

View on arXiv

About the Author

Guilherme A.

Guilherme A.

Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.

Connect on LinkedIn