Skip to Content

Explainable Reinforcement Learning Agent for Air Traffic Control

ATC RL Agent Hero Image

FieldDetails
DomainAviation
Assurance GoalExplainability

Overview

Project Bluebird is an EPSRC Prosperity Partnership between the Alan Turing Institute, NATS (the UK’s en route air navigation service provider), and the University of Exeter. The programme has developed BluebirdDT, a probabilistic Digital Twin of the London Area Control Centre (LACC) that integrates over 20 million flights of NATS operational data.1 BluebirdDT has been used to train and evaluate AI agents for en route air traffic management—the high-altitude phase of flight between departure and arrival airports.

A reinforcement learning (RL) agent has been trained using Proximal Policy Optimisation (PPO) to manage aircraft within simulated en route sectors.2 The agent learns to recommend clearances (i.e. instructions to aircraft to turn or change altitude) in order to maintain safe separation between aircraft while keeping them close to their planned routes.

The agent has completed simulation-based qualification using the Machine Basic Training (MBT) framework — a human-in-the-loop assessment methodology that adapts NATS’ regulator-certified training curriculum for evaluating AI agents, using certified ATCO assessors to grade agent performance against the same competency standards applied to human trainees.3 This case study considers a plausible near-future scenario in which, having passed MBT qualification, the agent enters a supervised operational trial in a designated en route sector within the LACC. During the trial, the agent operates as a decision-support tool: it monitors live traffic and recommends clearances to a supervising air traffic controller (ATCO), who retains full authority to accept, modify, or override any recommendation. The ATCO remains responsible for all clearances issued to pilots.

Because RL agents learn policies through trial-and-error interaction within a simulated environment, rather than being explicitly programmed with rules, their decision-making can be inherently opaque. An ATCO supervising the agent during the trial would need to understand why it recommended a particular clearance to judge whether the recommendation was safe and appropriate. A developer investigating a failure or unexpected recommendation would need to understand the internal factors driving the decision. A regulator evaluating the trial would need evidence that the agent’s behaviour is explainable in terms that map to established operational standards. This case study explores the challenge of assuring explainability across these different audiences and needs.

System Description

What the System Does

The RL agent provides decision support to a supervising ATCO by:

  • continuously monitoring the positions, headings, altitudes, and flight plan routes of all aircraft in the designated trial sector;
  • identifying potential conflicts between aircraft and recommending clearances (heading changes, altitude changes) to maintain safe separation;
  • presenting recommended clearances to the supervising ATCO for review and approval;
  • updating its situational model based on the ATCO’s decisions, whether the recommendation was accepted, modified, or overridden; and
  • recommending that aircraft be returned to their planned routes once a conflict has been resolved.

The ATCO issues all clearances to pilots via the standard voice communication loop. The agent does not communicate directly with pilots.

How It Works

The agent’s operation spans two distinct phases. During training (in simulation), the agent learns a policy through repeated interaction with the BluebirdDT environment. During inference (at trial runtime), the trained policy is fixed and applied to live traffic data.

Training (simulation)

  1. State Observation: At each time step, the agent receives a state vector containing information about aircraft in the sector (e.g. relative bearings, distances, vertical separations, time since last action, and each aircraft’s deviation from its planned route).
  2. Policy Evaluation: A neural network (the “policy”) maps the current state to a probability distribution over possible actions. The agent samples from this distribution to select an action.
  3. Reward Feedback: The agent receives reward signals reflecting multiple objectives — maintaining safe separation, minimising route deviation, and avoiding excessive or oscillatory commands. These signals shape the policy over many thousands of training episodes.

Inference (operational trial)

  1. State Observation: The agent receives the same state vector as in training, but now derived from live surveillance data (radar, flight plan updates) rather than simulation. The quality and completeness of this data stream is a critical dependency (see Deliberative Prompt 3).
  2. Policy Evaluation: The trained policy maps the current state to a recommended clearance. The policy is frozen — it does not update during the trial.
  3. Value Estimation: A separate value function (the “critic” network) estimates how much future reward the agent expects from the current state. This provides a real-time confidence signal that can be surfaced to the supervising ATCO and trial monitoring systems.
  4. Recommendation Presentation: The recommended clearance is presented to the supervising ATCO, who decides whether to issue it to the pilot, modify it, or override it entirely.

Key Technical Details

AspectDetails
Model ArchitectureFully-connected neural network with two hidden layers of 64 neurons each (approximately 8,320 parameters), using ReLU activation functions. Separate policy (actor) and value (critic) networks
Training AlgorithmProximal Policy Optimisation (PPO) — a policy gradient method that updates the policy in small, stable steps to avoid catastrophic performance changes
Training EnvironmentBluebirdDT simulation environments modelling en route sectors within the LACC, using probabilistic trajectory prediction informed by over 20 million flights of NATS operational data
State RepresentationRelative bearings, distances, vertical separations, time since last action, route deviation — a compact vector representation rather than raw radar imagery
Action SpaceHeading changes (turn left/right) and altitude changes (climb/descend), composed into operationally natural compound clearances
Reward FunctionMulti-objective: centreline tracking (route adherence), separation maintenance (safety), action damping (avoiding excessive or oscillatory commands)
Explainability MethodsUnder investigation — this is a central challenge of the case study. Candidate approaches include reward-component analysis (understanding learned behavioural priorities), value function monitoring (real-time confidence signal), policy entropy (a measure of how certain the agent is about what to do), and input perturbation methods (testing how the policy responds to changes in specific aircraft features). Attention-based architectures, which provide intrinsic interpretability by revealing which aircraft the policy attends to for each decision, are a promising future direction being explored within the Bluebird programme4
Pre-Trial QualificationMachine Basic Training (MBT) assessment: formative and summative exercises evaluated by certified ATCO assessors against competency standards for safety, controlling, planning, and coordination3

Deployment Context

  • Scope: Supervised operational trial in a designated en route sector within the London Area Control Centre (LACC)
  • Role: Decision support — the agent recommends clearances; the supervising ATCO retains full authority to accept, modify, or override any recommendation
  • Environment: Live operational airspace (single sector), with BluebirdDT used for pre-trial qualification and ongoing parallel validation
  • Users: Operational ATCOs, with trial oversight from NATS safety management and CAA observers
  • Human Oversight: The ATCO is always in command. Enhanced monitoring during the trial includes a dedicated safety officer and real-time recording of all agent recommendations and ATCO responses
  • Validation Framework: Pre-trial: MBT qualification (formative and summative assessment by certified ATCO assessors3). During trial: continuous monitoring against safety, efficiency, and procedural metrics; periodic human-in-the-loop assessment reviews
  • Trial Status: This case study describes a plausible near-future scenario. The RL agent and the systems it depends on are currently in simulation-based research and evaluation

Stakeholders

StakeholderInterestConcern
Air Traffic Controllers (ATCOs)Understanding agent recommendations well enough to make informed accept/override decisionsCannot calibrate trust if the agent’s reasoning is opaque; risk of over-reliance on recommendations or of dismissing them without engagement
ATCO Examiners and Training AssessorsAssessing whether the agent meets the same competency standards applied to human traineesMBT provides a pre-trial qualification methodology3, but how should competency be monitored on an ongoing basis during the trial?
Trial Engineering TeamDebugging failures, understanding generalisation gaps, improving reward designPolicy networks are opaque; reward-driven behaviour may satisfy metrics without genuinely safe reasoning
NATS (Air Navigation Service Provider)Safe, efficient, and operationally acceptable agent behaviour; trial evidence supporting the case for wider deploymentReputational and safety risk if agents produce unexplainable recommendations that erode controller trust
Civil Aviation Authority (CAA)Evidence that the system meets safety and explainability requirements; trial data to inform a future regulatory frameworkNo established certification pathway for RL agents in safety-critical ATC roles; CAP2970 provides emerging but incomplete guidance5
Trial Oversight BoardDefining trial entry/exit criteria, monitoring safety thresholds, deciding whether to continue, modify, or halt the trialNo precedent exists for RL agent trials in live ATC; difficulty defining appropriate go/no-go criteria
Airline Operators and PilotsSafe and predictable handling of their aircraft in controlled airspaceDuring the trial, pilots interact only with the human ATCO; pilots may be unaware of the agent’s role, raising questions about transparency
Passengers and PublicConfidence that AI involvement in ATC does not compromise safetyDifficulty understanding what role AI plays and what safeguards exist
Academic Research CommunityAdvancing understanding of explainable RL in safety-critical domainsPublication pressure may incentivise overstating agent capability relative to published evidence

Regulatory Context

No established regulatory pathway exists for deploying RL-based decision support in live UK airspace. The supervised operational trial described in this case study represents a plausible intermediate step between simulation-based research and any future certification. The trial itself is designed to generate the operational evidence — on safety, explainability, and human-machine interaction — that would be needed to inform a future regulatory framework. The following regulations and guidance documents are relevant to both the trial and any subsequent deployment.

  • CAA CAP493 (Manual of Air Traffic Services): The primary operational standard for ATC in the UK. Defines controller responsibilities, separation standards, and the clearances that controllers may issue. During the trial, all clearances are issued by the human ATCO in accordance with CAP493; the agent’s recommendations must be consistent with CAP493 procedures, and any explanation of those recommendations must be interpretable in CAP493 terms5
  • CAA CAP2970 (Artificial Intelligence in Aviation): The most directly relevant emerging framework for the trial. Addresses requirements for transparency, explainability, and human oversight of AI-based decision support in aviation6
  • EASA AI Concept Paper: The European Union Aviation Safety Agency’s framework for certifying AI in aviation. Distinguishes between Level 1 (human-assisted AI) and Level 2 (human-AI collaboration), with increasing explainability requirements at higher levels. The trial described in this case study sits at the boundary of these two levels. Relevant as a benchmark even outside EU jurisdiction7
  • EU AI Act: Classifies AI systems by risk level. ATC applications are likely to be classified as high-risk, requiring conformity assessments, documentation of training data, and ongoing monitoring. Relevant for any future European deployment
  • UK GDPR and Data Protection Act 2018: Relevant to the extent that NATS operational data (including flight tracks and controller actions) is used for training, and that trial operations may involve processing of participant and operational data
  • FAA Roadmap for Artificial Intelligence Safety Assurance: The US Federal Aviation Administration’s framework for AI assurance in aviation. While not directly applicable in the UK, it provides a comparative benchmark for international alignment8
  • NATS Internal Safety Management System: NATS’ own safety governance framework, which any operational trial would need to satisfy before reaching the regulator

Explainability Considerations

Explanation for Whom? The Audience Problem

A central challenge for explainability in this context is that different audiences need fundamentally different kinds of explanation. An ATCO supervising the agent during the trial needs rapid, operationally grounded justifications (e.g. why was this clearance recommended for this aircraft, why now, why this magnitude?). A developer investigating an unexpected recommendation needs access to internal model states, such as which reward components dominated, what the value function estimated, and where the policy was uncertain. A regulator evaluating the trial needs structured evidence linking the agent’s behaviour to established competency standards. These are not variations of the same explanation; they may require entirely different methods and representations.

A key question for the assurance case is whether a unified explanation framework can serve all audiences, or whether layered, audience-specific approaches are necessary.

Reward Function Transparency

The agent’s behaviour is shaped by a multi-objective reward function combining centreline tracking, separation maintenance, and action damping. Different weightings of these objectives produce qualitatively different agent behaviours. If the agent recommends what appears to be an unnecessarily aggressive turn, the explanation may lie not in the current state but in the reward function’s relative weighting of route adherence versus safety margin. Understanding why the agent learned to behave this way, rather than just what it recommended in this instance, requires transparency about the reward function design. This is primarily a developer-facing concern, but it also has regulatory implications: can an authority be satisfied that the reward function encodes the right priorities? The operational trial provides a valuable opportunity to validate whether reward-driven behaviour aligns with the expectations of experienced ATCOs in live traffic conditions.

Value Function as Confidence Signal

The value function (critic network) estimates the expected future reward from any given state. A sudden drop in the value estimate may signal that the agent perceives a deteriorating situation — effectively, a loss of confidence. This could be a useful real-time signal for the supervising ATCO and trial monitoring systems, but it requires careful interpretation. A low value estimate does not mean the agent will fail; it means the agent expects less reward. The relationship between value estimates and operationally meaningful concepts like “safety margin” or “sector complexity” is not straightforward and would need empirical validation before it could be relied upon for operational oversight. The trial provides an opportunity to calibrate this relationship against real operational outcomes.

Monitoring versus Controlling

Although the trial is explicitly designed as decision-support — with the ATCO retaining full authority — there is a well-documented risk that effective automation shifts human operators from active controlling to passive monitoring over time, even in advisory systems. Decades of human factors research in aviation and other safety-critical domains shows that passive monitoring degrades situational awareness, creating a risk that controllers fail to maintain the mental model of traffic that would enable them to intervene quickly and correctly if the agent produces a poor recommendation. Explanations are sometimes proposed as a solution (e.g. the agent narrates its reasoning, keeping the controller engaged). But explanations themselves impose a cognitive load, and processing them competes with the monitoring task. There is a genuine tension between providing enough explanation to maintain situational awareness and providing so much that it becomes a distraction. This tension is particularly acute in ATC, where decision timescales are measured in seconds and the operational tempo is high.

Communicability and the Voice Loop

ATC relies on voice communication between controllers and pilots. Every clearance is read back by the pilot and confirmed by the controller. This communication loop serves multiple functions:

  • it ensures the pilot has correctly understood the instruction;
  • it provides a record for other controllers monitoring the frequency; and
  • it supports the controller’s own situational awareness.

During the trial, the ATCO communicates with pilots in the normal way — the agent does not participate in the voice loop. However, the agent’s recommendations must be presented to the ATCO in a form that can be seamlessly translated into standard voice clearances. If the agent’s recommendations are expressed in terms that do not map naturally to CAP493 phraseology, or if the ATCO must mentally translate between the agent’s representation and the language of the voice loop, this creates friction that may slow decision-making or introduce errors. The communicability of agent recommendations is not only a technical challenge but a human factors and procedural one.

Governance, Documentation, and Audit

Explainability is not only about technical methods that probe the model’s internal states. It also requires that the context around the model — why it was built this way, what data shaped its behaviour, what design decisions were made and by whom — is documented to a standard that enables independent audit. For a supervised operational trial, this includes traceable design rationale for the reward function and its weightings, provenance of the training data (which sectors, which traffic conditions, which time periods were represented in the BluebirdDT simulations), version control of the trained policy, and structured records of pre-trial qualification outcomes. Without this documentation, even technically sound explainability methods cannot be placed in their proper context: knowing which features drove a recommendation is less useful if there is no record of why those features were included in the state representation or how the reward function was designed to weight them. The trial oversight board and CAA observers need access to this documentation layer to evaluate the trial independently — and it must be maintained as a living record throughout the trial, not produced after the fact.

Assurance Focus

The assurance case should demonstrate that:

The RL agent’s behaviour during the supervised operational trial is explainable to the degree necessary for operational supervision by ATCOs, developmental oversight by the trial engineering team, and regulatory evaluation by aviation authorities.

Deliberative Prompts

  1. An ATCO supervising a human trainee can ask “why did you do that?” and receive an answer in operational language. What would the equivalent interaction look like with an RL agent providing decision support, and is it achievable with current explainability methods?
  2. If the agent consistently performs well on safety and efficiency metrics but its developers cannot fully explain why it makes specific recommendations, should it be considered safe for operational use? How should the assurance case handle the gap between performance evidence and explanatory evidence?
  3. The agent was trained on clean, complete state vectors in simulation, but at trial runtime it depends on live surveillance data streams (radar feeds, flight plan updates) that may be incomplete, noisy, delayed, or temporarily unavailable. How should the assurance case address the quality and reliability of runtime data? If a data feed degrades mid-session — for example, a radar gap or a stale flight plan — the agent’s recommendations may become unreliable without the agent or the ATCO having any obvious signal that this has occurred. What continuous validation and monitoring mechanisms are needed to detect data quality issues in real time, and how should the system behave when data integrity cannot be assured? More broadly, how should the assurance case address whether explainability methods and monitoring mechanisms validated in simulation remain reliable under live operational conditions, where traffic patterns, data quality, and controller behaviour may differ from the training environment?
  4. What governance structures and documentation standards are needed to ensure that the trial can be independently audited — both during and after the trial — and that decisions about the agent’s design, training, and operational boundaries are traceable? Who has authority to halt the trial, and on what basis? How should the assurance case address the gap between the organisational oversight needed for a first-of-its-kind RL trial in live airspace and the governance frameworks that currently exist within NATS and the CAA?
  5. If the agent’s reward function encodes incorrect or incomplete priorities (e.g., underweighting a rare but critical safety scenario), the agent’s behaviour will be consistently wrong in ways that are difficult to detect through explainability methods alone. How should the assurance case address the trustworthiness of the reward function itself?
  6. Human controllers undergo a structured competency assessment where they must demonstrate and explain their decision-making. The MBT framework has been used to assess rules-based and optimisation-based agents against the same competency standards.3 Should the RL agent be assessed against this same framework, and if so, what counts as an “explanation” from a non-human agent? Does the MBT framework need adaptation for RL-specific challenges such as policy opacity and reward-driven behaviour?

Suggested Strategies

S1. Argument Over Reward-Component Attribution

Decompose the multi-objective reward into its constituent components (centreline tracking, separation maintenance, action damping) and analyse how the balance between these components shapes the policy’s learned behavioural priorities. Because RL agents optimise for cumulative future reward, reward decomposition is most useful at the policy level — understanding whether the agent has learned to systematically prioritise safety margins over route efficiency, or vice versa, and how changes to reward weightings produce qualitatively different behaviours. This provides a developer-facing explanation of why the agent behaves the way it does and supports regulatory review of whether the reward function encodes appropriate priorities.

S2. Argument Over Cognitive Compatibility and Controller Workload

Ensure that the explanations provided to the supervising ATCO support — rather than degrade — situational awareness and operational decision-making. This strategy addresses the tension between providing enough explanation for trust calibration and providing so much that it becomes a cognitive burden competing with the controller’s monitoring task. Evidence includes workload assessments, simulation trials measuring ATCO response times with and without explanations, and analysis of how explanation timing, modality, and complexity interact with the operational tempo of en route ATC and the voice communication loop.

S3. Argument Over Runtime Monitoring and Operational Envelope

Continuously monitor the agent’s confidence and input distribution during the trial, and characterise in advance where the agent behaves competently and where it does not. Pre-trial, this involves mapping the operational envelope using the MBT curriculum, structured probe scenarios, and comparative baseline testing against rules-based agents. At runtime, out-of-distribution detection, value-function monitoring, and policy entropy provide real-time signals that the agent is operating within its validated boundaries — or that it is not, triggering enhanced human oversight or reversion to manual control.

S4. Argument Over Governance and Oversight Framework

Establish that the trial has appropriate governance structures: clear roles and responsibilities, defined entry and exit criteria, escalation paths, incident response procedures, and periodic review mechanisms. This strategy argues that even where individual explanations are imperfect, the organisational and procedural wrapper around the trial ensures that problems are detected, investigated, and acted upon. Evidence includes documentation standards (design rationale, training data provenance, known limitations), monitoring of ATCO override patterns, and structured review processes involving the trial oversight board and CAA observers.

The following techniques from the TEA Techniques library  may be useful when gathering evidence for this assurance case:

  • Integrated Gradients  — Attribute the agent’s output to specific input features (e.g. which aircraft’s relative position most influenced a turn recommendation), supporting developer debugging and regulatory evidence of feature relevance
  • Partial Dependence Plots  — Visualise how the agent’s policy responds to changes in individual input features (e.g. how turn probability varies with separation distance), supporting developer understanding of learned behaviours. Note: assumes feature independence when averaging, so should be complemented with instance-level analysis where features are correlated
  • Permutation Importance  — Assess which input features the agent relies on most by measuring performance degradation when each feature is randomly shuffled; validates that the agent uses operationally meaningful features (e.g. separation distance) rather than spurious correlations
  • Contrastive Explanation Method  — Generate contrastive explanations of the form “the agent recommended turning aircraft A left because aircraft B was closing from the right; had aircraft B been 5 nautical miles further away, no turn would have been recommended”. Contrastive explanations are argued to be the most cognitively natural form of explanation and align with how ATCOs reason about traffic conflicts
  • Prototype and Criticism Models  — Identify representative traffic scenarios where the agent behaves as expected (prototypes) and scenarios where its behaviour is unexpected or atypical (criticisms), helping ATCOs build intuition about the agent’s behavioural tendencies through concrete examples
  • Human-in-the-Loop Safeguards  — Design structured checkpoints where the supervising ATCO reviews and approves agent recommendations before they are acted upon, with defined intervention criteria and escalation paths
  • Out-of-Distribution Detector for Neural Networks  — Detect traffic scenarios that fall outside the agent’s training distribution, flagging situations where the agent’s recommendations should not be trusted. Note: originally designed for supervised classification; application to RL policy networks requires adaptation of the temperature scaling approach to action probability distributions
  • Runtime Monitoring and Circuit Breakers  — Continuous surveillance of agent metrics (recommendation rates, override frequencies, value-function trends) with automated protective actions when thresholds are exceeded, supporting both real-time safety and post-trial analysis
  • Safety Envelope Testing  — Systematically evaluate the agent’s performance at operational boundaries — high traffic density, unusual geometries, degraded data quality — to characterise where it can be trusted and where enhanced human oversight is needed
  • Model Cards  — Standardised documentation recording the agent’s architecture, training process, performance characteristics, known limitations, and intended use, enabling independent audit and regulatory review
  • Model Development Audit Trails  — Immutable records of all design decisions, reward function changes, training configurations, and evaluation results throughout the agent’s development lifecycle, providing evidence of due diligence for the trial oversight board and CAA

Further Reading

Footnotes

  1. Pepper, N., Keane, A., Hodgkin, A., Gould, D., Henderson, E., Lauritsen, L., Vlahos, C., Ath, G. D., Everson, R., Cannon, R., Sierra-Castro, A., Korna, J., Carvell, B. J., & Thomas, M. (n.d.). A Probabilistic Digital Twin of UK en Route Airspace. In AIAA SCITECH 2026 Forum. American Institute of Aeronautics and Astronautics. https://doi.org/10.2514/6.2026-1794 

  2. Carvell, B., Ath, G. D., Benjamin, E., & Everson, R. (2026, January 12). Online Action-Stacking Improves Reinforcement Learning Performance for Air Traffic Control. AIAA SCITECH 2026 Forum. https://doi.org/10.2514/6.2026-2746 

  3. Carvell, B., Thomas, M., Pace, A., Dorney, C., De Ath, G., Everson, R., Pepper, N., Keane, A., Tomlinson, S., & Cannon, R. (2026). Human-in-the-Loop Testing of AI Agents for Air Traffic Control with a Regulated Assessment Framework. arXiv:2601.04288. 2 3 4 5

  4. Brittain, M. W., Alvarez, L. E., & Breeden, K. (2024). Improving Autonomous Separation Assurance through Distributed Reinforcement Learning with Attention Networks. In Proceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24), 22857–22863. https://doi.org/10.1609/aaai.v38i21.30321 

  5. https://www.caa.co.uk/data-and-publications/publications/documents/content/cap-493/  2

  6. https://www.caa.co.uk/data-and-publications/publications/documents/content/cap2970/ 

  7. https://www.easa.europa.eu/en/document-library/general-publications/easa-artificial-intelligence-concept-paper-issue-2 

  8. https://www.faa.gov/media/82891