You are required to read and agree to the below before accessing a full-text version of an article in the IDE article repository.

The full-text document you are about to access is subject to national and international copyright laws. In most cases (but not necessarily all) the consequence is that personal use is allowed given that the copyright owner is duly acknowledged and respected. All other use (typically) require an explicit permission (often in writing) by the copyright owner.

For the reports in this repository we specifically note that

  • the use of articles under IEEE copyright is governed by the IEEE copyright policy (available at
  • the use of articles under ACM copyright is governed by the ACM copyright policy (available at
  • technical reports and other articles issued by M‰lardalen University is free for personal use. For other use, the explicit consent of the authors is required
  • in other cases, please contact the copyright owner for detailed information

By accepting I agree to acknowledge and respect the rights of the copyright owner of the document I am about to access.

If you are in doubt, feel free to contact

Usage of more transparent and explainable conflict resolution algorithm: air traffic controller feedback



Christophe Hurter , Augustin Degas , Arnaud Guibert , Nicolas Durand , Mir Riyanul Islam, Shaibal Barua, Mobyen Uddin Ahmed, Shahina Begum, Stefano Bonelli , Giulia Cartocci , Gianluca Di Flumeri , Gianluca Borghini , Pietro Aricò , Fabio Babiloni

Publication Type:

Conference/Workshop Paper


Conference of the European Association for Aviation Psychology



Recently, Artificial intelligence (AI) algorithms have received increasable interest in various application domains including in Air Transportation Management (ATM). Different AI in particular Machine Learning (ML) algorithms are used to provide decision support in autonomous decision-making tasks in the ATM domain e.g., predicting air transportation traffic and optimizing traffic flows. However, most of the time these automated systems are not accepted or trusted by the intended users as the decisions provided by AI are often opaque, non-intuitive and not understandable by human operators. Safety is the major pillar to air traffic management, and no black box process can be inserted in a decision-making process when human life is involved. To address this challenge related to transparency of the automated system in the ATM domain, we investigated AI methods in predicting air transportation traffic conflict and optimizing traffic flows based on the domain of Explainable Artificial Intelligence (XAI). Here, AI models’ explainability in terms of understanding a decision i.e., post hoc interpretability and understanding how the model works i.e., transparency can be provided for air traffic controllers. In this paper, we report our research directions and our findings to support better decision making with AI algorithms with extended transparency.


author = {Christophe Hurter and Augustin Degas and Arnaud Guibert and Nicolas Durand and Mir Riyanul Islam and Shaibal Barua and Mobyen Uddin Ahmed and Shahina Begum and Stefano Bonelli and Giulia Cartocci and Gianluca Di Flumeri and Gianluca Borghini and Pietro Aric{\`o} and Fabio Babiloni},
title = {Usage of more transparent and explainable conflict resolution algorithm: air traffic controller feedback},
editor = {Elsevier},
volume = {62},
month = {October},
year = {2022},
booktitle = {Conference of the European Association for Aviation Psychology },
url = {}