Recently, Artificial intelligence (AI) algorithms have received increasable interest in various application domains including in Air Transportation Management (ATM). Different AI in particular Machine Learning (ML) algorithms are used to provide decision support in autonomous decision-making tasks in the ATM domain e.g., predicting air transportation traffic and optimizing traffic flows. However, most of the time these automated systems are not accepted or trusted by the intended users as the decisions provided by AI are often opaque, non-intuitive and not understandable by human operators. Safety is the major pillar to air traffic management, and no black box process can be inserted in a decision-making process when human life is involved. To address this challenge related to transparency of the automated system in the ATM domain, we investigated AI methods in predicting air transportation traffic conflict and optimizing traffic flows based on the domain of Explainable Artificial Intelligence (XAI). Here, AI models’ explainability in terms of understanding a decision i.e., post hoc interpretability and understanding how the model works i.e., transparency can be provided for air traffic controllers. In this paper, we report our research directions and our findings to support better decision making with AI algorithms with extended transparency.
Dettaglio pubblicazione
2022, Transportation Research Procedia, Pages 270-278 (volume: 66)
Usage of more transparent and explainable conflict resolution algorithm: Air traffic controller feedback (04b Atto di convegno in volume)
Hurter C., Degas A., Guibert A., Durand N., Ferreira A., Cavagnetto N., Islam M. R., Barua S., Ahmed M. U., Begum S., Bonelli S., Cartocci G., Flumeri G. D., Borghini G., Babiloni F., Aricó P.
keywords