Explainable AI for Autonomous Decision-Making in Unmanned Systems: A Systematic Review

C. S. Madhumathi, Hoeun Lee

Abstract


The increasing deployment of AI-driven autonomous unmanned systems in safety-critical applications demands transparency in decision-making processes to enable human oversight, facilitate debugging, ensure regulatory compliance, and build operator trust. This systematic review comprehensively examines explainable AI (XAI) methods for autonomous decision-making across perception, planning, and control functions in unmanned systems. We categorize XAI approaches into three paradigms: post-hoc explanation methods that interpret black-box models through saliency maps, attention visualization, and counterfactual analysis; intrinsically interpretable models including decision trees, rule-based systems, and linear models with inherent transparency; and hybrid architectures combining neural network performance with symbolic reasoning for explainability. The review analyzes explanation modalities spanning visual explanations highlighting image regions influencing perception decisions, textual explanations providing natural language justifications for actions, and interactive explanations enabling operator queries about system reasoning. Particular emphasis is placed on domain-specific requirements: real-time explanation generation for time-critical decisions, explanation fidelity ensuring accurate representation of actual decision processes, and explanation utility measured by operator comprehension and trust calibration. We examine evaluation methodologies for XAI including human subject studies assessing explanation quality, computational metrics measuring explanation consistency, and task-based evaluations quantifying performance improvements from explanations. Application-specific implementations are analyzed for autonomous vehicles explaining trajectory decisions, UAVs justifying target classifications, and multi-agent systems clarifying coordination strategies. The review identifies critical research gaps including lack of standardized evaluation frameworks, limited investigation of explanation effects on long-term operator performance, and insufficient attention to explanation security against adversarial manipulation.

Keywords


explainable AI, autonomous systems, interpretability, decision-making transparency.

References


Biran, Or, and Courtenay Cotton. “Explanation and Justification in Machine Learning: A Survey”. In IJCAI-17 Workshop on Explainable AI (XAI), 8:8–13, 2017. Available at https://www.cs.columbia.edu/~orb/papers/xai_survey_paper_2017.pdf.

Carvalho, Diogo V., Eduardo M. Pereira, and Jaime S. Cardoso. “Machine Learning Interpretability: A Survey on Methods and Metrics”. Electronics 8, no. 8 (2019). https://doi.org/10.3390/electronics8080832.

Doshi-Velez, Finale, and Been Kim. “Towards A Rigorous Science of Interpretable Machine Learning”. arXiv [Stat.ML], 2017. arXiv. Available at http://arxiv.org/abs/1702.08608.

Guidotti, Riccardo, Anna Monreale, Salvatore Ruggieri, Franco Turini, Fosca Giannotti, and Dino Pedreschi. “A Survey of Methods for Explaining Black Box Models”. ACM Comput. Surv. 51, no. 5 (August 2018). https://doi.org/10.1145/3236009.

Lipton, Zachary C. “The Mythos of Model Interpretability”. Commun. ACM 61, no. 10 (September 2018): 36–43. https://doi.org/10.1145/3233231.

Lundberg, Scott M., and Su-In Lee. “A Unified Approach to Interpreting Model Predictions”. In Advances in Neural Information Processing Systems, edited by I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, Vol. 30. Curran Associates, Inc., 2017. Available at https://proceedings.neurips.cc/paper_files/paper/2017/file/8a20a8621978632d76c43dfd28b67767-Paper.pdf.

Murdoch, W. James, Chandan Singh, Karl Kumbier, Reza Abbasi-Asl, and Bin Yu. “Definitions, Methods, and Applications in Interpretable Machine Learning”. Proceedings of the National Academy of Sciences 116, no. 44 (2019): 22071–80. https://doi.org/10.1073/pnas.1900654116.

Ribeiro, Marco Tulio, Sameer Singh, and Carlos Guestrin. ““Why Should I Trust You?”: Explaining the Predictions of Any Classifier”. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 1135–44. KDD ‘16. New York, NY, USA: Association for Computing Machinery, 2016. https://doi.org/10.1145/2939672.2939778.

Rosenfeld, Avi, and Ariella Richardson. “Explainability in Human–Agent Systems”. Autonomous Agents and Multi-Agent Systems 33, no. 6 (1 November 2019): 673–705. https://doi.org/10.1007/s10458-019-09408-y.

Schmidt, Philipp, and Felix Biessmann. “Quantifying Interpretability and Trust in Machine Learning Systems”. arXiv [Cs.LG], 2019. arXiv. Available at http://arxiv.org/abs/1901.08558.


Refbacks

  • There are currently no refbacks.




Creative Commons License
This work is licensed under a Creative Commons Attribution 3.0 License.