Abstract
As Connected Vehicles (CVs) increasingly depend on deep learning-based Intrusion Detection Systems (IDS), the need for models that are both accurate and interpretable has become essential. This paper explores the use of Explainable Artificial Intelligence (XAI) techniques to improve the transparency of a Convolutional Neural Network (CNN)- based IDS trained on the CICIoV2024 dataset. We evaluate four widely adopted XAI methods—SHAP, LIME, Integrated Gradients, and Grad- CAM—by examining their ability to explain predictions across various cyberattack scenarios, including spoofing and denial-of-service (DoS) at- tacks on CAN bus traffic. Our results show that SHAP and Integrated Gradients effectively highlight key features, with SHAP assigning up to
0.16 contribution to specific class decisions. LIME provided near-perfect agreement with the model’s predictions in local explanations, while Grad- CAM offered visual insights aligned with convolutional activations. The CNN model achieved 98.3% classification accuracy on the CICIoV2024 test set. These findings offer practical recommendations for selecting XAI tools in automotive cybersecurity and contribute to building trustworthy, explainable IDS for intelligent transportation systems.
0.16 contribution to specific class decisions. LIME provided near-perfect agreement with the model’s predictions in local explanations, while Grad- CAM offered visual insights aligned with convolutional activations. The CNN model achieved 98.3% classification accuracy on the CICIoV2024 test set. These findings offer practical recommendations for selecting XAI tools in automotive cybersecurity and contribute to building trustworthy, explainable IDS for intelligent transportation systems.
| Original language | English |
|---|---|
| Title of host publication | Knowledge Management and Acquisition for Intelligent Systems |
| Subtitle of host publication | 21st Principle and Practice of Data and Knowledge Acquisition Workshop, PKAW 2025, Wellington, New Zealand, November 17–18, 2025, Proceedings |
| Editors | Shiqing Wu, Weihua Li, Xiwei Xu, Yanbin Liu |
| Publisher | Springer Nature |
| Pages | 162-176 |
| ISBN (Electronic) | 9789819545759 |
| ISBN (Print) | 9789819545742 |
| DOIs | |
| Publication status | Published - 10 Nov 2025 |
| Event | 21st Principle and Practice of Data and Knowledge Acquisition Workshop: PKAW 2025 - Wellington, New Zealand Duration: 17 Nov 2025 → 18 Nov 2025 |
Publication series
| Name | Communications in Computer and Information Science |
|---|---|
| Publisher | Springer Nature |
| Volume | 2768 |
| ISSN (Print) | 1865-0929 |
| ISSN (Electronic) | 1865-0937 |
Conference
| Conference | 21st Principle and Practice of Data and Knowledge Acquisition Workshop |
|---|---|
| Country/Territory | New Zealand |
| City | Wellington |
| Period | 17/11/25 → 18/11/25 |
Keywords
- Explainable AI
- Intrusion Detection
- Connected Vehicles
- Convolutional Neural Networks