2025
ACHILLES: A Machine Learning Framework for Explainable and Generalized Automotive Intrusion Detection System
Authorea Preprints
An explainable and generalized automotive intrusion detection framework that directly addresses trustworthiness and explainability in vehicle cybersecurity.
2025
Enhancing Intrusion Detection in CPS and IIoT with Lightweight Explainable AI Models
WFCS 2025
Explainable AI for industrial and cyber-physical intrusion detection, directly aligned with trustworthy and interpretable AI for mission-critical systems.
2025
Evaluating Trust-Related Principles in an Implemented Distributed Edge AI System
SNCNW 2025
A direct trust-oriented paper that operationalizes and evaluates trust principles in a real distributed edge AI implementation.
2024
IIoT Intrusion Detection Using Lightweight Deep Learning Models on Edge Devices
WFCS 2024
A precursor paper that supports the later trustworthy and explainable intrusion-detection work for industrial edge environments.
2022
Experimental Analysis of Trustworthy In-Vehicle Intrusion Detection System Using eXplainable Artificial Intelligence (XAI)
IEEE Access 10
The clearest direct match for the website's Trustworthy AI focus, combining trustworthy intrusion detection with explainability.