Safety Implications of Explainable Artificial Intelligence in End-to-End Autonomous Driving
CoRR(2024)
Abstract
The end-to-end learning pipeline is gradually creating a paradigm shift in
the ongoing development of highly autonomous vehicles, largely due to advances
in deep learning, the availability of large-scale training datasets, and
improvements in integrated sensor devices. However, a lack of interpretability
in real-time decisions with contemporary learning methods impedes user trust
and attenuates the widespread deployment and commercialization of such
vehicles. Moreover, the issue is exacerbated when these cars are involved in or
cause traffic accidents. Such drawback raises serious safety concerns from
societal and legal perspectives. Consequently, explainability in end-to-end
autonomous driving is essential to enable the safety of vehicular automation.
However, the safety and explainability aspects of autonomous driving have
generally been investigated disjointly by researchers in today's state of the
art. In this paper, we aim to bridge the gaps between these topics and seek to
answer the following research question: When and how can explanations improve
safety of autonomous driving? In this regard, we first revisit established
safety and state-of-the-art explainability techniques in autonomous driving.
Furthermore, we present three critical case studies and show the pivotal role
of explanations in enhancing self-driving safety. Finally, we describe our
empirical investigation and reveal potential value, limitations, and caveats
with practical explainable AI methods on their role of assuring safety and
transparency for vehicle autonomy.
MoreTranslated text
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined