A Sparse Spatiotemporal Transformer for Detecting Driver Distracted Behaviors

2023-01-0835

04/11/2023

Features
Event
WCX SAE World Congress Experience
Authors Abstract
Content
At present, the development of autonomous driving technology is still immature, and there is still a long way until fully driverless vehicles. Therefore, the state of the driver is still an important factor affecting traffic safety, and it is of great significance to detect the driver’s distracted behavior. In the task of driver distracted behavior detection, some characteristics of driver behavior in the cockpit can be further utilized to improve the detection performance. Compared with general human behaviors, driving behaviors are confined to enclosed space and are far less diverse. With this in mind, we propose a sparse spatiotemporal transformer which extracts local spatiotemporal features by segmenting the video at the low level of the model, and filters out local key spatiotemporal information associated with larger attention values based on the attention map in the middle layer, so as to enhance the high-level global semantic features. Experiments are conducted on a public driver behavior detection dataset (Drive&Act), and the generalization ability of the proposal is evaluated with a dataset collected. Results show that the sparse spatiotemporal transformer devised in this study can obtain robust global semantic features via retaining key local spatiotemporal information while reducing the computational burden, and therefore achieves a high accuracy for the driver distracted behavior detection.
Meta TagsDetails
DOI
https://doi.org/10.4271/2023-01-0835
Pages
10
Citation
Wang, P., Yin, Z., Nie, L., and Zhai, X., "A Sparse Spatiotemporal Transformer for Detecting Driver Distracted Behaviors," SAE Technical Paper 2023-01-0835, 2023, https://doi.org/10.4271/2023-01-0835.
Additional Details
Publisher
Published
Apr 11, 2023
Product Code
2023-01-0835
Content Type
Technical Paper
Language
English