This content is not included in your SAE MOBILUS subscription, or you are not logged in.
In-Vehicle Occupant Head Tracking Using aLow-Cost Depth Camera
ISSN: 0148-7191, e-ISSN: 2688-3627
Published April 03, 2018 by SAE International in United States
This content contains downloadable datasetsAnnotation ability available
Analyzing dynamic postures of vehicle occupants in various situations is valuable for improving occupant accommodation and safety. Accurate tracking of an occupant’s head is of particular importance because the head has a large range of motion, controls gaze, and may require special protection in dynamic events including crashes. Previous vehicle occupant posture studies have primarily used marker-based optical motion capture systems or multiple video cameras for tracking facial features or markers on the head. However, the former approach has limitations for collecting on-road data, and the latter is limited by requiring intensive manual postprocessing to obtain suitable accuracy. This paper presents an automated on-road head tracking method using a single Microsoft Kinect V2 sensor, which uses a time-of-flight measurement principle to obtain a 3D point cloud representing objects in the scene at approximately 30 Hz. Vehicle passenger motions were recorded during hard braking and rapid lane changes. The dynamic head orientation and location data were obtained by fitting a subject-specific 3d head model to the depth data from each frame. Results were validated using a marker-based tracking system based on video images from multiple views. The results showed that the proposed method and system provides efficient and accurate in-vehicle head tracking using a single low-cost depth camera. Extensions of this method have broad applications for study of vehicle occupant dynamics, and with advances in technology may be applicable to routine use in production vehicles.
CitationPark, B., Jones, M., Miller, C., Hallman, J. et al., "In-Vehicle Occupant Head Tracking Using aLow-Cost Depth Camera," SAE Technical Paper 2018-01-1172, 2018, https://doi.org/10.4271/2018-01-1172.
Data Sets - Support Documents
|[Unnamed Dataset 1]|
- Cicchino, J., “Effectiveness of Forward Collision Warning and Autonomous Emergency Braking Systems in Reducing Front-to-Rear Crash Rates,” AAP, 2017.
- Gottfried, J.-M, Nair, R., Meister, S., Garbe, C.S. et al. , “Time of Flight Motion Compensation Revisited.” In Image Processing (ICIP), 2014 IEEE International Conference on, pp. 5861-5865. IEEE, 2014.
- Huber, P., Kirschbichler, S., Prüggler, A., and Steidl, T., “Passenger Kinematics in Braking, Lane Change and Oblique Driving Maneuvers”. IRCOBI, 2015.
- Jones, L.H.M., Miller, S.C., Ebert, S., Bonifas A., et al. , “Tracking Occupant Head Movements during Braking Events”. Proceedings of the 4th International Digital Human Modeling Conference. Montreal, Canada, 2017.
- Kirschbichler, S., Huber, P., Prüggler, A., Steidl, T., et al. , “Factors Influencing Occupant Kinematics During Braking and Lane Change Maneuvers in a Passenger Vehicle. IRCOBI, 2014.
- Kusano & Gabler , “Comparison of Expected Crash and Injury Reduction from Production Forward Collision and Lane Departure Warning Systems”. TIP 2015.
- Ólafsdóttir, J.M., Östh, J.K., Davidsson, J., and Brolin, K.B., “Passenger Kinematics and Muscle Responses in Autonomous Braking Events with Standard and Reversible Pre-tensioned Restraints”. IRCOBI, 2013.
- Östh, J., Ólafsdóttir, J.M., Davidsson, J., and Brolin, K., “Driver Kinematic and Muscle Responses in Braking Events with Standard and Reversible Pre-tensioned Restraints: Validation Data for Human Models. Stapp Car Crash Journal, 57, 2013.
- Paul, V. and Jones, M., “Rapid Object Detection Using a Boosted Cascade of Simple Features.” In Computer Vision and Pattern Recognition, 2001. CVPR 2001. Proceedings of the 2001 IEEE Computer Society Conference on, Vol. 1, pp. I-I. IEEE, 2001.
- Park, B.-K., Lumeng, J.C., Lumeng, C.N., Ebert, S.M. et al. , “Child Body Shape Measurement Using Depth Cameras and a Statistical Body Shape Model,” Ergonomics 58(2):301-309, 2015.