Invited Presentation Padfield Best Young Paper Award at ERF 2024: Multisensory Integration Model for Combined Visual-Haptic 2D Localization
F-0081-2025-0154
5/20/2025
- Content
-
This paper describes a combined visual and haptic localization experiment that addresses the area of multi-modal cueing. The aim of the present investigation is to characterize accuracy and precision of tactile cue-ing in the peri-personal space (PPS), the space around the body in which sensory information is perceived as meaningful (Ref. 1). Outcomes of the unimodal (visual and haptic) and multi-modal (combined visual-haptic) localizations are used to make predictions about the multimodal integrative phenomenon. In the localization experiment, participants are presented with visual, haptic, or multimodal target cues using the body-centered reference frame and are instructed to indicate the corresponding hypothetical target location in space using a mouse pointer in an open-loop feedback condition.
- Citation
- Fischer, M., Saetti, U., Godfroy-Cooper, M., Fischer, D., et al., "Invited Presentation Padfield Best Young Paper Award at ERF 2024: Multisensory Integration Model for Combined Visual-Haptic 2D Localization," Vertical Flight Society 81st Annual Forum and Technology Display, Virginia Beach, Virginia, May 20, 2025, https://doi.org/10.4050/F-0081-2025-0154.