TY - JOUR
T1 - Active binaural localization of multiple sound sources
AU - Zhong, Xuan
AU - Sun, Liang
AU - Yost, William
N1 - Funding Information:
This research was supported by a grant from the Air Force Office of Scientific Research (AFOSR) ( FA9550-12-1-0312 ) awarded to William Yost. We would like to thank Yi Zhou, Steve Helms-Tillery and Michael Dorman for their assistance in this research effort. Pilot studies were presented at the 169th Meeting of the Acoustical Society of America in May 2015 [51–54] .
Publisher Copyright:
© 2016 Elsevier B.V.
PY - 2016/11
Y1 - 2016/11
N2 - Sound source localization serves as a significant capability of autonomous robots that conduct missions such as search and rescue, and target tracking in challenging environments. However, localization of multiple sound sources and static sound source tracking in self-motion are both difficult tasks, especially when the number of sound sources or reflections increase. This study presents two robotic hearing approaches based on a human perception model (Wallach, 1939) that combines interaural time difference (ITD) and head turn motion data to locate sound sources. The first method uses a fitting-based approach to recognize the changing trends of the cross-correlation function of binaural inputs. The effectiveness of this method was validated using data collected from a two-microphone array rotating in a non-anechoic environment, and the experiments reveal its ability to separate and localize up to three sound sources of the same spectral content (white noise) at different azimuth and elevation angles. The second method uses an extended Kalman filter (EKF) that estimates the orientation of a sound source by fusing the robot's self-motion and ITD data to reduce the localization errors recursively. This method requires limited memory resources and is able to keep tracking the relative position change of a number of static sources when the robot moves. In the experiments, up to three sources can be tracked simultaneously with a two-microphone array.
AB - Sound source localization serves as a significant capability of autonomous robots that conduct missions such as search and rescue, and target tracking in challenging environments. However, localization of multiple sound sources and static sound source tracking in self-motion are both difficult tasks, especially when the number of sound sources or reflections increase. This study presents two robotic hearing approaches based on a human perception model (Wallach, 1939) that combines interaural time difference (ITD) and head turn motion data to locate sound sources. The first method uses a fitting-based approach to recognize the changing trends of the cross-correlation function of binaural inputs. The effectiveness of this method was validated using data collected from a two-microphone array rotating in a non-anechoic environment, and the experiments reveal its ability to separate and localize up to three sound sources of the same spectral content (white noise) at different azimuth and elevation angles. The second method uses an extended Kalman filter (EKF) that estimates the orientation of a sound source by fusing the robot's self-motion and ITD data to reduce the localization errors recursively. This method requires limited memory resources and is able to keep tracking the relative position change of a number of static sources when the robot moves. In the experiments, up to three sources can be tracked simultaneously with a two-microphone array.
KW - Binaural localization
KW - Extended Kalman filter (EKF)
KW - Sound source localization
KW - Spatial hearing
UR - http://www.scopus.com/inward/record.url?scp=84994430021&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84994430021&partnerID=8YFLogxK
U2 - 10.1016/j.robot.2016.07.008
DO - 10.1016/j.robot.2016.07.008
M3 - Article
AN - SCOPUS:84994430021
SN - 0921-8890
VL - 85
SP - 83
EP - 92
JO - Robotics and Autonomous Systems
JF - Robotics and Autonomous Systems
ER -