@inproceedings{fe85c385135c42f5b740217a763deb3e,
title = "Direction-Aware Speaker Beam for Multi-Channel Speaker Extraction",
abstract = "SpeakerBeam is a state-of-the-art method for extracting a speech signal of target speaker from a mixture using an adaption utterance. The existing multi-channel SpeakerBeam utilizes the spectral features of the signals with the ignorance of the spatial discriminability of the multi-channel processing. In this paper, we tightly integrate spectral and spatial information for target speaker extraction. In the proposed scheme, a multi-channel mixture signal is firstly filtered into a set of beamformed signals using fixed beam patterns. An attention network is then designed to identify the direction of the target speaker and to combine the beamformed signals into an enhanced signal dominated by the target speaker energy. Further, SpeakerBeam inputs the enhanced signal and outputs the mask of the target speaker. Finally, the attention network and SpeakerBeam are jointly trained. Experimental results demonstrate that the proposed scheme largely improves the existing multi-channel SpeakerBeam in low signal-to-interference ratio or same-gender scenarios.",
keywords = "Fixed beamforming, Jointly training, Multi-channel signal processing, Speaker extraction",
author = "Guanjun Li and Shan Liang and Shuai Nie and Wenju Liu and Meng Yu and Lianwu Chen and Shouye Peng and Changliang Li",
note = "Publisher Copyright: Copyright {\textcopyright} 2019 ISCA; 20th Annual Conference of the International Speech Communication Association: Crossroads of Speech and Language, INTERSPEECH 2019 ; Conference date: 15-09-2019 Through 19-09-2019",
year = "2019",
doi = "10.21437/Interspeech.2019-1474",
language = "English",
volume = "2019-September",
series = "Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH",
pages = "2713--2717",
booktitle = "20th Annual Conference of the International Speech Communication Association: Crossroads of Speech and Language, INTERSPEECH 2019",
}