Monaural voiced speech separation with multipitch tracking

Wei Jiang, Pengfei Hu, Shan Liang, Wenju Liu*, Zhanlei Yang

*Corresponding author for this work

Research output: Chapter in Book or Report/Conference proceedingConference Proceedingpeer-review

Abstract

Separating voiced speech from its mixtures with interferences in monaural condition is not only an important but also challenging task. As multipitch tracking can enable much better performance of speech separation for CASA systems, we propose a new multipitch determination algorithm, which can be used under various kinds of noise conditions. In the process of multipitch estimation, a new representation method is utilized together with maximum support constraint and harmonic completeness constraint. The proposed approach can reliably detect up to two pitches in each frame. Sequential grouping is performed based on a new target pitch tracking strategy. System evaluations show that our algorithm leads to significantly better speech separation results than previous ones.

Original languageEnglish
Title of host publicationPattern Recognition - Chinese Conference, CCPR 2012, Proceedings
Pages564-571
Number of pages8
DOIs
Publication statusPublished - 2012
Externally publishedYes
Event2012 5th Chinese Conference on Pattern Recognition, CCPR 2012 - Beijing, China
Duration: 24 Sept 201226 Sept 2012

Publication series

NameCommunications in Computer and Information Science
Volume321 CCIS
ISSN (Print)1865-0929

Conference

Conference2012 5th Chinese Conference on Pattern Recognition, CCPR 2012
Country/TerritoryChina
CityBeijing
Period24/09/1226/09/12

Keywords

  • computational auditory scene analysis
  • multipitch determination
  • voiced speech separation

Fingerprint

Dive into the research topics of 'Monaural voiced speech separation with multipitch tracking'. Together they form a unique fingerprint.

Cite this