[1]LUO Yuan,TONG Kaiguo,ZHANG Yi,et al.Sound source separation of a multi voice environment based on human ear listening properties[J].CAAI Transactions on Intelligent Systems,2012,7(2):121-128.
Copy
CAAI Transactions on Intelligent Systems[ISSN 1673-4785/CN 23-1538/TP] Volume:
7
Number of periods:
2012 2
Page number:
121-128
Column:
学术论文—机器感知与模式识别
Public date:
2012-04-25
- Title:
-
Sound source separation of a multi voice environment based on human ear listening properties
- Author(s):
-
LUO Yuan; TONG Kaiguo; ZHANG Yi; XING Wuchao; CHEN Kai; CHEN Hongsong; HE Chunjiang; CHEN Jun
-
Research Center of Intelligent System and Robot, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
-
- Keywords:
-
multivoice source environment; human ear listening properties; interaural time difference; interaural level difference; sound source separation
- CLC:
-
TP311
- DOI:
-
-
- Abstract:
-
Inspired by acoustics, an integrated voice separation model simulating the central auditory system was established to process a voice by imitating the listening properties of human ears. First, multispectral analysis of voice signals was carried out by a peripheral auditory model. Next, a coincidence neuron model was established to extract the features of voice signals. Last, the voices were separated in the cell model of the brain inferior colliculus. Compared to the majority of speech recognition models that can only be used in a single sound source and lownoise environment, this model is a good choice. Experimental results show that the model can separate voices in a multisound source environment, thus having a high robustness. With further research, speech separation models based on human ear listening properties will have a wide range of applications.