@article{Elhilali2008JASA,
abstract = {Sound systems and speech technologies can benefit greatly from a deeper understanding of how the auditory system, and particularly the auditory cortex, is able to parse complex acoustic scenes into meaningful auditory objects and streams under adverse conditions. In the current work, a biologically plausible model of this process is presented, where the role of cortical mechanisms in organizing complex auditory scenes is explored. The model consists of two stages: (i) a feature analysis stage that maps the acoustic input into a multidimensional cortical representation and (ii) an integrative stage that recursively builds up expectations of how streams evolve over time and reconciles its predictions with the incoming sensory input by sorting it into different clusters. This approach yields a robust computational scheme for speaker separation under conditions of speech or music interference. The model can also emulate the archetypal streaming percepts of tonal stimuli that have long been tested in human subjects. The implications of this model are discussed with respect to the physiological correlates of streaming in the cortex as well as the role of attention and other top-down influences in guiding sound organization.},
author = {Elhilali, Mounya and Shamma, Shihab A},
doi = {10.1121/1.3001672},
isbn = {1520-8524(ELECTRONIC)},
issn = {1520-8524},
journal = {The Journal of the Acoustical Society of America},
number = {6},
pages = {3751--71},
title = {{A cocktail party with a cortical twist: how cortical mechanisms contribute to sound segregation}},
url = {http://www.ncbi.nlm.nih.gov/pubmed/19206802 http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=PMC2676630},
volume = {124},
year = {2008}
}