Published March 2018 | Version v1
Book section

Deep neural network based multichannel audio source separation

Contributors

Others:

Description

This chapter presents a multichannel audio source separation framework where deep neural networks (DNNs) are used to model the source spectra and combined with the classical multichannel Gaussian model to exploit the spatial information. The parameters are estimated in an iterative expectation-maximization (EM) fashion and used to derive a multichannel Wiener filter. Different design choices and their impact on the performance are discussed. They include the cost functions for DNN training, the number of parameter updates, the use of multiple DNNs, and the use of weighted parameter updates. Finally, we present its application to a speech enhancement task and a music separation task. The experimental results show the benefit of the multichannel DNN-based approach over a single-channel DNN-based approach and the multichannel nonnegative matrix factorization based iterative EM framework.

Abstract

International audience

Additional details

Identifiers

URL
https://inria.hal.science/hal-01633858
URN
urn:oai:HAL:hal-01633858v1

Origin repository

Origin repository
UNICA