z-logo
open-access-imgOpen Access
Robust Non‐negative Matrix Factorization with β ‐Divergence for Speech Separation
Author(s) -
Li Yinan,
Zhang Xiongwei,
Sun Meng
Publication year - 2017
Publication title -
etri journal
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.295
H-Index - 46
eISSN - 2233-7326
pISSN - 1225-6463
DOI - 10.4218/etrij.17.0115.0122
Subject(s) - non negative matrix factorization , matrix decomposition , divergence (linguistics) , matrix (chemical analysis) , rank (graph theory) , multiplicative function , computer science , algorithm , dependency (uml) , noise (video) , sparse matrix , factorization , pattern recognition (psychology) , mathematics , artificial intelligence , combinatorics , mathematical analysis , linguistics , eigenvalues and eigenvectors , physics , philosophy , materials science , image (mathematics) , composite material , quantum mechanics , gaussian
This paper addresses the problem of unsupervised speech separation based on robust non‐negative matrix factorization (RNMF) with β ‐divergence, when neither speech nor noise training data is available beforehand. We propose a robust version of non‐negative matrix factorization, inspired by the recently developed sparse and low‐rank decomposition, in which the data matrix is decomposed into the sum of a low‐rank matrix and a sparse matrix. Efficient multiplicative update rules to minimize the β ‐divergence‐based cost function are derived. A convolutional extension of the proposed algorithm is also proposed, which considers the time dependency of the non‐negative noise bases. Experimental speech separation results show that the proposed convolutional RNMF successfully separates the repeating time‐varying spectral structures from the magnitude spectrum of the mixture, and does so without any prior training.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here