TY - GEN
T1 - Complex NMF with the generalized Kullback-Leibler divergence
AU - Kameoka, Hirokazu
AU - Kagami, Hideaki
AU - Yukawa, Masahiro
N1 - Funding Information:
This work was supported by JSPS KAKENHI Grant Numbers 26730100, 15K06081, 15K13986, 15H02757.
Publisher Copyright:
© 2017 IEEE.
PY - 2017/6/16
Y1 - 2017/6/16
N2 - We previously introduced a phase-aware variant of the non-negative matrix factorization (NMF) approach for audio source separation, which we call the 'Complex NMF (CNMF).' This approach makes it possible to realize NMF-like signal decompositions in the complex time-frequency domain. One limitation of the CNMF framework is that the divergence measure is limited to only the Euclidean distance. Some previous studies have revealed that for source separation tasks with NMF, the generalized Kullback-Leibler (KL) divergence tends to yield higher accuracy than when using other divergence measures. This motivated us to believe that CNMF could achieve even greater source separation accuracy if we could derive an algorithm for a KL divergence counterpart of CNMF. In this paper, we start by defining the notion of the 'dual' form of the CNMF formulation, derived from the original Euclidean CNMF, and show that a KL divergence counterpart of CNMF can be developed based on this dual formulation. We call this 'KL-CNMF'. We further derive a convergence-guaranteed iterative algorithm for KL-CNMF based on a majorization-minimization scheme. The source separation experiments revealed that the proposed KL-CNMF yielded higher accuracy than the Euclidean CNMF and NMF with varying divergences.
AB - We previously introduced a phase-aware variant of the non-negative matrix factorization (NMF) approach for audio source separation, which we call the 'Complex NMF (CNMF).' This approach makes it possible to realize NMF-like signal decompositions in the complex time-frequency domain. One limitation of the CNMF framework is that the divergence measure is limited to only the Euclidean distance. Some previous studies have revealed that for source separation tasks with NMF, the generalized Kullback-Leibler (KL) divergence tends to yield higher accuracy than when using other divergence measures. This motivated us to believe that CNMF could achieve even greater source separation accuracy if we could derive an algorithm for a KL divergence counterpart of CNMF. In this paper, we start by defining the notion of the 'dual' form of the CNMF formulation, derived from the original Euclidean CNMF, and show that a KL divergence counterpart of CNMF can be developed based on this dual formulation. We call this 'KL-CNMF'. We further derive a convergence-guaranteed iterative algorithm for KL-CNMF based on a majorization-minimization scheme. The source separation experiments revealed that the proposed KL-CNMF yielded higher accuracy than the Euclidean CNMF and NMF with varying divergences.
KW - Audio source separation
KW - Complex NMF
KW - generalized Kullback-Leibler (KL) divergence
KW - non-negative matrix factorization (NMF)
UR - http://www.scopus.com/inward/record.url?scp=85023775492&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85023775492&partnerID=8YFLogxK
U2 - 10.1109/ICASSP.2017.7952117
DO - 10.1109/ICASSP.2017.7952117
M3 - Conference contribution
AN - SCOPUS:85023775492
T3 - ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
SP - 56
EP - 60
BT - 2017 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2017 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2017 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2017
Y2 - 5 March 2017 through 9 March 2017
ER -