菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-14
📄 Abstract - Audio Source Separation in Reverberant Environments using $β$-divergence based Nonnegative Factorization

In Gaussian model-based multichannel audio source separation, the likelihood of observed mixtures of source signals is parametrized by source spectral variances and by associated spatial covariance matrices. These parameters are estimated by maximizing the likelihood through an Expectation-Maximization algorithm and used to separate the signals by means of multichannel Wiener filtering. We propose to estimate these parameters by applying nonnegative factorization based on prior information on source variances. In the nonnegative factorization, spectral basis matrices can be defined as the prior information. The matrices can be either extracted or indirectly made available through a redundant library that is trained in advance. In a separate step, applying nonnegative tensor factorization, two algorithms are proposed in order to either extract or detect the basis matrices that best represent the power spectra of the source signals in the observed mixtures. The factorization is achieved by minimizing the $\beta$-divergence through multiplicative update rules. The sparsity of factorization can be controlled by tuning the value of $\beta$. Experiments show that sparsity, rather than the value assigned to $\beta$ in the training, is crucial in order to increase the separation performance. The proposed method was evaluated in several mixing conditions. It provides better separation quality with respect to other comparable algorithms.

顶级标签: audio machine learning model training
详细标签: audio source separation nonnegative factorization beta-divergence multichannel wiener filtering reverberant environments 或 搜索:

基于$β$散度的非负分解在混响环境下的音频源分离 / Audio Source Separation in Reverberant Environments using $β$-divergence based Nonnegative Factorization


1️⃣ 一句话总结

这篇论文提出了一种在混响环境中分离多个音频信号的新方法,它通过一种名为$β$散度的非负分解技术来估计信号参数,并利用先验信息提升分离效果,实验表明该方法能有效控制分解的稀疏性,从而获得比其他算法更好的分离质量。

源自 arXiv: 2604.12480