Next Article in Journal
Polarization-Sensitive Patterning of Azopolymer Thin Films Using Multiple Structured Laser Beams
Next Article in Special Issue
Listen to the Brain–Auditory Sound Source Localization in Neuromorphic Computing Architectures
Previous Article in Journal
Research on Self-Noise Characteristics of Nine Types of Seismometers Obtained by PDF Representation Using Continuous Seismic Data from the Malingshan Seismic Station, China
Previous Article in Special Issue
Three-Dimensional Acoustic Device for Testing the All-Directional Anisotropic Characteristics of Rock Samples
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved Speech Spatial Covariance Matrix Estimation for Online Multi-Microphone Speech Enhancement

School of Electrical Engineering and Computer Science, Gwangju Institute of Science and Technology, Buk-gu, Gwangju 61005, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(1), 111; https://doi.org/10.3390/s23010111
Submission received: 10 October 2022 / Revised: 10 December 2022 / Accepted: 19 December 2022 / Published: 22 December 2022
(This article belongs to the Special Issue Advanced Technology in Acoustic Signal Processing)

Abstract

:
Online multi-microphone speech enhancement aims to extract target speech from multiple noisy inputs by exploiting the spatial information as well as the spectro-temporal characteristics with low latency. Acoustic parameters such as the acoustic transfer function and speech and noise spatial covariance matrices (SCMs) should be estimated in a causal manner to enable the online estimation of the clean speech spectra. In this paper, we propose an improved estimator for the speech SCM, which can be parameterized with the speech power spectral density (PSD) and relative transfer function (RTF). Specifically, we adopt the temporal cepstrum smoothing (TCS) scheme to estimate the speech PSD, which is conventionally estimated with temporal smoothing. Furthermore, we propose a novel RTF estimator based on a time difference of arrival (TDoA) estimate obtained by the cross-correlation method. Furthermore, we propose refining the initial estimate of speech SCM by utilizing the estimates for the clean speech spectrum and clean speech power spectrum. The proposed approach showed superior performance in terms of the perceptual evaluation of speech quality (PESQ) scores, extended short-time objective intelligibility (eSTOI), and scale-invariant signal-to-distortion ratio (SISDR) in our experiments on the CHiME-4 database.

1. Introduction

Speech enhancement is essential to ensure the satisfactory perceptual quality and intelligibility of speech signals in many speech applications, such as hearing aids and speech communication with mobile phones and hands-free systems [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43]. Currently, devices with multiple microphones are popular, which has enabled multi-microphone speech enhancement, exploiting spatial information as well as spectro-temporal characteristics of the input signals [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48]. One of the most popular approaches to multi-microphone speech enhancement may be spatial filtering in the time–frequency domain, which aims to extract a target speech signal from multiple microphone signals contaminated by background noise and reverberation by suppressing sounds from directions other than the target direction [6,7,8,9,10,11].
There have been various types of spatial filters with different optimization criteria [6,7,8,9,10]. Among them, the minimum mean square error (MMSE) criterion for speech spectra estimation led to the multi-channel Wiener filter (MWF), which has shown decent performance [10,12,21,22]. It has been shown that the MWF solution can be decomposed into the concatenation of the minimum-variance distortionless-response (MVDR) beamformer and the single-channel postfilter [11,12]. Spatial filters often require the estimation of acoustic parameters such as the relative transfer function (RTF) between the microphones and the speech and noise spatial covariance matrices (SCMs), which should be estimated from the noisy observations.
For applications such as speech communication and hearing aids, time delays are crucial, and thus an online algorithm is required for multi-microphone speech enhancement. The work in [25] extended the single-channel minima controlled recursive averaging (MCRA) framework [49,50] for noise estimation in the multi-channel case by introducing the multi-channel speech presence probability (SPP) [24]. In [26], a coherence-to-diffusion ratio (CDR) based a priori SPP estimator under the expectation and maximization (EM) framework was proposed to improve the robustness in nonstationary noise scenarios. In [25,26], the speech SCM was estimated with the maximum likelihood (ML) approach, while the multi-channel decision-directed (DD) estimator was proposed in [29]. In [27], the recursive EM (REM) algorithm, which performs an iterative estimation for the latent variables and model parameters in the current frame, was exploited by defining the exponentially weighted log-likelihood of the data sequence. The speech SCM was decomposed into the speech power spectral density (PSD) and RTF under the rank-1 approximation, and these components were estimated by an ML approach using the EM algorithm in [27].
In this paper, we propose an improved speech SCM estimation for online multi-microphone speech enhancement. First, we adopt the temporal cepstrum smoothing (TCS) approach [51] to estimate the speech PSD, which has not yet been tried in multi-channel cases. Furthermore, we propose an RTF estimator based on time difference of arrival (TDoA) estimation using the cross-correlation method. Finally, we propose refining the acoustic parameters by exploiting the clean speech spectrum and clean speech power spectrum estimated in the first pass. The experimental results show that the proposed speech enhancement framework exhibited improved performance in terms of the perceptual evaluation of speech quality (PESQ) scores, extended short-time objective intelligibility (eSTOI), and scale-invariant signal to distortion ratio (SISDR) for the CHiME-4 database. Additionally, we performed an ablation study to understand how each sub-module contributed to the performance improvement.
The remainder of this paper is organized as follows. Section 2 briefly introduces the previous work on multi-microphone speech enhancement depending on various classes of approaches and then summarizes the main contributions of our proposal. Section 3 reviews the previous MMSE multi-channel speech enhancement approach and explains the conventional speech and noise SCM estimation. Section 4 presents the proposed speech SCM estimation based on the novel speech PSD and RTF estimators. Section 5 outlines the experimental results that demonstrate the superiority of the proposed method compared with the baseline in terms of speech quality and intelligibility. Finally, a conclusion is provided in Section 6.

2. Previous Work and Contributions

Recently, many approaches to multi-microphone speech enhancement have been proposed. In [33], the estimation of the speech PSD reduces to seek a unitary matrix and the square roots of PSDs based on the factorization of the speech SCM. The RTF estimate was recursively updated based on these estimates. They also proposed a desmoothing of the generalized eigenvalues to maintain the non-stationarities of estimated PSDs. Furthermore, these parameter estimates were then exploited for a Kalman filter-based speech separation algorithm [35]. In the context of sound field analysis, ref. [34] proposed a masking scheme under the non-negative tensor factorization model and [36] exploited the sparse representation in a spherical harmonic domain. The work in [37] proposed a multi-channel non-negative factorization algorithm in the ray space transform domain.
Deep-learning-based approaches have also been proposed, which can be categorized into several types. One is the combination of deep learning with conventional beamforming methods, in which the deep neural networks (DNNs) are employed to implement beamforming [38,39]. In [38], the complex spectral mapping approach was proposed to estimate the speech and noise SCMs. In contrast, ref. [39] reformulated the MVDR beamformer as a factorized form associated with two complex components and estimated them using a DNN, instead of estimating the parameters of the MVDR beamformer. The other approach is neural beamforming, in which a DNN directly learns the relationship between multiple noisy inputs and outputs in an end-to-end way [40,41,42,43]. In [40], they defined spatial regions and proposed a non-linear filter that suppresses signals from the undesired region while preserving signals from the desired region. In [41], the authors proposed an end-to-end system to estimate the time-domain filter-and-sum beamformer coefficient using a DNN. This approach was later replaced with implicit filtering in latent space [42]. In [43], they built a causal neural filter comprising modules for fixed beamforming, beam filtering, and residual refinement in the beamspace domain.
One of the popular approaches that adapt the spatial filter according to the dynamic acoustic condition is the informed filter, which is computed by utilizing the instantaneous acoustic parametric information [15,16,17,18]. Refs. [15,16] exploited the instantaneous direction of arrival (DoA) estimates to find the time-varying RTF used to construct the spatial filter, and [18] formulated a Bayesian framework under the DoA uncertainty. In [19], the eigenvector decomposition was applied to the estimated speech SCM to extract the steering vectors, which were used for the MVDR beamformer. The aforementioned approaches often adopted classical techniques such as ESPRIT [52] or MUSIC [53] for DoA estimation, which may be improved by incorporating more sophisticated sound localization [47,48].
Another set of studies focus on the estimation of the acoustic parameters. An EM algorithm [14] was employed to perform a joint estimation of the signals and acoustic parameters. While clean speech signals were obtained in the E-step, the PSDs of signals, RTF, and SCMs were estimated in the M-step. As the previous EM algorithm processed all of the signal samples at once, REM algorithms [27,28] overcame these issues by carrying out frame-wise iterative processing to handle online scenarios. For the speech PSD estimation, ref. [32] proposed an instantaneous PSD estimation method based on generalized principal components to preserve the non-stationarity of speech signals. For the RTF estimation, previous approaches mainly exploited the sample SCMs [46]. The covariance subtraction (CS) approaches [44,45] estimated the RTF by taking the normalized first column of the SCM obtained by the subtraction of the noisy speech SCM and noise SCM, assuming that the rank of the speech SCM was one. On the other hand, the covariance whitening (CW) approaches [30,54] normalized the dewhitened principal eigenvector of the whitened noisy input SCM to obtain the RTF.
In this paper, we propose an improved speech SCM estimation method for the online multi-microphone speech enhancement system based on the MVDR beamformer–Wiener filter factorization. The main contributions of our proposals are as follows:
  • A speech PSD estimator based on the TCS scheme to take the knowledge on the speech signal in the cepstral domain into account;
  • An RTF estimator based on the TDoA estimate to take advantage of the information from all frequency bins, especially when the signal-to-noise ratio (SNR) is low;
  • The refinement of the acoustic parameter estimates by exploiting the clean speech spectrum and clean speech power spectrum estimated in the first pass.

3. MMSE Multi-Microphone Speech Enhancement

3.1. Signal Model

Suppose that there is an array of M microphones in a noisy and reverberant room. Assuming that a single speech source and noises are additive, the observed microphone signals are given as
y ( l , k ) = g ( l , k ) S 1 ( l , k ) + v ( l , k ) = s ( l , k ) + v ( l , k )
where y ( l , k ) = [ Y 1 ( l , k ) , Y 2 ( l , k ) , , Y M ( l , k ) ] T , s ( l , k ) = [ S 1 ( l , k ) , S 2 ( l , k ) , , S M ( l , k ) ] T , and v ( l , k ) = [ V 1 ( l , k ) , V 2 ( l , k ) , . . . , V M ( l , k ) ] T , in which Y m ( l , k ) , S m ( l , k ) , and V m ( l , k ) are the short-time Fourier transform (STFT) coefficients of the microphone signal, clean speech, and background noises, including reverberations at the mth microphone, respectively, and g ( l , k ) = [ 1 , g 2 ( l , k ) , , g M ( l , k ) ] T is the RTF vector for the direct path from the desired speech source to the microphones. We assume that S m ( l , k ) and V m ( l , k ) are uncorrelated as in [16], although early reflections may disrupt this assumption. The SCM for the input signal y ( l , k ) , Φ y ( l , k ) , is given by
Φ y ( l , k ) = E [ y ( l , k ) y H ( l , k ) ] = Φ s ( l , k ) + Φ v ( l , k ) ,
where E [ · ] denotes mathematical expectation, and Φ s ( l , k ) = E [ s ( l , k ) s H ( l , k ) ] and Φ v ( l , k ) = E [ v ( l , k ) v H ( l , k ) ] are the SCMs of s ( l , k ) and v ( l , k ) , respectively.

3.2. MWF and MVDR–Wiener Filter Factorization

The objective of multi-microphone speech enhancement is to estimate clean speech S 1 ( l , k ) from the noisy observation y ( l , k ) , and we assume that prior knowledge on the location of the source or RTF is not available. One of the popular approaches is the MWF, which is a linear MMSE estimator for clean speech S 1 ( l , k ) , i.e.,
S ^ 1 ( l , k ) = w m w f H ( l , k ) y ( l , k ) ,
where w m w f ( l , k ) denotes the MWF described as [6]
w m w f ( l , k ) = Φ v 1 ( l , k ) Φ s ( l , k ) 1 + tr ( Φ v 1 ( l , k ) Φ s ( l , k ) ) e 1 ,
where e 1 = 1 0 1 × M 1 T , in which 0 is a zero vector, and tr [ · ] denotes the trace of a matrix. It is noted that only the noise and speech SCM, Φ v ( l , k ) and Φ s ( l , k ) , need to be estimated to implement the MWF. Previous work often adopted the multi-channel MCRA approach for noise SCM estimation, whereas ML estimation was employed for speech SCM estimation [25,26].
The MWF can be decomposed into the MVDR beamformer, w m v d r , and a single-channel Wiener postfilter, w w i e n e r , as [11,12]
w m w f = w m v d r · w w i e n e r ,
which makes it possible to consider the spatial filtering depending on the RTF g and the energy-based postfiltering w w i e n e r separately. Note that the frame and frequency indices are omitted for notational convenience.
Let the output of the MVDR beamformer be Z, i.e.,
Z = w m v d r H y ,
where the MVDR beamformer is given as
w m v d r = Φ v 1 g g H Φ v 1 g .
With the distortionless constraint of the MVDR beamformer, the beamformer output can be expressed as [27]
Z = S 1 + O ,
where O is assumed to follow the Gaussian distribution with variance
ϕ o = ( g H Φ v 1 g ) 1 .
The clean speech spectrum can be obtained by applying the single-channel Wiener filter to the beamformer output Z, as
S ^ 1 = ϕ s ϕ s + ϕ o · Z ,
where ϕ s = E [ | S 1 | 2 ] is the speech PSD at the first microphone.
Figure 1 illustrates the block diagram of the multi-microphone speech enhancement system based on the MVDR–Wiener filter factorization. The noisy speech y is processed by the MVDR beamformer and the Wiener filter sequentially, for which acoustic parameters Φ v , g, ϕ s , and ϕ o need to be estimated. Existing methods for parameter estimation are present in the next subsection.

3.3. Speech and Noise SCM Estimation

As for the estimation of the SCM of noise, Φ v , the multi-channel MCRA approach [25] is widely used, which is given as
Φ ^ v ( l , k ) = α ˜ v ( l , k ) Φ ^ v ( l 1 , k ) + ( 1 α ˜ v ( l , k ) ) y ( l , k ) y H ( l , k ) ,
where α ˜ v ( l , k ) = λ + p ( H 1 ( l , k ) | y ( l , k ) ) ( 1 λ ) is an SPP-dependent smoothing parameter with a constant 0 < λ < 1 . This method updates Φ v more when the SPP is low and vice versa. The a posteriori SPP p ( H 1 | y ) can be obtained using Bayes’ rule as
p ( H 1 | y ) = p ( H 1 ) p ( y | H 1 ) p ( H 0 ) p ( y | H 0 ) + p ( H 1 ) p ( y | H 1 ) ,
where H 0 and H 1 denote the hypotheses for speech absence and presence, respectively, and p ( y | H 0 ) and p ( y | H 1 ) are modeled as complex multivariate Gaussian distributions, as follows:
p ( y | H 0 ) = 1 π M det [ Φ v ] exp y H Φ v 1 y
p ( y | H 1 ) = 1 π M det [ Φ v + Φ s ] exp { y H [ Φ v + Φ s ] 1 y } ,
in which det [ · ] denotes the determinant of a matrix. Then, p ( H 1 | y ) becomes [24]
p ( H 1 | y ) = 1 + p ( H 0 ) p ( H 1 ) ( 1 + ξ ) exp { β 1 + ξ } 1 ,
where ξ = tr ( Φ v 1 Φ s ) , β = y H Φ v 1 Φ s Φ v 1 y , and p ( H 1 ) = 1 p ( H 0 ) is the a priori SPP, which can be estimated using the CDR-based [26] or DNN-based [27] method.
The speech SCM is usually estimated with the ML approach, which is defined as [25,26]
Φ ^ s m l = Φ ^ y Φ ^ v ,
where Φ ^ y is obtained by recursive smoothing as
Φ ^ y ( l , k ) = λ Φ ^ y ( l 1 , k ) + ( 1 λ ) y ( l , k ) y H ( l , k ) .
Under the rank-1 approximation for the clean speech SCM, Φ ^ s can be further refined using the decomposition of Φ s , with the speech PSD and RTF given by [8]
Φ s = ϕ s g g H .
Adopting the covariance subtraction (CS) approach, which extracts the normalized first column vector of the ML estimator of the speech SCM Φ ^ s m l , the estimator for the RTF is given as [46]
g ^ c s = Φ ^ s m l e 1 e 1 T Φ ^ s m l e 1 ,
where the denominator represents the speech PSD, i.e.,
ϕ ^ s c s = e 1 T Φ ^ s m l e 1 ,
in which the superscript c s indicates the CS approach.
In the REM framework [27], the ML estimator for the RTF based on the observed noisy speech is obtained as [27]
g ^ m l ( l , k ) = τ = 1 l λ l τ p ( H 1 ( τ , k ) | y ( τ , k ) ) y ( τ , k ) S ^ 1 * ( τ , k ) τ = 1 l λ l τ p ( H 1 ( τ , k ) | y ( τ , k ) ) S ^ 1 ( τ , k ) S ^ 1 * ( τ , k ) ,
where the summations in the numerator and the denominator can be computed with recursive averaging. The numerator can be thought of as the estimate of the cross-correlation between y and S ^ 1 in (10), r ^ y s , given by
r ^ y s ( l , k ) = λ r ^ y s ( l 1 , k ) + ( 1 λ ) p ( H 1 ( l , k ) | y ( l , k ) ) y ( l , k ) S ^ 1 * ( l , k ) ,
and the denominator can be considered to be the estimate of the speech PSD obtained by the recursive smoothing of the estimated clean speech power spectrum,
ϕ ^ s t s ( l , k ) = λ ϕ ^ s t s ( l 1 , k ) + ( 1 λ ) | S 1 | 2 ^ ( l , k ) ,
where the superscript t s indicates it is a temporally smoothed estimate, and | S 1 | 2 ^ ( l , k ) is the MMSE estimator of | S 1 | 2 under the speech presence uncertainty given by
| S 1 | 2 ^ = E [ | S 1 | 2 | Z ] = p ( H 1 | Z ) · E [ | S 1 | 2 | Z , H 1 ] = p ( H 1 | y ) · ( ( ϕ s ϕ s + ϕ o ) 2 · | Z | 2 + ϕ s ϕ o ϕ s + ϕ o ) ,
where we let p ( H 1 | Z ) = p ( H 1 | y ) , as in [27]. With r ^ y s in (22) and ϕ ^ s t s in (23), g ^ m l in (21) can be expressed as
g ^ m l = r ^ y s ϕ ^ s t s .

4. Proposed Speech SCM Estimation

Figure 2 illustrates the block diagram of the proposed speech enhancement system. As in [25,26,27], the estimation of the speech and relevant statistical parameters is performed twice for each frame, which was shown to be effective for online speech enhancement. In this paper, we propose an improved method for speech SCM estimation, i.e., speech PSD estimation and RTF estimation with a rank-1 approximation, using the speech enhancement system described in Figure 2. Note that the proposed modules are highlighted with red boxes.
In the first pass, we exploit the noisy input y ( l ) in the current frame and the noise SCM estimate Φ ^ v ( l 1 ) obtained in the previous frame to estimate the acoustic parameters in the current frame and perform beamforming and postfiltering, as explained in Section 3.2.
The ML estimate of the speech PSD at the first microphone using an instantaneous estimate of the PSD of input noisy signal can be obtained as
ϕ ^ s m l ( l , k ) = max | Y 1 ( l , k ) | 2 ϕ ^ v ( l 1 , k ) , ϕ ^ s m i n ,
where ϕ ^ v ( l 1 , k ) is the ( 1 , 1 ) th component of Φ ^ v ( l 1 , k ) , and ϕ ^ s m i n is a certain minimum value for the speech PSD estimate, which is set as ξ m i n ϕ ^ v ( l 1 , k ) with a tunable parameter ξ m i n . To estimate the speech PSDs, the ML estimation with temporal smoothing has been commonly used as described in (16) and (17) [25,26,27]. However, this approach occasionally results in undesired temporal smearing of speech [51]. In this paper, we propose to apply TCS [51] to ϕ ^ s m l in (26). TCS is a selective temporal smoothing technique in the cepstral domain motivated by the observation that, although the excitation component resides in a limited number of cepstral coefficients dependent on the pitch frequency, the speech spectral envelope is well-represented by the cepstral bins with low indices [55]. Specifically, the TCS consists of the following procedure: First, the cepstrum of ML speech PSD estimate ϕ ^ s m l , c e p s ( l , q ) is computed by the inverse discrete Fourier transform (IDFT) of ϕ ^ s m l . Next, the selective smoothing is applied to ϕ ^ s m l , c e p s ( l , q ) , in which the cepstral bins that are less relevant to speech are smoothed more and those representing the spectral envelope and fundamental frequency are less smoothed. Finally, the discrete Fourier transform is used to convert ϕ ^ s c e p s ( l , q ) into the TCS-based speech PSD estimate in the spectral domain ϕ ^ s t c s ( l , k ) . The bias compensation for the reduced variance due to the cepstral smoothing can be found in [56], and a detailed description of the adaptation of the smoothing parameters and the fundamental frequency estimation is given in [51]. In this paper, we denote the aforementioned procedure of TCS as an operation:
ϕ ^ s t c s , f ( l ) = TCS ( ϕ ^ s m l ( l ) ) ,
in which the superscript f indicates that this is the estimate in the first pass.
In this paper, we model the RTF vector g as a relative array propagation vector, which depends on the DoA [16]. Note that the conventional approaches in [27,44] estimate the RTF for each frequency using the input statistics in the frequency bin, ignoring the inter-frequency dependencies. In the presence of heavy noise, the accurate estimation of the RTF may become difficult, and thus it would be beneficial to estimate TDoA by utilizing the input signal in all frequency bins and to reconstruct the RTF using the simplest model. The TDoA for the desired speech can be obtained from the estimate of the cross-PSD of the desired speech, ϕ s 1 m ( l , k ) = E [ S 1 ( l , k ) S m * ( l , k ) ] , using the cross-correlation method [57]. The TDoA estimate τ m between the first and the mth microphones is given by
τ m ( l ) = arg max τ γ 1 m ( τ ) , γ 1 m ( τ ) k = 1 K ϕ ^ s 1 m ( l , k ) · e j 2 π k τ / K ,
in which ϕ ^ s 1 m ( l , k ) is the estimate of ϕ s 1 m ( l , k ) . Then, the TDoA-based RTF estimator can be obtained as
g ^ t d o a ( l , k ) = exp { j 2 π k τ ¯ ( l ) / K } ,
where τ ¯ = [ τ 1 , τ 2 , . . . , τ M ] T .
In the first pass, the cross-PSD estimate ϕ ^ s 1 m can be obtained by taking the ( 1 , m ) element of the ML speech SCM estimate Φ ^ s m l ( l , k ) in (16) as
ϕ ^ s 1 m f ( l , k ) = e m T Φ ^ s m l ( l , k ) e 1 ,
where e m = 0 ( m 1 ) 1 0 ( M m ) T in which 0 n is an all-zero vector of length n; g ^ t d o a , f can be computed using (28) and (29) with ϕ ^ s 1 m f , and Φ ^ s f can be obtained as in (18) using ϕ ^ s t c s , f ( l ) in (27) and g ^ t d o a , f . The noise SCM is estimated with the multi-channel MCRA approach in (11) utilizing p ( H 1 | y ) in (15) computed with Φ ^ s f and Φ ^ v . Then, we can compute the beamformer output Z in (6) and ϕ o in (9), and the estimate for the speech spectrum, S ^ 1 , can be obtained as in (10).
In the second pass, we estimate the acoustic parameters again by additionally utilizing the estimates for the clean speech spectrum, clean speech power spectrum, and a posteriori SPP, computed in the first pass. These refined parameters are in turn used to estimate the clean speech once again.
To refine the estimate of the speech PSD, we apply the TCS to the clean speech power spectrum estimate | S 1 | 2 ^ in (24) as
ϕ ^ s t c s , r ( l ) = TCS ( | S 1 | 2 ^ ( l ) ) ,
in which the superscript r indicates it is the refined estimate in the second pass. As | S 1 | 2 ^ would be less affected by the noise compared with the ϕ ^ s m l by virtue of beamforming and the MMSE estimation, ϕ ^ s t c s , r ( l ) would be more accurate than ϕ ^ s t c s , f ( l ) . As for the RTF estimation, r ^ y s in (22) is evaluated with S ^ 1 in (10), as in [27]. Instead of using r ^ y s divided by the estimate of the speech PSD in the first microphone to obtain the RTF, as in [27], we again estimate the RTF based on the TDoA; ϕ ^ s 1 m can be computed by extracting the mth element of r ^ y s as
ϕ ^ s 1 m r ( l , k ) = e m T r ^ y s ( l , k ) ,
in contrast to (30). The TDoA-based RTF estimate in the second pass, g ^ t d o a , r , can be obtained through (28) and (29) with ϕ ^ s 1 m r . As in the first pass, Φ ^ s r is computed with ϕ ^ s t c s , r in (31) and g ^ t d o a , r , and p ( H 1 | y ) in (15) is updated with Φ ^ s r . Then, p ( H 1 | y ) and Φ ^ v are obtained again using (15) and (11), and then the beamformer output Z and ϕ o are updated using (6) in (9). The final clean speech estimate S ^ 1 is obtained by (10) using g ^ t d o a , r , Φ ^ v , and ϕ ^ s t c s , r . The whole procedure of the proposed online multi-microphone speech enhancement method is summarized in Algorithm 1.
Algorithm 1 Proposed multi-microphone speech enhancement algorithm with improved speech SCM estimation.
 1:
Inputs: y for all frames
 2:
Output: S ^ 1 for all frames
 3:
Initialize variables and parameters
 4:
for each frame do
 5:
     Compute p ( H 1 ) using CDR-based [26] or DNN-based [27] method
 6:
     (First pass)
 7:
     Compute ϕ ^ s t c s , f via (26) and (27)
 8:
     Compute g ^ t d o a , f via (16), (28)–(30)
 9:
     Estimate p ( H 1 | y ) and Φ ^ v via (11), (15) and (18)
10:
   Beamformer: Compute Z and ϕ o via (6) and (9)
11:
   Postfilter: Compute S ^ 1 via (10)
12:
   (Second pass)
13:
   Compute ϕ ^ s t c s , r via (24) and (31)
14:
   Compute g ^ t d o a , r via (22), (28), (29) and (32)
15:
   Estimate p ( H 1 | y ) and Φ ^ v via (11), (15) and (18)
16:
   Beamformer: Compute Z and ϕ o via (6) and (9)
17:
   Postfilter: Compute S ^ 1 via (10)
18:
end for

5. Experiments

5.1. Experimental Settings

To demonstrate the superiority of the proposed algorithm, we conducted a set of experiments to evaluate the performance of the multi-microphone speech enhancement on the simulated set in the CHiME-4 database [58]. In this database, a mobile tablet device with six microphones was used for recording, of which the three microphones numbered 1, 2, and 3 were located in the top left, center, and right with an inter-microphone distance of approximately 10 cm each, while the other three microphones numbered 4, 5, and 6 were placed in the bottom left, center, and right, respectively [58]. The vertical distance between pairs of microphones was approximately 19 cm [58]. All microphones were located on the frontal surface, except for microphone 2. The bus (BUS), cafe (CAF), pedestrian area (PED), and street junction (STR) types of noise were used, and the SNR was between 0 and 15 dB. The training set consisted of 7138 utterances spoken by 83 speakers, whereas the development and evaluation sets were 1640 utterances and 1320 utterances, respectively, from 4 different speakers. The sampling rate for the signals used in the experiments was 16 kHz, and the square-root Hann window was applied to a 32 ms signal with a 16 ms frame shift. The 512-point DFT was applied to the windowed signal. The reference channel for the algorithms and evaluations was microphone 5, located at the bottom center of the device.
We set the a posteriori SPP, p ( H 1 | y ) , to zero for the first 10 frames instead of computing it using (15) based on the assumption that the speech would be absent in the initial periods, which helped the fast stabilization of the algorithm, as in [27]. To mitigate speech distortion at the expense of increased residual noise [59], the lower bounds for the p ( H 1 | y ) to compute | S 1 | 2 ^ in (24) and the Wiener gain in (10) were configured to 0.5 and 18 dB, respectively. The parameter values for λ and ξ m i n were set to be 0.9 and −10 dB, respectively. For the TCS schemes in (27) and (31), we followed the procedure in [51], employing the same parameter values except for the constant smoothing parameter, α ¯ const , which was determined empirically as
α ¯ const ( q ) = 0.1 if q { 0 , , 2 } 0.5 if q { 3 , , 19 } 0.95 if q { 20 , , 256 } ,
in which q is the quefrency index.
For the DNN-based a priori SPP estimation, we adopted the DNN architecture in [27], which consisted of a uni-directional long short-term memory (LSTM) layer of 512 dimensions, followed by three fully-connected layers of 256 dimensions. The activation functions were the rectified linear unit (ReLU) for the first three layers and sigmoidal activation for the last layer, which produced a 257-dimensional output vector. The number of dimensions of the DNN output is 257. The input for the DNN was the noisy log magnitude spectrum at the reference microphone, and the training target was binary for each bin, which was set by thresholding the instantaneous SNR [13].

5.2. Experimental Results

To demonstrate the superior performance of the proposed speech enhancement method, we evaluated the wideband PESQ score [60], eSTOI [61], and SISDR [62]. As we focused on the online framework in which the algorithm only uses the current and previous audio samples for frame-wise processing, the online algorithms designed in this way were chosen for the baseline methods. Depending on the a priori SPP estimator, we compared the performance of the proposed method using the ML framework with the MWF in [26] when the CDR-based a priori SPP estimator [26] was adopted, whereas the REM approach [27] was used for performance comparison when the DNN-based a priori SPP estimator [27] was employed for the proposed algorithm. As in [27], two versions of the REM approach using the Wiener postfilter and Kalman postfilter, denoted by DNN-REMWF and DNN-REMKF, were included in the experiment. The configuration parameters for the compared methods were set as in the original papers.
Table 1, Table 2 and Table 3 show the average PESQ score, eSTOI, and SISDR for each method depending on the noise type, respectively. The proposed method with the CDR-based a priori SPP estimator, CDR-Proposed, outperformed the previous approach in [26] by 0.39 in terms of the average PESQ score, 0.022 in terms of the eSTOI, and 4.3 dB in terms of the SISDR on average, respectively. With the DNN-based a priori SPP estimator, the proposed method, DNN-Proposed, improved the performance of DNN-REMKF by 0.21 in terms of the average PESQ score, 0.017 in terms of the eSTOI, and 1.2 dB in terms of the SISDR on average, respectively. Table 4 shows the PESQ scores, eSTOIs, and SISDRs for the baselines and the proposed method depending on the SNR. As the SNRs for the utterances in the evaluation set of the CHiME-4 database are distributed as in Figure 3, we divided the evaluation set into three groups depending on the SNR: low SNR less than 6.5 dB, medium SNR between 6.5 and 8.5 dB, and high SNR over 8.5 dB. It can be seen that all the measures were improved in all SNR ranges, and the performance improvements were more pronounced in low SNRs. From the results, we may conclude that the proposed speech SCM estimation approach could improve the performance of the multi-microphone speech enhancement method, regardless of the adoption of the DNN for the a priori SPP estimation.

5.3. Ablation Study

Additionally, we carried out an ablation study to analyze how much each module in the proposed system contributed to the performance improvement. We propose the speech PSD estimator, ϕ ^ s t c s , s in (31), and the RTF estimator, g ^ t d o a , s in (29). The previous approaches were the speech PSD estimator using recursive smoothing, ϕ ^ s t s in (23), and the ML estimator of the RTF g ^ m l in (25). The performances of the systems replacing the proposed modules one by one with conventional modules are summarized in Table 5. DNN-REMWF is also included, which uses ϕ ^ s t s , g ^ m l , and the Wiener postfilter, but adopts a different noise SCM estimator, derived from the EM framework.
The proposed system with conventional speech PSD and RTF estimators, ϕ ^ s t s + g ^ m l , showed the same average PESQ score and improved eSTOI and SISDR compared with DNN-REMWF [27]. Among the systems in the same framework, the introduction of the proposed speech PSD estimator improved the average PESQ scores by relatively large differences of 0.12 and 0.19, whereas it did not result in increased eSTOIs and SISDRs. On the other hand, employing the proposed RTF estimator improved all three metrics. From the results, we may conclude that both the proposed speech PSD and the RTF estimators contributed to the performance improvement.

5.4. Computational Complexity

Additionally, we have compared the computational complexity of the baseline and proposed methods in terms of the normalized processing time for the MATLAB implementation of the methods. The processing times for each algorithm, normalized by the processing time of the proposed algorithm, are given in Table 6. In this experiment, the a priori SPP was estimated by a DNN for all cases. As they depend on implementation details and settings such as the number of microphones, sampling frequency, and the dimensions of the DFT, the numbers given in the table should only be used as a rough indication. To see how much the refinement in the second pass incurred additional computational burden, the proposed method without the second pass (denoted as woSP) is included. From the table, it can be seen that the computational complexity of the proposed method was higher than those for MWF [26] and REMWF [27], but less than that of REMKF [27].

6. Conclusions

Multi-microphone speech enhancement exploits spatial information and spectro-temporal characteristics to reduce noise from the input. The online algorithms are required for the applications sensitive to time delays such as speech communication and hearing aids. In this paper, we propose an improved estimator of the speech SCM for online multi-microphone speech enhancement. Using the decomposition of the speech SCM under a rank-1 approximation, we propose an improved estimator for the speech PSD and RTF. For speech PSD estimation, we adopt the TCS scheme, which exploits knowledge on the speech signal in the cepstral domain to provide a better estimate of the speech PSD compared with the ML estimate. The RTF is estimated based on the TDoA estimate summarizing the information from all frequency bins. These estimators are evaluated once with input statistics and refined with an estimated clean speech spectrum and power spectrum obtained in the first pass. Our proposed speech enhancement method showed an improved speech enhancement performance in terms of the PESQ score, eSTOI, and SISDR in various noise environments for the CHiME-4 dataset, compared with other online multi-microphone speech enhancement algorithms.
Future work may include the incorporation of other spatial cues such as the inter-channel level differences on top of the inter-channel phase differences [47] into the RTF estimation without resorting to the far-field assumption. We may also investigate a deep learning approach to estimate acoustic parameters such as the speech and noise PSDs and RTF in a causal manner in the MVDR–Wiener filter factorization framework.

Author Contributions

Conceptualization, M.K., S.C. and J.W.S.; methodology, M.K. and J.W.S.; software, M.K. and H.S.; validation, M.K., S.C., H.S. and J.W.S.; formal analysis, J.W.S.; investigation, M.K. and S.C.; resources, J.W.S.; data curation, M.K. and H.S.; writing—original draft preparation, M.K.; writing—review and editing, J.W.S.; visualization, M.K.; supervision, J.W.S.; project administration, J.W.S.; funding acquisition, J.W.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Research Foundation of Korea, grant number NRF-2019R1A2C2089324, and the MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program (IITP-2021-0-01835) supervised by the IITP (Institute of Information & Communications Technology Planning & Evaluation).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Vary, P.; Martin, R. Digital Speech Transmission: Enhancement, Coding and Error Concealment; John Wiley & Sons: Chichester, UK, 2006. [Google Scholar]
  2. Kates, J.M. Digital Hearing Aids; Plural Publishing: San Diego, CA, USA, 2008. [Google Scholar]
  3. Rabiner, L.; Juang, B.H. Fundamentals of Speech Recognition; Prentice-Hall, Inc.: Hoboken, NJ, USA, 1993. [Google Scholar]
  4. Kim, M.; Shin, J.W. Improved Speech Enhancement Considering Speech PSD Uncertainty. IEEE/ACM Trans. Audio Speech Lang. Process. 2022, 30, 1939–1951. [Google Scholar] [CrossRef]
  5. Kim, M.; Song, H.; Cheong, S.; Shin, J.W. iDeepMMSE: An improved deep learning approach to MMSE speech and noise power spectrum estimation for speech enhancement. Proc. Interspeech 2022, 2022, 181–185. [Google Scholar]
  6. Benesty, J.; Chen, J.; Huang, Y. Microphone Array Signal Processing; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  7. Gannot, S.; Vincent, E.; Markovich-Golan, S.; Ozerov, A. A consolidated perspective on multimicrophone speech enhancement and source separation. IEEE/ACM Trans. Audio Speech Lang. Process. 2017, 25, 692–730. [Google Scholar] [CrossRef] [Green Version]
  8. Souden, M.; Benesty, J.; Affes, S. On optimal frequency-domain multichannel linear filtering for noise reduction. IEEE Trans. Audio Speech Lang. Process. 2010, 18, 260–276. [Google Scholar] [CrossRef]
  9. Markovich-Golan, S.; Gannot, S.; Cohen, I. A weighted multichannel Wiener filter for multiple sources scenarios. In Proceedings of the 2012 IEEE 27th Convention of Electrical and Electronics Engineers in Israel, Eilat, Israel, 14–17 November 2012; pp. 1–5. [Google Scholar]
  10. Doclo, S.; Spriet, A.; Wouters, J.; Moonen, M. Speech distortion weighted multichannel Wiener filtering techniques for noise reduction. In Speech Enhancement; Springer: Berlin/Heidelberg, Germany, 2005; pp. 199–228. [Google Scholar]
  11. Balan, R.; Rosca, J. Microphone array speech enhancement by Bayesian estimation of spectral amplitude and phase. In Proceedings of the Sensor Array and Multichannel Signal Processing Workshop Proceedings, Rosslyn, VA, USA, 6 August 2002; pp. 209–213. [Google Scholar]
  12. Thüne, P.; Enzner, G. Maximum-likelihood approach with Bayesian refinement for multichannel-Wiener postfiltering. IEEE Trans. Signal Process. 2017, 65, 3399–3413. [Google Scholar] [CrossRef]
  13. Heymann, J.; Drude, L.; Haeb-Umbach, R. A generic neural acoustic beamforming architecture for robust multi-channel speech processing. Comput. Speech Lang. 2017, 46, 374–385. [Google Scholar] [CrossRef]
  14. Schwartz, O.; Gannot, S.; Habets, E.A. An expectation-maximization algorithm for multimicrophone speech dereverberation and noise reduction with coherence matrix estimation. IEEE/ACM Trans. Audio Speech Lang. Process. 2016, 24, 1495–1510. [Google Scholar] [CrossRef]
  15. Thiergart, O.; Taseska, M.; Habets, E.A. An informed MMSE filter based on multiple instantaneous direction-of-arrival estimates. In Proceedings of the 21st IEEE European Signal Processing Conference (EUSIPCO 2013), Marrakech, Morocco, 9–13 September 2013; pp. 1–5. [Google Scholar]
  16. Thiergart, O.; Taseska, M.; Habets, E.A. An informed parametric spatial filter based on instantaneous direction-of-arrival estimates. IEEE/ACM Trans. Audio Speech Lang. Process. 2014, 22, 2182–2196. [Google Scholar] [CrossRef]
  17. Taseska, M.; Habets, E.A. Informed spatial filtering for sound extraction using distributed microphone arrays. IEEE/ACM Trans. Audio Speech Lang. Process. 2014, 22, 1195–1207. [Google Scholar] [CrossRef]
  18. Chakrabarty, S.; Habets, E.A. A Bayesian approach to informed spatial filtering with robustness against DOA estimation errors. IEEE/ACM Trans. Audio Speech Lang. Process. 2018, 26, 145–160. [Google Scholar] [CrossRef]
  19. Higuchi, T.; Ito, N.; Araki, S.; Yoshioka, T.; Delcroix, M.; Nakatani, T. Online MVDR beamformer based on complex Gaussian mixture model with spatial prior for noise robust ASR. IEEE/ACM Trans. Audio Speech Lang. Process. 2017, 25, 780–793. [Google Scholar] [CrossRef]
  20. Jin, Y.G.; Shin, J.W.; Kim, N.S. Spectro-temporal filtering for multichannel speech enhancement in short-time Fourier transform domain. IEEE Signal Process. Lett. 2014, 21, 352–355. [Google Scholar] [CrossRef]
  21. Serizel, R.; Moonen, M.; Van Dijk, B.; Wouters, J. Low-rank approximation based multichannel Wiener filter algorithms for noise reduction with application in cochlear implants. IEEE/ACM Trans. Audio Speech Lang. Process. 2014, 22, 785–799. [Google Scholar] [CrossRef] [Green Version]
  22. Wang, Z.; Vincent, E.; Serizel, R.; Yan, Y. Rank-1 constrained multichannel Wiener filter for speech recognition in noisy environments. Comput. Speech Lang. 2018, 49, 37–51. [Google Scholar] [CrossRef] [Green Version]
  23. Schwartz, O.; Gannot, S.; Habets, E.A. Joint maximum likelihood estimation of late reverberant and speech power spectral density in noisy environments. In Proceedings of the 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; pp. 151–155. [Google Scholar]
  24. Souden, M.; Chen, J.; Benesty, J.; Affes, S. Gaussian model-based multichannel speech presence probability. IEEE Trans. Audio Speech Lang. Process. 2010, 18, 1072–1077. [Google Scholar] [CrossRef]
  25. Souden, M.; Chen, J.; Benesty, J.; Affes, S. An integrated solution for online multichannel noise tracking and reduction. IEEE Trans. Audio Speech Lang. Process. 2011, 19, 2159–2169. [Google Scholar] [CrossRef]
  26. Taseska, M.; Habets, E.A. Nonstationary noise PSD matrix estimation for multichannel blind speech extraction. IEEE/ACM Trans. Audio Speech Lang. Process. 2017, 25, 2223–2236. [Google Scholar]
  27. Martín-Doñas, J.M.; Jensen, J.; Tan, Z.H.; Gomez, A.M.; Peinado, A.M. Online Multichannel Speech Enhancement Based on Recursive EM and DNN-Based Speech Presence Estimation. IEEE/ACM Trans. Audio Speech Lang. Process. 2020, 28, 3080–3094. [Google Scholar] [CrossRef]
  28. Schwartz, O.; Gannot, S. A recursive expectation-maximization algorithm for online multi-microphone noise reduction. In Proceedings of the 2018 IEEE 26th European Signal Processing Conference (EUSIPCO), Rome, Italy, 3–7 September 2018; pp. 1542–1546. [Google Scholar]
  29. Jin, Y.G.; Shin, J.W.; Kim, N.S. Decision-directed speech power spectral density matrix estimation for multichannel speech enhancement. J. Acoust. Soc. Am. 2017, 141, EL228–EL233. [Google Scholar] [CrossRef] [Green Version]
  30. Markovich, S.; Gannot, S.; Cohen, I. Multichannel eigenspace beamforming in a reverberant noisy environment with multiple interfering speech signals. IEEE Trans. Audio Speech Lang. Process. 2009, 17, 1071–1086. [Google Scholar] [CrossRef]
  31. Hwang, S.; Kim, M.; Shin, J.W. Dual microphone speech enhancement based on statistical modeling of interchannel phase difference. IEEE/ACM Trans. Audio Speech Lang. Process. 2022, 30, 2865–2874. [Google Scholar] [CrossRef]
  32. Dietzen, T.; Moonen, M.; van Waterschoot, T. Instantaneous PSD Estimation for Speech Enhancement based on Generalized Principal Components. In Proceedings of the 2020 IEEE 28th European Signal Processing Conference (EUSIPCO), Amsterdam, The Netherlands, 18–21 January 2021; pp. 191–195. [Google Scholar]
  33. Dietzen, T.; Doclo, S.; Moonen, M.; van Waterschoot, T. Square root-based multi-source early PSD estimation and recursive RETF update in reverberant environments by means of the orthogonal Procrustes problem. IEEE/ACM Trans. Audio Speech Lang. Process. 2020, 28, 755–769. [Google Scholar] [CrossRef]
  34. Mitsufuji, Y.; Takamune, N.; Koyama, S.; Saruwatari, H. Multichannel blind source separation based on evanescent-region-aware non-negative tensor factorization in spherical harmonic domain. IEEE/ACM Trans. Audio Speech Lang. Process. 2020, 29, 607–617. [Google Scholar] [CrossRef]
  35. Dietzen, T.; Doclo, S.; Moonen, M.; van Waterschoot, T. Integrated sidelobe cancellation and linear prediction Kalman filter for joint multi-microphone speech dereverberation, interfering speech cancellation, and noise reduction. IEEE/ACM Trans. Audio Speech Lang. Process. 2020, 28, 740–754. [Google Scholar] [CrossRef] [Green Version]
  36. Pezzoli, M.; Cobos, M.; Antonacci, F.; Sarti, A. Sparsity-Based Sound Field Separation in The Spherical Harmonics Domain. In Proceedings of the 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 1051–1055. [Google Scholar]
  37. Pezzoli, M.; Carabias-Orti, J.J.; Cobos, M.; Antonacci, F.; Sarti, A. Ray-space-based multichannel nonnegative matrix factorization for audio source separation. IEEE Signal Process. Lett. 2021, 28, 369–373. [Google Scholar] [CrossRef]
  38. Wang, Z.Q.; Wang, P.; Wang, D. Complex spectral mapping for single-and multi-channel speech enhancement and robust ASR. IEEE/ACM Trans. Audio Speech Lang. Proc. 2020, 28, 1778–1787. [Google Scholar] [CrossRef] [PubMed]
  39. Kim, H.; Kang, K.; Shin, J.W. Factorized MVDR Deep Beamforming for Multi-Channel Speech Enhancement. IEEE Signal Process. Lett. 2022, 29, 1898–1902. [Google Scholar] [CrossRef]
  40. Markovic, D.; Defossez, A.; Richard, A. Implicit Neural Spatial Filtering for Multichannel Source Separation in the Waveform Domain. arXiv 2022, arXiv:2206.15423. [Google Scholar]
  41. Luo, Y.; Chen, Z.; Mesgarani, N.; Yoshioka, T. End-to-end microphone permutation and number invariant multi-channel speech separation. In Proceedings of the 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 6394–6398. [Google Scholar]
  42. Luo, Y.; Mesgarani, N. Implicit Filter-and-Sum Network for End-to-End Multi-Channel Speech Separation. In Proceedings of the Interspeech, Brno, Czech Republic, 30 August–3 September 2021; pp. 3071–3075. [Google Scholar]
  43. Liu, W.; Li, A.; Wang, X.; Yuan, M.; Chen, Y.; Zheng, C.; Li, X. A Neural Beamspace-Domain Filter for Real-Time Multi-Channel Speech Enhancement. Symmetry 2022, 14, 1081. [Google Scholar] [CrossRef]
  44. Cohen, I. Relative transfer function identification using speech signals. IEEE Trans. Speech Audio Process. 2004, 12, 451–459. [Google Scholar] [CrossRef] [Green Version]
  45. Varzandeh, R.; Taseska, M.; Habets, E.A. An iterative multichannel subspace-based covariance subtraction method for relative transfer function estimation. In Proceedings of the 2017 Hands-free Speech Communications and Microphone Arrays (HSCMA), San Francisco, CA, USA, 1–3 March 2017; pp. 11–15. [Google Scholar]
  46. Zhang, J.; Heusdens, R.; Hendriks, R.C. Relative acoustic transfer function estimation in wireless acoustic sensor networks. IEEE/ACM Trans. Audio Speech Lang. Process. 2019, 27, 1507–1519. [Google Scholar] [CrossRef]
  47. Pak, J.; Shin, J.W. Sound localization based on phase difference enhancement using deep neural networks. IEEE/ACM Trans. Audio Speech Lang. Proc. 2019, 27, 1335–1345. [Google Scholar] [CrossRef]
  48. Song, H.; Shin, J.W. Multiple Sound Source Localization Based on Interchannel Phase Differences in All Frequencies with Spectral Masks. In Proceedings of the Interspeech, Brno, Czech Republic, 30 August–3 September 2021; pp. 671–675. [Google Scholar]
  49. Cohen, I.; Berdugo, B. Noise estimation by minima controlled recursive averaging for robust speech enhancement. IEEE Signal Process. Lett. 2002, 9, 12–15. [Google Scholar] [CrossRef]
  50. Cohen, I. Noise spectrum estimation in adverse environments: Improved minima controlled recursive averaging. IEEE Trans. Speech Audio Process. 2003, 11, 466–475. [Google Scholar] [CrossRef] [Green Version]
  51. Breithaupt, C.; Gerkmann, T.; Martin, R. A novel a priori SNR estimation approach based on selective cepstro-temporal smoothing. In Proceedings of the 2008 IEEE International Conference on Acoustics, Speech and Signal Processing, Las Vegas, NV, USA, 31 March–4 April 2008; pp. 4897–4900. [Google Scholar]
  52. Roy, R.; Kailath, T. ESPRIT-estimation of signal parameters via rotational invariance techniques. IEEE Trans. Acoust. Speech Signal Process. 1989, 37, 984–995. [Google Scholar] [CrossRef] [Green Version]
  53. Schmidt, R. Multiple emitter location and signal parameter estimation. IEEE Trans. Antennas Propag. 1986, 34, 276–280. [Google Scholar] [CrossRef] [Green Version]
  54. Markovich-Golan, S.; Gannot, S.; Kellermann, W. Performance analysis of the covariance-whitening and the covariance-subtraction methods for estimating the relative transfer function. In Proceedings of the 2018 26th European Signal Processing Conference (EUSIPCO), Rome, Italy, 3–7 September 2018; pp. 2499–2503. [Google Scholar] [CrossRef]
  55. Noll, A.M. Cepstrum pitch determination. J. Acoust. Soc. Am. 1967, 41, 293–309. [Google Scholar] [CrossRef]
  56. Gerkmann, T.; Martin, R. On the statistics of spectral amplitudes after variance reduction by temporal cepstrum smoothing and cepstral nulling. IEEE Trans. Signal Process. 2009, 57, 4165–4174. [Google Scholar] [CrossRef]
  57. Knapp, C.; Carter, G. The generalized correlation method for estimation of time delay. IEEE Trans. Acoust. Speech Signal Process. 1976, 24, 320–327. [Google Scholar] [CrossRef] [Green Version]
  58. Vincent, E.; Watanabe, S.; Nugraha, A.A.; Barker, J.; Marxer, R. An analysis of environment, microphone and data simulation mismatches in robust speech recognition. Comput. Speech Lang. 2017, 46, 535–557. [Google Scholar] [CrossRef] [Green Version]
  59. Gerkmann, T.; Breithaupt, C.; Martin, R. Improved a posteriori speech presence probability estimation based on a likelihood ratio with fixed priors. IEEE Trans. Audio Speech Lang. Process. 2008, 16, 910–919. [Google Scholar] [CrossRef]
  60. International Telecommunication Union. Wideband Extension to Recommendation P.862 for the Assessment of Wideband Telephone Networks and Speech Codec; International Telecommunication Union: Geneve, Switzerland, 2007. [Google Scholar]
  61. Jensen, J.; Taal, C.H. An algorithm for predicting the intelligibility of speech masked by modulated noise maskers. IEEE/ACM Trans. Audio Speech Lang. Process. 2016, 24, 2009–2022. [Google Scholar] [CrossRef]
  62. Le Roux, J.; Wisdom, S.; Erdogan, H.; Hershey, J.R. SDR–half-baked or well done? In Proceedings of the 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 626–630. [Google Scholar]
Figure 1. Block diagram of the multi-microphone speech enhancement system based on the MVDR–Wiener filter factorization.
Figure 1. Block diagram of the multi-microphone speech enhancement system based on the MVDR–Wiener filter factorization.
Sensors 23 00111 g001
Figure 2. Block diagram of the proposed multi-microphone speech enhancement system.
Figure 2. Block diagram of the proposed multi-microphone speech enhancement system.
Sensors 23 00111 g002
Figure 3. Normalized histogram of the SNRs in dB scale for the utterances in the evaluation set of the CHiME-4 database.
Figure 3. Normalized histogram of the SNRs in dB scale for the utterances in the evaluation set of the CHiME-4 database.
Sensors 23 00111 g003
Table 1. The average PESQ scores for the different algorithms depending on the noise type.
Table 1. The average PESQ scores for the different algorithms depending on the noise type.
MethodNoise TypeAvg.
BUSCAFPEDSTR
Noisy1.321.241.261.281.27
CDR-MWF [26]1.931.741.861.741.82
CDR-Proposed2.352.152.182.172.21
DNN-REMWF [27]2.102.012.122.032.07
DNN-REMKF [27]2.132.032.152.072.10
DNN-Proposed2.432.242.282.292.31
Table 2. The average eSTOIs (x100) for the different algorithms depending on the noise type.
Table 2. The average eSTOIs (x100) for the different algorithms depending on the noise type.
MethodNoise TypeAvg.
BUSCAFPEDSTR
Noisy71.066.068.867.268.2
CDR-MWF [26]83.179.782.480.281.4
CDR-Proposed85.183.483.782.483.6
DNN-REMWF [27]83.782.083.983.283.2
DNN-REMKF [27]84.382.684.583.983.8
DNN-Proposed86.085.385.985.085.5
Table 3. The average SISDRs (in dB) for the different algorithms depending on the noise type.
Table 3. The average SISDRs (in dB) for the different algorithms depending on the noise type.
MethodNoise TypeAvg.
BUSCAFPEDSTR
Noisy6.797.778.606.857.51
CDR-MWF [26]9.689.4410.699.759.89
CDR-Proposed14.2514.5514.5613.4014.19
DNN-REMWF [27]14.4014.1814.9614.6414.54
DNN-REMKF [27]14.7114.4215.2514.9914.84
DNN-Proposed15.9016.0716.3415.8316.04
Table 4. The PESQ scores, eSTOIs, and SISDRs for the baselines and the proposed method depending on the SNR.
Table 4. The PESQ scores, eSTOIs, and SISDRs for the baselines and the proposed method depending on the SNR.
MethodPESQ ScoreeSTOI (×100)SISDR (in dB)
(−,6.5)(6.5,8.5)(8.5,)(−,5)(6.5,8.5)(8.5,)(−,6.5)(6.5,8.5)(8.5,)
Noisy1.221.261.3363.367.174.25.047.489.74
CDR-MWF [26]1.661.831.9576.780.786.58.819.8010.97
CDR-Proposed2.072.222.3380.683.286.912.3513.9516.14
DNN-REMWF [27]1.842.072.2679.382.887.212.8014.4116.28
DNN-REMKF [27]1.872.102.2980.083.487.713.0714.7016.60
DNN-Proposed2.152.322.4482.785.288.514.3815.8317.78
Table 5. The PESQ scores, eSTOIs, and SISDRs averaged over all noise types for the proposed method with DNN-based a priori SPP estimation by replacing the proposed sub-modules with conventional ones, one by one.
Table 5. The PESQ scores, eSTOIs, and SISDRs averaged over all noise types for the proposed method with DNN-based a priori SPP estimation by replacing the proposed sub-modules with conventional ones, one by one.
MethodPESQeSTOISISDR
DNN-REMWF [27]2.0783.214.54
ϕ ^ s t s + g ^ m l 2.0784.515.79
ϕ ^ s t s + g ^ t d o a , s 2.1285.416.07
ϕ ^ s t c s , s + g ^ m l 2.1983.815.52
ϕ ^ s t c s , s + g ^ t d o a , s (DNN-Proposed)2.3185.516.04
Table 6. Comparison of the normalized processing time when the a priori SPP was obtained by a DNN.
Table 6. Comparison of the normalized processing time when the a priori SPP was obtained by a DNN.
MWF [26]REMWF [27]REMKF [27]Proposed (woSP)Proposed
Process. time0.6140.6921.1320.7041.000
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, M.; Cheong, S.; Song, H.; Shin, J.W. Improved Speech Spatial Covariance Matrix Estimation for Online Multi-Microphone Speech Enhancement. Sensors 2023, 23, 111. https://doi.org/10.3390/s23010111

AMA Style

Kim M, Cheong S, Song H, Shin JW. Improved Speech Spatial Covariance Matrix Estimation for Online Multi-Microphone Speech Enhancement. Sensors. 2023; 23(1):111. https://doi.org/10.3390/s23010111

Chicago/Turabian Style

Kim, Minseung, Sein Cheong, Hyungchan Song, and Jong Won Shin. 2023. "Improved Speech Spatial Covariance Matrix Estimation for Online Multi-Microphone Speech Enhancement" Sensors 23, no. 1: 111. https://doi.org/10.3390/s23010111

APA Style

Kim, M., Cheong, S., Song, H., & Shin, J. W. (2023). Improved Speech Spatial Covariance Matrix Estimation for Online Multi-Microphone Speech Enhancement. Sensors, 23(1), 111. https://doi.org/10.3390/s23010111

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop