Professional Documents
Culture Documents
www.elsevier.com/locate/sigpro
PII: S0165-1684(16)30132-3
DOI: http://dx.doi.org/10.1016/j.sigpro.2016.06.021
Reference: SIGPRO6181
To appear in: Signal Processing
Received date: 21 February 2016
Revised date: 23 May 2016
Accepted date: 22 June 2016
Cite this article as: Laid Chergui and Saad Bouguezel, A New Pre-Whitening
Transform Domain LMS Algorithm and its Application to Speech Denoising,
Signal Processing, http://dx.doi.org/10.1016/j.sigpro.2016.06.021
This is a PDF file of an unedited manuscript that has been accepted for
publication. As a service to our customers we are providing this early version of
the manuscript. The manuscript will undergo copyediting, typesetting, and
review of the resulting galley proof before it is published in its final citable form.
Please note that during the production process errors may be discovered which
could affect the content, and all legal disclaimers that apply to the journal pertain.
A New Pre-Whitening Transform Domain LMS
Algorithm and its Application to Speech Denoising
Abstract
In this paper, we propose a new pre-whitening transform domain LMS algorithm. The main idea is to
introduce a pre-whitening using a simple finite impulse response decorrelation filter of order one before
applying the transform to reinforce its decorrelation. The resulting algorithm has the advantage of using any
transform even with low decorrelation. This advantage can be exploited to consider transforms having lower
computational and structural complexities than those of the classical transforms. For this purpose, we also
investigate the use of other transforms, namely the parametric Fourier and Hartley transforms. This
investigation is accomplished by studying the eigenvalue spreads obtained by a given parametric transform
and then finding the value of the parameter corresponding to the minimum eigenvalue spread, which is
equivalent to the best mean square error (MSE) convergence behaviour. This approach provides new
attractive transforms for the proposed algorithm. Moreover, we consider the adaptive speech denoising as an
application to evaluate the performance of the proposed algorithm. The comparisons between the proposed
and conventional algorithms for different transforms are perform in terms of the computational complexity,
MSE convergence speed, reached steady state level, residual noise in the denoised signal, steady state
Keywords: Adaptive filtering, parametric transforms, pre-whitening, speech denoising, transform domain LMS algorithm.
1. Introduction
The least mean square (LMS) algorithm is the mostly used in the adaptive filtering for its simplicity and
robustness [1]. However, it has a slow convergence in the case of highly correlated input signals [2,3]. This
is due to the fact that the autocorrelation matrix of the input signal has a large eigenvalue spread. To
overcome this problem by reducing the eigenvalue spread, whitening or decorrelated adaptive algorithms
have been proposed for time-domain LMS [4-6]. In [4], the authors proposed a joint decorrelation of both
the input and error signals. The decorrelation is achieved in the time domain using an adaptive decorrelation
filter based on the concept of prediction. The resulting decorrelated normalised LMS (NLMS) structure has
the advantage of improving the mean square error (MSE) convergence speed and steady state compared to
the conventional LMS and NLMS algorithms. The decorrelation of the input signal can also be achieved by
an orthogonal transformation followed by power normalization [7]. These two operations have led to a new
configuration named transform-domain LMS (TDLMS) adaptive filters, which outperform the time-domain
LMS adaptive filters in terms of the MSE convergence speed and steady state [8-10]. The orthogonal
transforms such as the discrete Fourier transform (DFT), the discrete Hartley transform (DHT) and the
discrete cosine transform (DCT) have been used in the TDLMS [7]. The resulting adaptive filters have been
named as DFT-LMS, DHT-LMS and DCT-LMS, respectively [11]. The convergence speed of these filters
depends on the used transform [8,11,12]. In general, the DCT-LMS filter presents a convergence
performance better than those of the DFT-LMS and DHT-LMS filters [11]. This is mainly due to the fact
that the DCT is suboptimal in terms of decorrelation [13,14]. Therefore, it is highly desirable to use in the
TDLMS adaptive filters a transform having better decorrelation. However, the existing transforms are fixed
and hence it is not possible to increase their decorrelation ability. Since the introduction of a decorrelation in
the time-domain LMS has brought interesting improvements, it is important to investigate the introduction
of such a decorrelation in the TDLMS. However, to the best of authors knowledge, this investigation has
enhance the quality and intelligibility of the voice and reduce communication fatigue in modern
communication systems such as mobile phones, hands-free telephony and voice-controlled systems, which
are generally used in noisy environments [15]. Most of the audio frequency acoustic noises, such as
computer fan noise and noise from people and cars, have low-frequency spectra and hence are colored
[15,16]. These noises often corrupt the speech signal, which is also colored. Therefore, it is important to
In this paper, we propose a new pre-whitening TDLMS (PW-TDLMS) algorithm. It maintains the
structure of the conventional TDLMS algorithm and introduces a pre-whitenning before applying the
transform. This is a novel and interesting strategy to reinforce the decorrelation of the used fixed transform
in the TDLMS. The proposed pre-whitening is achieved by using a simple finite impulse response (FIR)
decorrelation filter of order one based on a fixed prediction concept. The resulting PW-TDLMS has the
advantage of using any transform even with low decorrelation. For this purpose, we investigate the use of
other transforms such as the parametric DFT and DHT transforms [17] and study the performance of the
proposed PW-TDLMS algorithm in terms of MSE convergence speed and steady state. We carry out this
comparative study by considering adaptive speech denoising as an application of the proposed algorithm
and show the simulation results of the proposed PW-TDLMS and TDLMS for different transforms.
Moreover, we compute the eigenvalue spread of the autocorrelation matrix obtained after applying the
parametric DFT or DHT transform and power normalization in the case of highly correlated Markov-1
noise, which is equivalent to the first order autoregressive (AR) process [11], for different values of the
independent parameter of the transform. The transform with the value of the parameter corresponding to a
good compromise between the eigenvalue spread and computational complexity is selected for the proposed
PW-TDLMS. It should be noted that the eigenvalue spreads in the case of the DCT and classical DFT and
DHT. In Section 3, we analyze the TDLMS adaptive filter by considering the stability, steady state and
convergence performances in the case of a first order AR process. For the convergence analysis, we review
the eigenvalue spreads in the cases of the DFT, DHT and DCT, and then find the eigenvalue spreads in the
cases of the parametric DFT and DHT. The proposed PW-TDLMS adaptive filter is developed and
compared with the conventional TDLMS adaptive filter in Section 4 in terms of the eigenvalue spreads in
the cases of the DCT, DFT, DHT, parametric DFT and DHT transforms. Section 5 presents the
computational complexities of the DCT-LMS and proposed parametric DHT-based PW-LMS algorithms. In
order to compare the performance of the proposed PW-TDLMS with that of the TDLMS for different
transforms, we consider in Section 6 the speech denoising application for the cases of speech-like and real
speech signals. The simulation results and comparisons are given therein in terms of the MSE convergence
speed, reached steady state level, residual noise in the denoised signal, steady state excess MSE,
misadjustment and output SNR. Some concluding remarks are given in Section 7.
[17] as
[ ] (2)
. Since the parametric transform defined by (1) or (3) possesses three parameters that can arbitrarily be
chosen from the complex plane, a large number of new transforms with different features can be obtained. If
the three parameters are chosen as , and , then the parametric transform
reduces to the classical DFT. If the values of the three parameters are arbitrarily chosen from the unit circle,
then the parametric transform reduces to a three-parameter unitary transform. An interesting special case of
this transform is obtained in [17] when , with being an arbitrary real-valued parameter,
and . Let be the matrix operator of this unitary transform, which is nothing but the classical
DFT parameterized by one parameter and then denoted by . The matrix operator of the one-
(4)
matrix. It is obvious that is the classical DFT matrix and thus, is the classical DHT matrix, and
LMS
In the TDLMS adaptive filter presented in Fig. 1, the correlated tap delayed input vector
transform matrix , where N is the filter length. The resulting vector is less correlated then the
vector . The transformed vector is then power normalized using the diagonal matrix defined
in [9,11,19] as
[ ] (5)
where is the power estimate of the ith input , which can be computed using the recursive
equation given by
| | (6)
with the smoothing factor being a constant value between 0 and 1. The power normalization at each time
sample k leads to a time varying step size . The weight vector is then adapted as
(7)
where is a constant small-value introduced to avoid the overflow when the diagonal elements of are
close to zeros, is the step size parameter and denotes the complex conjugate of . The output error is
given by
We use the stability condition found in [11] for which the authors assume that the form of (7) is a mean-
(9)
It is clear that this stability condition on the step size depends only on the filter length N and therefore, the
The excess mean square error (EMSE) and the misadjustment (M) are among the mostly used adaptive
| |
where, in the case of the adaptive noise cancellation, is the residual error, which is the difference
between the original speech signal sample and the filtered speech signal sample , and is the number
of samples used in the estimation of the . The steady state excess mean square error is the
result of computing the average of for k after the algorithm reaches the steady state. Then, the
is defined as
( )
where is the total number of samples and P the number of samples after which the algorithm reaches the
(12)
where is defined as
( ) | |
The noise reduction performance is evaluated by the estimate of the output signal to noise ratio (SNR)
(| | | | )
As discussed earlier, the LMS algorithm suffers from a slow convergence in the case of a highly
correlated input signal, i.e., in the case where the autocorrelation matrix of the input signal has a large
eigenvalue spread. Therefore, we analyse in this subsection the effect of different transforms on reducing
the eigenvalue spread of the autocorrelation matrix of the transformed and power normalised signal, which
has a straight impact on the convergence of a TDLMS algorithm. This analysis is performed by modeling
the different noise sources by Markov 1 process, which is a simple class of signals, very general and
practical [4,5,9,17,21]. Markov 1 process, which is equivalent to the first order AR process [11], can be
generated by passing a white Gaussian noise through a single pole low-pass filter. The
autocorrelation matrix of the input signal obtained from the first order AR process of correlation
( )
In [9,22,23], the authors show that, for a very large value of (theoretically, for N tending to infinity),
the eigenvalue of are the values of the power spectrum of evaluated at uniformly distributed points on
( ) ( )
A high value of increases the correlation of the signal and therefore, it increases the eigenvalue spread
of the autocorrelation matrix . The autocorrelation matrix obtained after transformation and power
normalization of the input signal is not Toeplitz, and the above theory cannot be used to find the
eigenvalue spread. The asymptotic distribution of the eigenvalues of can be derived by solving the
denotes the complex conjugate transpose operation. In [9], the problem of solving (17) is simplified to
where is the DFT or DCT matrix and is an identity matrix. The eigenvalue spread of obtained after
( )
In [11,18], the authors showed that the eigenvalue spread of achieved by the DHT-LMS tends to
| | | | for , where | | denotes the absolute value. It is clear that, for [ , the
eigenvalue spread of achieved by the DHT-LMS is the same as that achieved by the DFT-LMS. In [11],
the authors showed that, for ], the eigenvalue spreads of achieved by the DFT-LMS and the
DFT-LMS and the DHT-LMS is not deteriorated in the case of negative values of , whereas that of the
DCT-LMS is deteriorated.
In this subsection, we study the eigenvalue spreads of for the cases of the and . We show
that these eigenvalue spreads depend on the parameter and can similarly be computed using (18) for
being the or matrix. This interesting propriety is exploited in this work to analyze these
eigenvalue spreads by numerical calculations for different values of and = 16 and then find the value of
that leads to the minimum eigenvalue spread that can be achieved by the -LMS and -LMS,
which is equivalent to the best MSE convergence behaviour. Fig. 2 shows the eigenvalue spreads of
achieved by the -LMS with a first order AR process for different values of the coefficient and filter
length It is clear from this figure that the minimum eigenvalue spread can be achieved for the
parameter in the interval [ ] Fig. 3 shows the eigenvalue spreads of achieved by the -
LMS and -LMS for = 16 and . It is seen from this figure that the eigenvalue spread of
achieved by the -LMS is slightly lower than that achieved by the -LMS. Some other results
obtained from numerical calculations are given in Table 1 for the case of and different values of .
It is clear from this table that the value of the parameter corresponding to the lowest eigenvalue spread
depends on the value of the correlation coefficient . For = 0.9 or 0.8, the parameter is . For
= 0.7, 0.6 and 0.5, the parameter is . It can also be seen from Table 1 that the difference
between the minimum eigenvalue spreads, for a specific value of and a specific transform corresponding
to the various values of the parameter , is not meaningful and hence, the MSEs of the TDLMS algorithm
based on the corresponding transforms would have similar convergence behavior. Moreover, in [17],
Bouguezel et al found that the and lead in Wiener filtering to MSEs similar to that of the
classical DFT and DHT, respectively, and showed that the former reduce the computational complexity
70 = 0.9 70 DFT-LMS
= 0.8 DHT-LMS
60 = 0.7
60
= 0.6
= 0.5
Eigenvalue Spreads of SN
50 50
Eigenvalue Spreads of SN
40 40
30 30
20 20
10 10
0 0
2 - 0 2 -
Fig. 2. Eigenvalue spreads of SN achieved by DFT-LMS algorithm for N = 16, Fig. 3. Eigenvalue spreads of SN achieved by DFT-LMS or DHT-LMS
and some values of . algorithm for N = 16, and .
Table 1
Minimum eigenvalue spreads of for the various transforms with =16 and some values of .
Noise path
Pre-whitening
TDLMS
Filter g
Proposed PW-TDLMS
In this section, we propose a new pre-whitening TDLMS (PW-TDLMS) adaptive noise canceller by
introducing a pre-whitening filter at the input of the TDLMS as presented in Fig. 4. The purpose of this
filter is to obtain a decorrelated version of the signal at the input of TDLMS. This has a direct impact
on further reducing the eigenvalue spread of the autocorrelation matrix obtained after transformation and
power-normalization of the signal in TDLMS and consequently on accelerating the convergence speed
and minimizing the reached steady state of the MSE of the TDLMS. The enhancement in the performance
of the TDLMS improves the quality of the output signal sample , which is the estimation of the original
signal sample to be recovered. The decorrelated signal at each time sample k can be obtained as
filter and [ ] is a finite impulse response (FIR) prediction filter. Thus, the impulse
Without loss of generality, we consider the case of and real-valued coefficients, i.e., [ ]
(22)
Let us now generate the input signal of the system given by Fig. 4 and used in (22) from a first order
AR process with a correlation coefficient that ensures a high eigenvalue spread of given by (15) as
(23)
where is a zeros mean random white Gaussian noise. Equation (23) can be rearranged as
( )
( )
It is clear from (25) that the signal can become completely white when leads to zero, i.e.,
tends to . In the case where the input signal is assumed to be highly correlated, the correlation
coefficient of the corresponding autocorrelation matrix would be close to unity. Therefore, perfect
pre-whitening can asymptotically be achieved by choosing the value of close to unity. For other values of
, , and unknown values of , some decorrelation of the input signal can be achieved, which
can then be exploited to reinforce the decorrelation that can be obtained by the considered transform in the
proposed PW-TDLMS.
Since the above condition on the filter coefficient depends on the coefficient of the input, which is in
practice not known, it is important to find an appropriate interval of possible values of that
asymptotically satisfy the condition without a prior knowledge on the exact value of , and thus ensure
acceptable decorrelation. For this purpose, we study the eigenvalue spreads of achieved by the DFT-
LMS, DHT-LMS, DFT-/6-LMS, DHT-/6-LMS and DCT-LMS without and with a pre-whitening filter in
terms of . The results of this study are given in Figs. 5, 6, 7, 8 and 9 for = 16, [ ] and 0.5,
0.6, 0.7, 0.8 and 0.9. Note that we have used the notation DFT-LMS, DHT-LMS, DFT-/6-LMS, DHT-/6-
LMS and DCT-LMS for the case of the conventional TDLMS and PW-DFT-LMS, PW-DHT-LMS, PW-
DFT-/6-LMS, PW-DHT-/6-LMS and PW-DCT-LMS for the proposed PW-TDLMS. Figs. 5, 6, 7, 8 and 9
show that the eigenvalue spreads of achieved by the proposed PW-TDLMS for different transforms and
[ ] decrease below the corresponding eigenvalue spreads achieved by the DCT-LMS in the cases
of , 0.8, 0.7, 0.6, 0.5, respectively, and approach to 1 when the coefficient leads to , i.e., the
input signal becomes completely decorrelated. In order to find the interval for that is appropriate to all
the cases, we adopt an averaging scheme and hence suggest the average interval [ ]. Therefore, the
middle value in this interval can allow the proposed PW-TDLMS to achieve a good reduction in
the eigenvalue spreads of for different transforms and for any value of [ .
DFT-/6-LMS DFT-LMS DFT-/6-LMS DFT-LMS
Proposed PW- DFT-/6-LMS Proposed PW- DFT-LMS Proposed PW- DFT-/6-LMS Proposed PW- DFT-LMS
DHT-/6-LMS DHT-LMS DHT-/6-LMS DHT-LMS
15 Proposed PW- DHT-/6-LMS Proposed PW- DHT-LMS Proposed PW- DHT-/6-LMS Proposed PW- DHT-LMS
DCT-LMS Proposed PW- DCT-LMS DCT-LMS Proposed PW- DCT-LMS
8
Eigenvalue Spread of SN
Eigenvalue Spread of SN
10 6
0 0
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
4
Eigenvalue Spread of SN
Eigenvalue Spread of SN
2
2
0 0
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
Fig. 7. Eigenvalue spreads of achieved by the TDLMS and proposed PW- Fig. 8. Eigenvalue spreads of achieved by the TDLMS and proposed PW-
TDLMS for different transforms and for N =16, = 0.7 and different values of TDLMS for different transforms and for N =16, = 0.6 and different values of
. .
DFT-/6-LMS DFT-LMS
Proposed PW- DFT-/6-LMS Proposed PW- DFT-LMS
DHT-/6-LMS DHT-LMS
Proposed PW- DHT-/6-LMS Proposed PW- DHT-LMS
DCT-LMS Proposed PW- DCT-LMS
Eigenvalue Spread of SN
4
0
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
Different values of a1
Fig. 9. Eigenvalue spreads of achieved by the TDLMS and proposed PW-
TDLMS for different transforms and for N =16, = 0.5 and different values of
.
In order to show that the average interval [ ] is appropriate for a random selection of the values
of regardless the values of the coefficient of the input signal, we present in Table 2 the eigenvalue
spreads of achieved by the TDLMS and the proposed PW-TDLMS for the various transforms. It is clear
from this table that, for a given transform including the DCT, the eigenvalue spreads of obtained in this
interval by the proposed PW-TDLMS are less than the eigenvalue spreads obtained by the corresponding
conventional TDLMS. This table also shows the importance of the proposed pre-whitening in rendering the
performance of the DHT-LMS, DFT-LMS, DFT-/6-LMS or DHT-/6-LMS closer to that of the DCT-LMS,
which is known for its high efficiency. Therefore, the proposed PW-TDLMS based on the DHT-/6, which is
a real-to-real transform and has reduced computational and structural complexities compared to the DCT
[17], would be an excellent alternative to the DCT-LMS algorithm. Table 2 also confirms that the suggested
value is convenient for all the considered cases and ensures a good performance for the pre-
whitening filter independently of the degree of the correlation of the input signal. Therefore, we suggest a
pre-whitening filter with the impulse response [ ] for the proposed PW-TDLMS.
Table 2
Some values of the eigenvalue spreads of achieved by the TDLMS and the proposed PW-TDLMS algorithms for various
transforms, N = 16, some values of and [ ].
0.9 0.8 0.7 0.6 0.5
Eigenvalue spreads of 186 60 27 14.5 8.5
with a pre-whitening filter [ ]
[ ] [ ] [ ] [ ] [ ]
PW-DCT-LMS [ ] [ ] [ ] [ ] [ ]
PW-DHT-LMS [ ] [ ] [ ] [ ] [ ]
PW-DFT-LMS [ ] [ ] [ ] [ ] [ ]
achieved by
PW-DHT-/6-LMS [ ] [ ] [ ] [ ] [ ]
PW-DFT-/6-LMS [ ] [ ] [ ] [ ] [ ]
with a pre-whitening filter
PW-DCT-LMS 1.44 1.29 1.09 1.09 1.31
Eigenvalue Spreads of
5. Computational complexity
According to [24], the computational complexity of the conventional TDLMS (without including the
complexity of the used transform) is multiplications and additions for each iteration. The
computational complexity of the proposed PW-TDLMS is increased by only one addition and one
multiplication due to the first order pre-whitening filter. Therefore, by including the complexity of the
transform, the computational complexities of the DCT-LMS and proposed PW- -LMS are given in
Table 3 for one iteration and different values of , where , and are the numbers of
multiplications, bit-shifts and additions, respectively. It should be mentioned that we have used for the DCT
the complexity required by the fast algorithm developed in [25] and for the the complexity
required by the algorithm reported in [17]. It is seen from Table 3 that the proposed PW- -LMS
outperforms the DCT-LMS in terms of the computational complexity. Moreover, it is well-known that the
structural complexity of fast DHT is lower than that of fast DCT. Therefore, the structural complexity of the
proposed PW- -LMS, which is based on the that has a structure similar to that of the
DHT, is lower than that of the DCT-LMS.
Table 3
Computational complexities of the DCT-LMS and proposed - -LMS.
DCT-LMS - -LMS
6. Simulation results
In this section, we present a set of simulations to test and compare the performance of the proposed PW-
TDLMS with that of the TDLMS using the DCT, DFT, DHT, DFT -/6 and DHT-/6 transforms. The
Let us first test and evaluate the performance of the TDLMS and proposed PW-TDLMS algorithms by
taking the signal as a stationary speech-like signal obtained by passing a white Gaussian noise with a
variance through an 11th order AR filter. The input signal of both the noise path and the
adaptive filter is considered as a colored noise generated from a first order AR process (AR (1)) with
and in order to evaluate the efficiency of the proposed algorithm for AR (1) input noise signals
with different eigenvalue spreads of as shown in Table 2. The signal obtained at the output of is
then used for corrupting with an SNR equals to 0 dB. The length, step size, smoothing factor and
regularization parameter of the adaptive filter are N = 16, , = 0.9 and = 2.510-2, respectively.
The proposed PW-TDLMS algorithm is used for the various transforms with the pre-whitening filter
(26)
Figs. 11 and 12 show the results of this test in terms of the MSE convergence of the TDLMS and the
proposed PW-TDLMS algorithms for the various transforms with and , respectively. It can be
seen from Fig. 11, which corresponds to a relatively lower eigenvalue spread of of the AR(1) input noise
signal, that the proposed PW-TDLMS algorithm performs well with the suggested pre-whitening filter
[ ] and outperforms the TDLMS algorithm for the various transforms in terms of the MSE
convergence speeds and reached steady state levels. Similar conclusion can be made from Fig. 12, which
corresponds to an AR (1) input noise with a high eigenvalue spread of . Figs. 11 and 12 show also that
the performance obtained using the suggested pre-whitening filter [ ] approaches that obtained
It has been shown in [4] that the decorrelated NLMS (DNLMS-NLMS) algorithm is better than the
NLMS algorithm. Thus, we compare in Fig. 13 the MSE convergence of the proposed PW-DCT-LMS
algorithm with that of the former by considering the speech-like signal corrupted by AR(1) colored noise.
The experiments are performed for the following three different cases:
Case 1: The weight vector adaptation step size is the same for both the algorithms.
Case 2: The adaptation step sizes of the DNLMS-NLMS are fixed to achieve an MSE steady state similar to
It is seen from Fig. 13 that the proposed PW-DCT-LMS algorithm significantly outperforms the
-30
-30
DCT-LMS
0.6 DFT-LMS
DHT-LMS
-35
-35
DFT-/6-LMS
-35 DHT-/6-LMS
0.4 PW-DCT-LMS (a1=0.65)
PW-DFT-LMS (a1=0.65)
PW-DHT-LMS (a1=0.65)
0.2 -35 PW-DFT-/6-LMS (a1=0.65)
MSE (dB)
-40
-40
-40 PW-DHT-/6-LMS (a1=0.65)
Tap weights
PW-DCT-LMS (a1=0.5)
PW-DFT-LMS (a1=0.5)
0 PW-DHT-LMS (a1=0.5)
-45
-45
-40
-35 PW-DFT-/6-LMS (a1=0.5)
-45 PW-DHT-/6-LMS (a1=0.5)
-0.2
-50
-50
-35
-45
-40
-50
-0.4
-55
-55
-35
-40
-50
-45 0 1000 2000 3000 4000 5000 6 000
1 3 5 7 9 11 13 15 -55
1000 2000 3000 4000 5000 6000
Tap Number
Iterations
Fig. 10. Impulse response of the noise path . -35
-40
-45
-55
-50
Fig. 11. MSE convergence of the TDLMS and proposed PW-TDLMS with
suggested and optimal values of for various transforms in the case of a
speech-like signal, N = 16 and AR (1) colored noise with .
-40
-45
-50
-55
-45
-50
-55
-30
-30
-20
-20
DCT-LMS
DFT-LMS ProposedPW-DCT-LMS
Proposed PW-DCT-LMS
DHT-LMS
DFT-/6-LMS -50 -25
-55 -25 DNLMS-NLMS( (
DNLMS-NLMS
-35
-35
-35 DHT-/6-LMS
PW-DCT-LMS (a1=0.65)
-35 DNLMS-NLMS( (
DNLMS-NLMS
PW-DFT-LMS (a1=0.65) -30
-30
DNLMS-NLMS( (
PW-DHT-LMS (a1=0.65) DNLMS-NLMS
PW-DFT-/6-LMS (a1=0.65) -55 -35
-35
-40
-40
-40 PW-DHT-/6-LMS (a1=0.65) -40
-35
-35
MSE (dB)
PW-DCT-LMS (a1=0.9)
MSE (dB)
MSE (dB)
-50
-50 -45
-35
-35
-45
-40
-50 -50
-40
-50
-50
-45
-35
-55
-55
-35 -50
-40
-40
-50
-45
-55 1000 2000 3000 4000 5000 6000 7000 8000 -55
-55
1000 2000 3000 4000 5000 6000 7000 8000
1000 2000 3000 4000
Iterations
5000 6000 7000 8000 -55
-45
-35 1000 2000 3000 4000 5000 6000 7000 8000
-50
-40
Iterations Iterations
-35
-40 -55
-45
-45
-55
-50
Fig. 12. MSE convergence of the TDLMS and proposed PW-TDLMS with -35
Fig. -40
13.
-50 MSE convergence of the DNLMS-NLMS and proposed PW-DCT-
suggested and optimal values of for various transforms in the case of a LMS-55 algorithms for a speech-like signal corrupted by AR (1) colored noise
speech-like signal, N = 16 and AR (1) colored noise with . -45
with and N = 16.
-40
-45
-50
-55 -50
-40
-55
-45
-50
-45
-50
-55 -55
-45
-50
-55
-50
-55 -50
-55
-55 -55
6.2 Test on real speech signal
The results obtained in Subsection 6.1 for the case of a speech-like signal also confirm that the proposed
PW-TDLMS algorithm, for the various transforms including the DHT-/6, outperforms the DCT-LMS
algorithm in terms of the convergence of the MSE. Moreover, as shown in Section 5, the proposed PW-
DHT-/6-LMS has reduced computational and structural complexities compared to the DCT-LMS algorithm.
Therefore, we carry out in this subsection a complete comparison between the performance of the proposed
PW-DHT-/6-LMS and that of the DCT-LMS by considering a clean speech signal taken from the noisy
speech corpus (NOIZEUS) described in [26]. The clean speech signal used corresponds to the sentence 24
A cruise in warm waters in a sleek yacht is fun. The speech signal is recycled four times in order to get a
signal with sufficient duration for the simulations and then corrupted by a test noise. The test noises, namely
Babble, Street, Train, and Airport, are taken from the LABROSA web site [27]. These noises are passed
through the noise path given by (26) and added to the clean speech signal with different SNR levels: 0, 5,
10 and 15 dB. A portion of the clean speech signal is presented in Fig. 14(a), whereas its corresponding
noisy speech is given in Fig. 14(b) for Street noise with SNR=0 dB. The noise cancellation in the cases of
the DCT-LMS and proposed PW-DHT-/6-LMS is carried out by considering the filter length N = 16,
, and the step sizes for the babble noise and for the other noises. A
portion of the resulting enhanced speech after convergence is presented in Fig. 15 for each of the two
algorithms.
The results corresponding to this test are shown in Figs. 16-21 and Table 4. It is clear from Fig.16 that the
proposed PW-DHT-/6-LMS algorithm outperforms the DCT-LMS algorithm in terms of the convergence
speed and steady state level achieved by the EMSE. Fig. 17 shows the residual noises in the denoised
signals at the outputs of the proposed PW-DHT-/6-LMS and DCT-LMS algorithms. It is clear from this
figure that the residual noise in the signal at the output of the proposed PW-DHT-/6-LMS algorithm is
weaker than that in the signal at the output of the DCT-LMS algorithm and the quality of the corresponding
denoised signal is better than that of the DCT-LMS, which is also confirmed by listening tests. Figs. 18, 19,
20 and 21 present the output SNRs evaluated at the outputs of the proposed PW-DHT-/6-LMS and DCT-
LMS algorithms for different input SNR levels and different noises. It is seen from these figures that the
proposed PW-DHT-/6-LMS algorithm achieves noticeably higher SNRs compared to the DCT-LMS
algorithm. Table 4 shows that the proposed PW-DHT-/6-LMS algorithm outperforms the DCT-LMS
algorithm in terms of the steady state excess mean square error EMSEss, Misadjustment M and output SNRs
0.4 0.4
Amplitude
0.2
Amplitude
0.2
(a)
(a)
0 0
-0.2 -0.2
-0.4 -0.4
7 8 9 10 11 12 13 7 8 9 10 11 12 13
Time (Seconds) Time (Seconds)
0.4 0.4
0.2
Amplitude
0.2
Amplitude
(b)
(b)
0
0
-0.2 -0.2
-0.4 -0.4
7 8 9 10 11 12 13 7 8 9 10 11 12 13
Time (Seconds) Time (Seconds)
Fig. 14. A portion of the: (a) clean speech signal, (b) noisy speech signal Fig. 15. A portion (after convergence) of the enhanced speech signal, in the
corrupted by Street noise with SNR = 0 dB. case of Street noise and SNR = 0 dB, at the output of the: (a) DCT-LMS
adaptive filter, (b) proposed PW-DHT-/6-LMS adaptive filter.
-25 0.4
DCT-LMS
-30 0.2
Amplitude
PW-DHT-/6-LMS
(a)
-35 0
-40 -0.2
EMSE (dB)
-45 -0.4
0 1 2 3 4 5 6 7 8 9 10 11 12 13
Time (Seconds)
-50
0.4
-55
0.2
Amplitude
-60
(b)
0
-65
-0.2
-70 -0.4
0 1 2 3 4 5 6 7 8 9 10 11 12 13 0 1 2 3 4 5 6 7 8 9 10 11 12 13
Time (Seconds) Time (Seconds)
Fig. 16. EMSE of the proposed PW-DHT-/6-LMS and DCT-LMS
algorithms for speech signal and Street noise with SNR = 0 dB. Fig. 17. Residual noise in the signal at the output, with Street noise and SNR
= 0 dB, of (a) the DCT-LMS, (b) the proposed PW-DHT-/6-LMS.
30 30
DCT-LMS DCT-LMS
Proposed PW-DHT-/6 -LMS Proposed PW-DHT-/6 -LMS
25 LMS 25 LMS
20 20
15 15
10 10
5 5
0 0
0 5 10 15 0 5 10 15
Input SNR (dB) Input SNR (dB)
Fig. 18. Output SNRs for the proposed PW-DHT-/6-LMS and DCT-LMS Fig. 19. Output SNRs for the proposed PW-DHT-/6-LMS and DCT-LMS
algorithms with Airport noise and different input SNR levels. algorithms with Street noise and different input SNR levels.
30 35
DCT-LMS DCT-LMS
Proposed PW-DHT-/6 -LMS Proposed PW-DHT-/6 -LMS
25 LMS 30 LMS
25
20
Output SNR (dB)
Output SNR (dB)
20
15
15
10
10
5
5
0 0
0 5 10 15 0 5 10 15
Input SNR (dB)
Input SNR (dB)
Fig. 20. Output SNRs for the proposed PW-DHT-/6-LMS and DCT-LMS Fig. 21. Output SNRs for the proposed PW-DHT-/6-LMS and DCT-LMS
algorithms with Babble noise and different input SNR levels. algorithms with Train noise and different input SNR levels.
Table 4
Steady state excess mean square error , Misadjustment and output SNR obtained by the DCT-LMS and proposed PW-
DHT-/6-LMS algorithms for P = 60000 and J = 200.
7. Conclusion
In this paper, a new pre-whitening transform domain LMS (PW-TDLMS) algorithm has been developed
by introducing a pre-whitening using a simple finite impulse response decorrelation filter of order one
before applying the transform to reinforce its decorrelation. It has been shown that the proposed PW-
TDLMS algorithm significantly decreases the eigenvalue spreads compared to the conventional TDLMS for
all considered transforms, namely the DCT, DFT, DHT, and parametric DFT and DHT. In both the
algorithms, the DCT has been found to be the best. In addition, we have studied the eigenvalue spreads in
the cases of the parametric DFT and DHT for different values of the parameters and found that the DHT-/6
presents a good compromise between the eigenvalue spreads and computational complexity. In order to
evaluated the performance of the proposed PW-TDLMS algorithm and compare it to that of the
conventional TDLMS, we have considered the speech denoising application and used the DHT-/6 in the
former and the DCT in the latter. All the obtained simulation results confirm the high efficiency and
superiority of the proposed PW-TDLMS algorithm compared to the conventional TDLMS algorithm in
terms of the mean square error convergence speed, reached steady state level, residual noise in the denoised
signal, steady state excess mean square error, misadjustment and output SNR. Moreover, we have shown
that the computational and structural complexities of the proposed PW-DHT-/6-LMS are lower than those
of the conventional DCT-LMS. Therefore, the former would be an excellent alternative to the latter for
References
[1] S. Haykin, Adaptive Filter Theory, 3 rd ed., Prentice-Hall, 1996.
[2] D. I. Kim, P. D. Wilde, Performance analysis of signed self-orthogonalizing adaptive lattice filter, IEEE Transactions on
Circuits and Systems II: Analog and Digital Signal Processing 47 (11) (2000) 1227-1237. doi: 10.1109/82.885130.
[3] E.V. Kuhn, J. E. Kolodziej, R. Seara, Analysis of the TDLMS algorithm operating in a nonstationary environment, Digital
Signal Processing 45 (2015) 69-83. doi: 10.1016/j.dsp.2015.05.013.
[4] S. Gazor, T. Liu, Adaptive filtering with decorrelation for coloured AR environments, IEE Proceedings - Vision, Image and
Signal Processing 152 (6) (2005) 806-818. doi: 10.1049/ip-vis:20045260.
[5] S. C. Douglas, A. Cichocki, S. Amari, Self-whitening algorithms for adaptive equalization and deconvolution, IEEE
Transactions on Signal Processing 47 (4) (1999) 1161-1165. doi: 10.1109/78.752617.
[6] M. Mboup, M. Bonnet, N. Bershad, LMS coupled adaptive prediction and system identification: a statistical model and
transient mean analysis, IEEE Transactions on Signal Processing 42 (10) (1994) 2607-2615. doi: 10.1109/78.324727.
[7] J. Lu, X. Qiu, H. Zou, A modified frequency-domain block LMS algorithm with guaranteed optimal steady-state
performance, Signal Processing 104 (2014) 27-32. doi: 10.1016/j.sigpro.2014.03.029.
[8] S. Narayan, A. Peterson, M. Narasimha, Transform domain LMS algorithm, IEEE Transactions on Acoustics, Speech, and
Signal Processing 31 (3) (1983) 609-615. doi: 10.1109/TASSP.1983.1164121.
[9] F. Beaufays, Transform-domain adaptive filters: an analytical approach, IEEE Transactions on Signal Processing 43 (2)
(1995) 422-431. doi: 10.1109/78.348125.
[10] K. Mayyas, New transform-domain adaptive algorithms for acoustic echo cancellation, Digital Signal Processing 13 (3)
(2003) 415-432. doi:10.1016/S1051-2004(03)00004-6.
[11] S. Zhao, Z. Man, S. Khoo, H. R. Wu, Stability and Convergence Analysis of Transform-Domain LMS Adaptive Filters With
Second-Order Autoregressive Process, IEEE Transactions on Signal Processing 57 (1) (2009) 119-130. doi:
10.1109/TSP.2008.2007618.
[12] B. F. Boroujeny, S. Gazor, Selection of orthonormal transforms for improving the performance of the transform domain
normalised LMS algorithm, IEE Proceedings F - Radar and Signal Processing 139 (5) (1992) 327-335. doi: 10.1049/ip-f-
2.1992.0046.
[13] N. Ahmed, T. Natarajan, K. R. Rao, Discrete Cosine Transform, IEEE Transactions on Computers C-23 (1) (1974) 90-93.
doi: 10.1109/T-C.1974.223784.
[14] D.I. Kim, P.D. Wilde, Performance analysis of the DCT-LMS adaptive filtering algorithm, Signal Processing 80 (8) (2000)
1629-1654. doi:10.1016/S0165-1684(00)00098-0.
[15] S.V. Vaseghi, Advanced Digital Signal Processing and Noise Reduction, 3rd ed., Wiley, 2006.
[16] K. Ngo, T.V. Waterschoot, M.G. Christensen, M. Moonen, S.H. Jensen, Improved prediction error filters for adaptive
feedback cancellation in hearing aids, Signal Processing 93 (11) (2013) 3062-3075. doi:10.1016/j.sigpro.2013.03.042.
[17] S. Bouguezel, M. O. Ahmad, M. N. S. Swamy, New Parametric Discrete Fourier and Hartley Transforms, and Algorithms for
Fast Computation, IEEE Transactions on Circuits and Systems I: Regular Papers 58 (3) (2011) 562-575. doi:
10.1109/TCSI.2010.2072151.
[18] M. Kamenetsky, Accelerating the convergence of the LMS adaptive algorithm, Ph.D. dissertation, Stanford Univ., Stanford,
CA, 2005.
[19] J. J. Shynk, Frequency-domain and multirate adaptive filtering, IEEE Signal Processing Magazine 9 (1) (1992) 14-37. doi:
10.1109/79.109205.
[20] Z. Ramadan, A. Poularikas, An adaptive noise canceller using error Nonlinearities in the LMS adaptation, Proceedings of the
IEEE Southeast Con. (2004) 359-364. doi:10.1109/SECON.2004.1287943.
[21] C.H. Chen, Signal and Image Processing for Remote Sensing, Taylor & Francis, 2007.
[22] U. Grenander, G. Sego, Toeplitz Forms and Their Applications, 2nd ed., New York: Chelsea. 1984.
[23] R. M. Gray, Toeplitz and circulant matrices: II, Tech. Rep. 6504-1, Inform. Syst. Lab., Stanford Univ., 1977.
[24] R. C. Bilcu, P. Kuosmanen, K. Egiazarian, A transform domain LMS adaptive filter with variable step-size, IEEE Signal
Processing Letters 9 (2) (2002) 51-53. doi: 10.1109/97.991136.
[25] M. Puschel, Cooley-Tukey FFT like algorithms for the DCT, Acoustics, Speech, and Signal Processing, IEEE International
Conference on Acoustics, Speech, and signal processing 2 (2003) 501-504. doi: 10.1109/ICASSP.2003.12024133
[26] Y. Hu, P.C. Loizou, Subjective comparison and evaluation of speech enhancement algorithms, Speech Communication 49
(78) (2007) 588-601. doi:10.1016/j.specom.2006.12.0063
Highlights
A novel pre-whitening transform domain LMS algorithm is proposed.
This pre-whitening is introduced before applying the transform to reinforce its decorrelation using a
simple finite impulse response decorrelation filter of order one.
In adaptive speech denoising, the proposed algorithm outperforms the conventional algorithms for
different transforms in terms of the computational complexity, mean square error convergence
speed, reached steady state level, residual noise in the denoised signal, steady state excess mean
square error, misadjustment and output SNR.