BASIS EXPANSION ADAPTIVE FILTERS FOR TIME-VARYING SYSTEM IDENTIFICATION
Luca
Rugini
and Geert
Leus
Delft University of Technology
Faculty of Electrical
Engineering, Mathematics, and Computer Science
Mekelweg 4, 2628 CD Delft, The
Netherlands
{I.rugini,g.leus} @tudelft.nl
ABSTRACT block adaptive (OBA) and the OBAshifting (OBAS) algorithms [9].
However, the tracking performance of all these block algorithms Inthis paper, we extend the concept of block adaptive filters to what
Hih
suers
the'opeatinse
theyigorethe
we call basis
expansion
adaptive
filters.
While in blockadaptive
ehighly
suffers from the averagngoperation,
sincethey
ignore
the filtersthe system is assumed to be constant within ablock, our ba-to
vercom
the
tracking
problem
oconventioablo
dap-sisexpansion adaptivefilters modelthe time variation of the systemTi
flers,
me
try
toaxlit
asible
od
fortentime
arato
withinablock by a set ofbasis functions. This allows us to improvewthi
ailock.Oet
exaple
sithbae
model (BEM) thetracking performance
ofblockadaptive
filtersconsiderably.
wihnabokWe.n
exml.
ste ai xasonmdl(E)theus
tracking
pefoh grmadinc
oblok
eadaptive
filters,
clthonside
.Wex
where
the
timevariability
isexpressed
as a linearcombination
offocus onstochasticgradient type ofadaptive filters,
although
basis
functions,which could becomplex exponentials, polynomials,
and so on [10, 11, 12, 13].
Specifically,
the authors of [12] propose Index Terms- Basis expansion model, block adaptive filters, an RLS algorithm that combines recursive block processing with alineartime-varying systems polynomialmodel for the time variability.
Inthis paper, we propose block adaptive stochastic gradient al-1. INTRODUCTION gorithms that
incorporate
aBEMforthe time variation of thesystem
impulseresponse. The proposed algorithm, called basis expansion Adaptivefiltering is a widespread and largely investigated option to LMS (BE-LMS), is a generalization of many already-known adap-tackletheproblemofsystem identification [1]. Inthe last decades, tive algorithms like normalized LMS (NLMS), OBA, and OBAS [9]. many adaptivealgorithms have been proposed ranging from the sim- We showby simulations that our proposed BE-LMS is able to track ple stochastic gradient type of algorithms, lead by the least mean the system time variation better than the NLMS, while maintaining squares (LMS) family, to more elaboratedalgorithmslikethose be- good convergence properties in case of overlapping blocks. To even longing tothe recursive least squares (RLS) class [1, 2, 3]. further improve the tracking capabilities, we also propose another Acommonfeature of most ofthese algorithms is that they have BEM-based adaptive algorithm that has a nice connection with the beeninitially derived assuming a time-invariant system model. In matrix generalization of the momentum LMS (MLMS) algorithm case of atime-varying model, adaptive algorithms like LMStry to [14].
trackthe time variation of the system impulse response, as analyzed
in[4, 5, 6]. However,this tracking is done on a sample-by-sample 2. SYSTEMMODEL
basis, byusing limited informationabout the past. This past
informa-tion canbe better taken into accountbyusing ablock of data rather The system under consideration isalineartime-varying system with than one single sample. As a consequence, to track time-varying additivenoise,whoseinput-outputrelation isexpressedby
systems, ablock-by-block algorithmcouldpotentiallybe more ben- L-1
eficial than asample-by-sample approach, especially if the blocks L- 1
areoverlapping.
31n
=h0,1xn
l+en, (1)Traditionally, block adaptive filteringtechniques have been de- 1-0
velopedtoreducecomplexity in the time-invariant case,byexploit- where yn is the output,
hn,1
is thelinear time-varying system im-ing fast convolutionmethods. For example, the block LMS (BLMS) pulse response, assumed finite with L taps, Xn is the input, anden
algorithm [3, 7] uses an average gradient over ablock of samples is the additive noise. Given Xn andyn,our goal is to find alinear toestimatetheimpulseresponse. Incaseofnon-overlappingblocks, time-varyingfilter
hn,1
such thatBLMShas approximately the same steady-state performance of LMS
[7,8], withcomputationalsavings due tothe shared operations. How- L- 1
ever, ifthe blocks areoverlapping, as inthe sliding-window LMS Yn
=E hn,lXn-1
(SW-LMS), the convergence and the steady-state performance of l=0
blockadaptivealgorithms improve, atthe expense of an increased
complexity caused by the repeated processing of the same data[3]. isascloseaspossible to yn. The filter
h,l
can then beconsideredThe convergence of both BLMS and SW-LMS can further be im- as an estimate of the system impulseresponse
hn,1.
provdbyallwingatme-vryig stp sze, eadng t th optmum To describe the proposed adaptive algorithms, we will first
re-provd b alowiga imevaringstepsiz, ladig totheoptmum shape thesystem input-output relation, and then we will make use This research was supported in part by NWO-STW under the VIDI pro- of the limitedtime-varying behavior of the system. Let us start by gram (DTC.6577)and the VICI program (DTC.5893). rewriting (1) asyn=xnThn,+en, whereXn
[xn, *.
*X*r
i-L+lland
hn,:
= [h,o,..
h,hn,L -]T. Stackingyn
over ablock of N By performingthis minimization over the BEM coefficients c(k) ofsamples every K samples, we obtain the filter h(k), we obtain the Wiener solution
y(k)
XT(k)h(k)
+e(k),
(2)
iOpt(k) =E(BHX* (k)XT (k)B)-lE(BHX* (k)y(k)).
wherewherey(k)
y(k)
=[yan
[YkK,
* *Y:kK+N-]T,
X(k)
=diag(XkK,
,LN x
yk-diTn
d(kK=h
T Clearly, in the absence of a basis expansion modeling error, theXk,K+N 1) iS anLN\ X N\ h(k ocU equaln atl, c= k
T T block-diagonal matrix, T " Wiener solutionCOpt (k)is
equal
to the actual BEM coefficientsc(k)
,hkK+N-1,:]
,ande(k)
=[ekK,
ekK+N-]T
.Hence,we of the system impulse response h(k). To find this solution adap-actually formulate a sliding-window data model, where K deter-tively,
we follow a standard stochastic gradient descentapproach,
mines how manysamples
we shift theN-length
window in everyleading
towhat we will label the basis expansion least mean squaresstep.
(BE-LMS)
algorithm.
The BE-LMSalgorithm
isdescribedby
By exploiting now the limited time-varying behavior of the
sys-tem,each tap of the system impulse responsehn,1 canbe modeled, c(k + 1) =(k) +,J(k)BHX*(k)
(k),
(6) over a timerange n C {kK ...IkK + N- 1},as asuperpositionofQ < N functions
{bn-kK,q}lQ-l
weighted by the coefficients where8(k)
is a possiblytime-varying step size. In ourBE-LMS,we{cq,i
(k)}qQjjq,
as expressed byimplicitly
assumethat theoptimal
8i(k)
isused. Morespecifically,
q=O
we canderivethe step size that maximizes the convergence speed inQ-1 asimilar manner as in[9].This maximization leads to
hn,1
=E
bn-kcK,qCq,1
(k).
HX .q=O k-
llBHX*(k)J(k)
E 2This model is usually referred to as the basis expansion model (BEM) ()
XT(k)BBI
HX*(k)e(k)j
2[10, 11].Inmatrix-vectornotation, the BEM becomes In practice, a small positive constant £
is
added to thedenominatorh.:,i(k)
=Bci
(k) (3) of8(k)
toprevent apossiblezero ofthe denominator [2].We now show the connections of our BE-LMSalgorithmwith where
h:,i(k)
=[hkK,l,
hkK+N-1,lIT,
[B]m,q
=bm,q
where thevastliteratureonadaptive algorithms.
First ofall,
it is easytothe indices start from zero, and
cl(k)
=[co,i(k),
...,CQl,l]. seethat when the block size N = 1, the BEM can notreally beNotethat the basis matrix B is independent of the time shift kK.
exploited
andonly Q
= 1 ispossible. In this case, the BE-LMSFor simplicity reasons, weassume that B is an isometry, i.e.,BHB boils down to the standard NLMS algorithm, andtothe standard
=
IQ.
Goodchoices for the BEM functions could be polynomi- LMSalgorithmif in addition the step size iskeptconstant. Anotherals orcomplex exponentials, leading to a polynomial (POL) BEM interesting connection occurs when we model the systemimpulse [12] or a complex exponential (CE) BEM[10, 11],respectively. In response asbeingconstantwithinawindow of Nsamples. In this these cases, B is derived as an orthonormalization of a matrix B, case,wehavea
single
BEMfunction, i.e., Q= 1, andwemodel itaswhere
[B]m,q
=(m
- N/2) forthe POL-BEM and[B]m,q
a constant function, i.e., B [1/,...N
1/I NT.It is then easyj27,mq/(KN)
forthe CE-BEM, with s a positive integer that con- to show that the BE-LMS coincides with the OBASalgorithm[9], trolsthe frequency separation of the basis functions [13]. and with the SW-LMS algorithmif in addition thestepsize iskeptWe canexpress
h(k)
ash(k)
=[h7
,hT
K+Nl,:IT constant [3]. In thespecialcasewhere K= N, i.e.,werecompute[IN 0 io
,.
IN®X
iL-]
[hTo(k),.
***,
hT(k)]T,
whereiT
is the a new set of BEM coefficients every window of N samples, theBE-lthcolumnofIL,with indices starting from zero, and X represents LMS becomes the OBA algorithm [9], and the BLMS algorithm in the Kroneckerproduct. We now use(3) toobtain
h(k)
=[IN
x caseM(k)
=io,
. ..,
IN
XiL-1](IL
XB)c(k)
=[B
®io,..
., B ®iLl]c(k)
= The OBAS or SW-LMS have basically been introduced toim-Bc(k), where c(k) = [c
T(k),
.*,
CT_i(k)]T
and B = [B x provethe convergence behavior of NLMS andLMS,
but,
because ofio, . .,B xiLi]. Finally, we can rephrase (2) as the large window size N, they suffer from a worse tracking
perfor-mance. The BE-LMS for
Q
> 1 solves thisproblem by
modeling
y(k) = XT(k)Bc(k) + e(k). (4) the time variation of the systemimpulseresponsewithin theconsid-ered windowwith a BEM. On the otherhand, the OBA and BLMS Since thesystem
impulse
responsehn,1
is assumedtofollowa algorithms were mainly introduced to further reduce the complex-BEM, we put a similar BEM constraint also onthe filterhn,1,
thereby ity of NLMS and LMS, but they naturally suffer inconvergence and reducing the number of unknowns that we want to estimate. Hence, tracking due to the largeshift K. The BE-LMS for Q > 1expe-followingsimilar steps asbefore,weobtain riences a similarconvergence and tracking behavior when the shift
A~(k)
=XT(k)(k)fi =X (k)13,&(k),parameter
K islarge. Tofurther improve the performance ofBE-^T(k)
= XT(k)h(k) =XT(k)Bd(k), (5)LMS,
wealso proposetoincludesomeprior
knowledge
onthe time-where:i(k),
h(k),
andi(k)
represent the estimated versions ofvarying
behavior ofc(k),
whichactually
leadstoaBE-LMSalgo-y(k), h(k),
andc(k),
respectively.
rithm withanacceleration featurethat will be calledmomentum.3. BASIS EXPANSION LEAST MEANSQUARES 4. BASIS EXPANSION MOMENTUM LMS
Obviously
the time-varying behavior of
c(k)
isdetermined
bythe
yi(k)
in (5) as close as possibleto y(k) in (4), given X(k) and y(k) time variation of the system. However, part of the transition from Basically, we want to minimize E( .(k) l2), wheree.(k)is theerror c(k- 1) to c(k) alsodepends on how the BEM functions of twosuc-signal given by: cessive blocks are related. In other words, when the blocks are
over-lapping, h(k) shares some elements with h(k -1), which clearly
e(k) =y(k)
-y(k) =y(k) -XT(k)Bd(k).
introduces a relationship between the BEM coefficients c(k) andc(k - 1). Therefore, by defining J as an N x N shift matrix, 5. SIMULATION RESULTS with ones along the Kth upper diagonal and zeros elsewhere, we
canmodelthe transition fromh(k- 1) toh(k)as Inorderto testtheproposedblockadaptive algorithms,weconsider the estimation of a linear time-varying channel driven by a white
h(k) = (J xIL)h(k- 1) +g(k), input sequence of quadrature phase-shift keying (QPSK) symbols,
withg(k) [OTN 1 hT hhT 1]T.Thus, assumed knownatthe receiver. Ina more
practical
case,only pilot
wet
write(N-K)LX
17vkK+N,: h(k+1)K+N-1,
symbols would be known, and the data symbols would be used inwe can write a decision-directed way. We assume a multipath FIR channel with
c(k) =BH(J XI)Bc(k 1)+BHg(k) =Ac(k 1)+d(k) L - 8 independent taps following a Jakes' Doppler spectrum. The
additive noise en is assumed white and Gaussian. The signal-to-where A = BH(J 0 IL)B andd(k) = BHg(k). Looking at noise ratio (SNR) is equal to 30 dB. For all the adaptive algorithms,
theright-hand-side ofthe equation above, the first term can be in- £ = 0.0001
is
added to the step-size denominator. To compareterpreted
as aprediction
ofc(k),
whereas the second term canbe theperformance
oftheadaptive
algorithms,
we lookatthemean-viewedasthe
corresponding prediction
error. Hence,by
exploiting
squared
error(MSE)
ofthe channel estimate. Inthe MSEcomputa-thisrelationshipin(4), weobtain tion, we include only the new elements and exclude those elements
already computed
in theprevious overlapping
block. As aconse-y(k) =
XT(k)BAc(k
- 1) +XT(k)Bd(k)
+ e(k). quence,wedefine the MSEasNow we can consider
XT(k)BAc(k
-1)
as abias ony(k).
By MSE= IE(11f(k)
-h(k)]L(N-K)+l:LN|
1)
subtractingthis bias term,weobtain KL
y(k) -XT(k)BAc(k- 1) =XT(k)Bd(k) +e(k), (7) where [X]a:b stands for the subvector of x from the ath to the bth
element.
which is similar to the input-output relation expressedby(4). As a Figure 1 shows the MSE as a function of the maximum Doppler consequence, we canupdatetheprediction error
d(k)
ina similar spread, normalized to the sampling period, when N=64 and K =way asbefore, leadingto 1. Toobtain the simulated MSE, the results are averaged over 10
I I HT longchannel realizations, after transient effects. Because of the
gra-d(k
+1)
=d(k)
+II(k)BHX*(k)E(k),
(8) dient averaging, the OBAS algorithm (or equivalently the BE-LMSalgorithm with Q = 1) outperforms the NLMS algorithmin the
where
e(k)
is defined as before with time-invariant case (no Doppler spread). However, the performance of OBAS rapidly degrades forincreasing Doppler spread, becausec(k)
=Ac(k- 1) +d(k). (9) morebasis functions are necessary to model the timevariability of A .~~~~~~~~~~the
channel. Similarly, the performance of the NLMS algorithm sig-Itisworthnotingthatc(k)
in(9)depends onthe true model coeffi-nificantly
Senslwhen
theter
variation ie
alessithan
cientsc(k
-1),
andhence isnotavailable. As aconsequence, wefoOaS.
worsenswhen thetsme varlatgon
increases,
but less thanreplace~ ~
c.
it'siae
A)
by eso ~k-1,laigt for OBAS. This iS due to the smallerlag
with respectto OBAS.replace c
-1bietadeOn
the contrary, the BE-LMSalgorithm with
Q > 1 is able toc(k)
=A(k-1)
+a(k)k
(10) better track the channelvariability,
and hence itsperformance
im-proveswith respect to the otheralgorithms. Obviously, for increas-Togetmore
insight
intothisupdating
formula, letusplug (10)
intoing
Doppler
spread,
the bestperformance
is obtainedby
BE-LMS(8) to obtain
with
anincreasing
number ofbasis functions
Q,which
allows tocapturethe increasing timevariability. Thisexplainswhy Q = 2is
c(k
+ 1) -Ac(k)
= (k) -Ac(k
- 1) +JJ(k)BHX*(k)E(k),
optimal only
forarelatively
small timevariability,
andworsensforhigherDoppler spreads.
orequivalently Figure 2 illustratesthe steady-state performance of the proposed
BE-LMS for time-invariantchannels as a function of the block size
c(k
+1)
=(k)
+,J(k)BHX*(k)E(k)
+A(c(k)
-c(k
-1)).
N, when K = 1. Asexpected, anincrease ofthe block sizepro-duces an improvement for all the block-based algorithms. When the Clearly, this equation is formally similar to (6), but with the addi- blocksize issufficiently large,the BE-LMS algorithmswithQ > 1 tional term
the~~~~~~~~~~~
A(c(k)-cd(k-1)).
Such aproceduregreatlyresembles areable
tooutperform
the NLMSalgorithm.
momntu
algorithm.[4
tooupefrmthS)M
able hc asitouethemomentumLMS
(MLMS)
algorithm
[14],
whichwasintroduced InFigure
3,theprediction
effect ofthe BE-MLMSalgorithmisto
speed
up the convergence and thetracking
of LMS.This iswhy
displayed using
asimulationexample,
which compares the realpartwelabel thisalgorithmasthe basisexpansionmomentumLMS(BE- ofonetapofthetruechannel with that of the estimated channel. In
MLMS)
algorithm. Inthe simulations section we will show that in this case, N 256, K 64,and the normalizedDoppler spread
some cases the BE-MLMS
algorithm
exhibits abettertracking
be- isequalto0.002. It isevidentthat the BE-MLMScould
be auseful
havior than BE-LMS. alternative to the BE-LMS in order to track a time-varying channel
It should be observed that astraightforward application ofthe when K > 1. momentumapproach tothe BE-LMS would lead to a different
al-gorithm with respect to our BE-MLMS. Indeed,inthe standard
mo-mentum approach, the step size is constant and, more importantly, 6. CONCLUSIONS the fixed matrix A is replaced by a scalar design parameter
ae
[14].A good value for
ae
would be rather difficult to determine; however, In this paper, we have developed some block adaptive algorithms that in the proposed set-up, we do not have to make this difficult choice, are suitable for the identification of linear time-varying systems. The and an intuitively pleasing value is selected for the matrix A. proposed BE-LMS algorithm, which extends the OBAS approach10U - 0 8 | |~U b True channel
-E- BE-LMS (POL, Q=3) - BE-MLMS (POL, Q = 3) 0.6-10 0.4- H-10~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 10~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~0 -a BE-LMS(POL,Q=2) -0.2-/v BE-LMS(CE,Q=3,K =2)
E o-4- BE-LMS(CE,Q=3,K=4) - 0 . 4 _ _ _l_l_ l _l_ _
io-4
~~~~~~~~~~~~~~~~~~~~~-0.4-0 1 2 3 4 5 6 7 8 0 100 200 300 400 500 600 Normalized Doppler spread x 10-3 Time index
Fig. 1. MSEas afunction ofthe normalized Doppler spread. Fig. 3. Example of channel estimate by the BE-MLMS algorithm.
o
2XNS.
[4] E.Eweda,"ComparisonofRLS, LMS,andsignalgorithmfortrack-OBAS 12 ingrandomly time-varying channels," IEEE Trans.Signal Process.,
BE-LMS
(CE,
Q=3,
K =2)
vol.42,pp.2937-2944,Nov. 1994.[5] B. Farhang-Boroujeny and S. Gazor, "Performance of LMS-based
adaptivefilters intrackingatime-varying plant,"IEEETrans.Signal
10-3 Process., vol. 44, pp. 2868-2871, Nov. 1996.
[6] S.Haykin,A. H.Sayed,J. R.Zeidler,P.Yee,andP.C.Wei,"Adaptive
trackingof linear time-variantsystemsbyextended RLSalgorithms",
U) ^ <IEEETrans.SignalProcess.,vol.45,pp. 1118-1128,May1997.
[7] G. A.Clark,S. K.Mitra, andS. R.Parker,"Blockimplementationof
X4
adaptivedigital filters,"IEEETrans.Acoustics, Speech, SignalPro-10 cess.,vol.29,pp.744-752, June 1981.
[8] A.Feuer,"Performanceanalysisofthe blockleastmeansquare algo-rithm,"IEEETrans.CircuitsandSystems,vol.32, pp. 960-963, Sept.
1985.
[9] W. B.Mikhael andF. H.Wu,"FastalgorithmsforblockFIRadaptive
2 2-55 digital filtering,"IEEE Trans. Circuits andSystems,vol.34, pp.
1152-32 64 128 192 256 1160, Oct. 1987.
Block size
[10] M. K. TsatsanisandG. B.Giannakis,"Modelling and equalizationof Fig.2.MSEas afunction ofthe blocksize inthetime-invariantcase. rapidly fadingchannels,"Int.J. Adapt. Control Signal Process., vol.
10,pp. 159-176, Mar. 1996.
[11] G. B. Giannakis andC.Tepedelenlioglu, "Basis expansion models anddiversity techniques forblind identification andequalization of to time-varying scenarios, is able to track the system impulse re- time-varying channels," Proc. IEEE, vol. 86, pp. 1969-1986, Oct. sponse by conveniently exploiting a basis expansion model of the 1998.
time variability. As aresult, the BE-LMS algorithm outperforms [12] D. K.Borah and B. D. Hart, "Frequency-selective fading channel esti-other adaptive algorithms like NLMS and OBAS. Inaddition, we mation with a polynomial time-varying channel model," IEEE Trans. have pointed out how to exploit the block overlapping in order to ob- Commun., vol. 47, pp. 862-873, June 1999.
tamnamomentum-like block
adaptive
algorithm
called BE-MLMS. [13] G. Leus, "Ontheestimation
ofrapidly time-varying
channels,"Proc.Additional features ofthe proposed algorithms, such as the choice oftheEuropean Signal Processing Conference (EUSIPCO 2004), pp. ofbasis functions and theirnumber,aswellas atheoreticalanalysis 2227-2230, Vienna, Austria, Sept. 2004.
of theexcessMSE,arecurrentlyunderinvestigation. [14] S. RoyandJ. J.Shynk, "Analysis of themomentumLMSalgorithm," IEEETrans.Acoustics, Speech, SignalProcess., vol. 38,pp.
2088-7. REFERENCES 2098,Dec. 1990.
[1] 5.Haykin, Adaptive Filter Theory, 2nd Ed., Prentice-Hall, 1991. [2] A. H. Sayed, Fundamentals ofAdaptive Filtering, Wiley, 2003. [3] G.-O. Glentis, K. Berberidis, and S. Theodoridis, "Efficient least
squares adaptive algorithms for FIR transversal filtering,"IEEE Sig-nal Process. Mag., vol. 16, pp. 13-41, July 1999.