iii Using the stationarized observation data ˆyk, the signal detection is based on the model ˆyk = ˆsk +wk, 4 where ˆskis the modified signal.. Stationarization of Observation Data Recal
Trang 2Ohsumi & Yamaguchi (2006) to estimate the time-delay of signals in nonstationary random
noise, incorporated with the Wigner distribution-based maximum likelihood estimation
In this paper the signal detection problem is investigated using the stationarization approach
to nonstationary data The model of the corrupting noise is given by an ARMA(p, q) model
with unknown time-varying coefficients These coefficient parameters are estimated from the
(original) observation data by the Kalman filter
2 Problem Statement
Let{ y(k)} be the (scalar) observation data taken at sampling time instant t k(k = 1, 2,· · · ),
and assume that it can be expressed as
y(k) =s(k) +n(k) (k=1, 2,· · · ), (1)
where s(·)is a signal to be detected, whose form is surely known, and is assumed to exist
in a brief interval if it exists; and n(·)is the nonstationary random noise In consequence,
the observation data{ y(k)}becomes nonstationary, but its trend time series is assumed to be
removed by the process
where Y(k)is the original data received by the receiver; ∆Y(k) = Y(k)− Y(k −1); and d
indicates the order
In this paper the random noise n(k) is assumed to be given as the output of ARMA(p, q)
model with time-varying coefficient parameters:
where w(·) is the white Gaussian noise with zero-mean and variance parameter σ2;{ α i(·)}
and{ β j(·)}are slowly and smoothly varying parameters to be specified
Then our purpose is to propose a method of detecting the signal s(k)from the noisy
observa-tion data{ y(k)}
The procedure taken in this paper is as follows:
(i) First, based on the noise model (3), coefficient functions{ α i(·)}and{ β j(·)}are estimated
using Kalman filter from the observation data{ y(k)}
(ii) Using the estimates{ ˆα i(·)}and{ ˆβ j(·)} obtained in (i), the observation data y(k)is
modi-fied to become stationary This procedure is called the stationarization of observation data.
(iii) Using the stationarized observation data ˆy(k), the signal detection is based on the model
ˆy(k) = ˆs(k) +w(k), (4)
where ˆs(k)is the modified signal Equation (4) is familiar in the conventional signal detection
problem where the noise is stationary
3 Stationarization of Observation Data
Recalling the assumption that the duration of the signal s(k)is short, neglect the signal in the
observation data and consider the signal-free case, i.e., y(k) =n(k), then the observation data
y(k)is expressed by (1) and (3) as follows:
In order to estimate the time-varying parameters{ α i(k)}and{ β j(k)}in (5), suppose that they
change from step k − 1 to k under random effects { e ·(k)} Define vectors
− α p(k)
β1(k)
− e p(k)
e p+1(k)
1,· · · , τ2
p+q.Then, Eq (5) is expressed formally as
ν () = y () − Hˆ() ˆx ( | −1) (12)and
Here, ˆx ( | −1)and P ( | −1)are the one-step prediction and its covariance matrix computed
by Kalman filter for the past interval
Trang 3Ohsumi & Yamaguchi (2006) to estimate the time-delay of signals in nonstationary random
noise, incorporated with the Wigner distribution-based maximum likelihood estimation
In this paper the signal detection problem is investigated using the stationarization approach
to nonstationary data The model of the corrupting noise is given by an ARMA(p, q) model
with unknown time-varying coefficients These coefficient parameters are estimated from the
(original) observation data by the Kalman filter
2 Problem Statement
Let{ y(k)} be the (scalar) observation data taken at sampling time instant t k(k = 1, 2,· · · ),
and assume that it can be expressed as
y(k) =s(k) +n(k) (k=1, 2,· · · ), (1)
where s(·)is a signal to be detected, whose form is surely known, and is assumed to exist
in a brief interval if it exists; and n(·)is the nonstationary random noise In consequence,
the observation data{ y(k)}becomes nonstationary, but its trend time series is assumed to be
removed by the process
where Y(k) is the original data received by the receiver; ∆Y(k) = Y(k)− Y(k −1); and d
indicates the order
In this paper the random noise n(k) is assumed to be given as the output of ARMA(p, q)
model with time-varying coefficient parameters:
where w(·) is the white Gaussian noise with zero-mean and variance parameter σ2;{ α i(·)}
and{ β j(·)}are slowly and smoothly varying parameters to be specified
Then our purpose is to propose a method of detecting the signal s(k)from the noisy
observa-tion data{ y(k)}
The procedure taken in this paper is as follows:
(i) First, based on the noise model (3), coefficient functions{ α i(·)}and{ β j(·)}are estimated
using Kalman filter from the observation data{ y(k)}
(ii) Using the estimates{ ˆα i(·)}and{ ˆβ j(·)} obtained in (i), the observation data y(k)is
modi-fied to become stationary This procedure is called the stationarization of observation data.
(iii) Using the stationarized observation data ˆy(k), the signal detection is based on the model
ˆy(k) = ˆs(k) +w(k), (4)
where ˆs(k)is the modified signal Equation (4) is familiar in the conventional signal detection
problem where the noise is stationary
3 Stationarization of Observation Data
Recalling the assumption that the duration of the signal s(k)is short, neglect the signal in the
observation data and consider the signal-free case, i.e., y(k) =n(k), then the observation data
y(k)is expressed by (1) and (3) as follows:
In order to estimate the time-varying parameters{ α i(k)}and{ β j(k)}in (5), suppose that they
change from step k − 1 to k under random effects { e ·(k)} Define vectors
− α p(k)
β1(k)
− e p(k)
e p+1(k)
1,· · · , τ2
p+q.Then, Eq (5) is expressed formally as
ν () = y () − Hˆ() ˆx ( | −1) (12)and
Here, ˆx ( | −1)and P ( | −1)are the one-step prediction and its covariance matrix computed
by Kalman filter for the past interval
Trang 4It is a simple exercise to show that the statistical properties of ν m(·) is the same as that of w(·),
i.e., E { ν m(k)} = 0 and E {| ν m(k)|2} = σ2(for proof, see Appendix) Then, instead of (8) we
have the expression,
y(k) =Hˆ(k)x(k) +w(k) (14)The procedure for computing ˆH(k)is stated as follows:
(i) Preliminaries: Assume for the past k (<0)that{ ν m(−1), ν m(−2),· · · , ν m(− q)}are set
appro-priately (may be set all zero), and preassign ˆx(0| −1), ˆP(0| −1)and ˆH(0)as initial values
Repeat Steps (ii) and (iii) for = k − q, k − q+1,· · · , k −1 to obtain ˆH(k) In computing (12)
and (13), ˆx ( | −1)and P ( | −1)are computed by the Kalman filter (e.g., Jazwinski, 1970):
Kalman filter constructed for (7) and (14) (whose form is the same as (15)-(19) replacing
by the present k) Under the basic assumption that the coefficient parameters vary slowly and
smoothly, they can be treated like constants in an interval I k around the current time k Write
them as ˆα ik and ˆβ jk in I k Replacing the past{ w(k − j)}in (5) by the statistically equivalent
sequence{ ν m(k − j)} , define the sequence ˆy(k)by
which implies that the sequence{ ˆy(k)} is stationary because w(k) is the stationary white
noise
4 Signal Detection
After obtained the estimates of coefficient parameters, the observation process (14) may be
written using estimates as
Note that (4)bisis familiar to us as the mathematical model for the detection problem of signals
in stationary noise (e.g., Van Trees, 1968).
Now, consider the binary hypotheses: H1: ˆy(k) =ˆs(k) +w(k), and H0: ˆy(k) =w(k), and let ˆY k
be the stationarized observation data taken up to k, ˆY k={ ˆy (), =1, 2,· · · , k } Since the
ad-ditive noise w(k)is white Gaussian sequence with zero-mean and variance σ2, the
likelihood-ratio function Λ(k) =p { ˆY k | H1} / ˆY k | H0}is evaluated as follows:
Λ(k) =
k
∏
=1(2π)−1exp
Trang 5It is a simple exercise to show that the statistical properties of ν m( ·) is the same as that of w(·),
i.e., E { ν m(k)} = 0 and E {| ν m(k)|2} = σ2(for proof, see Appendix) Then, instead of (8) we
have the expression,
y(k) =Hˆ(k)x(k) +w(k) (14)The procedure for computing ˆH(k)is stated as follows:
(i) Preliminaries: Assume for the past k (<0)that{ ν m(−1), ν m(−2),· · · , ν m(− q)}are set
appro-priately (may be set all zero), and preassign ˆx(0| −1), ˆP(0| −1)and ˆH(0)as initial values
Repeat Steps (ii) and (iii) for = k − q, k − q+1,· · · , k −1 to obtain ˆH(k) In computing (12)
and (13), ˆx ( | −1)and P ( | −1)are computed by the Kalman filter (e.g., Jazwinski, 1970):
Kalman filter constructed for (7) and (14) (whose form is the same as (15)-(19) replacing
by the present k) Under the basic assumption that the coefficient parameters vary slowly and
smoothly, they can be treated like constants in an interval I k around the current time k Write
them as ˆα ik and ˆβ jk in I k Replacing the past{ w(k − j)}in (5) by the statistically equivalent
sequence{ ν m(k − j)} , define the sequence ˆy(k)by
which implies that the sequence{ ˆy(k)} is stationary because w(k) is the stationary white
noise
4 Signal Detection
After obtained the estimates of coefficient parameters, the observation process (14) may be
written using estimates as
Note that (4)bisis familiar to us as the mathematical model for the detection problem of signals
in stationary noise (e.g., Van Trees, 1968).
Now, consider the binary hypotheses: H1: ˆy(k) = ˆs(k) +w(k), and H0: ˆy(k) =w(k), and let ˆY k
be the stationarized observation data taken up to k, ˆY k={ ˆy (), =1, 2,· · · , k } Since the
ad-ditive noise w(k)is white Gaussian sequence with zero-mean and variance σ2, the
likelihood-ratio function Λ(k) =p { ˆY k | H1} / ˆY k | H0}is evaluated as follows:
Λ(k) =
k
∏
=1(2π)−1exp
Trang 6The top of Fig.1 depicts a sample path of the observation data{ Y(k)}generated by calculating
the output of the ARMA(4, 1)-model:
trend was removed by setting d=1 For the Kalman filter (15)∼(19), the parameters are set
kstep
-20 -10 0 10 20
kstep
Fig 2 The trend-removed data y(k)(top) and the stationarized observation data ˆy(k) tom)
Trang 7(bot-The top of Fig.1 depicts a sample path of the observation data{ Y(k)}generated by calculating
the output of the ARMA(4, 1)-model:
trend was removed by setting d=1 For the Kalman filter (15)∼(19), the parameters are set
kstep
-20 -10 0 10 20
kstep
Fig 2 The trend-removed data y(k)(top) and the stationarized observation data ˆy(k)tom)
Trang 8Fig 3 Log-likelihood function L(k).
as Q=diag{0.05, 0.05, 0.05, 0.05, 0.05} and σ2=40 It should be noted that from Fig 2 the
observation data is well stationarized and that even in this figure the signal emerges from the
background noise
Figure 3 shows the result of signal detection by the current log-likelihood ratio function L(k)
Clearly, it exhibits a salient peak around the true time instant k = 300 and this shows the
existence of the signal
(ii) Experiment 2.
Efficacy of the signal detector proposed in this paper is also tested for the pulse signal
Figure 4 depicts observation data and embedded three pulses Random noise n(k)is
gener-ated by the same manner of previous simulation with same coefficients α i(k)and β(k) As a
signals s(k), a train of pulses with same magnitude is considered:
Figure 5 depicts trend-removed data and stationarized data ˆy(k) The trend was also removed
by setting d=1 The parameters of Kalman filter are set as the same of previous experiment
Figure 6 shows the result of signal detection Clearly, log-likelihood ratio function L(k)has large value around each time when each pulse exists Thus the signal detection is wellsucceeded
-200 -100 0 100 200
kstep
-200 -100 0 100 200
kstep
Fig 4 A sample path of the observation data Y(k)(top) and the pulse signal s(k)(bottom)
Trang 9Fig 3 Log-likelihood function L(k).
as Q=diag{0.05, 0.05, 0.05, 0.05, 0.05} and σ2=40 It should be noted that from Fig 2 the
observation data is well stationarized and that even in this figure the signal emerges from the
background noise
Figure 3 shows the result of signal detection by the current log-likelihood ratio function L(k)
Clearly, it exhibits a salient peak around the true time instant k = 300 and this shows the
existence of the signal
(ii) Experiment 2.
Efficacy of the signal detector proposed in this paper is also tested for the pulse signal
Figure 4 depicts observation data and embedded three pulses Random noise n(k)is
gener-ated by the same manner of previous simulation with same coefficients α i(k)and β(k) As a
signals s(k), a train of pulses with same magnitude is considered:
Figure 5 depicts trend-removed data and stationarized data ˆy(k) The trend was also removed
by setting d=1 The parameters of Kalman filter are set as the same of previous experiment
Figure 6 shows the result of signal detection Clearly, log-likelihood ratio function L(k)has large value around each time when each pulse exists Thus the signal detection is wellsucceeded
-200 -100 0 100 200
kstep
-200 -100 0 100 200
kstep
Fig 4 A sample path of the observation data Y(k)(top) and the pulse signal s(k)(bottom)
Trang 10as introduced in this paper will have potential ability to treat the nonstationary noise orobservation data in the signal processing.
Appendix.Proof of Statistical Equivalence Between{w(k) }and{ν m(k) }
The mean of the modified innovation sequence ν m(k)is clearly zero Indeed,
E{ ν m(k)} = c(k)E{ ν(k)}
= c(k)E{ y(k)− Hˆ(k)ˆx(k | k −1)}
Trang 11as introduced in this paper will have potential ability to treat the nonstationary noise orobservation data in the signal processing.
Appendix.Proof of Statistical Equivalence Between{w(k) }and{ν m(k) }
The mean of the modified innovation sequence ν m(k)is clearly zero Indeed,
E{ ν m(k)} = c(k)E{ ν(k)}
= c(k)E{ y(k)− Hˆ(k)ˆx(k | k −1)}
Trang 12Here, recalling that y(k)is given by the form (14), we have
Haykin, S & Bhattacharya, T K (1997) Modular learning strategy for signal detection in a
nonstationary environment IEEE Trans Signal Processing, Vol.45, No.6, pp.1619-1637
Haykin, S & Thomson, D J (1998) Signal detection in a nonstationary environment
reformu-lated as an adaptive pattern classification problem Proc of the IEEE, Vol.86, No.11,
pp.2325-2344
Ijima, H., Ohsumi, A & Okui, R (2006) A method of detection of signals corrupted by
non-stationary random noise via stationarization of the data, Trans IEICE, Fundamentals
of Electronics, Communications and Computer Sciences, Vol J89-A, No.6, pp.535-543 (in
Japanese)
Ijima, H., Ohsumi, A & Yamaguchi, S (2006) Nonlinear parametric estimation for signals in
nonstationary random noise via stationarization and Wigner distribution, Proc 2006 Int Symp Nonlinear Theory and its Applic (NOLTA 2006), Bologna, Italy, pp.851-854
Ijima, H., Okui, R & Ohsumi, A (2005) Detection of signals is nonstationary random noise via
staionarization and stationary test, Proc IEEE Workshop on Statistical Signal Processing (SSP’05), Bordeaux, France, Paper ID 68
Jazwinski A H (1970) Stochastic Processes and Filtering Theory, Academic Press, New York Van Trees, H L (1968) Detection, Estimation, and Modulation Theory, Part I, John Wiley
Trang 13Direct Design of Infinite Impulse Response Filters based on Allpole Filters
This chapter presents a new framework to design different types of IIR filters based on the
general technique for maximally flat allpole filter design The resulting allpole filters have
some desired characteristics, i.e., desired degree of flatness and group delay, and the desired
phase response at any prescribed set of frequency points Those characteristics are important
to define the corresponding IIR filters The design includes both real and complex cases
In that way we develop a direct design method for linear-phase Butterworth-like filters, using
the same specification as in traditional analog-based IIR filter design The design includes the
design of lowpass filters as well as highpass filters The designed filters can be either real or
complex The design of liner-phase two-band filter banks is also discussed
Additionally, we discussed the designs of some special filters such as Butterworth-like filters
with improved group delay, complex wavelet filters, and fractional Hilbert transformers
Finally, we addressed a new design of IIR filters based on three allpass filters As a result we
propose a new design of lowpass filters with a desired characteristic based on the complex
allpole filters
Closed form equations for the computation of the filter coefficients are provided All design
techniques are illustrated with examples
1 Introduction
The design of allpole filters has been attractive in the last years due to some promising
appli-cations, like the design of allpass filters (Chan et al., 2005; Lang, 1998; Pun & Chan, 2003;
Se-lesnick, 1999; Zhang & Iwakura, 1999), the design of orthogonal and biorthogonal IIR wavelet
filters (Selesnick, 1998; Zhang et al., 2001; 2000; 2006), the design of complex wavelets
(Fernan-des et al., 2003), the (Fernan-design of half band filters (Zhang & Amaratunga, 2002), the filter bank
design (Kim & Yoo, 2003; Lee & Yang, 2004; Saramaki & Bregovic, 2002), the fractional delay
filter design (Laakso et al., 1996), the fractional Hilbert transform (Pei & Wang, 2002), notch
filters (Joshi & Roy, 1999; Pei & Tseng, 1997; Tseng & Pei, 1998), among others The majority
of the methods use some approximation of the desired phase in the least square sense and
minimax sense
The allpole filters with maximally flat phase response characteristic have been specially
attrac-tive due to promising applications, like the design of IIR filters (Selesnick, 1999), the design
14
Trang 14of orthogonal and biorthogonal IIR wavelet filters (Selesnick, 1998; Zhang et al., 2001; 2000;
2006), the design of complex wavelets (Fernandes et al., 2003), the design of half band filters
(Zhang & Amaratunga, 2002), the fractional delay filter design (Laakso et al., 1996) and the
fractional Hilbert transform design (Pei & Wang, 2002)
This chapter presents a new design of real and complex allpole filters with the given phase,
group delay, and degree of flatness, at any desired set of frequency points The main
moti-vation of this work is to get some new promising cases related with the applications of
max-imally flat allpole filters In that way, using the proposed extended allpole filter design, we
introduced some new special cases
The rest of the chapter is organized as follows Section 2 establishes the general equations for
maximally flat real and complex allpole filters The discussion of the proposed design is given
in Section 3 for both, real and complex cases Different special cases of the general allpole
filter design is discussed in Section 4 Finally, Section 5 presents some applications of the
proposed allpole filter design, i.e., linear-phase Butterworth-like filter, Butterworth-like filters
with improved group delay, complex wavelet filters, fractional Hilbert transformers, and new
IIR filters based on three allpass filters
2 Equations for Maximally Flat Allpole Filter
We derive here equations for real and complex allpole filters both of order N, delay τ, and
degree of flatness K, at a given set of frequency points.
We consider that an allpole filter of order N is given by,
In general, the filter coefficients f n , n=1, , N, are complex, i.e., f n= f Rn+j f In where f Rn
and f In are the real and imaginary parts of f n , respectively Obviously, if f In=0, we obtain
real coefficients
The phase responses of D(z)and F(z)are related by
where φ α is the phase of α, and φ D(ω)and φ F(ω)are the phases of D(z)and F(z), respectively
The corresponding group delay is the negative derivative of the phase, as shown in (4)
where F∗(ejω), is the complex conjugate of F(ejω)
Using (4) and (6) the corresponding group delay G(ω)can be expressed as
Notice that for each frequency point ω l , we have K l+2 equations (see (10)) and 2N unknown
coefficients A consistent set of linear equations (10) is obtained if the following condition issatisfied,
(
K2
2 +1
)+ ⋅ ⋅ ⋅ +
Trang 15of orthogonal and biorthogonal IIR wavelet filters (Selesnick, 1998; Zhang et al., 2001; 2000;
2006), the design of complex wavelets (Fernandes et al., 2003), the design of half band filters
(Zhang & Amaratunga, 2002), the fractional delay filter design (Laakso et al., 1996) and the
fractional Hilbert transform design (Pei & Wang, 2002)
This chapter presents a new design of real and complex allpole filters with the given phase,
group delay, and degree of flatness, at any desired set of frequency points The main
moti-vation of this work is to get some new promising cases related with the applications of
max-imally flat allpole filters In that way, using the proposed extended allpole filter design, we
introduced some new special cases
The rest of the chapter is organized as follows Section 2 establishes the general equations for
maximally flat real and complex allpole filters The discussion of the proposed design is given
in Section 3 for both, real and complex cases Different special cases of the general allpole
filter design is discussed in Section 4 Finally, Section 5 presents some applications of the
proposed allpole filter design, i.e., linear-phase Butterworth-like filter, Butterworth-like filters
with improved group delay, complex wavelet filters, fractional Hilbert transformers, and new
IIR filters based on three allpass filters
2 Equations for Maximally Flat Allpole Filter
We derive here equations for real and complex allpole filters both of order N, delay τ, and
degree of flatness K, at a given set of frequency points.
We consider that an allpole filter of order N is given by,
In general, the filter coefficients f n , n=1, , N, are complex, i.e., f n= f Rn+j f In where f Rn
and f In are the real and imaginary parts of f n , respectively Obviously, if f In=0, we obtain
real coefficients
The phase responses of D(z)and F(z)are related by
where φ α is the phase of α, and φ D(ω)and φ F(ω)are the phases of D(z)and F(z), respectively
The corresponding group delay is the negative derivative of the phase, as shown in (4)
where F∗(ejω), is the complex conjugate of F(ejω)
Using (4) and (6) the corresponding group delay G(ω)can be expressed as
Notice that for each frequency point ω l , we have K l+2 equations (see (10)) and 2N unknown
coefficients A consistent set of linear equations (10) is obtained if the following condition issatisfied,
(
K2
2 +1
)+ ⋅ ⋅ ⋅ +