throbber

`
`
`
`
`
`Micro Motion
`
`Micro Motion 1046
`
`1
`
`

`

`DIGITAL SIGNAL PROCESSING
`Principles, Algorithms, and Applications
`
`Second Edition
`
`2
`
`

`

`DIGITAL SIGNAL
`
`
`
`PROCESSING
`Principles, Algorithms,
`
`and Applications
`
`Second Edition
`
`John G. Proakis
`
`Northeastern University
`
`Dimitris G. Manolakis
`
`Worcester Polytechnic Institute
`
`Macmillan Publishing Company
`NEW YORK
`
`Maxwell Macmillan Canada
`TORONTO
`
`Maxwell Macmillan International
`NEWYORK OXFORD SINGAPORE SYDNEY
`
`
`
`3
`
`

`

`Editor(s): John Griffin
`Production Supervisor: Ron Harris
`Production Managcr: Sandra E. Moore
`Text Designer: Eileen Burke
`
`
`
`Copyright © l992 by Macmillan Publishing Company. a division of Macmillan. Inc.
`Printed in the United States of America
`
`All rights reserved. No part of this book may be reproduced or
`transmitted in any form or by any means. electronic or mechanical.
`including photocopying. recording. or any information storage and
`retrieval system. without permission in writing from the publisher.
`
`Earlier edition entitled Introduction to Digital Signal Processing.
`copyright © I988 by Macmillan Publishing Company.
`
`Macmillan Publishing Company is part
`of the Maxwell Communication Group of Companies.
`
`Macmillan Publishing Company
`Roe/Third Avenue. New York. New York l0022
`
`Maxwell Macmillan Canada. Inc.
`l200 Eglinton Avenue East
`Suite 200
`
`Don Mills. Ontario M3C 3Nl
`
`Library of Congress Cataloging in Publication Data
`
`Rev. ed. of: Introduction to digital signal processing. © I988,
`Includes bibliographical references and index.
`ISBN 0—02-3968l5—X
`I. Manolakis. Dimitris
`l. Signal processing—~Digital techniques.
`II. Proakis. John G. Introduction to digital signal processing.
`6.
`III. Title.
`TK5 l 02.5.P6773
`I992
`62l.382'2~—dc20
`'
`
`91— l5536
`CIP
`
`Printing:l2345678
`
`Year:234567890l
`
`4
`
`

`

`- PREFACE
`
`
`
`X.
`
`This book has resulted from our teaching of undergraduate and graduate—level courses
`in digital signal processing over the past several years. In this book we present the
`fundamentals of discrete-time signals, systems, and modern digital processing algo—
`rithms and applications for students in electrical engineering, computer engineering,
`and computer science. As written, the book is suitable for either a one-semester or a
`two-semester undergraduate—level course in discrete systems and digital signal proc-
`essing. It is also intended for use in a one—semester first—year graduate-level course in
`digital sig'r‘ial processing.
`A balanced coverage is provided of both theory and practical applications. Included
`are software implementations of digital filters and digital signal processing algorithms.
`Appropriately designed programs in FORTRAN are provided to aid the student in
`implementing and exercising each algorithm immediately after exposition. A large
`number of well-designed problems and a number of computer experiments are also
`provided to help the student in mastering the subject matter. A solutions manual is
`available for the benefit of the instructor and can be obtained from the publisher.
`It is assumed that the student in electrical and computer engineering has had under—
`graduate courses in advanced calculus (including ordinary differential equations), and
`linear systems for continuous—time signals, including an introduction to the Laplace
`transform. Although the Fourier series and Fourier transforms of periodic and ape—
`riodic signals are described in Chapter 3, we expect that many students may have had
`this material in a prior course.
`In Chapter 1 we describe the operations involved in the analog-to-digital conversion
`of an analog signal. The process of sampling a sinusoid is described in some detail
`and the problem of aliasing is explained. Signal quantization and digital—to—analog
`conversion are also described in general terms in Chapter 1, but the analysis is pre-
`sented in subsequent chapters.
`Chapter 2 is devoted entirely to the characterization and analysis of liner time—
`invariant (shift-invariant) discrete-time systems and discrete-time signals in the time
`domain. The convolution sum is derived and systems are categorized according to the
`duration of their impulse response as finite-duration impulse response (FIR) and
`infinite—duration impulse response (IIR). Linear time—invariant systems characterized
`by difference equations are presented and the solution of difference equations with
`initial conditions is obtained. The chapter concludes with a treatment of discrete—time
`correlation.
`
`Chapter 3 treats the analysis of signals and systems in the frequency domain. Fouri-
`er series and the Fourier transform are presented for both continuous—time and discrete—
`time signals. Linear time—invariant (LTI) discrete systems are characterized in the
`frequency domain by their frequency response function and their response to periodic
`
`5
`
`V
`
`
`
`5
`
`

`

`vi Preface
`
`and aperiodic signals is determined. The frequency response characteristics of ideal
`frequency selection filters are also described.
`The z-transform is introduced in Chapter 4. Both the bilateral and the unilateral
`z-transform are presented, and methods for determining the inverse z-transform are
`described. Use of the z-transform in the analysis of linear time—invariant systems is
`illustrated, and the important properties of systems, such as causality and stability, are
`related to z-domain characteristics.
`
`Chapter 5 is concerned with the analysis and design of discrete-time systems in the
`frequency domain. LTI systems are characterized either by their z-transform or by the
`Fourier transform, and their characteristics are related to the location of the poles and
`zeros of their system function. A number of important types of digital systems are
`described, including resonators, notch filters, comb filters, all—pass filters, and oscil-
`lators. The design of some simple FIR and IIR filters is also considered. In addition,
`the student is introduced to the concepts of minimum—phase, mixed—phase, and max-
`imum-phase systems and the problem of deconvolution.
`Chapter 6 is focused on the sampling of signals in the time and frequency domains.
`In this chapter we establish the sampling theorem for bandlimited continuous—time
`signals, both lowpass and bandpass. Also described in some detail are A/D and D/A
`conversion techniques, including oversampling A/D and D/A converters. The last
`major topic in this chapter is a discussion of sampling of the spectrum of a discrete-
`time signal. This discussion leads to the discrete Fourier transform (DFT) and its use
`in frequency analysis.
`Chapter 7 treats the realization of IIR and FIR systems. This treatment includes
`direct-form, cascade, parallel, lattice, and lattice-ladder realizations. The chapter in—
`cludes a treatment of state-space analysis and structures for discrete—time systems, and
`examines quantization effects in a digital implementation of FIR and IIR systems.
`Techniques for design of digital FIR and IIR filters are presented in Chapter 8. The
`design techniques include both direct design methods in discrete time and techniques
`that involve the conversion of analog filters into digital filters by various transfor-
`mations. Also treated in this chapter is the design of FIR and IIR filters by least-
`squares methods.
`“"-
`Chapter 9 treats the computation of the DFT. Radix—2, radix-4, and split-radix fast
`Fourier transform (FFT) algorithms are derived and applications of the FFT algorithms
`to the computation of convolution and correlation are also described. The Goertzel
`algorithm and the chirp-z transform are introduced as two methods by which the DFT
`can be computed via linear filtering.
`Chapter 10 provides an in-depth treatment of sampling-rate conversion and its
`applications to multirate digital signal processing. In addition to describing decimation
`and interpolation by integer factors, a method of sampling-rate conversion by an ar-
`bitrary factor is presented. Several applications to multirate signal processing are pre—
`sented, including the implementation of digital filters, subband coding of speech sig—
`nals, transmultiplexing, and oversampling A/D and D/A converters.
`Linear prediction and optimum linear (Wiener) filters are topics treated in Chapter
`11. Also included in this chapter are descriptions of the Levinson-Durbin algorithm
`and Schur algorithm for solving the normal equations, as well as AR lattice and ARMA
`lattice-ladder, filters.
`
`Power spectrum estimation is the main topic of Chapter 12. Our coverage of this
`subject includes a description of nonparametric and model-based (parametric) methods
`for power spectrum estimation. Also described are eigen-decomposition—based meth—
`ods, including MUSIC and ESPRIT.
`At Northeastern University, we have used the first six chapters of this book for a
`
`
`
`one—se
`
`For th
`It is a
`
`juncti(
`A (
`discre
`
`then 1:
`introd
`filter (
`In :
`
`ters pi
`may r
`9, fol]
`we hi;
`and pr
`Thl
`
`who 1
`We V
`for as
`H. Le
`
`sugge
`appre
`the m
`We a;
`radix
`Proak
`
`Ackl
`
`We i
`
`gestit
`Engii
`Wins
`State
`
`(D613
`East
`
`Com;
`ence
`
`B. M
`2500
`
`
`
`6
`
`

`

`
`
`i of ideal
`
`tnilateral
`form are
`{stems is
`
`Iiiity. are
`
`ns in the
`
`Jr by the
`oles and
`terns are
`1d oscil-
`1ddilion,
`nd max-
`
`lomains.
`)us-timc
`
`ind D/A
`Fhe last
`iiscrete-
`1 its use
`
`.ncludes
`
`lpter in—
`ms, and
`terns.
`r 8. The
`
`hniques
`ransfor-
`
`y least-
`
`.dix fast
`orithms
`ioerlzcl
`he DPT
`
`and its
`imalion
`I an ar—
`
`are pre-
`:ch sig-
`
`fhapter
`gorithm
`ARMA
`
`of this
`uethods
`l meth-
`
`kfora
`
`
`
`Preface vii
`
`one—semester (junior~level) course in discrete systems and digital signal processing.
`For this purpose we have introduced some simple filter design methods in Chapter 5.
`It is also possible to include the radix-2 decimation—in—time FFT algorithm in con-
`junction with the discussion in Chapter 6 on computation of the BET.
`A one-semester senior-level course for students who have had prior exposure to
`discrete systems may use the material in Chapters 1 through 4 for a quick review and
`then proceed to cover Chapters 5 through 9. In such a course the students may be
`introduced to the we of software packages that are now readily available for digital
`filter design.
`In a first-year graduate-level course in digital signal processing, the first five chap-
`ters provide the student with a good review of discrete-time systems. The instructor
`may move quickly through most of this material and then cover Chapters 6 through
`9, followed by either Chapters‘lO and 11 or by Chapters 11 and 12. In such a course
`we highly-recommend the use of available software packages for digital filter design
`and power spectrum estimation.
`The. authors are indebted to a number of graduate students and faculty colleagues
`who provided valuable suggestions and assistance in preparation of the manuscript.
`We wish to thank Mr. J. Lin for preparing the solutions manual, Dr. A. L. Kok
`for assisting in preparation of the figures, and Drs. J. Deller, V. Ingle, C. Keller,
`H. Lev-Ari. L. Merakos, P. Monticciolo, and M. Schetzen for their comments and
`suggestions on various drafts of the manuscript. We especially wish to express Our
`appreciation to Professors C. L. Nikias, H. J. Trussell, and S. G. Wilson, who reviewed
`the manuscript and provided many important suggestions that have been incorporated.
`We are also indebted to Dr. R. Price for informing us of the different types of split-
`radix FFT algorithms and related suggestions. Finally. we wish to thank Ms. Gloria
`Proakis for typing the entire manuscript.
`
`John G. Proakis
`
`Dimitris G. Manolakis
`
`Acknowledgments
`
`We would like to titanic the following reviewers, who made many valuable sug-
`gestions: Professor Michael Zoltowski (Dept. of Electrical Engineering; Electrical
`Engineering Bldg; Purdue University; West Lafayette,
`IN 47907); Professor
`Winser E. Alexander (Dept. of Electrical and Computer Engineering; North Carolina
`State University; Box 7911; Raleigh, NC 27695-7911); Professor John R. Deller
`(Dept. of Electrical Engineering; 260 Engineering Bldg; Michigan State University;
`East Lansing, MI 48824-1226); Professor Yorarn Bresler (Dept. of Electrical and
`Computer Engineering; University of Illinois#Urbana, Champlain; Coordinated Sci—
`ence Laboratory; 1101 West Springfield Avenue; Urbana, IL 61301); Professor Wasfy
`B. Mikhael (Dept. of Electrical Engineering; University of Central Florida; PO. Box
`25000; Orlando, FL 32316).
`
`7
`
`

`

`
`
`E12
`
`
`
`POWER SPECTRUM ESTIMATION
`
`In this chapter we are concerned with the estimation of the spectral CITHF'cl'CiL‘i‘lsii‘a\ of
`signals that are characterized as random processes. Many of the phenomena am occur
`in nature are best characterized statistically in terms of averages. For exan‘ag‘xte. me»
`teorological phenomena such as the fluctuations in air temperature and pressure are
`best characterized statistically as random processes. Thermal noise voltages '
`.
`
`in resistors and electronic devices are additional examples of physical signalu ,
`:1‘ 1‘ .
`well modeled as random processes.
`Due to the random fluctuations in such signals, we must adopt a statistical \iCVx»
`point, which deals with the average characteristics of random signals. In panicular.
`the autocorrelation function of a random process is the appropriate statisticaé art—rage
`that we will be concerned with for characterizing random signals in the time domain.
`and the Fourier transform of the autocorrelation function, which yields the power
`density spectrum, provides the transformation from the time domain to the frequent}
`domain.
`
`Power spectrum estimation methods have a relatively long history. For a histcrical
`perspective, the reader is referred to the paper by Robinson (1982) and the book by
`Marple (1987). Our treatment of this subject covers the classical power spectrum
`estimation methods based on the periodogram originally introduced by Scittrstcr
`(1898), and the modern model»based or parametric methods that originated with the
`work of Yule (1927). These methods were subsequently developed and applied by
`Walker (l93l), Bartlett (l948), Parzen (1957), Blackman and Tukey (1958;. Burg
`(1967), and others. We also describe the method of Capon (l969) and methods based
`on eigenanalysis of the data correlation matrix.
`
`12.1 ESTIMATION OF SPECTRA FROM FINITE-
`DURATION OBSERVATIONS OF SIGNALS
`
`The basic problem that we consider in this chapter is the estimation of the power
`density spectrum of a signal from observation of the signal over a finite time human}
`As we will see. the finite record length of the data sequence is a major limitation Ht?
`the quality of the power spectrum estimate. When dealing with signals that are statis—
`tically stationary, the longer the data record, the better the estimate that can be 2:»-
`tracted from the data. On the other hand, if the signal statistics are nonstationztr}
`‘
`
`cannot select an arbitrarily long data record to estimate the spectrum. In such a c ,
`the length of the data record that we select is determined by the rapidity of the rim:
`variations in the signal statistics. Ultimately. our goal is to select as short a data record
`864
`
`8
`
`8
`
`

`

`12.1 / Estimation of Spectra from Finite—Duration Observations at Signals
`
`865
`
`as possible that will allow us to resolve spectral cl’iaracteristics of different signal
`components, contained in the data record. that have closely spaced spectra.
`One of the problems that we encounter with classical power spectrum estimation
`methods based on a tinite~length data record is distortion of the spectrum that we are
`attempting to estimate. This problem occurs in both the computation of the spectrum
`for a deterministic signal and the estimation of the power spectrum of a random signal
`Since it
`is easier to observe the effect of the finite length ot‘ the data record en a
`deterministic signal, we treat this case lirst. Thereafter. we consider oniy random
`signals and the estimation of their power spectra
`
`12.1.1 Computation of the Energy Density Spectrum
`
`Let us consider the computation ot‘ the spectrum of a deterministic signal from a finite
`sequence of data. The sequence .\'(/1) is usually the result of sampling a continnons~
`time signal .\‘(,(I) at some uniform sampling rate R. Our objective is to obtain an
`estimate of the true spectrum from a tinite~duiation sequence Alli).
`Recall that if,\‘(t) is a finite-energy signal, that is,
`
`E 2 f T i_\‘u(l‘)iz dr <J 7:
`
`then its Fourier transform exists and is given as
`
`XuiFl : f v\'|l([)(f"_l2’77/‘t (It
`
`From Parseval’s theorem we have
`
`E 2 f 7
`
`l-\1,(f)l3 dz : f
`
`«:4
`
`EXHtFig'3 d1”
`
`.
`
`(Elli)
`
`The quantity [Xutfll2 represents the distribution of signal energy as Lt timction of
`frequency, and hence it is called the energy density spectrum of the signal. that is.
`
`SHtF) : txutrritl
`
`t t 2. l .2;
`
`as described in Chapter 3. Thus the total energy in the signai is simpty the integral of
`S, \tF) over all F lie. the total area under 5‘ \tFil.
`It is also interesting to note that 8‘ \(F) may be viewed as the Fourier transform of
`another function. R_\,_\(T). called the atItor'nrrylutin/iflirtation of the tinite~energy signal
`\,,tt). defined as
`
`RHm : J
`
`‘
`
`.rfftririr + 7) dz
`
`indeed, it easily follows that
`
`l— Rmtm 13“"? in : S‘HtFi : tier/iii
`
`tiltfit
`
`tutti
`
`stl that Rut?) and 5‘ \tF) are a Fourier transform pair.
`Now suppose that we compute the energy density spectrum of the sigtml tuifl from
`it“: samples taken at the rate R samples per second. To ensure that there is no spectrai
`:tiiasing resulting from the sampling process. the signal is assumed to he pretittcred.
`w that. for practical purposes. its bandwidth is limited to 8 hertz. Then the sampling
`tierntency FX is selected such that E 1?» 38.
`
`
`
`9
`
`

`

`866 Ch. 12 / Power Spectrum Estimation
`
`The sampled version of .ra(r) is a sequence .r(n), — 00 < n < 00, which has a Fourier
`transform (voltage spectrum)
`ac
`
`‘
`
`X(w) = 2 x(n>e~fw"
`
`or, equivalently,
`
`00
`
`X(f) = E x(n)e”jz“f”
`
`(12.1.5)
`
`Recall that X(f) may be expressed in terms of the voltage spectrum of the analog
`signal ,1'a(t) as
`
`FX<f> = F5 2 XG(F - kFS)
`
`S
`
`(12.1.6)
`
`where f = F/F3 is the normalized frequency variable.
`In the absence of aliasing, within the fundamental range IFI S FS/Z, we have
`
`XG) = FSXa(F),
`
`S
`
`[Fl 5 F‘s/2
`
`(12.1.7)
`
`Hence the voltage spectrum of the sampled signal is identical to the voltage spectrum
`of the analog signal. As a consequence, the energy density spectrum of the sampled
`signal is
`2
`
`S .
`
`F
`F
`. — = X —
`
`
`
`(F5)
`A)(F5)
`We may proceed further by noting that the autocorrelation of the sampled signal,
`which is defined as
`cc
`
`
`
`= F3 IXa(F)|2
`
`(12.1.8)
`
`rim-(k) = 2 x*(n)x(n ‘+ k)
`n=~0¢
`
`(12.1.9)
`
`has the Fourier transform (Wiener—Khintchine theorem)
`
`51mm = 2 r,.,.(k)e~12rkf
`
`(12.1.10)
`
`Hence the energy density spectrum may be obtained by Fourier transforming the
`autocorrelation of the sequence {x(n)}.
`The relations above lead us to distinguish between two distinct methods for come
`puting the energy density spectrum of a signal xa(t) from its samples x(n). One is the
`direct method, which involves computing the Fourier transform of {x(n)}, and then
`
`SKA-(f) = |X(f)|2x
`
`
`
`2 x(n)e ”12“”
`11:: ~96
`
`2
`
`
`
`(12.l.ll)
`
`The second approach is called the indirect method because it requires two steps. First,
`the autocorrelation ij\._\.(k) is computed from x(n) and then the Fourier tranSform of the
`autocortelation is computed as in (12.1.10) to obtain the energy density SpCCII‘UUL
`
`10
`
`10
`
`

`

`12.1 / Estimation of Spectra from FiniteDuration Observations of Signals 867
`In practice. however. only the finite-duration sequence .r(n), 0 _<. n _<. N — l. is
`available for computing the spectrum of the signal. In effect. limiting the duration of
`the sequence .r(n) to N points is equivalent to multiplying .r(n) by a rectangular win-
`dow. Thus we have
`
`0 E II S N — l
`.\‘(n).
`_
`otherwise
`0
`H”) 2 “WM!” :
`From our discussion of FIR filter design based on the use of windows to limit the
`duration of the impulse response, we recall that multiplication of two sequences is
`equivalent
`to convolution of their voltage spectra. Consequently,
`the frequency-
`domain relation corresponding to (l2. H2) is
`
`W“) = W) Wm
`(l2.l.l3)
`[V2
`4/3 X(or)W(f — or) dd
`Recall from our discussion in Section 6.4.5 that convolution of the window function
`W(f) with X(f) smooths the spectrum X(f). provided that the spectrum W(f) is rela-
`tively narrow compared to X(f). But this condition implies that the window Mn) be
`sufficiently long (i.e.. N must be sufficiently large) such that W(f) is narrow compared
`to X(f). Even if W(f) is narrow compared to X(f). the convolution of X(f) with the
`sidelobes of W(f) results in sidelobe energy in X(f). in frequency bands where the
`true signal spectrum X(f) = 0. This sidelobe energy is called lea/(age. The following
`example illustrates the leakage problem.
`
`EXAMPLE 12.1.1
`
`
`A signal with (voltage) spectrum
`
`otherwise
`X(f) _ {0.
`is convolved with the rectangular window of length N = 6i. Determine the spectrum
`ofX(f) given by (l2.l.l3).
`
`if! S 0-1
`
`_
`
`1’
`
`Solution: The spectral characteristic W(f} for the length N 2 6i rectangular window
`is illustrated in Fig. 8.2(b). Note that the width ofthe main lobe of the window function
`is At» = 4w/6l or Af = 2/6l. which is narrow compared to X(f).
`The convolution of X(f) with W(f) is illustrated in Fig. l2. i. We note that energy
`has leaked into the frequency band 0.l < if! _<. 0.5. where X(f) = l). A part of this
`is due to the width of the main lobe in W(f). which causes a broading or smearing of
`.X’tf) outside the range if! _<. 0.l. However. the sidelobe energy in X(f) is due to the
`presence of the sidelobes of W(f). which are convolved with X(f). The smearing of
`Kit") for if! > 0.l and the sidelobes in the range 0.l S if! S 0.5 constitute the leakage.
`Just as in the case of FIR filter design. we can reduce sidelobe leakage by selecting
`windows that have low sidelobes. This implies that the windows have a smooth time-
`domain cutoff instead of the abrupt cutoff in the rectangular window. Although such
`Window functions reduce sidelobe leakage. they result in an increase in smoothing or
`broadening of the spectral characteristic X(f). For example. the use of a Blackman
`
`11
`
`11
`
`

`

`
`
`
`
`
`
`868 Ch. 12 / Power Spectrum Estimation
`A
`
`10l—
`0
`
`LA.00"l'TlTr
`
` 0L—9
`
`
`
`
`
`Energydensity(dB)
`
`
`_100LJJJA_LJIJJ.LULL[L1L1JLJJ_LJHLJ_[LLJLIJJJJ_LJJJ_1J_LLJ_LJ_>f
`0
`.1
`.2
`.3
`.4
`.5,
`Frequency (cycles/sample)
`
`FIGURE 12.1 Spectrum obtained by convolving an M = 61 rectangular window with
`the ideal lowpass spectrum in Example 12.1.1.
`
`window of length N = 61 in Example 12.1.1 results in the spectral characteristic X(f)
`shown in Fig. 12.2. The sidelobe leakage has certame been reduced, but the spectral
`width has been increased by about 50%.
`The broadening of the spectrum being estimated due to windowing is particularly
`a problem when we wish to resolve signals with closely spaced frequency components.
`For example, the signal with spectral characteristic X(f) = Xl(f) + X2(f), as shown
`in Fig. 12.3, cannot be resolved as two separate signals unless the width of the window
`function is significantly narrower than the frequency separation Af. Thus we observe
`that using smooth time—domain windows reduces leakage at the expense of a decrease
`in frequency resolution.
`It is clear from the discussion above that the energy density spectrum of the win—
`dowed sequence {x(n)} is an approximation of the desired spectrum of the sequence
`
`
`
`
`
`Energydensity(dB)
`
`
`
`0
`
`.l
`
`.2
`
`.3
`
`.4
`
`.5
`
`Frequency (cycles/sample)
`FIGURE 12.2 Spectrum obtained by convolving an M = 61 Blackman window with
`the ideal lowpass spectrum in Example 12.1.1.
`
`12
`
`12
`
`

`

`12.1 / Estimation of Spectra from Finite~Duration Observations of Signals 869
`le)
`
`Xilf)
`
`X20’)
`
`0
`
`‘1 M 1._
`
`*f
`
`0.5
`
`FIGURE 12.3 Two narrowband signal spectra.
`
`{x(n)}. The spectral density obtained from {2201)} is
`
`Sad) = |X(f)|2 =
`
`N—l
`2 fink—1'2““
`n=O
`
`2
`
`
`
`
`
`(12.1.14)
`
`The spectrum given by (12.1.14) can be computed numerically at a set of N fre—
`quency points by means of the DFT. Thus
`N—l
`
`,
`
`m) = 2 awe—WW
`n=O
`
`(12.1.15)
`
`Then
`
`and hence
`
`-
`
`|X(k)|2 = Sfi(f)lf=k/N = S4171)
`
`k
`
`k
`
`
`512.2(3) = Z i(n)e‘-’2"""/N
`
`N—l
`n=0
`
`2
`
`
`
`(12.116)
`
`(12.1.17)
`
`which is a distorted version of the true spectrum Sxx(k/N).
`
`12.1.2 Estimation of the Autocorrelation and Power Spectrum
`of Random Signals: The Periodogram
`
`The finite—energy signals considered in the preceding section possess a Fourier trans—
`form and were characterized in the spectral domain by their energy density spectrum.
`On the other hand, the important class of signals characterized as stationary random
`processes do not have finite energy and hence do not possess a Fourier transform.
`Such signals have finite average power and hence are characterized by a power density
`Spectrum. lf x(t) is a stationary random process, its autocorrelation function is
`
`WAT) = E[x*(t)x(t + 7)]
`
`(12.1.18)
`
`where E H denotes the statistical average. Then via the Wiener~Khintchine theorem,
`the power density spectrum of the stationary random process is the Fourier transform
`of the autocorrelation function, that is,
`
`00
`
`I‘Xx(F) = f w Yxx(T)€ ‘12” dz
`
`(12.1.19)
`
`13
`
`13
`
`

`

`870 Ch. 12 / Power Spectrum Estimation
`
`In practice, we deal with a single realization of the random process from which We
`estimate the power spectrum of the process. We do not know the true autocorrelation
`function vmfi), and as a consequence, we cannot compute the Fourier transform in
`(12.1.19) to obtain I’M(F). On the other hand, from a single realization of the random
`process we can compute the time—average autocorrelation function
`
`T0
`1
`R.T=— x*th+Tdt
`..(>
`27b
`_TO
`(> <
`>
`
`(
`
`1213‘
`-0)
`
`where 2T0 is the observation interval. If the stationary random process is ergodic in
`the first and second moments (mean and autocorrelation function), then
`
`vxxfi) = lim R“,(T)
`T —<)oo
`To
`O
`1
`=lim—f x*txt+Tdt
`“6&20 _% (>(
`>
`
`(12.1.21)
`
`This relation justifies the use of the time—average autocorrelation Rn.("r) as an es-
`timate of the statistical autocorrelation function vxx("r). Furthermore, the Fourier trans—
`form of Rxx("r) provides an estimate PXX(F) of the power density spectrum, that is,
`TO
`I
`RXX(T)€
`‘jZ‘TF" d1"
`To
`1
`To
`U x*(t)x(t + T) 422-12?” d’T
`fig —70
`‘70
`To
`
`2T0
`I—TO X( )6
`
`Pxx(F)
`
`_TO
`
`.
`
`(12.1.22)
`
`= _
`
`
`
`2
`
`t —j21'rFI dt
`
`The actual power density spectrum is the expected value of PXX(F) in the limit as
`To —> 00, that is,
`’”
`
`FXX(F) = lim E[PXX(F)]
`T ~—>oo
`
`'2 F
`To
`1
`o
`
`=limE—f xte‘l"’dt
`TO—éoc
`2T0
`_TU ( )
`
`2
`
`
`
`(12.1.23)
`
`From (12.1.20) and (12.1.22) we again note the two possible approaches to c0m~
`puting PXX(F), the direct method as given by (12.1.22) or the indirect method, in which
`we obtain RH(T) first and then compute the Fourier transform.
`We shall consider the estimation of the power density spectrum from samples of a
`single realization of the random process. In particular, we assume that xa(t) is sampled
`at a rate F5 > 23, where B is the highest frequency contained in the power density
`spectrum of the random process. Thus we obtain a finite—duration sequence x(n),
`0 S n S N — 1, by sampling xa(t). From these samples we may compute the time
`average autocorrelation sequence
`1 N—Iml—l
`
`r;x(m) = N
`2 x*(n)x(n + m), m = 0,1,...,N — 1
`— m n:
`[HO
`Z x*(n)x(n + m), m = —1, —2,..., 1 — N
`N — lml n=|m1
`
`(12.1.24)
`
`r,’.,.(m) =
`
`and then compute the Fourier transform
`
`14
`
`14
`
`

`

`12.1 / Estimation of Spectra from Finite-Duration Observations of Signals
`N—l
`
`871
`
`P410”) = E r.L..—(m)e“/2"f’"
`m= ——N+l
`
`"
`
`(12.1.25)
`
`The normalization factor N —-
`value
`
`[ml in (12.1.24) results in an estimate with mean
`
`E[44171)]
`
`
`1
`N‘imi
`= m0")
`
`N—Iml—l
`2 E[x*(n)x(n + m)]
`"=0
`
`(12.1.26)
`
`where yxx(m) is the true (statistical) autocorrelation sequence of x(n). Hence r,’m(m) is
`an unbiased estimate of the autocorrelation function yxx(m). The variance of the es—
`timate r,’CX(m) is approximately
`N
`GO
`t
`[Iv-[ml]2 ”:23.
`['ymm)!2 + 'yfivm - m)'yxx(n + m)]
`-~——-——
`which is a result given by Jenkins and Watts (1968). Clearly,
`
`var [réc(m)] %
`
`(12.1.27)
`
`lim var [r;x(m)] = 0
`Nam
`
`(12.1.28)
`
`provided that
`
`Do
`
`2 Mn)? < oo
`n=—m
`
`Since E[r)'cx(m)] = yxx(m) and the variance of the estimate converges to zero as
`N -—> 00, the estimate r;X(m) is said to be consistent.
`For large values of the lag parameter m, the estimate ryéx(m) given by ( 12.1.24) has
`a large variance, especially as m approaches N. This is due to the fact that fewer data
`points enter into the estimate for large lags. As an alternative to (12.1.24) we may use
`the estimate
`
`1 N—Iml—l
`rxx(m) = N E x*(n)x(n + m),
`1 ~50
`'
`rxx(m) = N
`x*(n)x(n + m),
`
`n=lml
`
`0 S m S N —-
`
`1
`
`m = —-l, —-2, ..,1 —- N
`
`(12.1.29)
`
`which has a bias of lmlyxx(m)/N, since its mean value is
`
`1 N“ 1m! “ 1
`E[rxx(m)] = N 20 E[x*(n)x(rt + m)]
`N _ imi
`imi
`= T MW) = < - 7V")VXX(M)
`However, this estimate has a smaller variance, given approximately as
`oo
`
`(12.1.30)
`
`var[rxx(m)] ~ _. 2 [Film-(””2 + FY34” _. mnyxUl + ’71)]
`N ,1: ——00
`
`(120-131)
`
`We observe that rxx(m) is asymptotically unbiased, that is,
`
`lim E[r_”(m)] = yxx(m)
`Nam
`
`(12.1.32)
`
`15
`
`15
`
`

`

`872 Ch. 12 / Power Spectrum Estimation
`
`and its variance converges to zero as N ——> 00. Therefore, the estimate 1““..(m) is also a
`consistent estimate of 'yxx(m).
`We shall use the estimate (”(m) given by (12.1.29) in our treatment of power
`spectrum estimation. The corresponding estimate of the power density spectrum is
`N—l
`
`P,.,.(f) =
`
`2
`m=—(N— 1)
`
`rm(m)e_j2"f’"
`
`(12.1.33)
`
`If we substitute for r“.(m) from (12.1.29) into (12.1.33), the estimate P“.(f) may also
`be expressed as
`
`1
`
`P“.(f) = N
`
`N—i
`
`2 X(I’I)€ —j2-rrfn
`n=0
`
`2
`
`
`
`1
`= N 1X(f)l2
`
`(12.1.34)
`
`where X(f) is the Fourier transform of the sample sequence x(n). This well—known
`form of the power density spectrum estimate is called the periodogram. It was origi—
`nally introduced by Schuster (1898) to detect and measure “hidden periodicities” in
`data.
`
`From (12.1.33), the average value of the periodograrn estimate Pxx(f) is
`N— 1
`N— 1
`
`E[P,\x(f)] : E[ 2
`ElPX..(f)l
`E
`
`m= —(N— 1)
`
`=—(N—1)
`N—l
`
`rxx(m)€—j2flfm:l : 3 2
`(I ~ %>vx(m)e"2“fm
`
`m=—(N—I)
`
`E[7'xx(m)]e_j2-rrfm
`
`(12.1.35)
`
`The interpretation that we give to (12.1.35) is that the mean of the estimated spectrum
`is the Fourier transform of the windowed autocorrelation function
`
`m
`
`fixgm) 2 <1 —— U>‘3/J,.,.(171)
`
`N
`
`(12.1.36)
`
`where the window function is the (triangular) Bartlett window. Hence the mean of the
`estimated spectrum is
`
`ElPx1-(f)l
`
`2 1...<m)e‘-"2"fm
`"1: —OO
`1/2
`
`{—1/2 I‘”((x)WB(f -— (x) dOL
`
`(12.1.37)
`
`where WB(f) is the spectral characteristic of the Bartlett window. The relation (1 1.1.37)
`illustrates that the mean of the estimated spectrum is the convolution of the true power
`density spectrum THU") with the Fourier transform WB(f) of the Bartlett window.
`Consequently, the mean of the estimated spectrum is a smoothed version of the true
`spectrum and suffers from the same spectral leakage problems which are due to the
`finite number of data points.
`We observe that the estimated spectrum is asymptotically unbiased, that is,
`N— 1
`m
`
`Um E[ E
`
`N—>3C
`
`m=—(N—1)
`
`r...<m)ewe] = 2 v...<m>eW!“ = F....<f>
`
`m=—:>c
`
`However, in general, the variance of the estimate PXX(f) does not decay to zero as
`N ——> 00. For example, when the data sequence is a Gaussian random process, the
`variance is easily shown to be (see Problem 12.4)
`
`16
`
`16
`
`

`

`12.1 / Estimation of Spectra from Finite—Duration Observations of Signals 873
`2
`Txx(f)[l + (N sin 2m,
`(
`-
`-
`2 W 12 1 38
`
`)
`
`VarlPXX(f)l
`
`z:
`
`which, in the limit as N —~> 00, becomes
`
`lim var[PXX(f)] = FEXQ‘)
`New
`
`(12.1.39)
`
`Hence we conclude that the periodogram is not a consistent estimate offhe true power
`density spectrum (i.e., it does not converge to the true power density spectrum).
`In summary, the estimated autocorrelation rxx(m) is a consistent estimate of the true
`autocorrelation function yxx(m). HOWever, its Fourier transform Pxx(f), the periodo—
`gram, is not a consistent estimate of the true power density spectrum. We observed
`that PXX(f) is an asymptotically unbiased estimate of I‘xx(f), but for a finite—duration
`sequence, the mean value of Pxx(f) contains a bias, which from (12.1.37) is evident
`as a distortion of the true power density spectrum. Thus the estimated spectrum suffers
`from the smoothing effects and the leakage embodied in the Bartlett window. The
`smoothing and leakage ultimately limit our ability to resolve closely spaced spectra.
`The problems of leakage and frequency resolution that we have described above,
`as well as the problem that the periodogram is not a consistent estimate of the power
`spectrum, provide the motivation for the power spectrum estimation methods de-
`scribed in Sections 12.2, 12.3, and 12.4. The methods described in Section 12.2 are
`classical nonparametric methods, which make no assumptions about the data se—
`quence.

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket