Details

Probability, Random Variables, and Random Processes


Probability, Random Variables, and Random Processes

Theory and Signal Processing Applications
1. Aufl.

von: John J. Shynk

116,99 €

Verlag: Wiley
Format: PDF
Veröffentl.: 17.09.2012
ISBN/EAN: 9781118393949
Sprache: englisch
Anzahl Seiten: 794

DRM-geschütztes eBook, Sie benötigen z.B. Adobe Digital Editions und eine Adobe ID zum Lesen.

Beschreibungen

<p><i>Probability, Random Variables, and Random Processes</i> is a comprehensive textbook on probability theory for engineers that provides a more rigorous mathematical framework than is usually encountered in undergraduate courses. It is intended for first-year graduate students who have some familiarity with probability and random variables, though not necessarily of random processes and systems that operate on random signals. It is also appropriate for advanced undergraduate students who have a strong mathematical background.</p> <p>The book has the following features:</p> <ul> <li>Several appendices include related material on integration, important inequalities and identities, frequency-domain transforms, and linear algebra. These topics have been included so that the book is relatively self-contained. One appendix contains an extensive summary of 33 random variables and their properties such as moments, characteristic functions, and entropy.</li> <li>Unlike most books on probability, numerous figures have been included to clarify and expand upon important points. Over 600 illustrations and MATLAB plots have been designed to reinforce the material and illustrate the various characterizations and properties of random quantities.</li> <li>Sufficient statistics are covered in detail, as is their connection to parameter estimation techniques. These include classical Bayesian estimation and several optimality criteria: mean-square error, mean-absolute error, maximum likelihood, method of moments, and least squares.</li> <li>The last four chapters provide an introduction to several topics usually studied in subsequent engineering courses: communication systems and information theory; optimal filtering (Wiener and Kalman); adaptive filtering (FIR and IIR); and antenna beamforming, channel equalization, and direction finding. This material is available electronically at the companion website.</li> </ul> <p><i>Probability, Random Variables, and Random Processes</i> is the only textbook on probability for engineers that includes relevant background material, provides extensive summaries of key results, and extends various statistical techniques to a range of applications in signal processing.</p>
PREFACE xxi <p>NOTATION xxv</p> <p><b>1 Overview and Background 1</b></p> <p>1.1 Introduction 1</p> <p>1.1.1 Signals, Signal Processing, and Communications 3</p> <p>1.1.2 Probability, Random Variables, and Random Vectors 9</p> <p>1.1.3 Random Sequences and Random Processes 11</p> <p>1.1.4 Delta Functions 16</p> <p>1.2 Deterministic Signals and Systems 19</p> <p>1.2.1 Continuous Time 20</p> <p>1.2.2 Discrete Time 25</p> <p>1.2.3 Discrete-Time Filters 29</p> <p>1.2.4 State-Space Realizations 32</p> <p>1.3 Statistical Signal Processing with MATLAB® 35</p> <p>1.3.1 Random Number Generation 35</p> <p>1.3.2 Filtering 38</p> <p>Problems 39</p> <p>Further Reading 45</p> <p><b>PART I Probability, Random Variables, and Expectation</b></p> <p><b>2 Probability Theory 49</b></p> <p>2.1 Introduction 49</p> <p>2.2 Sets and Sample Spaces 50</p> <p>2.3 Set Operations 54</p> <p>2.4 Events and Fields 58</p> <p>2.5 Summary of a Random Experiment 64</p> <p>2.6 Measure Theory 64</p> <p>2.7 Axioms of Probability 68</p> <p>2.8 Basic Probability Results 69</p> <p>2.9 Conditional Probability 71</p> <p>2.10 Independence 73</p> <p>2.11 Bayes’ Formula 74</p> <p>2.12 Total Probability 76</p> <p>2.13 Discrete Sample Spaces 79</p> <p>2.14 Continuous Sample Spaces 83</p> <p>2.15 Nonmeasurable Subsets of R 84</p> <p>Problems 87</p> <p>Further Reading 90</p> <p><b>3 Random Variables 91</b></p> <p>3.1 Introduction 91</p> <p>3.2 Functions and Mappings 91</p> <p>3.3 Distribution Function 96</p> <p>3.4 Probability Mass Function 101</p> <p>3.5 Probability Density Function 103</p> <p>3.6 Mixed Distributions 104</p> <p>3.7 Parametric Models for Random Variables 107</p> <p>3.8 Continuous Random Variables 109</p> <p>3.8.1 Gaussian Random Variable (Normal) 110</p> <p>3.8.2 Log-Normal Random Variable 113</p> <p>3.8.3 Inverse Gaussian Random Variable (Wald) 114</p> <p>3.8.4 Exponential Random Variable (One-Sided) 116</p> <p>3.8.5 Laplace Random Variable (Double-Sided Exponential) 119</p> <p>3.8.6 Cauchy Random Variable 122</p> <p>3.8.7 Continuous Uniform Random Variable 124</p> <p>3.8.8 Triangular Random Variable 125</p> <p>3.8.9 Rayleigh Random Variable 127</p> <p>3.8.10 Rice Random Variable 129</p> <p>3.8.11 Gamma Random Variable (Erlang for r ∈ N) 131</p> <p>3.8.12 Beta Random Variable (Arcsine for α = β = 1/2, Power Function for β = 1) 133</p> <p>3.8.13 Pareto Random Variable 136</p> <p>3.8.14 Weibull Random Variable 137</p> <p>3.8.15 Logistic Random Variable (Sigmoid for {μ = 0, α = 1}) 139</p> <p>3.8.16 Chi Random Variable (Maxwell–Boltzmann, Half-Normal) 141</p> <p>3.8.17 Chi-Square Random Variable 144</p> <p>3.8.18 F-Distribution 147</p> <p>3.8.19 Student’s t Distribution 149</p> <p>3.8.20 Extreme Value Distribution (Type I: Gumbel) 150</p> <p>3.9 Discrete Random Variables 151</p> <p>3.9.1 Bernoulli Random Variable 152</p> <p>3.9.2 Binomial Random Variable 154</p> <p>3.9.3 Geometric Random Variable (with Support Z+ or N) 157</p> <p>3.9.4 Negative Binomial Random Variable (Pascal) 160</p> <p>3.9.5 Poisson Random Variable 162</p> <p>3.9.6 Hypergeometric Random Variable 165</p> <p>3.9.7 Discrete Uniform Random Variable 167</p> <p>3.9.8 Logarithmic Random Variable (Log-Series) 168</p> <p>3.9.9 Zeta Random Variable (Zipf) 170</p> <p>Problems 173</p> <p>Further Reading 176</p> <p><b>4 Multiple Random Variables 177</b></p> <p>4.1 Introduction 177</p> <p>4.2 Random Variable Approximations 177</p> <p>4.2.1 Binomial Approximation of Hypergeometric 177</p> <p>4.2.2 Poisson Approximation of Binomial 179</p> <p>4.2.3 Gaussian Approximations 181</p> <p>4.2.4 Gaussian Approximation of Binomial 181</p> <p>4.2.5 Gaussian Approximation of Poisson 181</p> <p>4.2.6 Gaussian Approximation of Hypergeometric 183</p> <p>4.3 Joint and Marginal Distributions 183</p> <p>4.4 Independent Random Variables 186</p> <p>4.5 Conditional Distribution 187</p> <p>4.6 Random Vectors 190</p> <p>4.6.1 Bivariate Uniform Distribution 193</p> <p>4.6.2 Multivariate Gaussian Distribution 193</p> <p>4.6.3 Multivariate Student’s t Distribution 196</p> <p>4.6.4 Multinomial Distribution 197</p> <p>4.6.5 Multivariate Hypergeometric Distribution 198</p> <p>4.6.6 Bivariate Exponential Distributions 200</p> <p>4.7 Generating Dependent Random Variables 201</p> <p>4.8 Random Variable Transformations 205</p> <p>4.8.1 Transformations of Discrete Random Variables 205</p> <p>4.8.2 Transformations of Continuous Random Variables 207</p> <p>4.9 Important Functions of Two Random Variables 218</p> <p>4.9.1 Sum: Z = X + Y 218</p> <p>4.9.2 Difference: Z = X − Y 220</p> <p>4.9.3 Product: Z = XY 221</p> <p>4.9.4 Quotient (Ratio): Z = X/Y 224</p> <p>4.10 Transformations of Random Variable Families 226</p> <p>4.10.1 Gaussian Transformations 226</p> <p>4.10.2 Exponential Transformations 227</p> <p>4.10.3 Chi-Square Transformations 228</p> <p>4.11 Transformations of Random Vectors 229</p> <p>4.12 Sample Mean ¯X and Sample Variance S2 232</p> <p>4.13 Minimum, Maximum, and Order Statistics 234</p> <p>4.14 Mixtures 238</p> <p>Problems 240</p> <p>Further Reading 243</p> <p><b>5 Expectation and Moments 244</b></p> <p>5.1 Introduction 244</p> <p>5.2 Expectation and Integration 244</p> <p>5.3 Indicator Random Variable 245</p> <p>5.4 Simple Random Variable 246</p> <p>5.5 Expectation for Discrete Sample Spaces 247</p> <p>5.6 Expectation for Continuous Sample Spaces 250</p> <p>5.7 Summary of Expectation 253</p> <p>5.8 Functional View of the Mean 254</p> <p>5.9 Properties of Expectation 255</p> <p>5.10 Expectation of a Function 259</p> <p>5.11 Characteristic Function 260</p> <p>5.12 Conditional Expectation 265</p> <p>5.13 Properties of Conditional Expectation 267</p> <p>5.14 Location Parameters: Mean, Median, and Mode 276</p> <p>5.15 Variance, Covariance, and Correlation 280</p> <p>5.16 Functional View of the Variance 283</p> <p>5.17 Expectation and the Indicator Function 284</p> <p>5.18 Correlation Coefficients 285</p> <p>5.19 Orthogonality 291</p> <p>5.20 Correlation and Covariance Matrices 294</p> <p>5.21 Higher Order Moments and Cumulants 296</p> <p>5.22 Functional View of Skewness 302</p> <p>5.23 Functional View of Kurtosis 303</p> <p>5.24 Generating Functions 304</p> <p>5.25 Fourth-Order Gaussian Moment 309</p> <p>5.26 Expectations of Nonlinear Transformations 310</p> <p>Problems 313</p> <p>Further Reading 316</p> <p><b>PART II Random Processes, Systems, and Parameter Estimation</b></p> <p><b>6 Random Processes 319</b></p> <p>6.1 Introduction 319</p> <p>6.2 Characterizations of a Random Process 319</p> <p>6.3 Consistency and Extension 324</p> <p>6.4 Types of Random Processes 325</p> <p>6.5 Stationarity 326</p> <p>6.6 Independent and Identically Distributed 329</p> <p>6.7 Independent Increments 331</p> <p>6.8 Martingales 333</p> <p>6.9 Markov Sequence 338</p> <p>6.10 Markov Process 350</p> <p>6.11 Random Sequences 352</p> <p>6.11.1 Bernoulli Sequence 352</p> <p>6.11.2 Bernoulli Scheme 352</p> <p>6.11.3 Independent Sequences 353</p> <p>6.11.4 Bernoulli Random Walk 354</p> <p>6.11.5 Binomial Counting Sequence 356</p> <p>6.12 Random Processes 359</p> <p>6.12.1 Poisson Counting Process 359</p> <p>6.12.2 Random Telegraph Signal 365</p> <p>6.12.3 Wiener Process 368</p> <p>6.12.4 Gaussian Process 371</p> <p>6.12.5 Pulse Amplitude Modulation 372</p> <p>6.12.6 Random Sine Signals 373</p> <p>Problems 375</p> <p>Further Reading 379</p> <p><b>7 Stochastic Convergence, Calculus, and Decompositions 380</b></p> <p>7.1 Introduction 380</p> <p>7.2 Stochastic Convergence 380</p> <p>7.3 Laws of Large Numbers 388</p> <p>7.4 Central Limit Theorem 390</p> <p>7.5 Stochastic Continuity 394</p> <p>7.6 Derivatives and Integrals 404</p> <p>7.7 Differential Equations 414</p> <p>7.8 Difference Equations 422</p> <p>7.9 Innovations and Mean-Square Predictability 423</p> <p>7.10 Doob–Meyer Decomposition 428</p> <p>7.11 Karhunen–Lo`eve Expansion 433</p> <p>Problems 441</p> <p>Further Reading 444</p> <p><b>8 Systems, Noise, and Spectrum Estimation 445</b></p> <p>8.1 Introduction 445</p> <p>8.2 Correlation Revisited 445</p> <p>8.3 Ergodicity 448</p> <p>8.4 Eigenfunctions of RXX(τ ) 456</p> <p>8.5 Power Spectral Density 457</p> <p>8.6 Power Spectral Distribution 463</p> <p>8.7 Cross-Power Spectral Density 465</p> <p>8.8 Systems with Random Inputs 468</p> <p>8.8.1 Nonlinear Systems 469</p> <p>8.8.2 Linear Systems 471</p> <p>8.9 Passband Signals 476</p> <p>8.10 White Noise 479</p> <p>8.11 Bandwidth 484</p> <p>8.12 Spectrum Estimation 487</p> <p>8.12.1 Periodogram 487</p> <p>8.12.2 Smoothed Periodogram 493</p> <p>8.12.3 Modified Periodogram 497</p> <p>8.13 Parametric Models 500</p> <p>8.13.1 Autoregressive Model 500</p> <p>8.13.2 Moving-Average Model 505</p> <p>8.13.3 Autoregressive Moving-Average Model 509</p> <p>8.14 System Identification 513</p> <p>Problems 515</p> <p>Further Reading 518</p> <p><b>9 Sufficient Statistics and Parameter Estimation 519</b></p> <p>9.1 Introduction 519</p> <p>9.2 Statistics 519</p> <p>9.3 Sufficient Statistics 520</p> <p>9.4 Minimal Sufficient Statistic 525</p> <p>9.5 Exponential Families 528</p> <p>9.6 Location-Scale Families 533</p> <p>9.7 Complete Statistic 536</p> <p>9.8 Rao–Blackwell Theorem 538</p> <p>9.9 Lehmann–Scheff´e Theorem 540</p> <p>9.10 Bayes Estimation 542</p> <p>9.11 Mean-Square-Error Estimation 545</p> <p>9.12 Mean-Absolute-Error Estimation 552</p> <p>9.13 Orthogonality Condition 553</p> <p>9.14 Properties of Estimators 555</p> <p>9.14.1 Unbiased 555</p> <p>9.14.2 Consistent 557</p> <p>9.14.3 Efficient 559</p> <p>9.15 Maximum A Posteriori Estimation 561</p> <p>9.16 Maximum Likelihood Estimation 567</p> <p>9.17 Likelihood Ratio Test 569</p> <p>9.18 Expectation–Maximization Algorithm 570</p> <p>9.19 Method of Moments 576</p> <p>9.20 Least-Squares Estimation 577</p> <p>9.21 Properties of LS Estimators 582</p> <p>9.21.1 Minimum ξWLS 582</p> <p>9.21.2 Uniqueness 582</p> <p>9.21.3 Orthogonality 582</p> <p>9.21.4 Unbiased 584</p> <p>9.21.5 Covariance Matrix 584</p> <p>9.21.6 Efficient: Achieves CRLB 585</p> <p>9.21.7 BLU Estimator 585</p> <p>9.22 Best Linear Unbiased Estimation 586</p> <p>9.23 Properties of BLU Estimators 590</p> <p>Problems 592</p> <p>Further Reading 595</p> <p>A Note on Part III of the Book 595</p> <p><b>APPENDICES</b></p> <p>Introduction to Appendices 597</p> <p>A Summaries of Univariate Parametric Distributions 599</p> <p>A.1 Notation 599</p> <p>A.2 Further Reading 600</p> <p>A.3 Continuous Random Variables 601</p> <p>A.3.1 Beta (Arcsine for α = β = 1/2, Power Function for β = 1) 601</p> <p>A.3.2 Cauchy 602</p> <p>A.3.3 Chi 603</p> <p>A.3.4 Chi-Square 604</p> <p>A.3.5 Exponential (Shifted by c) 605</p> <p>A.3.6 Extreme Value (Type I: Gumbel) 606</p> <p>A.3.7 F-Distribution 607</p> <p>A.3.8 Gamma (Erlang for r ∈ N with (r ) = (r − 1)!) 608</p> <p>A.3.9 Gaussian (Normal) 609</p> <p>A.3.10 Half-Normal (Folded Normal) 610</p> <p>A.3.11 Inverse Gaussian (Wald) 611</p> <p>A.3.12 Laplace (Double-Sided Exponential) 612</p> <p>A.3.13 Logistic (Sigmoid for {μ = 0, α = 1}) 613</p> <p>A.3.14 Log-Normal 614</p> <p>A.3.15 Maxwell–Boltzmann 615</p> <p>A.3.16 Pareto 616</p> <p>A.3.17 Rayleigh 617</p> <p>A.3.18 Rice 618</p> <p>A.3.19 Student’s t Distribution 619</p> <p>A.3.20 Triangular 620</p> <p>A.3.21 Uniform (Continuous) 621</p> <p>A.3.22 Weibull 622</p> <p>A.4 Discrete Random Variables 623</p> <p>A.4.1 Bernoulli (with Support {0, 1}) 623</p> <p>A.4.2 Bernoulli (Symmetric with Support {−1, 1}) 624</p> <p>A.4.3 Binomial 625</p> <p>A.4.4 Geometric (with Support Z+) 626</p> <p>A.4.5 Geometric (Shifted with Support N) 627</p> <p>A.4.6 Hypergeometric 628</p> <p>A.4.7 Logarithmic (Log-Series) 629</p> <p>A.4.8 Negative Binomial (Pascal) 630</p> <p>A.4.9 Poisson 631</p> <p>A.4.10 Uniform (Discrete) 632</p> <p>A.4.11 Zeta (Zipf) 633</p> <p>B Functions and Properties 634</p> <p>B.1 Continuity and Bounded Variation 634</p> <p>B.2 Supremum and Infimum 640</p> <p>B.3 Order Notation 640</p> <p>B.4 Floor and Ceiling Functions 641</p> <p>B.5 Convex and Concave Functions 641</p> <p>B.6 Even and Odd Functions 641</p> <p>B.7 Signum Function 643</p> <p>B.8 Dirac Delta Function 644</p> <p>B.9 Kronecker Delta Function 645</p> <p>B.10 Unit-Step Functions 646</p> <p>B.11 Rectangle Functions 647</p> <p>B.12 Triangle and Ramp Functions 647</p> <p>B.13 Indicator Functions 648</p> <p>B.14 Sinc Function 649</p> <p>B.15 Logarithm Functions 650</p> <p>B.16 Gamma Functions 651</p> <p>B.17 Beta Functions 653</p> <p>B.18 Bessel Functions 655</p> <p>B.19 Q-Function and Error Functions 655</p> <p>B.20 Marcum Q-Function 659</p> <p>B.21 Zeta Function 659</p> <p>B.22 Rising and Falling Factorials 660</p> <p>B.23 Laguerre Polynomials 661</p> <p>B.24 Hypergeometric Functions 662</p> <p>B.25 Bernoulli Numbers 663</p> <p>B.26 Harmonic Numbers 663</p> <p>B.27 Euler–Mascheroni Constant 664</p> <p>B.28 Dirichlet Function 664</p> <p>Further Reading 664</p> <p>C Frequency-Domain Transforms and Properties 665</p> <p>C.1 Laplace Transform 665</p> <p>C.2 Continuous-Time Fourier Transform 669</p> <p>C.3 z-Transform 670</p> <p>C.4 Discrete-Time Fourier Transform 676</p> <p>Further Reading 677</p> <p>D Integration and Integrals 678</p> <p>D.1 Review of Riemann Integral 678</p> <p>D.2 Riemann–Stieltjes Integral 681</p> <p>D.3 Lebesgue Integral 684</p> <p>D.4 Pdf Integrals 688</p> <p>D.5 Indefinite and Definite Integrals 690</p> <p>D.6 Integral Formulas 692</p> <p>D.7 Double Integrals of Special Functions 692</p> <p>Further Reading 696</p> <p>E Identities and Infinite Series 697</p> <p>E.1 Zero and Infinity 697</p> <p>E.2 Minimum and Maximum 697</p> <p>E.3 Trigonometric Identities 698</p> <p>E.4 Stirling’s Formula 698</p> <p>E.5 Taylor Series 699</p> <p>E.6 Series Expansions and Closed-Form Sums 699</p> <p>E.7 Vandermonde’s Identity 702</p> <p>E.8 Pmf Sums and Functional Forms 703</p> <p>E.9 Completing the Square 704</p> <p>E.10 Summation by Parts 705</p> <p>Further Reading 706</p> <p>F Inequalities and Bounds for Expectations 707</p> <p>F.1 Cauchy–Schwarz and H¨older Inequalities 707</p> <p>F.2 Triangle and Minkowski Inequalities 708</p> <p>F.3 Bienaym´e, Chebyshev, and Markov Inequalities 709</p> <p>F.4 Chernoff’s Inequality 711</p> <p>F.5 Jensen’s Inequality 713</p> <p>F.6 Cram´er–Rao Inequality 714</p> <p>Further Reading 718</p> <p>G Matrix and Vector Properties 719</p> <p>G.1 Basic Properties 719</p> <p>G.2 Four Fundamental Subspaces 721</p> <p>G.3 Eigendecomposition 722</p> <p>G.4 LU, LDU, and Cholesky Decompositions 724</p> <p>G.5 Jacobian Matrix and the Jacobian 726</p> <p>G.6 Kronecker and Schur Products 728</p> <p>G.7 Properties of Trace and Determinant 728</p> <p>G.8 Matrix Inversion Lemma 729</p> <p>G.9 Cauchy–Schwarz Inequality 730</p> <p>G.10 Differentiation 730</p> <p>G.11 Complex Differentiation 731</p> <p>Further Reading 732</p> <p>GLOSSARY 733</p> <p>REFERENCES 743</p> <p>INDEX 755</p> <p><b>PART III Applications in Signal Processing and Communications</b></p> <p>Chapters at the Web Site <a href="http://www.wiley.com/go/randomprocesses">www.wiley.com/go/randomprocesses</a></p> <p><b>10 Communication Systems and Information Theory 771</b></p> <p>10.1 Introduction 771</p> <p>10.2 Transmitter 771</p> <p>10.2.1 Sampling and Quantization 772</p> <p>10.2.2 Channel Coding 777</p> <p>10.2.3 Symbols and Pulse Shaping 778</p> <p>10.2.4 Modulation 781</p> <p>10.3 Transmission Channel 783</p> <p>10.4 Receiver 786</p> <p>10.4.1 Receive Filter 786</p> <p>10.4.2 Demodulation 787</p> <p>10.4.3 Gram–Schmidt Orthogonalization 789</p> <p>10.4.4 Maximum Likelihood Detection 794</p> <p>10.4.5 Matched Filter Receiver 797</p> <p>10.4.6 Probability of Error 802</p> <p>10.5 Information Theory 803</p> <p>10.5.1 Mutual Information and Entropy 804</p> <p>10.5.2 Properties of Mutual Information and Entropy 810</p> <p>10.5.3 Continuous Distributions: Differential Entropy 813</p> <p>10.5.4 Channel Capacity 818</p> <p>10.5.5 AWGN Channel 820</p> <p>Problems 821</p> <p>Further Reading 824</p> <p><b>11 Optimal Filtering</b> <a href="http://www.wiley.com/go/randomprocesses"><b>www.wiley.com/go/randomprocesses</b></a> <b>825</b></p> <p>11.1 Introduction 825</p> <p>11.2 Optimal Linear Filtering 825</p> <p>11.3 Optimal Filter Applications 827</p> <p>11.3.1 System Identification 827</p> <p>11.3.2 Inverse Modeling 827</p> <p>11.3.3 Noise Cancellation 828</p> <p>11.3.4 Linear Prediction 828</p> <p>11.4 Noncausal Wiener Filter 829</p> <p>11.5 Causal Wiener Filter 831</p> <p>11.6 Prewhitening Filter 837</p> <p>11.7 FIR Wiener Filter 839</p> <p>11.8 Kalman Filter 844</p> <p>11.8.1 Evolution of the Mean and Covariance 846</p> <p>11.8.2 State Prediction 846</p> <p>11.8.3 State Filtering 848</p> <p>11.9 Steady-State Kalman Filter 851</p> <p>11.10 Linear Predictive Coding 857</p> <p>11.11 Lattice Prediction-Error Filter 861</p> <p>11.12 Levinson–Durbin Algorithm 865</p> <p>11.13 Least-Squares Filtering 868</p> <p>11.14 Recursive Least-Squares 872</p> <p>Problems 876</p> <p>Further Reading 879</p> <p><b>12 Adaptive Filtering</b> <a href="http://www.wiley.com/go/randomprocesses"><b>www.wiley.com/go/randomprocesses</b></a> <b>880</b></p> <p>12.1 Introduction 880</p> <p>12.2 MSE Properties 880</p> <p>12.3 Steepest Descent 889</p> <p>12.4 Newton’s Method 894</p> <p>12.5 LMS Algorithm 895</p> <p>12.5.1 Convergence in the Mean 899</p> <p>12.5.2 Convergence in the Mean-Square 901</p> <p>12.5.3 Misadjustment 906</p> <p>12.6 Modified LMS Algorithms 911</p> <p>12.6.1 Sign-Error LMS Algorithm 911</p> <p>12.6.2 Sign-Data LMS Algorithm 912</p> <p>12.6.3 Sign-Sign LMS Algorithm 914</p> <p>12.6.4 LMF Algorithm 914</p> <p>12.6.5 Complex LMS Algorithm 916</p> <p>12.6.6 “Leaky” LMS Algorithm 917</p> <p>12.6.7 Normalized LMS Algorithm 918</p> <p>12.6.8 Perceptron 920</p> <p>12.6.9 Convergence of Modified LMS Algorithms 922</p> <p>12.7 Adaptive IIR Filtering 923</p> <p>12.7.1 Output-Error Formulation 924</p> <p>12.7.2 Output-Error IIR Filter Algorithm 928</p> <p>12.7.3 Equation-Error Formulation 932</p> <p>12.7.4 Equation-Error Bias 933</p> <p>Problems 936</p> <p>Further Reading 939</p> <p><b>13 Equalization, Beamforming, and Direction Finding</b> <a href="http://www.wiley.com/go/randomprocesses"><b>www.wiley.com/go/randomprocesses</b></a> <b>940</b></p> <p>13.1 Introduction 940</p> <p>13.2 Channel Equalization 941</p> <p>13.3 Optimal Bussgang Algorithm 943</p> <p>13.4 Blind Equalizer Algorithms 949</p> <p>13.4.1 Sato’s Algorithm 949</p> <p>13.4.2 Constant Modulus Algorithm 950</p> <p>13.5 CMA Performance Surface 952</p> <p>13.6 Antenna Arrays 958</p> <p>13.7 Beampatterns 960</p> <p>13.8 Optimal Beamforming 962</p> <p>13.8.1 Known Look Direction 962</p> <p>13.8.2 Multiple Constraint Beamforming 964</p> <p>13.8.3 Training Signal 966</p> <p>13.8.4 Maximum Likelihood 968</p> <p>13.8.5 Maximum SNR and SINR 969</p> <p>13.9 Adaptive Beamforming 970</p> <p>13.9.1 LMS Beamforming 970</p> <p>13.9.2 Constant Modulus Array 970</p> <p>13.9.3 Decision-Directed Mode 973</p> <p>13.9.4 Multistage CM Array 974</p> <p>13.9.5 Output SINR and SNR 977</p> <p>13.10 Direction Finding 981</p> <p>13.10.1 Beamforming Approaches 981</p> <p>13.10.2 MUSIC Algorithm 984</p> <p>Problems 985</p> <p>Further Reading 989</p>
<p><b>JOHN J. SHYNK, PhD,</b> is Professor of Electrical and Computer Engineering at the University of California, Santa Barbara. He was a Member of Technical Staff at Bell Laboratories, and received degrees in systems engineering, electrical engineering, and statistics from Boston University and Stanford University.</p>
<p><b>An engineering perspective on probability and random processes</b></p> <p><i>Probability, Random Variables, and Random Processes</i> is a comprehensive textbook on probability theory for engineers that provides a more rigorous mathematical framework than is usually encountered in undergraduate courses. It is intended for first-year graduate students who have some familiarity with probability and random variables, though not necessarily of random processes and systems that operate on random signals. It is also appropriate for advanced undergraduate students who have a strong mathematical background.</p> <p>The book has the following features:</p> <ul> <li>Several appendices include related material on integration, important inequalities and identities, frequency-domain transforms, and linear algebra. These topics have been included so that the book is relatively self-contained. One appendix contains an extensive summary of 33 random variables and their properties such as moments, characteristic functions, and entropy.</li> <li>Unlike most books on probability, numerous figures have been included to clarify and expand upon important points. Over 600 illustrations and MATLAB plots have been designed to reinforce the material and illustrate the various characterizations and properties of random quantities.</li> <li>Sufficient statistics are covered in detail, as is their connection to parameter estimation techniques. These include classical Bayesian estimation and several optimality criteria: mean-square error, mean-absolute error, maximum likelihood, method of moments, and least squares.</li> <li>The last four chapters provide an introduction to several topics usually studied in subsequent engineering courses: communication systems and information theory; optimal filtering (Wiener and Kalman); adaptive filtering (FIR and IIR); and antenna beamforming, channel equalization, and direction finding. This material is available electronically at the companion website.</li> </ul> <p><i>Probability, Random Variables, and Random Processes</i> is the only textbook on probability for engineers that includes relevant background material, provides extensive summaries of key results, and extends various statistical techniques to a range of applications in signal processing.</p>

Diese Produkte könnten Sie auch interessieren:

Bandwidth Efficient Coding
Bandwidth Efficient Coding
von: John B. Anderson
PDF ebook
114,99 €
Bandwidth Efficient Coding
Bandwidth Efficient Coding
von: John B. Anderson
EPUB ebook
114,99 €