Details

Information Theory Meets Power Laws


Information Theory Meets Power Laws

Stochastic Processes and Language Models
1. Aufl.

von: Lukasz Debowski

100,99 €

Verlag: Wiley
Format: PDF
Veröffentl.: 02.11.2020
ISBN/EAN: 9781119625360
Sprache: englisch
Anzahl Seiten: 384

DRM-geschütztes eBook, Sie benötigen z.B. Adobe Digital Editions und eine Adobe ID zum Lesen.

Beschreibungen

<p><b>Discover new theoretical connections between stochastic phenomena and the structure of natural language with this powerful volume!</b></p> <p><i>Information Theory Meets Power Laws: Stochastic Processes and Language Models</i> presents readers with a novel subtype of a probabilistic approach to language, which is based on statistical laws of texts and their analysis by means of information theory. The distinguished author insightfully and rigorously examines the linguistic and mathematical subject matter while eschewing needlessly abstract and superfluous constructions.</p> <p>The book begins with a less formal treatment of its subjects in the first chapter, introducing its concepts to readers without mathematical training and allowing those unfamiliar with linguistics to learn the book’s motivations. Despite its inherent complexity, <i>Information Theory Meets Power Laws: Stochastic Processes and Language Models </i>is a surprisingly approachable treatment of idealized mathematical models of human language.</p> <p>The author succeeds in developing some of the theory underlying fundamental stochastic and semantic phenomena, like strong nonergodicity, in a way that has not previously been seriously attempted. In doing so, he covers topics including:</p> <ul> <li>Zipf’s and Herdan’s laws for natural language</li> <li>Power laws for information, repetitions, and correlations</li> <li>Markov, finite-state,and Santa Fe processes</li> <li>Bayesian and frequentist  interpretations of probability</li> <li>Ergodic decomposition, Kolmogorov complexity, and universal coding</li> <li>Theorems about facts and words</li> <li>Information measures for fields</li> <li>Rényi entropies, recurrence times, and subword complexity</li> <li>Asymptotically mean stationary processes</li> </ul> <p>Written primarily for mathematics graduate students and professionals interested in information theory or discrete stochastic processes, <i>Information Theory Meets Power Laws: Stochastic Processes and Language Models </i>also belongs on the bookshelves of doctoral students and researchers in artificial intelligence, computational and quantitative linguistics as well as physics of complex systems.</p>
<p>Preface ix</p> <p>Acknowledgments xiii</p> <p>Basic Notations xv</p> <p><b>1 Guiding Ideas 1</b></p> <p>1.1 The Motivating Question 1</p> <p>1.2 Further Questions About Texts 5</p> <p>1.3 Zipf’s and Herdan’s Laws 8</p> <p>1.4 Markov and Finite-State Processes 14</p> <p>1.5 More General Stochastic Processes 20</p> <p>1.6 Two Interpretations of Probability 23</p> <p>1.7 Insights from Information Theory 25</p> <p>1.8 Estimation of Entropy Rate 28</p> <p>1.9 Entropy of Natural Language 30</p> <p>1.10 Algorithmic Information Theory 35</p> <p>1.11 Descriptions of a Random World 37</p> <p>1.12 Facts and Words Related 43</p> <p>1.13 Repetitions and Entropies 47</p> <p>1.14 Decay of Correlations 52</p> <p>1.15 Recapitulation 54</p> <p><b>2 Probabilistic Preliminaries </b><b>57</b></p> <p>2.1 Probability Measures 59</p> <p>2.2 Product Measurable Spaces 63</p> <p>2.3 Discrete Random Variables 65</p> <p>2.4 From IID to Finite-State Processes 68</p> <p>Problems 73</p> <p><b>3 Probabilistic Toolbox </b><b>77</b></p> <p>3.1 Borel 𝜎-Fields and a Fair Coin 79</p> <p>3.2 Integral and Expectation 83</p> <p>3.3 Inequalities and Corollaries 87</p> <p>3.4 Semidistributions 92</p> <p>3.5 Conditional Probability 94</p> <p>3.6 Modes of Convergence 101</p> <p>3.7 Complete Spaces 103</p> <p>Problems 106</p> <p><b>4 Ergodic Properties </b><b>109</b></p> <p>4.1 Plain Relative Frequency 111</p> <p>4.2 Birkhoff Ergodic Theorem 116</p> <p>4.3 Ergodic and Mixing Criteria 119</p> <p>4.4 Ergodic Decomposition 125</p> <p>Problems 128</p> <p><b>5 Entropy and Information </b><b>131</b></p> <p>5.1 Shannon Measures for Partitions 133</p> <p>5.2 Block Entropy and Its Limits 139</p> <p>5.3 Shannon Measures for Fields 145</p> <p>5.4 Block Entropy Limits Revisited 155</p> <p>5.5 Convergence of Entropy 159</p> <p>5.6 Entropy as Self-Information 160</p> <p>Problems 163</p> <p><b>6 Equipartition and Universality </b><b>167</b></p> <p>6.1 SMB Theorem 169</p> <p>6.2 Universal Semidistributions 171</p> <p>6.3 PPM Probability 172</p> <p>6.4 SMB Theorem Revisited 178</p> <p>6.5 PPM-based Statistics 180</p> <p>Problems 186</p> <p><b>7 Coding and Computation </b><b>189</b></p> <p>7.1 Elements of Coding 191</p> <p>7.2 Kolmogorov Complexity 197</p> <p>7.3 Algorithmic Coding Theorems 207</p> <p>7.4 Limits of Mathematics 215</p> <p>7.5 Algorithmic Randomness 220</p> <p>Problems 225</p> <p><b>8 Power Laws for Information </b><b>229</b></p> <p>8.1 Hilberg Exponents 231</p> <p>8.2 Second Order SMB Theorem 238</p> <p>8.3 Probabilistic and Algorithmic Facts 241</p> <p>8.4 Theorems About Facts and Words 248</p> <p>Problems 255</p> <p><b>9 Power Laws for Repetitions </b><b>259</b></p> <p>9.1 Rényi–Arimoto Entropies 261</p> <p>9.2 Generalized Entropy Rates 266</p> <p>9.3 Recurrence Times 268</p> <p>9.4 Subword Complexity 272</p> <p>9.5 Two Maximal Lengths 280</p> <p>9.6 Logarithmic Power Laws 284</p> <p>Problems 289</p> <p><b>10 AMS Processes </b><b>291</b></p> <p>10.1 AMS and Pseudo AMS Measures 293</p> <p>10.2 Quasiperiodic Coding 295</p> <p>10.3 Synchronizable Coding 298</p> <p>10.4 Entropy Rate in the AMS Case 301</p> <p>Problems 304</p> <p><b>11 Toy Examples </b><b>307</b></p> <p>11.1 Finite and Ultrafinite Energy 309</p> <p>11.2 Santa Fe Processes and Alike 315</p> <p>11.3 Encoding into a Finite Alphabet 323</p> <p>11.4 Random Hierarchical Association 334</p> <p>11.5 Toward Better Models 345</p> <p>Problems 348</p> <p>Future Research 349</p> <p>Bibliography 351</p> <p>Index 365</p>
<p><b>ŁUKASZ DĘBOWSKI, P<small>H</small>D,</b> works at the Institute of Computer Science of the Polish Academy of Sciences in Poland. His doctorate is in mathematics and computer science and his primary research focus is in the areas of information theory and discrete stochastic processes. He is also interested in the theoretical properties of statistical and neural language models.
<p><b>Discover new theoretical connections between stochastic phenomena and the structure of natural language with this powerful volume</b> <p><i>Information Theory Meets Power Laws: Stochastic Processes and Language Models</i> presents readers with a novel subtype of a probabilistic approach to language, which is based on statistical laws of texts and their analysis by means of information theory. The distinguished author insightfully and rigorously examines the linguistic and mathematical subject matter while eschewing needlessly abstract and superfluous constructions. <p>The book begins with a less formal treatment of its subjects in the first chapter, introducing its concepts to readers without mathematical training and allowing those unfamiliar with linguistics to learn the book's motivations. Despite its inherent complexity, <i>Information Theory Meets Power Laws: Stochastic Processes and Language Models</i> is a surprisingly approachable treatment of idealized mathematical models of human language. <p>The author succeeds in developing some of the theory underlying fundamental stochastic and semantic phenomena, like strong nonergodicity, in a way that has not previously been attempted. In doing so, he covers topics including: <ul> <li>Zipf's and Herdan's laws for natural language</li> <li>Power laws for information, repetitions, and correlations</li> <li>Markov, finite-state, and Santa Fe processes</li> <li>Bayesian and frequentist interpretations of probability</li> <li>Ergodic decomposition, Kolmogorov complexity, and universal coding</li> <li>Theorems about facts and words</li> <li>Information measures for fields</li> <li>Rényi entropies, recurrence times, and subword complexity</li> <li>Asymptotically mean stationary processes</li> </ul> <p>Written primarily for mathematics graduate students and professionals interested in information theory or discrete stochastic processes, <i>Information Theory Meets Power Laws: Stochastic Processes and Language Models</i> also belongs on the bookshelves of doctoral students and researchers in artificial intelligence, computational and quantitative linguistics as well as physics of complex systems.

Diese Produkte könnten Sie auch interessieren:

Hot-Melt Extrusion
Hot-Melt Extrusion
von: Dennis Douroumis
PDF ebook
136,99 €
Hot-Melt Extrusion
Hot-Melt Extrusion
von: Dennis Douroumis
EPUB ebook
136,99 €
Kunststoffe
Kunststoffe
von: Wilhelm Keim
PDF ebook
99,99 €