Information Theory: A Concise Introduction
By Stefan Hollos and J. Richard Hollos
()
About this ebook
Books on information theory tend to fall into one of two extreme categories. There are large academic textbooks that cover the subject with great depth and rigor. Probably the best known of these is the book by Cover and Thomas. At the other extreme are the popular books such as the ones by Pierce and Gleick. They provide a very superficial introduction to the subject, enough to engage in cocktail party conversation but little else. This book attempts to bridge these two extremes.
This book is written for someone who is at least semi-mathematically literate and wants a concise introduction to some of the major concepts in information theory. The level of mathematics needed is very elementary. A rudimentary grasp of logarithms, probability, and basic algebra is all that is required. Two chapters at the end of the book provide a review of everything the reader needs to know about logarithms and discrete probability to get the most out of the book. Very little attention is given to mathematical proof. Instead the results are presented in a way that makes them almost obvious or at least plausible.
The book will appeal to anyone looking for a fast introduction to most of the major topics in information theory. An introduction that is concise but not superficial.
Stefan Hollos
Stefan Hollos is a physicist and electrical engineer by training, and enjoys anything related to math, physics, engineering and computing. He also enjoys creating music and visual art, and being in the great outdoors. He is the author of 18 books.
Read more from Stefan Hollos
Creating Rhythms Rating: 5 out of 5 stars5/5Creating Melodies Rating: 0 out of 5 stars0 ratingsNell: An SVG Drawing Language Rating: 0 out of 5 stars0 ratingsCoin Tossing: The Hydrogen Atom of Probability Rating: 0 out of 5 stars0 ratingsThe Enigma of the Crookes Radiometer Rating: 0 out of 5 stars0 ratings
Related to Information Theory
Related ebooks
Information Theory Rating: 0 out of 5 stars0 ratingsMathematical Foundations of Information Theory Rating: 4 out of 5 stars4/5Mathematical Logic Rating: 4 out of 5 stars4/5The Foundations of Statistics Rating: 0 out of 5 stars0 ratingsIntroduction to Mathematical Thinking: The Formation of Concepts in Modern Mathematics Rating: 4 out of 5 stars4/5Naive Set Theory Rating: 4 out of 5 stars4/5A Beginner's Guide to Mathematical Logic Rating: 3 out of 5 stars3/5Essential Algorithms: A Practical Approach to Computer Algorithms Rating: 5 out of 5 stars5/5Prelude to Mathematics Rating: 4 out of 5 stars4/5Markov Models: An Introduction to Markov Models Rating: 3 out of 5 stars3/5A Passion for Mathematics: Numbers, Puzzles, Madness, Religion, and the Quest for Reality Rating: 4 out of 5 stars4/5Fundamental Concepts of Abstract Algebra Rating: 5 out of 5 stars5/5Algorithmic Information Theory: Fundamentals and Applications Rating: 0 out of 5 stars0 ratingsExtremal Graph Theory Rating: 3 out of 5 stars3/5Finite Field Fun: A lightweight introduction to finite fields and their applications for engineers, computer scientists, and others Rating: 0 out of 5 stars0 ratingsAlgorithmic Probability: Fundamentals and Applications Rating: 0 out of 5 stars0 ratingsSimply Riemann Rating: 0 out of 5 stars0 ratingsFoundations of the Theory of Probability: Second English Edition Rating: 0 out of 5 stars0 ratingsA First Course in Graph Theory Rating: 5 out of 5 stars5/5Fundamentals of Number Theory Rating: 4 out of 5 stars4/5An Introduction to Stochastic Processes in Physics Rating: 3 out of 5 stars3/5Probability Theory: A Concise Course Rating: 4 out of 5 stars4/5Formal Languages And Automata Theory Rating: 0 out of 5 stars0 ratingsIntroduction to Combinatorial Analysis Rating: 5 out of 5 stars5/5Learning Probabilistic Graphical Models in R Rating: 0 out of 5 stars0 ratingsChaotic Dynamics of Nonlinear Systems Rating: 5 out of 5 stars5/5What Is Mathematical Logic? Rating: 3 out of 5 stars3/5Theory of Games and Statistical Decisions Rating: 4 out of 5 stars4/5Elementary Number Theory: An Algebraic Approach Rating: 0 out of 5 stars0 ratings
Information Technology For You
How to Write Effective Emails at Work Rating: 4 out of 5 stars4/5Data Analytics for Beginners: Introduction to Data Analytics Rating: 4 out of 5 stars4/5A Mind at Play: How Claude Shannon Invented the Information Age Rating: 4 out of 5 stars4/5Creating Online Courses with ChatGPT | A Step-by-Step Guide with Prompt Templates Rating: 4 out of 5 stars4/5An Ultimate Guide to Kali Linux for Beginners Rating: 3 out of 5 stars3/5Excel VBA: A Step-By-Step Tutorial For Beginners To Learn Excel VBA Programming From Scratch: 1 Rating: 4 out of 5 stars4/5ChatGPT: The Future of Intelligent Conversation Rating: 4 out of 5 stars4/5CompTIA A+ CertMike: Prepare. Practice. Pass the Test! Get Certified!: Core 1 Exam 220-1101 Rating: 0 out of 5 stars0 ratingsAWS Certified Cloud Practitioner: Study Guide with Practice Questions and Labs Rating: 5 out of 5 stars5/5COMPUTER SCIENCE FOR ROOKIES Rating: 0 out of 5 stars0 ratingsCompTia Security 701: Fundamentals of Security Rating: 0 out of 5 stars0 ratingsIntroduction to Information Systems: Information Technology Essentials, #1 Rating: 0 out of 5 stars0 ratingsPersonal Knowledge Graphs: Connected thinking to boost productivity, creativity and discovery Rating: 5 out of 5 stars5/520 Windows Tools Every SysAdmin Should Know Rating: 4 out of 5 stars4/5CompTIA Network+ CertMike: Prepare. Practice. Pass the Test! Get Certified!: Exam N10-008 Rating: 0 out of 5 stars0 ratingsCyber Security Consultants Playbook Rating: 0 out of 5 stars0 ratingsMachine Learning Interview Questions Rating: 5 out of 5 stars5/5Cybersecurity for Beginners : Learn the Fundamentals of Cybersecurity in an Easy, Step-by-Step Guide: 1 Rating: 0 out of 5 stars0 ratingsCybersecurity Playbook for Executives Rating: 0 out of 5 stars0 ratingsThe CISSP Fast-Track: Conquer the 8 Domains: CyberSecurity Rating: 0 out of 5 stars0 ratingsCompTIA ITF+ CertMike: Prepare. Practice. Pass the Test! Get Certified!: Exam FC0-U61 Rating: 5 out of 5 stars5/5The Rise of AI Income: Using Artificial Intelligence for Financial Success Rating: 5 out of 5 stars5/5Object-Oriented JavaScript: Create scalable, reusable high-quality JavaScript applications, and libraries Rating: 3 out of 5 stars3/5Character Expression: Using ChatGPT to Write Believable Emotions in Fiction Rating: 3 out of 5 stars3/5Practical Ethical Hacking from Scratch Rating: 5 out of 5 stars5/5Agile Business Architecture for Digital Transformation - V2 Rating: 5 out of 5 stars5/5The Ultimate Guide to ChatGPT Prompts: Tips, Tricks and Templates Rating: 2 out of 5 stars2/5
Reviews for Information Theory
0 ratings0 reviews
Book preview
Information Theory - Stefan Hollos
Information Theory: A Concise Introduction
Contents
Preface
Introduction
Number Guessing Game
Counterfeit Coins
Encoding Messages
Nonuniform Probabilities
Kraft-McMillan Inequality
Average Code Word Length
Huffman Coding
Arithmetic Coding
Entropy
Entropy of a Markov Chain
Principle of Maximum Entropy
Entropy of English
Channel Capacity
Channel Capacity and Gambling
Error Correction Coding
Repetition Codes
Parity Check Codes
Hamming Codes
Review of Logarithms
Review of Discrete Probability
References & Further Reading
Acknowledgments
About the Authors
Thank You
Preface
Books on information theory tend to fall into one of two extreme categories. There are large academic textbooks that cover the subject with great depth and rigor. Probably the best known of these is the book by Cover and Thomas. At the other extreme are the popular books such as the ones by Pierce and Gleick. They provide a very superficial introduction to the subject, enough to engage in cocktail party conversation but little else. This book attempts to bridge these two extremes.
This book is written for someone who is at least semi-mathematically literate and wants a concise introduction to some of the major concepts in information theory. The level of mathematics needed is very elementary. A rudimentary grasp of logarithms, probability, and basic algebra is all that is required. Two chapters at the end of the book provide a review of everything the reader needs to know about logarithms and discrete probability to get the most out of the book. Very little attention is given to mathematical proof. Instead we try to present the results in a way that makes them almost obvious or at least plausible.
We start in the introduction with a discussion of how information theory has its roots in the field of communication systems design. This leads to the question of how to quantify information and how a logarithmic measure is the most sensible. The concept of entropy is introduced at this point but only for the case where all messages are equally probable. The introduction ends with two examples of how information concepts come up in areas seemingly unrelated to communication. The first is a number guessing game and the second is the problem of finding a counterfeit coin.
The next chapter looks at the problem of encoding messages as efficiently as possible. This is the source coding or data compression problem. The idea of prefix free codes and the Kraft-McMillan inequality are introduced. It is shown how the entropy is a lower limit for the average length of a code word.
The following two chapters discuss specific coding techniques. The first is Huffman coding. Three detailed examples of constructing a Huffman code are worked out. Software for constructing Huffman codes can be found on the book’s website:
The next coding chapter discusses a powerful technique called arithmetic coding. This technique encodes a string of symbols as a single code word. For long strings of symbols it gets very close to the per symbol entropy.
There is a long chapter devoted just to the concept of entropy. How to calculate joint and conditional entropy is covered along with a detailed example of how to use them. There is a discussion of mutual information, what it means and how it is calculated. This is followed by a simple example of how to calculate the entropy of a Markov chain. The chapter ends with an elementary example using the principle of maximum entropy to infer a probability distribution.
Calculating the entropy of English is covered in the next chapter. It shows how to use the statistics of n-grams to get a series of increasingly accurate estimates for the entropy. It also shows how to use the statistics of words to estimate the entropy.
The next chapter covers channel capacity and the noisy channel coding theorem. It shows in general how to calculate channel capacity but only the capacity of a binary symmetric channel is worked out in detail. There is a brief discussion of the noisy channel coding theorem with no proofs. The chapter ends with an unusual example of the use of channel capacity in gambling and investing. This is called the Kelly gambling system.
The final chapter is a brief introduction to the topic of error correction or channel coding. Repetition codes, parity check codes, and Hamming codes are covered.
We hope this book is useful for someone looking for a fast introduction to most of the major topics in information theory. An introduction that is concise but not superficial.
Introduction
The field of information theory was created almost entirely by one man, an American mathematician and engineer named Claude Elwood Shannon. This is a man whose list of intellectual accomplishments is so impressive, you have to wonder if he’s not a figment of someone’s imagination.
Claude Elwood Shannon, the father of information theory.Claude Elwood Shannon - The Father of Information Theory.
Shannon was born in 1916 in Michigan, and grew up in the small town of Gaylord, Michigan. In 1936 he graduated from the University of Michigan with a bachelor degree in mathematics and another in electrical engineering. He then went on to graduate school at the Massachusetts Institute of Technology. His 1937 master’s thesis was on the use of Boolean algebra to design relay and switching circuits. This work provided the foundation for what would later become known as digital circuit design, which is the field of electrical engineering concerned with the design of digital computers and other kinds of digital circuits. His 1940 PhD thesis was on theoretical genetics.
After spending some time at the Institute for Advanced Study in Princeton, Shannon went to work at the Bell Telephone Labs. There he was exposed to problems in communication theory and cryptography which he worked on during World War II. In 1948 he published a paper in two parts in the Bell System Technical Journal titled A Mathematical Theory of Communication
. This paper is widely considered to be the founding document of information theory.
Shannon created information theory to solve communication problems. The goal of communication is to transmit information from one point in space/time to another point in space/time. Some common examples are radio broadcasts, the interaction of a web browser with a web server, and the storage of a file on disk for later access.
Communication problems fall roughly into two categories. First we have the source coding problem where the goal is to represent information as efficiently as possible. This is also known as data compression