Information Theory
π§© Information Theory
Information theory is a branch of applied mathematics and electrical engineering involving the quantification of information. It was founded by Claude Shannon in 1948 and has since become the foundation for digital communication and data science.
π’ 1. Entropy: The Measure of Uncertainty
Shannon Entropy ()
Entropy measures the average amount of information produced by a stochastic source of data.
- High Entropy: The outcome is highly uncertain (e.g., a fair coin toss).
- Low Entropy: The outcome is highly predictable (e.g., a biased coin).
Cross-Entropy
Used extensively as a loss function in machine learning, it measures the difference between two probability distributions (the true labels and the predicted probabilities).
π‘ 2. Information Gain and Divergence
Mutual Information ()
Measures the amount of information that can be obtained about one random variable by observing another.
- It is symmetric: .
- It is non-negative: .
KL-Divergence ()
Also known as relative entropy, it measures how one probability distribution diverges from a second, expected probability distribution .
π΄ 3. Coding Theory and Channel Capacity
Source Coding Theorem
Shannonβs source coding theorem establishes that, on average, the number of bits needed to represent the result of an uncertain event is given by its entropy.
- This provides the theoretical limit for lossless data compression.
Channel Capacity ()
The tightest upper bound on the rate of information that can be reliably transmitted over a communications channel. For a Gaussian channel with bandwidth and signal-to-noise ratio :