Entropy
In the previous section we saw that an irrational rotation on and the doubling map are essentially different dynamical systems because the former has lots of eigenfunctions while the latter has none. Let us further pursue the topic of distinguishing dynamical systems by analysing the shift map
on with respect to different coin measures and . Are the measure-preserving systems
distinguishable? Both are mixing, so both have no eigenfunctions. We cannot distinguish them by properties of their Koopman operators.
It turns out that these transformations are not the same, and that they can be distinguished by an invariant originating in statistical physics and information theory - entropy - that is not detected by the Koopman operator. For some of the history on the extent to which the Koopman operator determines the point dynamics, see this article.
Defining entropy
Fix a measure space with . By a partition of we mean a finite tuple
of sets in that are pairwise disjoint and cover . We think of a partition as classifying the points of according to some rule. For example
is a partition of classifying points according to their first two terms, and
is a partition of classifying points based on the first digits in their ternary expansions.
The measure assigns a size to each set in a given partition. If we take the view that sets of smaller size are less likely, we might therefore consider ourselves to have gained more or less information about a -random point depending on whether it is to be found in a smaller or larger set in the partition.
Definition
Given a partition
of a probability space the quantity
is called its entropy.
To make the above definition we take .
We imagine that the entropy of a partition tells us the amount of information we gain by finding out to which member of our partition a random point in belongs.
Example
The partition
has, for the fair coin measure an entropy of
and for any other partition of into four sets we have
by convexity.
Given a partition we can form a new partition
using . The new partition corresponds to performing the experiment corresponding to after one iteration of the dynamics. Since is measure-preserving the partitions and have the same entropy. How much information do we gain from performing the experiment corresponding to now and after one iteration of the dynamics? This is the same as performing the experiment corresponding to the partition
where the symbol is the join of two partitions.
It may be that entropy gained
by performing the experiment a second time is zero. On the other hand it could be quite large.
Example
For the partition we have
and its entropy for the fair coin measure is . Note that
and we can think of this difference as representing the extra bit needed to store the outcome of the experiment compared with the outcome of experiment .
Definition
Fix a measure-preserving transformation on a probability space . The quantity
is the entropy of for the partition .
The entropy of for the partition is the exponential growth rate of the amount of information obtained by repeatedly performing the experiment corresponding to after more and more iterations of the dynamics. We imagine that positivity of tells us something about the randomness of . If is positive then is unpredictable in the sense that no matter how many iterations of the experiment one carries out, there is still information to be gained by performing the experiment more often.
Example
Let us calculate the entropy of the full shift on for the partition and the fair coin measure . Since
is equal to the partition of into cylinder sets of length and each such cylinder has measure we see that
and therefore
is the entropy of with respect to .
The transformation may very well have different entropies with respect to different partitions of . To form an invariant of the transformation itself we look for the experiment that maximizes the entropy.
Definition
The entropy of a measure-preserving map on a probability space is the supermum
of all possible entropies of with respect to partitions that themselves have finite entropy.
The following important theeorem gives us a way to calculate the entropy of a measure-preserving transformations.
Theorem (Kolmogorov-Sinai)
If is a partition such that
then .
Since the cylinder sets generate the Borel σ-algebra on we can calculate the entropy of a measure-preserving transformation on using the partition . If is the coin measure then
and we conclude that, for not two values are with and with isomorphic measure-preserving transformations.