In What Sense is the Kolmogorov-Sinai Entropy a Measure for Chaotic Behaviour? - Bridging the Gap Between Dynamical Systems Theory and Communication Theory

Publication Year:
Usage 1698
Downloads 1698
Repository URL:
Frigg, Roman
artifact description
On an influential account, chaos is explained in terms of random behaviour; and random behaviour in turn is explained in terms of having positive Kolmogorov-Sinai entropy (KSE). Though intuitively plausible, the association of the KSE with random behaviour needs justification since the definition of the KSE does not make reference to any notion that is connected to randomness. I provide this justification for the case of Hamiltonian systems by proving that the KSE is equivalent to a generalized version of Shannon's communication-theoretic entropy under certain plausible assumptions. I then discuss consequences of this equivalence for randomness in chaotic dynamical systems.