Categories
Nevin Manimala Statistics

Equivalence of information production and generalised entropies in complex processes

PLoS One. 2023 Sep 6;18(9):e0290695. doi: 10.1371/journal.pone.0290695. eCollection 2023.

ABSTRACT

Complex systems with strong correlations and fat-tailed distribution functions have been argued to be incompatible with the Boltzmann-Gibbs entropy framework and alternatives, so-called generalised entropies, were proposed and studied. Here we show, that this perceived incompatibility is actually a misconception. For a broad class of processes, Boltzmann entropy -the log multiplicity- remains the valid entropy concept. However, for non-i.i.d. processes, Boltzmann entropy is not of Shannon form, -k∑ipi log pi, but takes the shape of generalised entropies. We derive this result for all processes that can be asymptotically mapped to adjoint representations reversibly where processes are i.i.d. In these representations the information production is given by the Shannon entropy. Over the original sampling space this yields functionals identical to generalised entropies. The problem of constructing adequate context-sensitive entropy functionals therefore can be translated into the much simpler problem of finding adjoint representations. The method provides a comprehensive framework for a statistical physics of strongly correlated systems and complex processes.

PMID:37672525 | DOI:10.1371/journal.pone.0290695

By Nevin Manimala

Portfolio Website for Nevin Manimala