Next Article in Journal
Towards Quality Assessment for Arbitrary Translational 6DoF Video: Subjective Quality Database and Objective Assessment Metric
Previous Article in Journal
Operational Meaning of Classical Fidelity and Path Length in Kubo–Mori–Bogoliubov Fisher Geometry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Differential and the Integral Value of Information

by
Raphael D. Levine
1,2,3
1
Institute of Chemistry, The Hebrew University of Jerusalem, Jerusalem 91904, Israel
2
Department of Molecular and Medical Pharmacology, David Geffen School of Medicine, University of California, Los Angeles, CA 90095, USA
3
Department of Chemistry and Biochemistry, University of California, Los Angeles, CA 90095, USA
Entropy 2025, 27(1), 43; https://doi.org/10.3390/e27010043
Submission received: 25 November 2024 / Revised: 26 December 2024 / Accepted: 4 January 2025 / Published: 7 January 2025
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
A quantitative expression for the value of information within the framework of information theory and of the maximal entropy formulation is discussed. We examine both a local, differential measure and an integral, global measure for the value of the change in information when additional input is provided. The differential measure is a potential and as such carries a physical dimension. The integral value has the dimension of information. The differential measure can be used, for example, to discuss how the value of information changes with time or with other parameters of the problem.

1. Introduction

The fathers of what we now call information theory were quite clear. Information is to be defined in a manner that is objective. Information is provided by an answer to a question. The value of the answer for a particular person is not part of the theory. To represent the uncertainty before the answer is given, one considers a random variable X. Take it that there are n different possible answers n ≥ 2. And that the random variable X can assume n different values x i ,   i = 1 , 2 , n , each one being a possible answer with a probability pi. The uncertainty about the answer ( entropy, H(X), of the random variable X) is reduced when additional information is provided. This reduction is the meaning of the term value as used in this paper. Our quantitative examination of the value of information is most influenced by the erudite discourse of this topic by Dunn and Golan [1]. There is, of course, the general result, introduced by Shannon [2], of the reduction in the uncertainty about the random variable X given another random variable Y. This concept is central to the celebrated channel capacity theorem of Shannon. In that context, the noisy channel that determines the conditional distribution of the output Y given an input X is fixed. It is the probability distribution of the input that one can vary. This central development of Shannon is extensively discussed in texts on information theory such as Ash [3] or Cover and Thomas [4] and many others. The general expression of Shannon for the information provided by Y about X
I X ; Y = H X H X | Y
is not limited to information transmission by a channel. H(X) is the uncertainty ( entropy) about the random variable X while H(X|Y) is the remaining uncertainty about X when Y is given.
In this note, I assume that the additional information on X is provided by one or more additional expectation values over the distribution of X. Say that we are given one additional expectation value Fl of an observable that takes the value f l x i when the answer is x i ,
F l = f l = i = 1 n f l x i p i
Before we are given this additional information, the random variable X is characterized by m expectation values, 1 m < n , that we denote as F 1 , F 2 , , F m . We take it that the different vectors of n components f 1 , f 2 , , f m are linearly independent. Otherwise, some of them are redundant and do not provide additional information about the distribution of X. By a similar reasoning, the vector f l , whose expectation value Equation (2) is used to specify the additional information, needs to be linearly independent of the m vectors f 1 to f m . Of course, we need to know that the m expectation values F 1 to F m are compatible, meaning that there is one or more distribution for which these are possible expectation values. This is also for the additional information, meaning the expectation value Fl. Our program, as implemented in the Results section, is to determine the value of the information on X provided by the additional information. I will discuss two views of the result, a differential and an integral value. For either, I take it that the prior information about X, namely, the values of the m expectations F 1 to F m , are kept constant. The differential form is a partial derivative, an infinitesimal change in the entropy of X upon an infinitesimal change in the value of the new information, the value of Fl, when all the other m expectation values are unchanged. The integral form is a change in the entropy due to a finite change in the value of Fl again, at constant value of the prior information. The distinction between a differential and an integral change is of course familiar in many other contexts of the exact sciences perhaps most notably so in quantum chemistry where the expression for differential change in the energy is known as the Hellmann–Feynman theorem [5]. One can also apply the theorem for changes in the dynamics, e.g., [6].
The presentation is organized as follows. Section 2, Methods, provides essential results of the maximum entropy formalism [7,8,9,10,11]. We use the m expectation values F 1 to F m as constraints on the entropy of the distribution of X. Among all distributions over the n different values x i ,   i = 1 , 2 , n that are consistent with the m expectation values F 1 to F m , m < n, we determine the (unique) distribution whose entropy is maximal. We take it that there is a feasible solution, meaning that the values of the m expectation values are such that there is one or more distribution with all probabilities non zero that reproduces these values. The expectation value of Fl is not imposed as a constraint on the distribution. Technically, the constraints are imposed by the Lagrange method of undetermined multipliers. The numerical value of the m Lagrange thus far undetermined multipliers λ 1 to λ m is determined at the last stage by the condition that the distribution reproduces the m expectation values. For the observable that takes the value f l x i on the i’th outcome whose probability is p i 0 , its expectation value as determined by the maximal entropy procedure subject to the m constraints is F l 0 = i = 1 n f l x i p i 0 . Our purpose, as already introduced above, and as discussed in technical detail in Section 3, Results, is to determine the amount of information provided when the expectation value F l is changed from F l 0 . Section 4, The Value of Information as a Potential, provides motivation for the possible implications of the expression of the value of information, specifically for when the new information changes with time, a case of particular relevance for systems not in equilibrium. Also examined is a more formal issue when the additional information depends on a parameter.

2. Methods

To have a compact derivation, we take the variable X to be such that its distribution is uniform when entropy is maximal and no additional information is available. Shannon and then many others have shown that if we take as an axiom that when the distribution is uniform the entropy is maximal, then with a few additional axioms, the entropy of X is H X = i = 1 n p i ln p i . The use of a natural logarithm just determines the units of entropy, nats in this case. In the physical sciences, it is usually the case that there is always some information, e.g., the conservation of energy, and so the distribution at maximal entropy is not uniform. The needed modification is well understood, and so we proceed with the most elementary case as above. As is emphasized early on, e.g., by Tolman [12], this is the case when the index i enumerates individual quantum states.
The distribution when the entropy is at a constrained maximum, i.e., at a maximum where the search for a maximum is subject to given m values F 1 to F m , m < n is [7,8,9,10,13]
p i 0 = exp k = 1 m λ k f k x i
The numerical values of the m Lagrange multipliers λ k are determined by the m implicit equations
F j = i = 1 n f j x i p i 0 = i = 1 n f j x i   exp k = 1 m λ k f k x i ,   j = 1 , 2 , , m
The distribution needs to be inherently normalized, meaning that i = 1 n p i 0 = 1 = 1 . So, either one of the observables is the identity, i.e., f j x i = 1   for   all   i   and   F j = 1 , or some linear combination of observables is the identity. Either way, the normalization is enforced so that 1 / F j = 0 . Since the values of the Lagrange multipliers are determined by the value of the m observables, it follows from i = 1 n p i 0 / F j = 0 and Equation (3) that
k = 1 m ( λ k / F j ) F k = 0
The entropy of the distribution of X is
H X = i = 1 n p i 0 ln p i 0 = k = 1 m λ k F k
Taking the partial derivative of H(X) wrt F j and using the implication of the normalization, we have the basic identity for the value of information
λ j = H X / F j
The value λ j is defined as a partial derivative when all the m-1 observables that are not F j have their value kept constant. The value carried dimension, those of 1 / F j or λ j F j , has the dimension of information, λ j F j = F j   H X / F j . It follows from Equation (7) that the entropy of X is a homogeneous first-order function of the m observables that are used to constrain the distribution at its maximal possible entropy
H X = k = 1 m F k   ( H X / F k )
Equation (8) generalizes a known result in thermodynamics [10,14,15,16]. The more general result, Equation (6), is that the entropy is the weighted sum of the values of the observables that define the state. It is a macroscale analog of the basic microscale definition of the entropy as the weighted sum of the surprisals, ln p i , of the different possible outcomes. We call the value, λ j , a potential because as was just shown, it is the latent ability of the observable conjugate to λ j to change the entropy.

3. Results

So far, we have examined the value of information when the value of one of the observables that defines the state is changed. In Equation (7), it is the change in the value of observable j. Next, we consider the value of the information provided when an observable Fl is changed, and that observable was not used previously to characterize the state. As already noted, such an observable was not used because the distribution we assigned at a given value of m other observables already correctly produces its current value F l 0 = i = 1 n f l x i p i 0 . This is a rather common situation. A well-known example is the Boltzmann distribution at thermal equilibrium. It is normalized and subject to the given (expectation) value of the energy of the different quantum states i. Given the mean, the distribution of quantum states correctly predicts the variance of the energy, which is the specific heat. Indeed, that was a very early success of the then new quantum theory.
Given the observable F l 0 , the value of making an infinitesimal change in its expectation is zero. This is because the distribution of X is at maximal entropy subject to the given expectations of m constraints. These values are to be held constant when we make an infinitesimal change in F l 0 . But linear variations about a stationary point of a function do not change it. A finite change, from F l 0 to Fl, does lead to a new distribution of maximal entropy, and Equation (3) is replaced by
p i = exp k = 1 m λ k f k x i λ l f l x i
We use primes to denote a distribution of maximal entropy subject to all the previous m constraints. F 1 to F m and to the new constraint F l , where in Equation (9) k runs from 1 to m, excluding l. The value of the m expectation values remains unchanged when we go from the distribution p l o , Equation (3), to the distribution p i , Equation (9), but the m Lagrange multipliers can change, and that is why they have a superscript prime,
λ j / F l = 2 H X / F j F l
To characterize the value of the new information, we use the Shannon fundamental result, Equation (1) plus the inequality i = 1 n p i ln p i / p i o 0 , where equality is iff p i = p i o for all i. Using the explicit Expressions (3) and (9) and noting that by construction, the m observables have the same expectation values for p i o   and   p i , we have
i = 1 n p i ln p i / p i o = i = 1 n p i o ln p i o i = 1 n p i ln p i = H X H X | Y = I X ; Y 0
H X is the entropy of the random variable X before the new information is provided, while the entropy H X | Y = i = 1 n p i ln p i is the entropy of X after the additional information Y, that is here the value of F l , is provided. As is to be expected on general grounds, the value of the new information is positive unless p i = p i o , meaning that the new information is not really informative as the distribution of X is unchanged.
Explicitly, the finite value of the new information is, using the expressions (3) and (9),
I X ; Y = i = 1 n p i ln p i / p i o = k = 1 m ( λ k λ k ) F k λ l ( F l F l o )
The differential change in the Lagrange multipliers due to the change in the value F l is provided by Equation (10). It is a vector with m component indexed by j. It is a vector orthogonal to the constraints, as shown in Equation (5).
Any information that is provided adiabatically has no value. In general, this follows from the Shannon definition, Equation (1), of the information provided by Y about X. I X ; Y = 0 when the distribution of X is unchanged when Y is given, H X | Y = H X . On the microscale, that is in terms of the individual outcomes, this is p x i | y j = p x i for all i and j. This shows that our use of the term adiabatic follows the conventional use in thermodynamics and mechanics: the elementary probabilities of the different outcomes do not change upon a change in the macro scale. But how can that be? We take a clue from a differential form of the first law of thermodynamics to write for an infinitesimal change in an observable
δ F k = i = 1 n f k x i δ p i + i = 1 n p i δ f k x i
It follows that if the addition of information is such that δ F k = i = 1 n p i δ f k x i δ f k , then the probabilities of the elementary events are unchanged and there is no value to the information provided. In the case of the first law of thermodynamics F is the energy, the f x i s are the energies of the individual states. Then, 〈 δ f 〉 is the work performed on or by the system, while δ f δ f   is the heat transfer, which is zero in an adiabatic change. Performing pure work on or by a system does not change the value of the information of its state. The transfer of heat does, as was early and clearly noted by Clausius [17].

4. The Value of Information as a Potential

The value of information, the Lagrange multiplier that is conjugate to the observable that is changes, is a potential. We here discuss it as a potential, see also [13,18], and then specifically discuss how the value changes with time in the special but important case of Hamiltonian dynamics. Given that the state of the system is of maximal entropy, the set of m values of the constraints and the set of m Lagrange multipliers that is conjugate are each an equally correct and useful characterization of the state. As discussed by Callen [14] (the first edition is much better in this particular respect), there is a Legendre transform relating the use of the two sets of variables. As clearly discussed by Callen, one can also usefully introduce intermediate characterizations, using some observables and the other variables being the rest of the Lagrange multipliers. The practical advantages of using certain thermodynamic Lagrange multipliers such as temperature or pressure are well recognized. In chemical problems, there are the chemical potentials of the different species. In general, it is the intensive character of the Lagrange multipliers that often makes them more convenient. We use intensive in its canonical thermodynamic terminology: intensive meaning independent of the actual amount as opposed to the extensive character of the mean value of the observables that will double when we double the number of systems.
The change in the value in a process is often a useful measure. Already noted is that the value is unchanged in an adiabatic process. How does the value change in time? The problem in making a definitive answer is that we only have mechanics as an agreed upon dynamical theory of change, and in mechanics, both classical and quantum mechanical, the processes are reversible. There is no dissipation. Realistically, we all recognize that there is dissipation. Still, what can one say about the rate of change in value in mechanics. I will use quantum mechanics and, as a practical point, I assume a Hilbert space of finite dimensions n. So, operators and in particular the density operator are represented by n by n matrices. It will be useful below to use the n2 operators E i j defined as a matrix where all elements are zero except the element in position i,j that is unity. Any n by n matrix can be expressed as a linear combination of these matrices. As a side comment, these operators close a Lie algebra, E i j , E k l = E i l δ j k E k j δ l i , where the square bracket is the commutator and the delta symbol is the usual Kronecker delta.
The most general form of an initial state density matrix is
ρ 0 = exp i , j = 1 n Λ i j 0 E i j
where the multipliers Λ i i are real, while, since E j i = E i j and the density needs to be Hermitian, the off diagonal coefficients must satisfy Λ j i = Λ i j * . The observables F k of Equation (3) and following can be expressed as linear combinations of the E i j s .
In quantum dynamics, the density matrix of an isolated system evolves in time under the action of a unitary evolution operator U t with the initial value U 0 = I . Since U is unitary and using a superscript dagger to denote a Hermitian conjugate, we can write
ρ t = U t exp i , j = 1 n Λ i j 0 E i j U t = exp i , j = 1 n Λ i j 0 U t E i , j U t
An initial density matrix of maximal entropy that is propagated in time remains a density matrix of maximal entropy with a different, time-dependent set of constraints [19]. These constraints can be shown to be time-dependent constants of the motion (see, e.g., [20]).
In a finite, n, dimensional Hilbert space, the time-dependent operators are n by n Hermitian matrices. So, they can all be written as linear combinations of the time-independent matrices E i j with time-dependent coefficients
É i j t U t E i , j U t = k l , = 1 n e i j , k l t E k l
Thereby, the density at time t, Equation (15), can be written as
ρ t = exp i , j = 1 n Λ i j 0 É i j t = exp k , l = 1 n Λ k l t E k l
where the multipliers evolve as
Λ k l t = i , j = 1 n Λ i j 0 e i j , k l t
The multipliers evolve contra-gradient to the observables. This is the most general equation of change for the values of the information provided by the basis observables E k l . The one cardinal assumption is that the dynamics are unitary as used in Equation (14). It is a strong assumption because unitary implies reversible, U t = U t = U 1 t . Even if U t is not unitary, as long as the dynamics keep the system confined to the n dimensional Hilbert space, it is still possible to expand the surprisal matrix, ln ρ t = k , l = 1 n Λ k l t E k l   in the complete basis of n2 observables, with time-dependent coefficients. Thereby, the Λ k l t s remain their significance as the time-changing values of the elementary observables E k l , but the system no longer evolves in a reversible manner.

Funding

This research received no external funding.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Acknowledgments

This paper is based on a plenary lecture at the Workshop on Information, Value, Modeling, and Inference held at the Info-Metrics Institute of the American University, September 2024. I thank the Workshop Co-Chairs Min Chen, Amos Golan, and Esfandiar Maasoumi for the invitation and the participants in the workshop for their comments.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Dunn, J.M.; Golan, A. Information and Information Processing across Disciplines. In Advances in Info-Metrics; Min Chen, J.M., Dunn Golan, A., Ullah, A., Eds.; Oxford University Press: Oxford, UK, 2020. [Google Scholar]
  2. Shannon, C. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  3. Ash, R.A. Information Theory; Dover: Mineola, NY, USA, 1990. [Google Scholar]
  4. Cover, T.M.; Thomas, J.A. Information Thoery; Wiley: New York, NY, USA, 1991. [Google Scholar]
  5. Levine, I.N. Quantum Chemsitry, 7th ed.; Pearsn: Bopston, MA, USA, 2013. [Google Scholar]
  6. Levine, R.D. An extended Hellmann-Feynman theorem. Proc. Roy. Soc. 1966, A294, 467–485. [Google Scholar]
  7. Jaynes, E.T. Information Theory and Statistical Mechanics. Phys. Rev. 1957, 106, 620–630. [Google Scholar] [CrossRef]
  8. Jaynes, E.T. Probability Theory: The Logic of Science; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  9. Levine, R.D.; Tribus, M. (Eds.) The Maximum Entropy Formalism; MIT Press: Cambridge, MA, USA, 1980. [Google Scholar]
  10. Tribus, M. Thermodynamics and Thermostatics: An Introduction to Energy, Information and States of Matter, with Engineering Applications; D. Van Nostrand Company: New York, NY, USA, 1961. [Google Scholar]
  11. Wichmann, E.H. Density Matrices Arising from Incomplete Measurements. J. Math. Phys. 1963, 4, 884–896. [Google Scholar] [CrossRef]
  12. Tolman, R.C. The Principles of Statistical Mechanics; Oxford University Press: London, UK, 1948. [Google Scholar]
  13. Katz, A. Principles of Statistical Mechanics: The Information Theory Approach; Freeman: San Francisco, CA, USA, 1967; p. 188. [Google Scholar]
  14. Callen, H.B. Thermodynamics and an Introduction to Thermostatistics, 2nd ed.; Wiley: New York, NY, USA, 1985. [Google Scholar]
  15. Mayer, J.E.; Mayer, M.G. Statistical Mechanics; Wiley: New York, NY, USA, 1940. [Google Scholar]
  16. McMillan, W.G.; Mayer, J.E. The Statistical Thermodynamics of Multicomponent Systems. J. Chem. Phys. 1945, 13, 276–305. [Google Scholar] [CrossRef]
  17. Clausius, R. Mechanical Theory of Heat; Macmillan: London, UK, 1879. [Google Scholar]
  18. Golan, A. Foundations of Info-Metrics: Modeling, Inference, and Imperfect Information; Oxford University Press: New York, NY, USA, 2017. [Google Scholar]
  19. Alhassid, Y.; Levine, R.D. Connection Between Maximal Entropy and Scattering Theoretic Analyses of Collision Processes. Phys. Rev. A 1978, 18, 89–116. [Google Scholar] [CrossRef]
  20. Levine, R.D. Dynamical Symmetries. J. Phys. Chem. 1985, 89, 2122–2129. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Levine, R.D. On the Differential and the Integral Value of Information. Entropy 2025, 27, 43. https://doi.org/10.3390/e27010043

AMA Style

Levine RD. On the Differential and the Integral Value of Information. Entropy. 2025; 27(1):43. https://doi.org/10.3390/e27010043

Chicago/Turabian Style

Levine, Raphael D. 2025. "On the Differential and the Integral Value of Information" Entropy 27, no. 1: 43. https://doi.org/10.3390/e27010043

APA Style

Levine, R. D. (2025). On the Differential and the Integral Value of Information. Entropy, 27(1), 43. https://doi.org/10.3390/e27010043

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy