Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences
You have accessResearch article

Dynamics of non-stationary processes that follow the maximum of the Rényi entropy principle

    Abstract

    We propose dynamics equations which describe the behaviour of non-stationary processes that follow the maximum Rényi entropy principle. The equations are derived on the basis of the speed-gradient principle originated in the control theory. The maximum of the Rényi entropy principle is analysed for discrete and continuous cases, and both a discrete random variable and probability density function (PDF) are used. We consider mass conservation and energy conservation constraints and demonstrate the uniqueness of the limit distribution and asymptotic convergence of the PDF for both cases. The coincidence of the limit distribution of the proposed equations with the Rényi distribution is examined.

    1. Introduction

    Entropy is actively used by many fields of science including physics, chemistry, computer science, biology, etc. [1]

    Today there are many different types of entropy in use, which often become the centre of discussions in both statistical physics and thermodynamics. The most famous is the Shannon entropy [2]:

    H(X)=iP(xi)logP(xi),1.1
    where X is a discrete random variable with possible values {x1,…,xn} and P is a probability mass function.

    In 1961, Rényi [3] introduced a generalized Shannon entropy as a one-parameter family of entropy

    H(X,β)=11βlog(iPβ(xi)),β1,β>0.1.2
    The Rényi entropy tends to the Shannon entropy when β1.

    Extension of the Rényi entropy for the continuous case can be defined as

    H(X,β)=11βlog(pβ(x)dx),β1,β>0,1.3
    where X is an absolutely continuous random variable with probability density function (PDF) p.

    The Rényi entropy is often taken as a typical measure of complexity to describe dynamical systems in physics, engineering and information theory [4]. The Rényi entropy applications are overviewed in [5].

    A variety of physical systems obey the famous maximum entropy (MaxEnt) principle: their entropy achieves maximum under constraints caused by other physical laws. Since 1957, when the seminal works of Jaynes were published [6]–[8], and until now [9]–[11], the MaxEnt principle has caused lively interest among researchers. For example, the MaxEnt for the Rényi entropy has been successfully applied to generalize the Thomas–Fermi model in [12].

    Although the states of maximum entropy are widely discussed in scientific articles and studies, the dynamics of evolution and transient behaviour of systems are still not well investigated.

    In this paper, we propose a set of equations that describe the dynamics of the PDFs for non-stationary processes that follow the maximum of the Rényi entropy principle.

    The speed-gradient (SG) principle [13]–[16] used here is originated from control theory. This principle has already been applied in [13],[15],[17] to derive the equations of dynamics for systems with a finite number of particles for the case of the maximum Shannon entropy principle. Systems with continuous probability distributions are considered in [18]. The dynamics of discrete systems for the Tsallis entropy is discussed in [19]. The SG principle for systems with a finite number of particles is studied in [15], where systems are emulated with the molecular dynamics method. We take a similar approach for systems with discrete and continuous probability distributions while researching the Rényi entropy. The derived equations describe the dynamics of non-stationary (transient) states and show the way and trajectory of a system tending to the state with maximum Rényi entropy.

    The well-known Fokker–Planck (FP) equations [20] describe the time evolution of the PDF. Jaynes's MaxEnt approach can also be applied to these equations [21],[22]. Another general form of time-evolution equations for non-equilibrium systems is known as GENERIC (general equation for the non-equilibrium reversible–irreversible coupling) [23],[24]. The relation between GENERIC and FP equations is established in [23]. It states that FP equations are a particular case of the GENERIC when a noise term is added into the GENERIC. Thus, if the FP equation is represented as a stochastic differential equation, from which fluctuations are eliminated, this equation matches the GENERIC equation [25].

    Following this, we can also claim that the SG principle matches the GENERIC (and thus it matches the FP equation) if a goal function is set as entropy and constraints are specified by energy [19]. Moreover, the SG principle is a more general case of the GENERIC because almost every smooth function can be taken as a goal function, not only entropy. The GENERIC and the SG principle relations are examined in §3.

    We propose an evolution law of the system in the following form:

    p˙=Γ(IΨ)pβ1,1.4
    where I is an identity operator, Ψ is a linear integral operator that does not depend on p and Γ>0 is a constant gain. All the solutions of equation (1.4) converge to the distribution that corresponds to the maximum value of the Rényi entropy.

    A way to find the distribution achieving the extreme value (maximum or minimum) of entropy within a given variational distance from any given distribution is proposed in [26]. An approximation of the probability distribution is built there based on new bounds of entropy distance. These bounds are also applied for the entropy estimation in [27].

    Jaynes's MaxEnt principle was successfully used for inductive inference in [28]. It is shown that, given information in the form of constraints on expected values, there is only one appropriate distribution which can be obtained by maximizing entropy. Inductive inference can be used in systems with non-stationary processes which satisfy the MaxEnt [29]. In [29], the notion of entropy dynamics (ED) is used. ED is a theoretical framework which combines inductive inference [28] and information geometry [30]. ED investigates the possibility of deriving dynamics from purely entropic arguments.

    Besides the Rényi entropy, more general forms of relative entropies and divergences can also be studied from a perspective of the SG principle, for example the Cressie–Read and the Csiszár–Morimoto conditional entropies (f-divergences) [31].

    This paper has the following structure. The next section describes the SG principle, accompanied with two examples. The third section examines the relations between the SG principle, the GENERIC and the FP equations. The fourth section introduces Jaynes's formalism; the Rényi distribution (RD) is derived from the maximum of the Rényi entropy principle. The fifth section gives an example of a dynamic system with discrete distribution of parameters that follows the maximum of the Rényi entropy principle (we consider cases with one and two constraints and derive equations for the transient states). The sixth section extends the results obtained in the fifth section for the case of the continuous PDF. The asymptotic stability of PDF dynamics is proved for the cases with one and two constraints.

    2. The speed-gradient principle

    There is a connection between the laws of control in technical systems and the laws of dynamics in physical systems. It is known that the methods for the synthesis of control algorithms allow one to derive the laws of dynamics for physical systems. In particular, the model of the dynamics for a number of physical systems can be derived based on the SG method with an appropriate choice of goal function.

    Consider the class of open physical systems whose dynamics can be described by the system of differential equations

    x˙=f(x,u,t),2.1
    where xCn is the system state vector and u is the vector of input (free) variables t≥0. The problem of modelling of the system can be formulated as finding the law of change (evolution) u(t) which meets a certain criterion of ‘naturalness’ for its behaviour and grants a set of properties observed in real physical systems to a generated model.

    Such formulations are well known in physics. Variational principles of systems models have long been recognized. They usually involve the task of an integral functional that characterizes the behaviour of the system [32]. Minimization of the functional defines the real possible trajectories of the system {x(t),u(t)} as points in the corresponding functional space. To explicitly specify the dynamics of the system a developed apparatus of the variations calculus is used.

    Methods of optimal control (e.g. Bellman dynamic programming, Pontryagin maximum principle, etc.) are the result of the development of classical variational calculus methods. And it can be used to build dynamic models of mechanical systems in Nature and society.

    Together with integral principles, differential local time principles have also been proposed, such as the Gauss principle of least constraint, the principle of minimum energy dissipation, etc. As noted by Planck [33], local principles have some advantages over integral ones, because they do not make the current state and the movement of the system dependent on later states and movements. Let us formulate another local variational principle based on the method of SG [14],[17], as follows.

    The speed-gradient principle.Only those possible movements of the system are realized (among all possible movements) for which the input variables change proportionally to the SG of a ‘goal’ functional.

    The SG principle offers researchers the choice of two types of systems dynamics models:

    • (A) models which follow the algorithm of the SG in differential form:

      u˙=ΓuQt˙;2.2

    • (B) models following the algorithm of the SG in finite form:

      u=ΓuQt˙,2.3

    where Qt˙ is the rate of change of the target functional along the trajectories of the system (2.1). We describe the application of the SG principle in the simplest (and most important) case when the class of models of dynamics is given by the relation:

    x˙=u.2.4
    Relation (2.4) means only that we are looking for a law of variation rates of the state variables of the system.

    In accordance with the SG principle, the goal functional Q(x) has to be determined first. Selection of Q(x) should be based on the physics of the real system and reflect the presence of a tendency to decrease the current value of Q(x(t)). After that, the law of dynamics can be written in the form (2.2) or (2.3).

    Let us illustrate the introduced SG principle with several examples.

    (a) Example 1: the motion of a particle in a potential field

    As a first example, consider the problem of describing the motion of a particle in a potential field. State variables here are the coordinates of the point x1,x2,x3 which form a vector x=(x1,x2,x3)T.

    We choose a smooth ‘goal’ functional Q(x) as the potential energy of a particle and derive the SG law in the differential form. We calculate the speed

    Q˙=tQ=[xQ(x)]Tu2.5
    and the SG
    uQ˙=xQ(x).2.6
    Then, choosing the diagonal positive definite gain matrix Γ=m−1I3, where m>0 is a parameter and I3 is the 3×3 identity matrix, we arrive at Newton's law u˙=m1xQ(x) or
    mx¨=xQ(x),
    where m is interpreted as the mass of a particle.

    Note that the SG laws with non-diagonal gain matrices Γ can be incorporated if a non-Euclidean metric in the space of inputs is introduced by the matrix Γ−1. Admitting dependence of the metric matrix Γ on x, one can obtain evolution laws for complex mechanical systems described by Lagrangian or Hamiltonian formalism.

    The SG principle applies not only to finite-dimensional systems but also to infinite-dimensional (distributed) ones. Particularly, x may be a vector of a functional space X and f(x,u,t) may be a nonlinear differential operator (in such a case, the solutions of (2.1) should be understood as generalized ones). We will omit the mathematical details for simplicity.

    (b) Example 2: the viscous fluid dynamics

    Let the infinite-dimensional state vector be formed of two functions: x=(v(⋅,t),p(⋅,t))T, where v(r,t)R3 is the velocity field of the three-dimensional fluid flow and p(r,t) is the pressure field.

    We introduce the ‘goal’ functional in the following way:

    Qt=Ωp(r,t)dr+v0Ω|rv(r,t)|2dr,2.7
    where v0>0 is a weight coefficient. Calculating functional (2.7) relative to (2.4) gives uQ˙t=rpv0Δv. The differential form of the SG principle is the Navier–Stokes equation, which describes the motion of a viscous fluid:
    ρvt(r,t)=rp(r,t)+v0Δv(r,t),
    where v=v0γ−1 is the viscosity factor and ρ=γ−1 is the density of the liquid.

    Note that the differential form of the SG laws often corresponds to reversible processes, while the finite form generates irreversible ones. For modelling of more complex dynamics, a combination of finite and differential SG laws may be useful.

    In a similar way, dynamical equations for many other mechanical, electrical and thermodynamic systems can be recovered. The SG principle applies to a broad class of physical systems subjected to potential and/or dissipative forces.

    This paper is aimed at application of the SG principle to entropy-driven systems.

    3. GENERIC and the speed-gradient principle

    The GENERIC time-evolution equation is formulated as [23]:

    dxdt=L(x)E(x)x+M(x)S(x)x,3.1
    where x is a state of a system, E is a total energy functional and S is an entropy functional.

    Let us show the relation between the GENERIC and SG equations. Assume that we have to maximize the entropy function S(x) of a system that has an additional constraint for a total energy E(x)=E=const. The Lagrangian for this case can be defined as:

    Λ(x,λ)=S(x)+λ(E(x)E),3.2
    where λ is a real Lagrange multiplier.

    Let us use the SG principle equation (2.3) for the Lagrangian (3.2), i.e. Qt=Λ,

    u=ΓuΛ˙(x,λ)=Γu[S˙+λE˙].3.3

    According to (2.5) and (2.6) we can write equation (3.3) as:

    u=Γ[xS(x)+λxE(x)].3.4

    We can see that the dynamics equation obtained from the SG principle (3.4) coincides with the GENERIC equation (3.1) for dx/dt=u, L(x)=−Γλ and M(x)=−Γ.

    GENERIC is based on two ‘potentials’ of total energy and entropy. The SG principle can use any smooth functional that has to be maximized (minimized) as a goal function. So it is not necessary to be only Lagrangian (3.2) or entropy functional. The SG principle can be treated as a more general approach to describe the dynamics of a system. Nevertheless, GENERIC is also a general equation. It uses parametrized matrices L(x) and M(x) that make it possible to use GENERIC for a wide range of time-evolution systems.

    The relation between GENERIC and FP equations is established in [23], which means that there is also a relation between the FP equations and the SG principle. So many different methods allow the researcher to select the one that is most convenient or more fully describes the problem being solved.

    4. Maximum entropy principle

    (a) Jaynes's MaxEnt

    Jaynes [6]–[8] proposed the approach which became one of the foundations for modern-day statistical physics.

    Let p(x) be an unknown PDF of a multi-dimensional random variable x. Suppose that we have to define it on the basis of certain system information. Consider a continuous case and suppose that there is a priori known information about some average values H¯m,

    H¯m=Hm(x)p(x)dx,m=1,,M.4.1

    Conditions (4.1) can be insufficient to derive p(x) in general. According to Jaynes, applying maximization of information entropy H(X) is the most objective method to define the PDF in this case.

    Lagrange multipliers are used to perform the maximum entropy search with additional conditions (4.1). This leads to

    p(x)=1Zexp(m=1MλmHm(x)),4.2
    where Z=exp(m=1MλmHm(x))dx and λm can be derived from conditions (4.1).

    These formulae show the match between the maximum information entropy and the Gibbs entropy in the case of equilibrium. So the information entropy can be identified with the thermodynamic entropy in this case.

    As the Rényi entropy is a generalization of the Shannon entropy, it seems natural to extend Jaynes's maximum entropy principle to the case of the Rényi entropy.

    (b) The Rényi distribution

    The RD that corresponds to the state with maximum Rényi entropy is proposed in [34],[35]. The RD is considered there for a system with a discrete probability distribution when two constraints are imposed: normalization constraint (mass conservation law) ipi=1 and a fixed average energy (total energy) constraint iHipi=E.

    If the Rényi entropy is used instead of the Shannon entropy in MaxEnt, then the equilibrium distribution (the RD) can be formulated as

    pi=1ZR(1λβ1β(HiE))1/(β1),4.3
    where ZR=i(1λ((β1)/β)(HiE))1/(β1) and λ is a Lagrange multiplier which can be derived from the total energy constraint.

    Following the constraints in form (4.1), the expression (4.3) for continuous distributions becomes

    p(x)=1ZR(1λβ1β(h(x)E))1/(β1),4.4
    where ZR=(1λ((β1)/β)(h(x)E))1/(β1)dx and λ is a Lagrange multiplier.

    5. The speed-gradient dynamics of the Rényi entropy maximization process

    We extend the approach introduced in [13] to the case of the Rényi entropy. Similar to [13] for the Shannon entropy, we consider a discrete system which consists of N identical particles distributed over m cells.

    In the case when the mass conservation constraint holds, it is true that i=1mNi=N. It can be normalized as

    i=1mNiN=1.5.1

    Particles can move from one cell to another. The steady-state and the transient behaviour of the system are both interesting for us. According to the MaxEnt principle, the limit behaviour of the system maximizes its entropy for the steady state when nothing else is known [7],[8].

    To obtain a transient mode behaviour, we apply the SG principle, choosing the Rényi entropy as the goal function to be maximized

    H(X,β)=11βlog(i=1mNiβ),5.2
    where β≥0, β≠1 and X=(N1,…,Nm)T is the state vector of the system.

    Assume that the motion is continuous in time and the numbers Ni are changing continuously. Then the law of motion can be represented as N˙i=ui,i=1,,m, where ui=ui(t) are control functions which have to be determined.

    The evaluation scheme is as follows. First the speed of entropy change is evaluated:

    H˙(β)=β1βi=1mNiβ1Ni˙i=1mNiβ.

    Then the gradient of the speed is evaluated with respect to the vector of controls ui considered as frozen parameters,

    uiH˙(β)=β1βNiβ1i=1mNiβ.

    And finally the actual controls are defined proportional to the projection of the SG to the surface of constraints (5.1),

    Ni˙=Γβ1βNiβ1i=1mNiβ+λ,
    where λ′ is a Lagrange multiplier chosen in order to fulfil the constraint (5.1).

    Now we can evaluate λ′:

    i=1mNi˙=0λ=Γmβ1βi=1mNiβ1i=1mNiβ.

    The final form of the system dynamics law is as follows:

    Ni˙=Γβ1βNiβ1i=1mNiβ+Γmβ1βi=1mNiβ1i=1mNiβ.5.3

    Let us find the equilibrium mode which corresponds to the asymptotic behaviour of the variables Ni. In this mode Ni˙=0. Based on (5.3), this means that mNiβ1=i=1mNiβ1. Hence all Ni are equal. According to constraint (5.1), we have that Ni=N/m. This result corresponds to RD (4.3) for the case of one constraint (i.e. λ=0 as there is no total energy constraint). It also corresponds to the maximum state of classical entropy and agrees with thermodynamics.

    (a) Equilibrium stability

    Let us examine the stability of the equilibrium mode. We introduce the Lyapunov function

    V(X,β)=Hmax(β)H(X,β).

    Evaluation of V˙ yields

    V˙(β)=H˙(X,β)=β1βi=1mNiβ1uii=1mNiβ=Γ(β1β)21(i=1mNiβ)2((i=1mNiβ1)2mi=1m(Niβ1)2).

    Consider the Cauchy–Bunyakovsky–Schwarz (CBS) inequality for two vectors a,bRm:

    (i=1maibi)2i=1mai2i=1mbi2.5.4

    According to the CBS inequality for vectors a=(N1β1,,Nmβ1) and b=(1,…,1), we have that V˙(β)0. Equality V˙(β)=0 holds if and only if all values Ni are equal. This is the maximum entropy state. Thus law (5.3) provides global asymptotic stability of the maximum entropy state. The physical meaning of this law is nothing but moving along the direction of the maximum entropy production rate (direction of the fastest entropy growth).

    (b) Total energy constraint

    The case of more than one constraint can be treated in the same way. Suppose that in addition to the mass conservation law (5.1) the energy conservation law also holds. Let Ei be the energy of the particle in the ith cell and the total energy does not change. The total energy constraint is

    E=i=1mNiNEi.5.5

    A new set of constraints can be formulated as

    i=1mui=0andi=1muiEi=0.5.6

    Then the evolution law should have the form

    ui=Γβ1βNiβ1i=1mNiβ+λ1Ei+λ2,5.7
    where λ1 and λ2 are Lagrange multipliers that can be defined by substitution of (5.7) into the set of constraints (5.6).

    The solutions for λ1 and λ2 are given by formulae

    λ1=Γβ1β1i=1mNiβ(i=1mNiβ1i=1mEimi=1mNiβ1Ei)mi=1mEi2(i=1mEi)2andλ2=Γβ1β1i=1mNiβ(i=1mNiβ1Eii=1mEii=1mNiβ1i=1mEi2)mi=1mEi2(i=1mEi)2.}5.8
    This solution is defined for (mi=1mEi2(i=1mEi)2)0. It holds for all cases except a degenerate case when all Ei are equal.

    The general form of the evolution law can be obtained by substitution of λ1 and λ2 from (5.8) into equation (5.7). In abbreviated form, we represent this law as

    N˙=Γβ1β(IA)Nβ1,5.9
    where N=(N1,…,Nm)T, I is an identity matrix and A is a symmetric m×m matrix defined as follows:
    aij=1m+E~iE~jE¯2(1/m)(1E¯)2,
    where E~i=Ei+(1/m)i=1mEi,1=(1,,1)T and E¯=(E1,,Em)T is a vector of energies.

    As before it can be shown that V˙(X,β)=Hmax(β)H(X,β) is a Lyapunov function and there is only one stable equilibrium state of the system in non-degenerate cases. We demonstrate this as:

    V˙(β)=H˙(X,β)=β1βi=1mNiβ1uii=1mNiβ.5.10
    We substitute λ1 and λ2 from (5.8) into equation (5.7) and then we substitute the expression for ui into (5.10). The resulting expression is
    V˙(β)=Γβ2(1β)21(i=1mNiβ)2×((mi=1mEiNiβ1i=1mEii=1mNiβ1)2mi=1mEi2(i=1mEi)2(mi=1m(Niβ1)2(i=1mNiβ1)2)).5.11

    We introduce a new scalar product function for two vectors as

    f,g=mimfigiimfiimgi.5.12

    For scalar product (5.12), the CBS inequality is true:

    f,g2f,fg,g.5.13
    An equality in (5.13) takes place when λ,μR:f=λg+μ.

    Using inequality (5.13) for vectors f=(N1β1,,Nmβ1)T and g=(E1,…,Em)T, we get for (5.11) that V˙(β)0. And V˙(β)=0 occurs only for the case when λ,μR:Niβ1=λEi+μ for all i. Due to (5.7) at the equilibrium state of the system, the following equalities hold:

    Γβ1βNiβ1i=1mNiβ+λ1Ei+λ2=0.5.14
    This means that, for the equilibrium state, the final distribution of points satisfies
    Niβ1=λEi+μ,
    where λ=λ1((1β)/Γβ)i=1mNiβ, μ=λ2((1β)/Γβ)i=1mNiβ for λ1 and λ2 defined in (5.8).

    (i) Final distribution and the Rényi distribution equivalence

    Let us show that the distribution in (5.14) is the RD (4.3). If we multiply (5.14) by Ni and sum up over i, we obtain

    Γβ1β+λ1NE+λ2N=0λ2=Γβ1βλ1E,Γ=ΓN.5.15
    From (5.14), we get that
    Ni=((λ1Ei+λ2)1βΓβi=1mNiβ)1/(β1).5.16

    Let us substitute Ni from (5.16) into (5.1). We get

    1N(1βΓβi=1mNiβ)1/(β1)i=1m(λ1Ei+λ2)1/(β1)=1(1βΓβi=1mNiβ)1/(β1)=Ni=1m(λ1Ei+λ2)1/(β1).5.17
    After substitution of (5.17) into (5.16), we get that
    NiN=(λ1Ei+λ2)1/(β1)i=1m(λ1Ei+λ2)1/(β1).5.18
    Then we substitute the expression for λ2 in (5.15) into (5.18) and obtain
    NiN=(1λ((β1)/β)(EiE))1/(β1)i=1m(1λ((β1)/β)(EiE))1/(β1),5.19
    where λ=λ1/Γ′. We can see that (5.5) coincides with the RD (5.19).

    It is evident that (5.19) satisfies the normalization constraint (5.1). Let us check that the second constraint for energy (5.5) is also satisfied.

    Let us substitute Ni from (5.16) into (5.5). Then we get

    1N(1βΓβi=1mNiβ)1/(β1)i=1m(λ1Ei+λ2)1/(β1)Ei=E(1βΓβi=1mNiβ)1/(β1)=NEi=1m(λ1Ei+λ2)1/(β1)Ei.5.20
    After substituting (5.20) into (5.16), we get that
    Ni=((λ1Ei+λ2)1/(β1))(NEi=1m(λ1Ei+λ2)1/(β1)Ei)EiNi=(λ1Ei+λ2)1/(β1)NEi=1m(λ1Ei+λ2)1/(β1)i=1mNiEi=NE,5.21
    which means that energy constraint (5.5) is true for (5.19).

    6. System with continuous probability density function

    Let us extend the same approach based on the SG principle for the case of continuous PDFs. Consider a system with a continuous distribution of possible states that evolves on a compact carrier. Distribution over states is characterized by PDF p(t,x), which is continuous everywhere except for a set with zero measure. It is true that

    Ωp(t,x)dx=1,t,6.1
    where Ω is a compact carrier.

    The Rényi entropy for a continuous PDF is defined as

    H(X,β)=11βlogΩpβ(t,x)dx,6.2
    where β≥0,β≠1.

    From constraint (6.1), it follows that

    Ωu(t,x)dx=0,6.3
    where u(t,x)=p˙(t,x).

    According to the SG principle, we calculate H˙:

    H˙(X,β)=β1βΩp˙(t,x)pβ1(t,x)dxΩpβ(t,x)dx.

    The gradient of H˙ by u is equal to

    uH˙(X,β)=β1βupβ1(t,x)Ωpβ(t,x)dx,u=β1βpβ1(t,x)Ωpβ(t,x)dx.6.4

    The SG principle of motion forms the evolution law:

    u=Γβ1βpβ1(t,x)Ωpβ(t,x)dx+λ,
    where Γ can be taken as a scalar value and the Lagrange multiplier λ′ is selected to satisfy the constraint (6.3),
    Ω(Γβ1βpβ1(t,x)Ωpβ(t,x)dx+λ)dx=0λ=βΓ(1β)mes(Ω)Ωpβ1(t,x)dxΩpβ(t,x)dx,6.5
    where mes(Ω)=Ω1dΩ.

    The final system dynamics equation has the following form:

    p˙=Γβ1βpβ1(t,x)Ωpβ(t,x)dx+βΓ(1β)mes(Ω)Ωpβ1(t,x)dxΩpβ(t,x)dx.6.6

    Equation (6.6) can be represented in the more general form

    p˙=ΓC(IΨ)pβ1,6.7
    where Ψ=Ω()dx/mes(Ω) is a linear operator which is invariant for p, I is an identity operator and C=(1β)Ωpβ(t,x)dx/β.

    (a) Equilibrium stability

    Let us investigate a stability of the obtained equilibrium equation (6.6). Consider function V(p)=HmaxH(β)0. The derivative of this function is

    V˙(p)=H˙(β)=β1βΩu(t,x)pβ1(t,x)dxΩpβ(t,x)dx.6.8

    After substitution of the expression for u from (6.6) to (6.8), we obtain:

    V˙(p)=Γβ2mes(Ω)(β1)2(Ωpβ(t,x)dx)2((Ωpβ1(t,x)dx)2mes(Ω)Ω(pβ1(t,x))2dx).6.9

    Then we use the CBS inequality in integral form

    |Ωf(x)g(x)dx|2(Ω|f(x)|2)(Ω|g(x)|2)6.10
    for functions f=pβ−1 and g=1. Taking into account that a scalar value Γ is positive, we get that V˙(p)0. It is known that equality in the CBS inequality is achieved when multiplicity occurs, i.e. f(x)=αg(x). In our case, V˙(p)=0 is true when pβ−1=α. This is possible only when p(t,x)=C=const. Using the constraint (6.1), we get that C=mes−1(Ω). This means that there is only one unique limit PDF p*=mes−1(Ω) for the equilibrium state of the system which evolves by evolution law (6.6). Also note that p* corresponds to RD (4.4) for a case with one constraint.

    (b) Asymptotic convergence

    To show an asymptotic convergence of all solutions to p*, we will use Barbalat's lemma [36].

    Lemma 6.1 (Barbalat's lemma)

    If differentiable functionf(t) has a finite limit fortand its derivativef˙(t)is uniformly continuous thenf˙(t)0fort0.

    Theorem 6.2

    For all PDFs defined by equation (6.6), it is true thatp(t,x)pfort.

    Proof.

    For the sake of simplicity, we define a notation for V in (6.9) as v(t)=V (p(t)). We will use Barbalat's lemma to show that v˙(t)0. We use v(t) as a function f(t) in Barbalat's lemma. Because of v(t)≥0 and v˙0, the function v(t) has a finite limit for t.

    It can be shown that v˙ is uniformly continuous. Consider the expression for |v¨(t)|,

    v¨(t)=ββ1Ω(u˙(t,x)pβ1(t,x)+(β1)u2(t,x)pβ2(t,x))dxΩpβdxβ2β1Ωu(t,x)pβ1(t,x)dxΩpβ1(t,x)u(t,x)dx(Ωpβ(t,x)dx)2.6.11

    Due to the constraint (6.1) and compactness of the carrier Ω, it can be shown that function |v¨(t)| is bounded. This leads us to the fact that v˙ is uniformly continuous.

    As all the necessary conditions of Barbalat's lemma for differentiable function v(t) are satisfied, we have that v˙(t)0 for t0.

    Taking into account that Ωfdx=(1,f) and Ωf2dx=f2, the expression for v˙ from (6.9) may be rewritten as

    v˙=Γβ2mes(Ω)(β1)21(pβ,1)2(1(1,pβ1)2mes(Ω)pβ12)=Γβ2mes(Ω)(β1)2pβ12(pβ,1)2(1cos2(α(t))),6.12
    where α(t)=cos(1,pβ1).

    If pβ120, then mest{r:pt(r)0}0. But this case conflicts with the constraint (6.1). Given (6.2) we obtain that α(t)0. This means that p^β11^, where p^β1 and 1^ are normalized values for pβ−1 and 1, respectively.

    It follows that p(t,x) tends to the stationary distribution. As explained earlier, this distribution is unique. Thus, ptp for t. ▪

    (c) Total energy constraint

    The constraint (6.1) can be interpreted as the mass conservation law on the space Ω. Consider a system with an additional constraint for the total energy conservation, i.e. a conservative case when energy does not depend on time. The new constraint may be described as

    Ωp(t,x)h(x)dx=E,6.13
    where E is the common energy of a system and h(x) is the density of energy.

    The equation for the dynamics can be defined in the form

    u=ΓuH˙(β)+λ1h+λ2.6.14

    Based on constraints (6.1) and (6.13), we can find expressions for Lagrange multipliers λ1 and λ2:

    λ1=Γβ(1β)Ωpβ(t,x)dxmes(Ω)Ωpβ1(t,x)h(x)dxΩpβ1(t,x)dxΩh(x)dxmes(Ω)Ωh(x)2dx(Ωh(x)dx)2andλ2=Γβ(1β)Ωpβ(t,x)dxΩpβ1(t,x)dxΩh2(x)dxΩpβ1(t,x)h(x)dxΩh(x)dxmes(Ω)Ωh(x)2dx(Ωh(x)dx)2.}6.15

    The above equations are valid when the denominator in both fractions is not equal to zero. If we use the CBS inequality for f=h and g=1, then the following inequality becomes true: |Ωhdx|2mes(Ω)Ωh2dx. This inequality becomes equality when h=const. This means that all energy levels coincide. This case is supposed to be degenerate and not considered here. Thus the expression mes(Ω)Ωh2dx(Ωhdx)2 is always true.

    The resulting equation for the dynamics can be obtained by substituting (6.4) and (6.15) into (6.14). It can be transformed to the brief form:

    p˙=ΓC(IΨ)pβ1,6.16
    where I is an identity operator and Ψ is a linear integral operator that is independent of p:
    Ψ=(1,)mes(Ω)+h~(h~,)h21mes(Ω)(1,h)2,6.17
    where h~=h(1/mes(Ω))Ωhdx,C=(1β)Ωpβ(t,x)dx/β.

    (i) Equilibrium stability

    Let us examine the equilibrium of the obtained equation (6.14). We use the same Lyapunov function V˙ as we have used in the previous section with only one constraint. For two constraints, the new expression for V˙ is

    V˙(p)=Γβ2mes(Ω)(β1)2(Ωpβ(t,x)dx)2(mes(Ω)Ωhpβ1dxΩhdxΩ(pβ1)2dx)2mes(Ω)Ωh2dx(Ωhdx)2Γβ2mes(Ω)(β1)2(Ωpβ(t,x)dx)2(mes(Ω)Ω(pβ1(t,x))2dx(Ωpβ1(t,x)dx)2).6.18
    We will prove that
    V˙(p)0.6.19

    Let us define a functional: ,:L2(Ω)×L2(Ω)R,f,gL2(Ω)

    f,g=mes(Ω)ΩfgdxΩfdxΩgdx.6.20

    The new functional has several useful properties (the proof of each property is provided in appendix A):

    • (1) linearity for the first argument f,g,hL2(Ω),λRλf+g,h=λf,h+g,h;

    • (2) symmetry ∀f,gL2(Ω)〈f,g〉=〈g,f〉;

    • (3) positiveness and the condition of zero value ∀fL2(Ω)〈f,f〉≥0, 〈f,f〉=0⇔f=μ=const.

    Let us prove inequality (6.19) based on properties 1–3.

    It is obvious that for any f,gL2(Ω) and λR it is true that f−λgL2(Ω). This function has property 3: 〈f−λg,f−λg〉≥0. Using properties 1 and 2, we obtain the quadratic inequality with respect to λ:

    0fλg,fλg=λ2g,g2λf,g+f,f.

    This inequality holds for any real λ. Hence the discriminant cannot be positive,

    D=4f,g24f,fg,g.
    Thus
    f,g2f,fg,g.6.21

    If the equality takes place in (6.21), then there exists a unique solution λR of equation 〈f−λg,f−λg〉=0. But then, by property 3, we have μR:fλg=μ1.

    Substituting f=pβ−1,g=h into inequality (6.21), we get

    (mes(Ω)Ωhpβ1dxΩhdxΩpβ1dx)2(mes(Ω)Ω(pβ1)2pdx(Ωpβ1dx)2)(mes(Ω)Ωh2pdx(Ωhdx)2),
    which implies inequality (6.19).

    Note that equality (6.19) holds if and only if

    λ,μR:pβ1=λh+μ.6.22

    According to the SG for H (6.4), expression (6.14) can be rewritten for the case of equilibrium as

    p(t,x)β1=Cλ1h(x)Γ+Cλ2Γ,6.23
    where C=(1β)Ωpβ(t,x)dx/β. This coincides with (6.22) for λ=λ1(C/Γ), μ2(C/Γ), where λ1 and λ2 are defined in (6.15). So this is the only PDF p*(β) for an equilibrium state.

    It can also be shown that p*(β) corresponds to the continuous form of the RD (4.4). To do this, we write expression (6.23) as

    p(t,x)=((λ1h(x)+λ2)1Γ(1β)βΩpβ(t,x)dx)1/(β1).6.24

    Following a similar approach to that used for a discrete case, expression (6.24) can be formulated as RD:

    p(t,x)=(1λ((β1)/β)(h(x)E))1/(β1)Ω(1λ((β1)/β)(h(x)E))1/(β1)dx,
    where λ=λ1/Γ.

    (ii) Asymptotic convergence

    We will prove asymptotic convergence similarly to the case with one constraint (see proposition 1).

    Theorem 6.3

    For all PDFs defined by equation (6.14), it is true thatp(t,x)pfort.

    Proof.

    To use Barbalat's lemma, we have to check the conditions under which the function v¨ is bounded. Based on the expression for v˙ in (6.18) and according to similar logic so that used in proposition 1, we can conclude that v¨ is bounded for the compact carrier Ω. According to Barbalat's lemma, it is true that

    v˙0,t.6.25

    We introduce a scalar product as f,f=mes(Ω)Ωf2(x)dx(Ωf(x)dx)2. Having ∥f2=〈f,f〉, expression (6.18) can be rewritten as

    v˙=Γβ2mes(Ω)(β1)2(Ωpβ(t,x)dx)2(pβ12(pβ1,h)2h2)=Γβ2pβ12mes(Ω)(β1)2(Ωpβ(t,x)dx)2(1(pβ1,h)2pβ12h2).6.26

    As v˙0, consider the case when pβ120 first. In the CBS inequality

    (Ωpβ1dx)2mes(Ω)Ω(pβ1)2pdx,
    the equality takes place only when pβ−1=α. We have previously demonstrated in (6.23) that equality v˙=0 holds only in the case when pβ−1h+μ, where λ and μ are constants. Then h must be a constant since the equality λh+μ=α. We assume this to be a degenerate case.

    Given (6.25), (6.26) and pβ120, we have that (pβ1,h)2/pβ12h21. This implies that p^β1h^, where p^β1 and h^ are normalized values for pβ−1 and h, respectively. Thus p tends to the only stationary distribution p*(β) since h does not depend on time. ▪

    7. Conclusion

    The Rényi entropy is widely used in communication and coding theory, quantum information theory, signal processing, data mining and many other areas [5],[4]. Stationary states which maximize Rényi entropy have already been well investigated. The MaxEnt principle defines the asymptotic behaviour of the system, but does not answer the question about how the system moves to this asymptotic behaviour.

    We have investigated non-stationary states of processes that follow the MaxEnt principle for the Rényi entropy. We have derived equations (5.3), (5.9), (6.6) and (6.16) which describe the dynamics of the PDF for the system that tends to the state with maximum Rényi entropy. Systems with discrete probability distribution and continuous PDFs were considered under mass conservation and energy conservation constraints. We have shown that the limit PDF p*(β) is unique and corresponds to the RD. We have also proved the convergence of PDFs with dynamics described by equations (5.3), (5.9), (6.6) and (6.16) to PDF p*(β), which corresponds to the state with the maximum value of the Rényi entropy.

    The key point of our approach is to use the SG method with the goal function chosen as the Rényi entropy of the process. The SG principle originates from control theory and it generates equations for the transient (non-stationary) states of the system's operation which help to track how the system evolves to the steady state.

    There are many other generalizations of the Shannon entropy, e.g. the Tsallis entropy [37], the Khinchin entropy [38], the Burbea–Rao entropy [39], the Cressie–Read family [40], the Burg entropy [41], f-divergences [31], etc. Investigation of these entropies based on the SG principle seems to be promising for further investigations. A comparative analysis of the dynamics for different kinds of entropy may be performed.

    Data accessibility

    All methods and corresponding data are given in the manuscript and are fully reproducible.

    Authors' contributions

    A.L.F. designed the study to use the SG principle for non-stationary processes which tend to maximize its entropy. D.S.S. derived the equations for the Rényi entropy and prepared the manuscript. Both authors gave final approval for publication.

    Competing interests

    We have no competing interests.

    Funding

    This work was supported by SPbU grant nos. 6.37.181.2014 and 6.38.230.2015. Formulation of the SG principle for the Rényi entropy (§5) was performed in IPME RAS and supported by the RSF (grant no. 14-29-00142).

    Appendix A. Extra materials

    Here, we prove the properties of functional 〈f,g〉 (6.20) from §5c(i).

    (1) Linearity in the first argument f,g,hL2(Ω),λRλf+g,h=λf,h+g,h.

    Proof.

    Using the linearity of the integral, we obtain

    λf+g,h=mes(Ω)Ω(λf+g)hdxΩ(λf+g)dxΩhdx=(mes(Ω)ΩλfhdxΩλfdxΩhdx)+(mes(Ω)ΩghdxΩgdxΩhdx)=λf,h+λg,h.
     ▪

    (2) Symmetry ∀f,gL2(Ω) 〈f,g〉=〈g,f〉.

    Proof.

    f,g=mes(Ω)ΩfgdxΩfdxΩgdx=mes(Ω)ΩgfdxΩgdxΩfdx=g,f.
     ▪

    (3) Positiveness and the condition of zero value ∀fL2(Ω) 〈f,f〉≥0, 〈f,f〉=0⇔f=μ=const.

    Proof.

    Let us consider a scalar multiplication f,g=Ωf(x)g(x)dΩ. The CBS inequality becomes true for: |(f,g)|2≤(f,f)(g,g), |(f,g)|2=(f,f)(g,g)μR:f=μg. Substituting g≡1, we get |Ωfdx|2(f,f)Ω1dx=mes(Ω)Ωf2dx. Thereby, it is true that f,f=mes(Ω)Ωf2dx(Ωfdx)20. Moreover, f,f=0|Ωfdx|2=(f,f)(1,1)μR:f=μ1, i.e. f=μ. ▪

    Footnotes

    Published by the Royal Society. All rights reserved.

    References