Next Article in Journal
Interacting Colour Strings Approach in Modelling of Rapidity Correlations
Previous Article in Journal
Fluctuations-Induced Quantum Radiation and Reaction from an Atom in a Squeezed Quantum Field
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Selected Topics of Social Physics: Equilibrium Systems †

by
Vyacheslav I. Yukalov
1,2
1
Bogolubov Laboratory of Theoretical Physics, Joint Institute for Nuclear Research, Dubna 141980, Russia
2
Instituto de Fisica de São Carlos, Universidade de São Paulo, CP 369, São Paulo 13560-970, Brazil
The article represents the first part of the review based on the lectures that the author gave during several years at the Swiss Federal Institute of Technology in Zürich (ETH Zürich).
Physics 2023, 5(2), 590-635; https://doi.org/10.3390/physics5020041
Submission received: 10 February 2023 / Revised: 11 May 2023 / Accepted: 16 May 2023 / Published: 2 June 2023
(This article belongs to the Section Statistical Physics and Nonlinear Phenomena)

Abstract

:
The paper gives an introduction to the physics approach to social systems providing the main definitions and notions used in the modeling of these systems. The behavior of social systems is illustrated by several quite simple, typical models. The present part considers equilibrium systems. Nonequilibrium systems will be presented in the second part of the review. The style of the paper combines the features of a tutorial and a survey, which, from one side, makes it simpler to read for nonspecialists aiming to grasp the basics of social physics, and from the other side, describes several rather recent original models containing new ideas that could be of interest to experienced researchers in the field. The selection of the material is limited and motivated by the author’s research interests.

1. Introduction

Nowadays, physics methods and models are often used in social science for characterizing the behavior of different social systems. It is useful to stress that physics provides methods and models for describing social systems, but it does not pretend to replace social science.
Under a social system, one implies a bound collective of a number of interacting agents. These can be various human societies, starting with families, professional groups, members of organizations, financial markets, country populations, etc., or these can be animal societies forming biological systems, e.g., wolf packs, fish shoals, bird flocks, horse herds, swarms of bees, ant colonies, and the like. Thus, the number, N, of agents in a society is more than one, and generally it is much larger than one: N 1 .
The collective of agents in a society is bound in the sense of having some features uniting the society members. For example, the members of a society can participate in a joint activity or share similar goals or beliefs. The common features uniting the members make the collective stable or metastable, that is, bound for a sufficiently long time that is much larger than the interaction time between the members. Agent interactions can be physical, economic, financial, and so on. Societies can be formed by force, as in the army and in prison, or they can be self-organized.
Social systems pertain to the class of complex systems. A system is complex if some of its properties are qualitatively different from the agglomerated properties of its parts. Complex systems can be structured, consisting of parts that are complex systems themselves. For instance, the human world is composed of countries that, actually, are also complex systems. In that sense, social systems are often hierarchical, containing several levels of complex systems. For example, the human world is made up of countries that are composed of social groups that form organizations that include individuals whose bodies are made of biological cells and whose brains making decisions are extremely complex systems.
To represent a complex system, the society needs to have the following properties. The number of agents composing the society must be large: N 1 ; however, this is not sufficient. If the members of the society are independent, the overall system can be characterized by a set of independent agents, this does not make such a system complex. In a complex society, its features are not just a sum of the features of the separate members. Thus, bees can fly, but a bee colony is not just a swarm of flying bees. A group of people walking separately is not a social system. A society becomes complex due to its agents’ interactions and mutual relations, which cause the nonadditivity of the society features. Strictly speaking, there is no generally accepted mathematical measure of social system complexity [1,2,3,4,5]. Complexity is rather a qualitative notion, assuming that a society is complex if it is stable and consists of many interacting agents, so the typical society features are not just an arithmetic average of the features of separate individuals.
For example, an ant colony is a complex system since the behavior of separate ants is regulated by their interactions and distribution of jobs; thus, ant queens lay eggs, worker ants form ant-hills and feed the queen, and winged males mate with the queen and die. Another example of a complex system is a bee colony, where a queen produces eggs; workers clean out the cells, remove debris, feed the brood, care for the queen, build combs, guard the entrance, ventilate the hive, and forage for nectar, pollen, propolis, and water; and drones fertilize the queen and die upon mating.
There are three interconnected problems in the description of complex social systems: how to model a society, how to investigate the model, and how it would be possible to regulate the society’s behavior. A model is to be simple enough and, at the same time, realistic. Too simple models may not describe reality, but too complicated models can distort reality because of the accumulation of errors in excessively complicated descriptions. Sometimes it happens that “less is more” [6]. So, it is desirable that models are neither trivial nor overcomplicated.
In this review, first, the basic ideas are described, allowing one to understand the general principles of constructing society models, so that, from one side, the model is not overcomplicated and, from another side, can identify the characteristic features of the considered society. Second, the main methods of investigating the system’s behavior are studied. Third, conditions are discussed, making it possible to find the desired properties of the society and the related optimal parameters of the models.
The layout of the present review is as follows. In Section 2, the principle of minimal information is discussed, which plays a pivotal role in establishing probability distributions for equilibrium and quasi-equilibrium systems. Section 3 discusses some quite simple, typical models of equilibrium social systems. Section 4 gives the basics of collective decision-making in a society. Finally, Section 5 concludes.
The material of this paper is based on the lectures that the author has been giving for several years at the Swiss Federal Institute of Technology in Zürich (ETH Zürich). Overall, the content constitutes a one-year course consisting of two parts, one devoted to equilibrium (or quasi-equilibrium) systems and the other to nonequilibrium systems. Following this natural separation, the presentation of the material is also split into the corresponding two parts. The present review deals with equilibrium social systems. The forthcoming second part will consider nonequilibrium social systems [7]. Being bound by the lecture frames, the content of the review is limited accordingly. The choice of the presented material is based on the research interests of the author.

2. Principle of Minimal Information

Actually, almost all models that intend to describe the behavior of equilibrium or quasi-equilibrium complex systems start with some extremization principles. These principles are widely spread in science as well as in life. Just as a joke, we may say that all life follows the extremization principle, which can be formulated as a “minimum of labor and maximum of pleasure”.
To model social systems in complicated situations where not all information is available, it is customary to resort to probabilistic descriptions, defining the probability distribution based on the principle of minimal information. This principle helps to develop the optimal description of a social system with limited information on its properties. The information is always limited since there are many agents and, in addition, their actions are often not absolutely rational. Experimental studies of brain activity well demonstrate that only a finite amount of information can be successfully processed by living beings [8]. The rationality of social individuals is always bound [9]. In addition, there exists random influences from the environment. Moreover, not all information is even necessary for a correct description, but excessive information can lead to the accumulation of errors and incorrect conclusions.

2.1. Information Entropy

Let us consider a social system of N agents enumerated by the index j = 1 , 2 , , N . Generally, the number of agents depends on time. Each j-th agent is marked by a set of characteristics,
σ j = { σ j α : α = 1 , 2 , , M j } .
The collection of the characteristics of all agents makes up the society configuration set,
X = { σ j : j = 1 , 2 , , N } .
The probability distribution, ρ ( σ ) , over the variables σ X is to be normalized,
σ ρ ( σ ) = 1 , 0 ρ ( σ ) 1 ,
where the sum over σ implies
σ = j = 1 N α = 1 M j .
If all variables are equiprobable, then the distribution is uniform,
ρ ( σ ) = 1 N tot N tot = j = 1 N M j .
A statistical social system is the triplet,
{ N , X , ρ ( σ ) } .
However, the probability distribution needs to be defined. For this purpose, one introduces the information entropy, S, which is a measure of the system uncertainty. The entropy is a measure of the inaccessible information, representing a functional of the distribution over the society’s characteristics,
S = S [ ρ ( σ ) ]
benefiting from the following properties.
(i) Continuity. It is a continuous function of the distribution, such that, for a small variation, δ ρ , in the latter, the entropy variation, δ S , is small:
S [ ρ ( σ ) + δ ρ ( σ ) ] S [ ρ ( σ ) ] δ S [ ρ ( σ ) ] δ ρ ( σ ) δ ρ ( σ ) .
(ii) Monotonicity. For a uniform distribution,
S 1 N tot < S 1 N tot + 1 .
(iii) Additivity. If a statistical system with the set of variables σ can be separated into two mutually independent subsystems with the variable sets σ 1 and σ 2 , such that
ρ ( σ ) = ρ ( σ 1 ) ρ ( σ 2 )
and
σ 1 ρ ( σ 1 ) = 1 , σ 2 ρ ( σ 2 ) = 1 ,
then the entropy of the total system is a sum of the subsystems entropies,
S [ ρ ( σ 1 ) ρ ( σ 2 ) ] = S [ ρ ( σ 1 ) ] + S [ ρ ( σ 2 ) ] .
Theorem 1
([10]). The unique functional, satisfying the above conditions, up to a positive constant factor, has the form:
S [ ρ ( σ ) ] = σ ρ ( σ ) ln ρ ( σ ) .
Here, the natural logarithm is used, but, generally, it is not essential which type of the logarithm base is employed since the entropy is defined up to a constant factor. The information entropy coincides with Gibbs’ entropy in statistical mechanics [11,12]. For the uniform distribution, one has:
S 1 N tot = ln N tot ,
which, visibly, is a monotonic increasing function of N tot . Generally, the entropy is in the range,
0 S [ ρ ( σ ) ] ln N tot .
The entropy is zero when there is just a single agent with a single characteristic, so that if N tot = 1 and the distribution is trivial, then
S [ 1 ] = 0 ( N tot = 1 ) .
The information entropy is a measure of the system’s uncertainty. In other words, one can say that entropy is a measure of inaccessible information.
When a statistical system with a distribution ρ ( σ ) is initially characterized by a trial likelihood distribution, ρ 0 ( σ ) , then the form,
I KL [ ρ ( σ ) ] = σ ρ ( σ ) ln ρ ( σ ) ρ 0 ( σ ) ,
is called the Kullback–Leibler information gain, relative entropy, or Kullback–Leibler divergence [13,14].

2.2. Information Functional

The information functional has to include the information gain (13). In addition, one should not forget that the probability distribution is to be normalized, as in Equation (3), hence
σ ρ ( σ ) 1 = 0 .
There can exist other information defining some average quantities, C i , by the condition,
σ ρ ( σ ) C i ( σ ) C i = 0 .
Overall, the information functional takes the form,
I [ ρ ( σ ) ] = σ ρ ( σ ) ln ρ ( σ ) ρ 0 ( σ ) + λ 0 σ ρ ( σ ) 1 + i λ i σ ρ ( σ ) C i ( σ ) C i .
The use of the Kullback–Leibler information gain (13) for deriving probability distributions is justified by the Shore–Johnson theorem [15].
Theorem 2
([15]). There exists only one distribution satisfying consistency conditions, and this distribution is uniquely defined by the minimum of the Kullback–Leibler information gain (13) under given constraints.
In the information functional (16), the coefficients λ 0 and λ i are the Lagrange multipliers, whose variation,
I [ ρ ( σ ) ] λ 0 = 0 , I [ ρ ( σ ) ] λ i = 0 ,
yields the normalization condition (14) and the expectation conditions (15).
Principle of minimal information. The probability distribution of an equilibrium statistical system (4) is defined as the minimizer of the information functional (16), thus satisfying the conditions:
δ I [ ρ ( σ ) ] δ ρ ( σ ) = 0 , δ 2 I [ ρ ( σ ) ] δ ρ ( σ ) 2 > 0 .
The minimization conditions give:
δ I [ ρ ( σ ) ] δ ρ ( σ ) = 1 + λ 0 + ln ρ ( σ ) ρ 0 ( σ ) + i λ i C i ( σ ) , δ 2 I [ ρ ( σ ) ] δ ρ ( σ ) 2 = 1 ρ ( σ ) ,
which leads to the distribution,
ρ ( σ ) = ρ 0 ( σ ) exp { i λ i C i ( σ ) } σ ρ 0 ( σ ) exp { i λ i C i ( σ ) } .
If there is no preliminary information on the trial distribution properties, so that all states are equiprobable, this implies that the trial distribution is uniform:
ρ 0 ( σ ) = 1 N tot .
Then, the sought distribution reads:
ρ ( σ ) = 1 Z exp i λ i C i ( σ ) ,
with the normalization factor, called partition function,
Z = σ exp i λ i C i ( σ ) .
In statistical mechanics, this is called the Gibbs’ distribution [11,12]. The principle of minimal information is equivalent to the conditional maximization of entropy. As is stated by Jaynes, the Gibbs’ distribution can be used for any complex system whose description is based on information theory [16,17].
It may happen that the probability distribution ρ ( σ , x ) depends on a parameter x that has not been uniquely prescribed. In such a case, this parameter can be chosen such that it follows the principle of minimal information. To this end, substituting the distribution (19) into the information function (16) results in
I [ ρ ( σ , x ) ] = R ( x ) i λ i C i ,
where
R ( x ) = ln Z ( x )
is relative information. Keeping in mind that the average quantities C i are fixed, one can see that the minimization of the information function with respect to the parameter x is equivalent to the minimization of the relative information (22), since
I [ ρ ( σ , x ) ] x = R ( x ) x = 0 , 2 I [ ρ ( σ , x ) ] x 2 = 2 R ( x ) x 2 > 0 .
In this way, the probability distribution can be uniquely defined [18].

2.3. Representative Ensembles

When the probability distribution is defined by the principle of minimal information, this does not mean that it is useful to have little information on the system. Vice versa, it is necessary to include in the information function all available relevant information on the system corresponding to the expected-value conditions. The principle of minimal information shows how to obtain an optimal description of a complex system while possessing minimal information on the latter. All available important information on the system must be taken into account.
Thus, a statistical system is described by the triplet (4), consisting of N system members, society set X, and the probability distribution ρ ( σ ) . The pair { X , ρ ( σ ) } is called a statistical ensemble. Observables are represented by real functions A ( σ ) = A * ( σ ) whose averages,
A ( σ ) = σ ρ ( σ ) A ( σ ) ,
are the observable quantities that can be measured. The collection of all available observable quantities { A ( σ ) } is a statistical state.
A statistical ensemble is termed representative when it provides the correct description for the system’s statistical state, so the theoretical expectation values of observable quantities accurately describe the corresponding measured values. To this end, it is necessary to include in the information function all relevant information in the form of additional constraints. Only then will the minimization of the information function produce a correct probability distribution.
The idea of representative ensembles goes back to Gibbs [11,12], and their importance is emphasized by Ter Haar [19,20]. The necessity of employing representative ensembles for obtaining reliable theoretical estimates has been analyzed, and the explanation of how the use of non-representative ensembles leads to incorrect results has been discussed in detail [21,22,23,24].

2.4. Arrow of Time

In general, the probability distribution, ρ ( σ , t ) , can depend on time t. An important question is: Why is time assumed to always increase? One often connects this with the second law of thermodynamics, according to which the entropy of an isolated system left to spontaneous evolution cannot decrease [25]. However, strictly speaking, by Liouville’s theorem, the entropy of a closed system remains constant in time (see, e.g., [26]). It is possible to infer that the arrow of time appears in quasi-isolated systems due to their stochastic instability [27,28,29,30,31]. Here, we show that there is quite a straightforward way of proving that the irreversibility of time can be connected with the non-decrease in information gain. First, one nees to recall the Gibbs’ inequality.
Gibbs’ inequality. For two non-negative functions, A ( σ ) 0 and B ( σ ) 0 , one has:
σ [ A ( σ ) ln A ( σ ) A ( σ ) ln B ( σ ) ] σ [ A ( σ ) B ( σ ) ] .
Proof. 
The proof follows from the inequality
ln A σ ) B ( σ ) 1 B ( σ ) A ( σ ) .
Consequence. The information gain caused by the transition from the probability ρ 0 ( σ ) to ρ ( σ ) is semi-positive:
σ ρ ( σ ) ln ρ ( σ ) ρ 0 ( σ ) 0 .
More examples of useful inequalities in information theory can be found in Ref. [32].
Now, let us consider the natural change in the probability distribution with time, starting from the initial value, ρ ( σ , 0 ) , to the final ρ ( σ , t ) at time t. Then, the following statement is valid.
Non-decrease in information gain. The information gain, due to the evolution of the probability distribution from the initial distribution ρ ( σ , 0 ) to the final ρ ( σ , t ) , does not decrease:
σ ρ ( σ , t ) ln ρ ( σ , t ) ρ ( σ , 0 ) 0 .
Thus, the non-decrease in the information gain with time can be connected with the direction of time.

3. Equilibrium Social Systems

This Section introduces the main notions required for modeling equilibrium social systems and studies some simple enough statistical models. In the long run, social systems are, strictly speaking, nonequilibrium. However, if the society does not experience external shocks during a period of time much longer than the typical interaction time between the society members, this society can be treated as equilibrium for that period of time. More details on the so-called social physics can be found in Refs. [33,34,35,36,37,38].

3.1. Free Energy

Equilibrium social systems can be treated by statistical theory as a particular application of the above notions of information theory and the principle of minimal information. Following the notation of Section 2, let us consider a society of N members, where each member is associated with characteristics (1). The considered society occupies a volume V.
Our aim is to study almost isolated societies with self-organization due to their internal properties. Of course, no real society can be absolutely isolated from its surroundings. The influence of surroundings is treated as stationary random perturbations or stationary noise. The perturbations can be produced by other societies and by natural causes, such as earthquakes, floods, droughts, epidemics, etc. The noise is considered stationary, which implies that an equilibrium situation is assumed. The influence of noise on the society is measured by temperature, T, which is a measure of noise intensity. In limiting cases, the absence of noise implies T = 0 , and extremely strong noise means T = . One often uses the inverse temperature,
β 1 T ,
which can be interpreted as the measure of a society’s isolation from random noise. Here, β = 0 means no isolation and extreme noise, while β = implies complete isolation and no noise.
An equilibrium system is conveniently characterized by a functional H ( σ ) , termed a Hamiltonian or harm. The expected value of the Hamiltonian is society’s energy,
E = σ ρ ( σ ) H ( σ ) H ( σ ) ,
which is also called society cost. In some economic and financial applications, it is termed disagreement or dissatisfaction, since higher energy assumes a more excited society.
The energy of noise is T S , with S being the information entropy. The free energy is part of the society’s energy, due to the society itself without the noise energy:
F = E T S .
In the limiting case of a zero temperature, when there is no noise, the free energy coincides with the society’s energy,
F = E ( T = 0 ) .
The probability distribution over the society characteristics is defined by the principle of minimal information, which, in the present notation, gives
ρ ( σ ) = 1 Z e β H ( σ ) ,
with the partition function
Z = σ e β H ( σ ) .
Substituting distribution (32) into entropy (9) leads to
S = β E + ln Z .
Comparing Equation (34) with Equation (30) results in the expression,
F = T ln Z .
It is straightforward to check that the entropy S can be represented as
S = F T .
The society pressure is defined as
P = F V .
The relation (37) implies that the region occupied by a society can be changed because of the pressure. Population density is
ρ N V .
Combining Equation (22) and (35) gives the equality
F ( x ) = T ln Z ( x ) = T R ( x ) .
Hence, the information function (21) and the free energy (39) are connected:
I [ ρ ( σ , x ) ] = β F ( x ) i λ i C i .
Since the values C i are fixed, one has:
I [ ρ ( σ , x ) ] x = 1 T F ( x ) x .
Therefore, the minimization of the information function over a parameter x is equivalent to the minimization of free energy over this parameter:
min x I [ ρ ( σ , x ) ] min x F ( x ) .

3.2. Society Stability

Considering society models, it is important to make sure that the society is stable. A society is stable if small variations in parameters do not drive it far from its initial state. Between several admissible statistical states, the system chooses that which provides the minimal free energy. The system is stable with respect to the variation in parameters if its free energy is minimal, so that
δ F = 0 , δ 2 F > 0 .
The variation in the intensity of noise, that is, of temperature, under a fixed volume, is characterized by specific heat,
C V = 1 N E T V = T N 2 F T 2 V .
With the free energy (35), this takes the form,
C V = var ( H ) N T 2 ,
where the variation in the Hamiltonian, H = H ( σ ) , means
var ( H ) H 2 H 2 .
The latter is non-negative; hence, the specific heat has to be non-negative.
The variation in volume, under a fixed temperature, is characterized by the compressibility,
ϰ T = 1 V V P T N = 1 V 2 F V 2 T N 1 .
The condition of stability requires that the specific heat, as well as compressibility, be positive and finite:
0 C V < , 0 ϰ T < .
In this way, the principle of minimal information agrees with the minimal free energy, which, in turn, assumes societal stability.
As an example of societal instabilities, one may mention the disintegration of a country as a result of a war or because of external economic pressures. Another example is the bankruptcy of a firm caused by changed financial conditions.

3.3. Practical Approaches

In order to accomplish quantitative investigations of society’s properties, two different approaches are employed: the network approach and the typical-agent approach.
The network approach, also called multi-agent modeling, is based on the following assumptions:
(i) The considered society consists of agents, or nodes, that are fixed at the lattice sites of a spatial (often two-dimensional) lattice.
(ii) The agents interact with each other when they are close to each other; commonly, the nearest-neighbor interactions are considered.
(iii) Due to rather complicated calculations, generally, one has to resort to numerical modeling with computers.
There exist many examples of networks, such as electric-current networks, models of magnetic and ferroelectric materials, neuron networks in brains, computer networks, etc.
The results of the network approach depend on lattice dimensionality modeling the society (whether one-, two-, or three-dimensional lattices are considered), lattice geometry (cubic, triangular, or another structure), and on the interaction type and range (long-range, short-range, or mid-range). The network approach is appropriate for small or just structured systems, with agents that can be treated as fixed at spatial points.
However, complex societies, such as human and biological societies, are not composed of agents tied to a spatial lattice; commonly, they are not attached to any fixed spatial locations or sites. The agent interactions are not of the nearest-neighbor type and can be independent of the distance between the agents. Interactions in a society can be either direct, when interacting with fixed neighbors, or can involve changing neighbors. Nowadays, interactions that do not depend on distance are widespread, such as through phone, Skype, WhatsApp, Telegram, and the like. There exist indirect interactions through letters and e-mails, reading newspapers and books, listening to the radio, and watching television. The majority of interactions are long-range, but not solely with nearest neighbors. Other biological societies also interact at large distances by means of their voices and smells.
Summarizing, complex societies, such as human or other biological societies, are formed by agents that are not fixed at spatial locations and can interact over long distances. This kind of society is better described by the typical-agent approach. In this approach, one reduces the problem to the consideration of the behavior of typical agents, representing a kind of average member of the society. For example, the typical interaction of agents, having the characteristics σ i and σ j and described by the term σ i σ j , is transformed into the expression,
σ i σ j = σ i σ j + σ i σ j σ i σ j ,
with i j . The average characteristic is
σ j = σ ρ ( σ ) σ j s .
Thus, the representation (49) becomes
σ i σ j = s σ i + s σ j s 2 ,
and the average interaction reads:
σ i σ j = σ i σ j = s 2 .
Then, the description of the interaction is reduced to the consideration of typical agents subject to the average influence of other typical agents. For complex social systems, the typical-agent approach is not merely simpler but also more correct.

3.4. Society Transitions

In general, one is interested not in the characteristics of single agents but in the general behavior of a society. For this purpose, one considers the mean arithmetic characteristic,
s ( σ ) = 1 N j = 1 N σ j .
The observable quantity is the average characteristic
s s ( σ ) = 1 N j = 1 N σ j .
The general behavior of a society can be associated with its average characteristics. When the property of a characteristic qualitatively changes, one speaks of a social phase transition. For example, it may happen that, under such a social transition, the average characteristic varies between zero and nonzero values. Conditionally, one can name the social state with | s | > 0 an ordered state, while the state where s = 0 , a disordered state. In that case, the average characteristic (54) is termed an order parameter. There can occur the following types of transitions when a system parameter, for instance, temperature, varies.
First-order transition. This type of transition happens when the order parameter at some point T 0 changes from nonzero to zero by a discontinuous jump:
s ( T 0 0 ) > 0 , s ( T 0 + 0 ) = 0 , s ( T 0 0 ) s ( T 0 + 0 ) .
Discontinuous social transitions can be associated with revolutions.
Second-order transition. In this case, the order parameter at a critical point T c changes continuously from nonzero to zero:
s ( T c 0 ) > 0 , s ( T c + 0 ) = 0 , s ( T c 0 ) = s ( T c + 0 ) .
Continuous transitions describe evolutions.
Gradual crossover. The order parameter does not become zero at a finite point, but at some crossover point, T c , it strongly diminishes and tends to zero only in the limit of large T:
s ( T c 0 ) = s ( T c + 0 ) > 0 ,
s ( T ) 0 ( T T c ) .
This transition corresponds to a smooth evolution.
There can occur more unusual situations when a society is not completely equilibrium [39], but for the description of equilibrium societies, the above three types of social transitions are sufficient.

3.5. Yes–No Model

This Section considers a simple enough model that is known in statistical physics, where it is called the Ising model [40,41,42,43]. This model is also often used in different applications for financial and economic problems [44]. This model is needed in order to illustrate in action the notions introduced in the previous sections, to exemplify the terminology associated with social systems, and to have the ground to study more complicated models in the following sections.
Suppose each agent of a society can have just two features, which are opinions that can be termed “yes” and “no”. Generally, this can be any decision with two alternatives. For instance, this can be voting for or against a candidate in elections, supporting or rejecting a suggestion in a referendum, buying or selling stocks in a market, etc. These two alternatives can be represented by the binary variable taking two possible values, e.g.:
σ j = 1 , no ; + 1 , yes .
The interaction, or mutual influence, between two members of the society is expressed as
H i j = J i j σ i σ j ,
with the value J i j = J j i being the intensity of the interaction. The case of agreement (collaboration) or disagreement (competition) among the members, respectively, corresponds to the values
J i j > 0 ( agreement , collaboration ) ,
J i j < 0 ( disagreement , competition ) .
The terminology comes from the fact that, under agreement, when J i j > 0 , the interaction energy is minimal for coinciding σ i and σ j , while in the case of disagreement, when J i j < 0 , the interaction energy is minimal for opposite σ i and σ j . When there is mutual agreement and both agents vote in the same way, either both “yes” or both “no”, the interaction energy is lower than when the agents vote differently.
The Hamiltonian of the system is
H = 1 2 i j N J i j σ i σ j .
In physics, this is called the Ising model; see the history of the model in Ref. [45].
In the typical-agent approach, the Hamiltonian reads:
H = J s j = 1 N σ j + 1 2 J s 2 N ,
where
J 1 N i j N J i j = i ( j ) N J i j .
Using the equalities,
exp β J s j = 1 N σ j = j = 1 N exp ( β J s σ j ) ,
σ j = 1 N exp ( β J s σ j ) = [ exp ( β J s ) + exp ( β J s ) ] N ,
one obtains the probability distribution,
ρ ( σ ) = j = 1 N ρ ( σ j ) ,
where
ρ ( σ j ) = exp ( β J s σ j ) exp ( β J s ) + exp ( β J s ) .
For the order parameter (54), one obtains the equation,
s = tanh ( β J s ) .
When there is mutual disagreement between the agents, so that J < 0 , the sole solution for the order parameter is s = 0 , which means a disordered state:
s 0 ( J < 0 ) .
However, when the members of the society are in mutual agreement, so that J > 0 , then there are two solutions to Equation (65). One solution is s = 0 , but the other solution is nonzero for T < T c = J . In order to choose the stable solution, one needs to find out which of them minimizes the free energy.
It is convenient to work with dimensionless, reduced free energy,
F ¯ ( s ) F N J ( J > 0 ) ,
and to measure temperature in units of J, keeping in mind a positive J. Then,
F ¯ ( s ) = 1 2 s 2 T ln [ 2 cosh ( β s ) ] .
The condition of the free-energy extremum is
F ¯ ( s ) s = s tanh ( β s ) = 0 ,
which gives the order parameter Equation (66). The condition of the free-energy minimum,
2 F ¯ ( s ) s 2 = tanh 2 ( β s ) > 0 ,
holds true for nonzero s. It is also not complicated to check that
F ¯ ( s ) < F ¯ ( 0 ) ( T < T c ) ,
where F ¯ ( 0 ) = T ln 2 . Hence, for temperatures lower than the critical temperature T c = J > 0 , the stable state corresponds to the nonzero value of the order parameter s.
At zero temperature ( T = 0 ) , with collaborating agents ( J > 0 ) , the society is completely ordered:
s = ± 1 ( T = 0 , J > 0 ) .
Although with an unspecified decision: either all deciding “yes” or all choosing the option “no”.
One can immediately notice that for any T below T c there are two solutions, positive and negative, both corresponding to the same free energy F ¯ ( s ) = F ¯ ( s ) that does not depend on the sign of the order parameter. This means that, with equal probability, the society can vote “yes” as well as “no”. In that sense, the situation is degenerate.

3.6. Enforced Ordering

The yes–no degeneracy can be lifted by imposing an ordering force that acts on the members of the society. The ordering force, or regulation force, acting on the members represents different regulations, such as governmental rules and laws, as well as the society’s traditions and habits. The society Hamiltonian, including the ordering force, takes the form,
H = 1 2 i j N J i j σ i σ j j = 1 N B j σ j .
In what follows, we assume that the regulation is uniform, that is, that the same regulations are applied to all members of the society. This translates into the condition that the ordering force is the same for all agents, such that
B j = B 0 .
In other words, the laws are the same for everyone.
Resorting to the typical-agent approach yields the Hamiltonian,
H = J ( s + h ) j = 1 N σ j + 1 2 J s 2 N ,
with the dimensionless force,
h B 0 J ( J > 0 )
and the order parameter,
s σ j .
We keep in mind the case of mutual agreement, where J > 0 .
In the typical-agent approach, the reduced free energy becomes
F ¯ ( s ) = 1 2 s 2 T ln 2 cosh s + h T ,
where temperature is measured in units of J. The order parameter (78) is given by the equation,
s = tanh s + h T .
The order parameter can also be defined as the derivative,
s = F ¯ ( s ) h .
How the society responds to the imposed regulations is described by the susceptibility
χ = s h = 2 F ¯ ( s ) h 2 ,
which can be represented as
χ = 1 T var ( s ( σ ) ) ,
with the variance
var ( s ( σ ) ) = s 2 ( σ ) s ( σ ) 2 .
Thus, the susceptibility is positive, which means that the imposed regulations increase the order. Explicitly, one finds:
χ = 1 s 2 T + s 2 1 .
From the order parameter Equation (80), it is seen that the sign of the order parameter is prescribed by the sign of the ordering force, so that s > 0 for h > 0 and s < 0 for h < 0 . By choosing the sign of h, it is possible to enforce either the ordering “yes” or the ordering “no”. For concreteness, we consider h > 0 . If the ordering force is extremely strong, then
s 1 ( h ) .
In that way, the imposed ordering force increases the order in the society and makes it more stable. For illustration, let us consider the case of no noise; hence, T = 0 . Then, the reduced free energy coincides with the reduced energy,
E ¯ E J N = 1 2 h ( T = 0 ) .
As one can see, the imposed force diminishes the societal energy, which assumes that the society should be more stable.

3.7. Command Economy

From Section 3.6, one may conclude that the stricter the regulations, the more ordered the society. That is, the larger the force h, the larger the order parameter s. The seeming conclusion could be that it is profitable to make the regulations as stringent as possible. Is this so? Let us consider, as an example, an economic society. The most strictly regulated type of economic organization is command economy or centrally planned economy. Is such over-regulated economics the most efficient kind?
The basic points of command economy are:
(i)
A centralized government owns most means of production and most businesses.
(ii)
The government controls production levels and distribution quotas.
(iii)
The government controls all prices and salaries.
The proclaimed advantages of command economy assume that regulatory decisions are made for the benefit of the whole society, there is no large economic inequality, and the economy is claimed to be more stable. However, in reality, the proclaimed catchwords contain many problems:
  • It is not always well defined what the benefits of the society are. Governmental decision-makers often make decisions in their own favor, but not in favor of the society, announcing their egoistic goals as the society’s objective needs.
  • It is actually impossible to formulate correct plans for all goods in the long term. Constant shortages of necessary goods and surpluses of unnecessary goods are a rule. Economy is in a permanent crisis.
  • Since the conditions for long-term future production cannot be exactly predicted, the plans are never accomplished. There is a necessity for correcting the plans, with additional spending for the plans’ corrections.
  • Since everything is planned in advance, it is almost impossible to introduce innovations that have not been planned. As a result, technological retardation is imminent.
  • The kinds of science to develop are also planned. Some sciences are mistakenly announced as unnecessary or wrong. Examples are cybernetics and genetics in the Soviet Union. This results in irreparable harm to the economy.
  • It is impossible to absolutely fairly distribute wealth. Those who distribute always take more for themselves. Consequently, people are dissatisfied. Bureaucratic corruption flourishes, leading to enormous economic losses.
  • Since wealth is rigidly distributed by the government, there is no reason to work hard. Labor becomes inefficient, with low productivity.
  • The necessity of having huge planning institutions consumes a large amount of economic means. Ineffective planning reduces the planners to the state of parasites, merely wasting resources.
  • The necessity of having a large number of controllers for implementing economic plans and controlling their accomplishment also reduces such controllers to parasites.
  • The suppression of economic freedom, ascribed to the needs of the state but often needed to protect the privileges of the country’s rulers, kills the motivation of people to work well.
  • As soon as there appears frustration of the suppressed citizens, there is a necessity to have excessively large police and regulating services supervising the society.
  • To support the order and punish those who are against it, who hesitate, or who could even be potentially dangerous, the government organizes massive suppression of the people, which results in large economic losses.
  • To realize unreasonable plans, the government practices mass arrests of innocent people to create slave labor. However, slavery is not economically efficient.
  • To distract the discontentment of the population from economic failure, a necessity arises to invent enemies, which requires a large army, which consumes a substantial amount of the country’s wealth.
  • In order to persuade people that everything is all right, it is necessary to organize propaganda through mass media, which results in ineffective spending of resources.
More details on societies with a command economy can be found in the literature [46]. Due to so many factors that make the economy of an over-regulated society ineffective, it does not seem feasible that enforcing regulations unlimitedly could result in an ideally ordered society. It seems that the model of a regulated society, described in the previous section, does not take into account some factors that prevent an indefinite increase in order by over-regulating the society.

3.8. Regulation Cost

The problem with regulation is that it requires the use of societal resources. Regulations are costly. The more strict regulations there are, the more resources required. All regulations imposed on a society are produced by a part of the same society, that is, by the society itself. Some members of the society act on other members. The regulation cost is the cost that the society has to pay in order to introduce the desired regulations. The regulation cost, keeping in mind the above, can be modeled by the expression
H reg = 1 2 i j N A i j B i B j σ i σ j ,
where A i j = A j i > 0 . The coefficients A i j describe the regulation efficiency. By the order of magnitude, they are proportional to the interaction, J i j , between the agents since regulations need to overcome mutual interactions in order to impose restrictions. Note that J i j and A i j are different types of interactions: one is a direct interaction not related to the process of ordering, and the other is an interaction inducing the ordering in the presence of an additional force. Different types of interactions, in general, are different.
Thus, the total Hamiltonian becomes
H = 1 2 i j N J i j σ i σ j i = 1 N B i σ i + 1 2 i j N A i j B i B j σ i σ j .
Let us again assume that the force acting on the society members is uniform in the sense of equality (75).
Now, the order parameter (78) differs from the expression,
M = F ¯ h = s 1 N i j N α i j h σ i σ j ,
with the notation,
α i j A i j J .
The expression (89) plays the role of the total order parameter, contrary to s, which, in the present case, is a partial order parameter.
In the typical-agent approach, Hamiltonian (88) reads:
H = ( J s + B 0 A B 0 2 s ) i = 1 N σ i + 1 2 J A B 0 2 s 2 N ,
where
A 1 N i j N A i j .
The reduced free energy takes the form,
F ¯ = 1 2 1 α h 2 s 2 T ln 2 cosh ( 1 α h 2 ) s + h T ,
in which
α 1 N i j N α i j = A J .
The parameters α and h are assumed to be positive.
The order parameter, s = σ j , can be obtained from the condition,
F ¯ s = 0
leading to the equation,
s = tanh ( 1 α h 2 ) s + h T .
However, in the presence of a Hamiltonian of a quadratic, with respect to the ordering force term, the quantity s does not define the total order that is given by the form (89), which yields the total order parameter
M = F ¯ h = s α h s 2 .
The susceptibility is defined by the derivative,
χ M h = 2 M h 2 ,
resulting in the expression,
χ = ( 1 s 2 ) ( 1 2 α h s ) 2 T + ( 1 s 2 ) ( α h 2 1 ) ,
which differs from the derivative,
s h = ( 1 s 2 ) ( 1 2 α h s ) T + ( 1 s 2 ) ( α h 2 1 ) .
The behavior of the social system at small and large regulating forces shows how the order parameters and free energy vary. The analysis of the asymptotic behavior demonstrates the relation between the system’s stability and its order. This relation can be rather complicated depending on the system parameters.
With a small ordering force and weak noise, when 0 < T < 1 , the order parameter s behaves as
s s 0 + 1 s 0 2 s 0 2 + T 1 h ( 0 < T < 1 , h 0 ) ,
where s 0 is the solution to the equation,
s 0 = tanh s 0 T .
The total order parameter, M, with a small ordering force and weak noise behaves as
M s 0 + 1 s 0 2 s 0 2 + T 1 α s 0 2 h ( 0 < T < 1 , h 0 ) ,
which shows that the order parameters can either increase or decrease with rising h, depending on the system parameters.
The free energy with a small ordering force and weak noise is
F ¯ s 0 2 2 T ln 2 cosh s 0 T s 0 h ( 0 < T < 1 , h 0 ) .
For T > 1 and weak h, the order parameters are
s 1 T 1 h 3 α ( T 1 ) 2 + T 3 ( T 1 ) 4 h 3 ,
M 1 T 1 h 3 α ( T 1 ) 2 + T 3 ( T 1 ) 4 + α ( T 1 ) 2 h 3 ( T > 1 , h 0 ) .
The free energy reads:
F ¯ T ln 2 1 2 ( T 1 ) h 2 ( T > 1 , h 0 ) .
When the regulating force is strong, then, at any temperature, the order parameters are
s 1 α h T 1 α 2 h 3 , M 2 ( T 1 ) α 2 h 3 ( h ) ,
and the free energy,
F ¯ T ln 2 1 2 α + T 1 2 α 2 h 2 ( h ) .
In the absence of noise, that is, when T = 0 , the free energy equals the energy. Considering the reduced energy,
E ¯ H N J = 1 2 s 2 h s + 1 2 α h 2 s 2 ,
at zero T, one obtains:
E ¯ = 1 2 h + 1 2 α h 2 ( T = 0 ) .
As one can see, the energy (110) with switching on regulations, first decreases, making the society more stable,
E ¯ 1 2 h ( T = 0 , h 0 ) ;
then, it reaches the minimum,
min h E ¯ = 1 2 1 + 1 α h = 1 α ,
after which it increases, making the society less and less stable:
E ¯ 1 2 α h 2 ( T = 0 , h ) .
Thus, in the absence of noise, the optimal regulation, making the society the most stable, corresponds to h = 1 / α , when the energy (112) is minimal. This implies that some amount of regulation is useful in stabilizing the society. However, the society should not be over-regulated. Over-regulation can make the society unstable.
In the presence of noise, the situation is much more complicated, depending on the noise strength and the value of the regulating force. The peculiarity of the system’s behavior is illustrated in Figure 1 and Figure 2. Figure 1 shows the behavior of the free energy F ¯ as a function of the regulation force h for different temperatures (noise intensity), when T < 1 in Figure 1a and when T > 1 in Figure 1b. Figure 2 shows the order parameter s as a function of the regulation force h for different temperatures: low ( T < 1 ) and high ( T > 1 ) . As one can see, at low temperatures, the free energy first decreases and then increases, similarly to the behavior of energy (109). The order parameter first increases but later decreases. That is, there exists an optimal regulation strength when the society is the most stable and at the same time the most ordered. In the presence of strong noise, the free energy decreases with h; however, the order parameter first increases but then decreases. So, the regulating force stabilizes the society but improves the order only at a limited value of h, while at a quite strong force, the order diminishes.

3.9. Fluctuating Groups

It is customary for society members to separate into several groups with different properties, so that the number of members in a group is not constant but varies with time. For example, these could be groups of opponents to the government, workers on strike, opposing organizations, and the like. Different groups of agents can be described by different order parameters. The groups can be localized in some spatial parts of the society or may move through the whole societal space. The groups are not permanent in their agent numbers and do not necessarily exist forever, but they can arise and vanish. In that sense, the groups fluctuate; they can appear, change, and then disappear.
Let the number of agents in a fluctuating group be N f and the characteristic time during which they do not essentially change be t f . Fluctuating groups are mesoscopic, at least in one of the following senses.
Mesoscopic in size:
1 N f N .
The number N f is much larger than can be used to define the group order parameter, and N f is much smaller than the total number of agents, N, in the society to be classified as a group inside that society.
Mesoscopic in time:
t int t f t exp ,
where t int is the characteristic interaction time between the agents and t exp is the observation (experiment) time during which the society is studied. The time t f has to be much larger than the interaction time in order that the groups as such could be formed. Additionally, t f is much smaller than the observation time for the groups to be classified as fluctuating. The general method of describing this kind of society that contains fluctuating groups is presented in this section.
Let us consider a snapshot of a society picture, where the spatial location of a j-th agent of a society is denoted by a vector a j . The overall society containing all its members is given by the collection,
G = a j : j = 1 , 2 , , N .
The society consists of several groups, each characterized by a specific feature. The features are enumerated by the index f = 1 , 2 , A group with an f-th feature is
G f = a j G f : j = 1 , 2 , , N f .
The union of all groups forming the whole society is
f G f = G , f N f = N .
The spatial locations of the groups is represented by the manifold indicator functions,
ξ f ( a j ) = 1 , a j G f , 0 , a j G f .
The collection of all indicator functions, showing the society configuration, is denoted as
ξ = { ξ f ( a j ) : a j G , f = 1 , 2 , } .
The indicator functions possess the properties:
f ξ f ( a j ) = 1 , j ξ f ( a j ) = N f ,
the first of which means that each agent pertains to some of the groups, while the second shows the number of agents in an f-th group.
In a realistic society, the agents can change their locations, so the location vector a j = a j ( t ) depends on time t. Hence, the societal configuration is also a function of time,
ξ ( t ) = { ξ f ( a j ( t ) ) : a j ( t ) G ; f = 1 , 2 , } .
Since t exp t int , the observable quantities describing the society correspond to the double average over the system variables and over time,
A ( σ , ξ ( t ) ) = σ 1 t exp 0 t exp ρ ( σ , ξ ( t ) ) A ( σ , ξ ( t ) ) d t .
The motion of the groups inside the society is generally so complicated that it is neither possible nor reasonable to follow the detailed movements of all agents, but the agent locations can be treated as random. Then, it is possible to interchange the average over time with the average over the random society configurations by the rule,
1 t exp 0 t exp d t D ξ .
Then, one needs to realize the functional integration over the manifold indicator functions. We do not describe the mathematical details of the integration that can be found in the reviews [47,48], but present the results.
The probability distribution ρ ( σ , ξ ) of a heterophase society, depending on the configuration, ξ , can be derived following Section 2.2. The probability is to be normalized:
σ ρ ( σ , ξ ) D ξ = 1 .
The average energy is given by the expression,
σ ρ ( σ , ξ ) H ( σ , ξ ) D ξ = E .
The information functional takes the form,
I [ ρ ( σ , ξ ) ] = σ ρ ( σ , ξ ) ln ρ ( σ , ξ ) ρ 0 ( σ , ξ ) D ξ
+ λ 0 σ ρ ( σ , ξ ) D ξ 1 + λ σ ρ ( σ , ξ ) H ( σ , ξ ) D ξ E .
Minimizing the information functional and assuming the uniformity of the trial distribution, ρ 0 ( σ , ξ ) = const , yields the probability,
ρ ( σ , ξ ) = 1 Z e β H ( σ , ξ ) ,
with the partition function,
Z = σ e β H ( σ , ξ ) D ξ .
The free energy reads:
F = T ln Z = T ln σ e β H ( σ , ξ ) D ξ .
Observable quantities are given by the averages,
A ( σ , ξ ) = σ ρ ( σ , ξ ) A ( σ , ξ ) D ξ .
Accomplishing the average over configurations allows us to define an effective Hamiltonian, H ˜ , by the relation,
e β H ˜ = e β H ( σ , ξ ) D ξ .
The effective Hamiltonian takes the form (with the direct summation ⨁),
H ˜ = f H f , H f = H f ( σ , w f ) ,
where the probability that an agent pertains to a j-th group is
w f N f N .
In other words, this is the fraction of the society agents belonging to the f-th group. Certainly, the normalization conditions are valid:
f w f = 1 , 0 w f 1 .
The group probabilities are the minimizers of the free energy,
F = T ln σ e β H ˜ ,
under the normalization conditions (135),
F = abs min F ( w 1 , w 2 , ) .
The observable quantities are given by the averages
A ( σ , ξ ) = f A f ( σ , w f ) ,
where
A f ( σ , w f ) = σ ρ f ( σ , w f ) A f ( σ , w f ) ,
with the effective probability distributions,
ρ f ( σ , w f ) = 1 Z f e β H f ( σ , w f ) ,
and the partition functions,
Z f = σ e β H f ( σ , w f ) .
The sum of direct summation ⨁ in Equation (133) is used instead of the simple sum in order to stress that the summands are defined on different configuration sets,
X f = { σ f j : j = 1 , 2 , , N f } ,
with the total society configuration set being the tensor product,
X f = f X f .
This is the general approach for describing heterophase societies with fluctuating groups. The approach reduces the consideration of quasi-equilibrium systems with mesoscopic group fluctuations to the description of effective equilibrium systems with a renormalized effective Hamiltonian. The mathematics of accomplishing functional integration over manifold indicator functions is described in reviews [21,47,48]. More mathematical details are presented in Refs. [49,50,51]. Section 3.10 just below gives an example of applying this approach.

3.10. Self-Organized Disorder

Let us consider the yes–no model of Section 3.5 modified to take account of fluctuating groups. For simplicity, two groups have to be kept in mind: one ordered and the other disordered. The ordered group consists of agents agreeing with each other and is characterized by a nonzero order parameter, while the disordered group is described by a zero order parameter to be defined in this Section below.
Each j-th agent from an f-th group is characterized by the features σ f j . In the yes–no model, these features are described by the variable σ f j = ± 1 . The total feature set for the whole system is the configuration set (143).
In addition to the interaction J i j > 0 , typical for self-organization in the yes–no model, one needs to include other interactions that would describe some disagreements between the agents. Let us denote the disordering interaction by U i j .
In the snapshot picture, the Hamiltonian with ordering and disordering interactions reads as
H ( σ , ξ ) = H 1 ( σ , ξ 1 ) H 2 ( σ , ξ 2 ) ,
with
H f ( σ , ξ f ) = 1 2 i j N ξ f ( a i ) ξ f ( a j ) U i j 1 2 i j N ξ f ( a i ) ξ f ( a j ) J i j σ f i σ f j .
Here, the manifold indicator functions show the belonging of the agents to the corresponding groups.
After averaging over group configurations, one obtains the effective Hamiltonian,
H ˜ = H 1 H 2 , H f = H f ( σ , w f ) ,
H f ( σ , w f ) = 1 2 w f 2 i j N ( U i j J i j σ f i σ f j ) .
The order parameter for the f-th group is
s f 1 N j = 1 N σ f j ,
which translates into the equation,
s f = tanh J w f 2 s f T .
Let the first group be ordered, so that
s 1 0 ,
while the second group be disordered, in the sense that
s 2 0 .
In the case of two groups, it is convenient to denote the fraction of agents in the ordered group and disordered group as
w 1 w , w 2 = 1 w .
Then, the necessary condition for the free-energy minimum is
F w = H ˜ w = 0 .
This results in the equation,
w = u s 2 2 2 u s 1 2 s 2 2 ,
in which the notation is used:
u U J , U 1 N i j N U i j , J 1 N i j N J i j .
Taking into account that the disordered group is described by the zero-order parameter s 2 = 0 , one has the fraction of agents in the ordered group:
w = u 2 u s 1 2 .
Due to the probability definition 0 w 1 , and because 0 s 1 1 , expression (155) exists only for sufficiently strong disordering interactions, such that u 1 .
Note that the second derivative of the free energy, being positive, leads to the inequality
2 F w 2 = 2 H ˜ w 2 β H ˜ w 2 > 0 ,
which yields the condition u > 1 / 2 .
The societal reduced energy,
E ¯ H ˜ J N ,
leads to
E ¯ = E ¯ 1 + E ¯ 2 , E f = 1 2 w f 2 u s f 2 .
For simplicity, let us study the case of no noise, so that T = 0 . Then, s 1 = 1 , and the energy becomes
E ¯ = w 2 u 1 2 + 1 2 w u ,
while the fraction of ordered agents is
w = u 2 u 1 .
Thus, the society’s energy is
E ¯ = u ( u 1 ) 2 ( 2 u 1 ) .
Now, one has to decide which society is more stable: that which contains fluctuating disordered groups or that which does not contain them? The more stable is the society whose free energy is lower. In the case of small noise, one needs to compare the corresponding expected energies.
If no fluctuating groups were allowed, then the ordered fraction would be exactly one ( w 1 ) . The energy of a completely ordered society is
E ord E ¯ ( w = 1 ) = u 1 2 ,
where w 1 . If the whole society contained only disordered agents ( w 0 ) , then the societal energy would be
E dis E ¯ ( w = 0 ) = 1 2 u .
The society with disordered group fluctuations is more stable when its energy is lower. Comparing the above expressions, one sees that
E ¯ < E ord < E dis ( u > 1 ) .
Thus, for competing interactions, satisfying the condition u > 1 , the energy of the society with fluctuating disorder is lower than that of the perfectly ordered society, which, in turn, is lower than the disordered society. That is, a completely ordered society is more stable than a disordered society. However, the most stable is the society with a mixture of fluctuating disorder. Thus, disorder fluctuations make the society more stable. In that sense, the disorder is self-organized.
The total order parameter is defined as the average of the expression
M ( σ , ξ ) = 1 N f j = 1 N ξ f ( a j ) σ f j ,
which gives
M = M ( σ , ξ ) = f w f s f .
When there is no noise, that is at zero temperature, one finds:
M = w s 1 = u 2 u 1 ( T = 0 ) .
Concluding, under sufficiently strong competition between the society’s agents, there spontaneously appear social disorder groups. When the disorder groups appear, they diminish the total society order. However, the existence of the disorder groups makes the society more stable. This is an example that shows that it is not always useful to try to realize a complete order in a society since the presence of some disorder can make the society more stable. Sometimes, societies that are more stable generate self-organized disorder.

3.11. Coexistence of Populations

Quite often, social systems consist of several groups of rather different people, so the groups are more or less stable and, on average, do not essentially change for a long time, contrary to fluctuating groups considered in the previous section. Each of the groups can be characterized by different typical agents. There are numerous examples of such societies. Thus, a society, forming a country, can often include groups of different nationalities or religions.
In a society composing a financial market, there are groups of fundamentalists and chartists. Fundamentalists base their decisions upon market fundamentals, such as interest rates, the growth or decline of the economy, companies’ performance, etc. Fundamentalists expect asset prices to move towards their fundamental values, hence, they either buy or sell assets that are assumed to be undervalued or, respectively, overvalued. Chartists, or technical analysts, look for patterns and trends in past market prices and base their decisions upon the extrapolation of these patterns. There exist as well contrarian groups that buy or sell contrary to the trend followed by the majority.
Different groups of people or other living beings possessing differing features are called populations. Let a society, inhabiting the volume V, contain several different populations enumerated by the index n = 1 , 2 , . In an n-th population, there are N n members. In what follows, we consider a general approach that describes whether the populations can or cannot coexist with each other. The populations can be of any origin. For concreteness, we can keep in mind the different population groups living in a country.
The populations can either peacefully live in the whole country, which can be called a mixed society, or can possess the tendency to separate, thus having no intention for joint coexistence. Our aim is to understand how one could quantify the conditions when different populations prefer to live together and when they wish to separate.
According to the general law, a society is more stable whose free energy is lower. That is, a mixed society living in the same country is more stable than separated populations existing separately from each other, if the free energy, F mix , of the mixed society is lower than the free energy, F sep , of the separated society:
F mix < F sep .
The free energies can be represented as
F mix = E mix T S mix , F sep = E sep T S sep .
Therefore, denoting the entropy of mixing
Δ S mix 1 N ( S mix S sep ) ,
one has the condition of stability of the united country:
E mix E sep < N T Δ S mix .
We need to write down the energy of the mixed and separated populations. To this end, let us denote the density of an n-th population by ρ n ( r ) and the interaction between the members of an m-th and n-th populations by Φ m n ( r ) , where r represents the member’s location. Then, the interaction energy for the mixed populations can be represented as
E mix = 1 2 m n V ρ m ( r ) Φ m n ( r r ) ρ n ( r ) d r d r .
Assuming a uniform distribution of the people across the country implies the unform densities,
ρ m ( r ) = N m V , ρ n ( r ) = N n V .
Thus, the interaction energy of a country with mixed populations reads as
E mix = 1 2 m n Φ m n N m N n V ,
where the quantity
Φ m n V Φ m n ( r ) d r
describes the average interaction strength between the members of m-th and n-th populations.
Now, let us consider the case of separated populations, where each population lives in a separate location characterized by the volume V n . Then, the energy of a separated country is the sum,
E sep = 1 2 n V n ρ n sep ( r ) Φ n n ( r r ) ρ n sep ( r ) d r d r .
Again keeping in mind that each population inside their part of the country is uniformly distributed gives the uniform densities,
ρ n sep ( r ) = N n V n .
Then, the energy of a separated country is
E sep = 1 2 n Φ ¯ n n N n 2 V n ,
where
Φ ¯ n n V n Φ n n ( r ) d r .
Using the identity,
V V m = n V m V n ,
reduces the energy of a separated country to
E sep = 1 2 m n Φ ¯ m m N m 2 N n V V m .
Any two different populations, although separated, exert pressure on each other through their common boundary. If the pressure of one of them is greater than that of the other, there is no equilibrium between the populations, but there can arise nonequilibrium movements, such as invasions and wars. The equilibrium coexistence of the two separated populations implies the equality of their pressures:
P m = P n P n = F sep V n .
By its meaning, this equality can be called a no-war condition. For not too strong noise, such that
F sep V m E sep V m = 1 2 Φ ¯ m m N m V m 2 ,
one finds the no-war condition in the form,
Φ ¯ m m Φ ¯ n n = N n V m N m V n 2 .
Equation (182) shows that two neighboring populations have no war and are in equilibrium with each other only when the signs of the effective interactions of the members inside each population are the same and the interactions satisfy the above conditions. An m-th population and a n-th population can be in equilibrium only when their internal interactions Φ m m and Φ n n are both either positive or negative. If one population has Φ m m > 0 , while the members of the other population have Φ n n < 0 , then there can be no equilibrium between such neighboring populations.
Taking account of the no-war condition (182) makes it possible to rewrite the energy of the separated country in the form,
E sep = 1 2 m n N m N n V Φ ¯ m m Φ ¯ n n .
Then, the condition of stability (171) for the mixed society yields the inequality,
m n N m N n 2 V Φ m n Φ ¯ m m Φ ¯ n n < N T Δ S mix .
With the identity,
N 1 N m n N m N n ,
one finds:
m n N m N n 2 V Φ m n Φ ¯ m m Φ ¯ n n 2 T ρ Δ S mix < 0 ,
where ρ = N / V is the average density of the total population. From Equation (185), the sufficient condition of stability for the mixed country becomes
Φ m n < Φ ¯ m m Φ ¯ n n + 2 T ρ Δ S mix .
The entropy of mixture can be represented as
Δ S mix = m n m ln n m n m N m N .
Thus, one obtains the condition of stability for the country with the mixed population, as compared to the country where the populations are separated,
Φ m n Φ ¯ m m Φ ¯ n n < 2 T ρ m n m ln n m .
When this condition is not met, the country with the mixed population is not stable and will disintegrate into several separate countries, each composed of just one population type. There exist numerous examples of countries that have disintegrated into several smaller countries. This destiny, for example, has happened with the Roman Empire, Makedonsky Empire, British Empire, French Empire, Austrian Empire, Ottoman Empire, Soviet Union, Yugoslavia, and Czechoslovakia.

3.12. Forced Coexistence

From history, it is known that empires are generally kept united not merely by the economic advantage of the countries composing them but also by force. Why is this possible, and what happens when the required force becomes too strong? Is it possible to keep different populations together inside one empire by increasing force? To answer these questions, it is necessary to consider the case of imposed regulations, including regulation cost.
Let the regulation force, acting on a member of an n-th population be f n ( r ) . The energy of a society with mixed populations, including the force applied to keep different populations together, can be written as
E mix = 1 2 m n V ρ m ( r ) Φ m n ( r r ) ρ n ( r ) d r d r n V f n ( r ) ρ n ( r ) d r +
+ 1 2 m n V ρ m ( r ) f m ( r ) A m n ( r r ) f n ( r ) ρ n ( r ) d r d r ,
where the last term characterizes the cost of supporting the regulation force, with A m n ( r ) being the strength of the interaction between the members of the society caused by the force applied to them.
It is possible to assume that the population densities are uniform and the force is constant, so that
ρ n ( r ) = N n V , f n ( r ) = f n .
Then, the societal energy reads:
E mix = 1 2 m n ( Φ m n + A m n f m f n ) N m N n V n f n N n ,
where
A m n V A m n ( r ) d r .
When the society is disintegrated into several independent countries, there is no need to apply force; hence, the energy of a society with separated populations is Equation (183). Using the identity,
n f n N n = 1 2 N m n N m N n ( f m + f n )
and following the same way as in Section 3.11, one obtains the sufficient condition for the society with mixed populations
Φ m n + A m n f m f n < Φ ¯ m m Φ ¯ n n + 2 ρ T Δ S mix + f m + f n 2 .
For brevity, condition (193) can be named the condition of empire stability. If the condition (193) does not hold, the empire disintegrates into separate countries.
The condition (193) shows that switching on the regulating force first increases the right-hand side of the inequality, thus making the mixed society more stable. However, quite a strong force makes the left-hand side of the inequality larger, thus leading to the instability of the empire and its disintegration. Hence, by enforcing reasonable regulations, it is possible to maintain the coexistence of populations even when, without such an enforcement, the country would disintegrate. However, the regulating force should not be quite strong. This explains why, for some time, an empire can exist as a united family of different populations. However, when enforcement is too costly, it becomes unsupportable, and disintegration becomes inevitable. Empires do not exist forever.
An empire with forced coexistence of different populations is almost always less stable than several independent countries with their own populations and weaker regulation forces. However, the unions of different populations can also exist when they are kept together not by force but by their collaborative interactions. A prominent example is Switzerland, where the German, French, and Italian cantons peacefully coexist, not being forced but due to the advantage of their mutual interactions.

4. Collective Decisions in a Society

The models considered above allow us to describe the general structures and states of social systems whose members exhibit specific behaviors. Thus, in the yes–no model, the members are assumed to make decisions, either “yes” or “no”, with regard to problems, say voting in elections or buying something. As one can see, the underlying process of any action is the process of taking decisions. One can say that actually all properties of a social system are caused by the decisions made by its members. This is why it is so important to have a general understanding of how decisions are made. In this Section, some models are considered describing how members of a society make decisions. Collective decisions are based on decisions made by individuals interacting with each other. Therefore, first of all, it is necessary to understand how decisions are made by individuals and then to model how they interact amongst themselves to reach a collective decision. Moreover, there are deep parallels between the processes of making decisions by individuals and by groups since even individual decision-making is a kind of collective decision-making accomplished by the neurons of a brain. Since the notions of decision-making theory are less known to the physically oriented audience, we give an overview below of the main literature and describe the basic points of the theory.

4.1. General Overview

Nowadays, the dominant theory describing the individual behaviors of decision-makers is expected utility theory. This theory was introduced by Bernoulli [52] when investigating the so-called St. Petersburg paradox. Von Neumann and Morgenstern [53] axiomatized this theory. Savage [54] integrated into the theory the notion of subjective probability. The power of the theory was demonstrated by Arrow [55], Pratt [56], and Rothschild and Stiglitz [57,58] in their studies on risk aversion. The flexibility of the theory for characterizing the attitudes of decision-makers toward risk was illustrated by Friedman and Savage [59] and Markowitz [60]. The expected utility theory has provided a mathematical basis for several fields of economics, finance, and management, including the theory of games, the theory of investment, and the theory of search [61,62,63,64,65,66,67,68,69,70].
Despite many successful applications of the expected decision theory, quite a number of researchers have discovered a large body of evidence that decision-makers, both human as well as animal, often do not obey the prescriptions of the theory and depart from this theory in a rather systematic way [71]. There then started to appear numerous publications, beginning with Allais [72], Edwards [73,74], and Ellsberg [75], that experimentally confirmed systematic deviations from the prescriptions of expected utility theory, leading to a number of paradoxes.
As examples of such paradoxes, it is possible to mention the Allais paradox [72], independence paradox [72], Ellsberg paradox [75], Kahneman–Tversky paradox [76], Rabin paradox [77], Ariely paradox [71], disjunction effect [78], conjunction fallacy [79,80], isolation effects [81], planning paradox [82], and dynamic inconsistency [83,84].
In order to avoid paradoxes, there have been many attempts to change the expected utility theory, which have been named non-expected utility theories. There are a number of such non-expected utility theories, among which we mention just a few of the best-known approaches. These are the prospect theory [73,76,85], weighted-utility theory [86,87,88], regret theory [89], optimism–pessimism theory [90], dual-utility theory [91], ordinal-independence theory [92], and quadratic-probability theory [93]. More detailed information on this topic can be found in the recent reviews by Camerer, Loewenstein, and Rabin [94] and by Machina [95].
Despite numerous attempts to resolve the expected utility theory, none of the suggested modifications can explain all of the paradoxes, as has been shown by Safra and Sigal [96]. The best that could be achieved is to fit the interpretation of just one or two of the paradoxes, with the other paradoxes remaining unexplained. Moreover, spoiling the structure of the expected utility theory results in the appearance of several inconsistencies. Accomplishing a detailed analysis, Al-Najjar and Weinstein [97,98] concluded that any variation in the expected utility theory “ends up creating more paradoxes and inconsistences than it resolves”.
An original interpretation was advanced by Bohr [99,100,101], who suggested that psychological processes could be described by resorting to quantum notions, such as interference and complementarity. Von Neumann [102] mentioned that the theory of quantum measurements could be interpreted as decision theory. These ideas were developed in quantum decision theory [103,104], on the basis of which the classical decision-making paradoxes could be explained. It has also been shown [105,106] that quantum decision theory can be reformulated into classical language without employing quantum formulas.
When decision-makers interact with each other, the process of decision-making becomes collective [107,108,109,110,111,112]. Then, each individual takes decisions based not solely on personal deliberations but also, to some extent, on imitating the actions of other members of the society. Sometimes, this imitation grows to the level of a herding effect. In the present chapter, we describe the main steps of how the process of decision-making develops, starting from individual decisions and progressing to collective decisions taken by a network of society members.

4.2. Utility Function

The primary elements in a problem of choice are some events, outcomes, consequences, or payoffs denoted as x. One considers a set of outcomes, a set of payoffs, a consumer set, or a field of events,
X = { x i : i = 1 , 2 , , N } .
Generally, payoffs, x i , can be either finite or infinite, and the payoff set as well can be finite or infinite. A mathematically correct definition of infinity is the limit of a sequence.
Payoffs or outcomes have to be measured in a common system of units, accomplished through a utility function u ( x ) , also called an elementary utility function, a pleasure function, a satisfaction function, or a profit function u ( x ) : X R . The utility function has to satisfy the following properties:
(i) It has to be nondecreasing, so that
u ( x 1 ) u ( x 2 ) ( x 1 x 2 ) .
If Equation (195) is a strict inequality, the function is termed strictly increasing.
(ii) It is often (although not always) taken to be concave, when
u ( α 1 x 1 + α 2 x 2 ) α 1 u ( x 1 ) + α 2 u ( x 2 ) ,
where
α 1 0 , α 2 0 , α 1 + α 2 = 1 .
It is called strictly concave, if the inequality (196) represents a strict inequality.
A twice differentiable function is nondecreasing if
u ( x ) d u ( x ) d t 0
and it is concave when
u ( x ) d 2 u ( x ) d x 2 0 .
The derivative u ( x ) defines the marginal utility function. According to the above properties, the marginal utility does not increase.
An important property of a utility function is its risk aversion. The degree of absolute risk aversion is measured [56] by the quantity,
r ( x ) u ( x ) u ( x ) .
The coefficient of relative risk aversion [55,56] is defined as
r ˜ ( x ) x r ( x ) = x u ( x ) u ( x ) .
For a concave utility function, the degree of risk aversion is non-negative; hence, the utility function is risk averse, r ( x ) 0 . This means that with increasing x, the growth of the utility function u ( x ) does not increase.
Actually, one often employs a linear utility function,
u ( x ) = k x ( k > 0 ) .
For function (199),
u ( x ) = k , u ( x ) = 0 ;
hence, the degree of risk aversion is zero, r ( x ) = 0 .
Another used form is a power-law utility function,
u ( x ) = k x γ ( k > 0 , 0 γ < 1 ) ,
assuming x 0 . For this case,
u ( x ) = k γ x 1 γ , u ( x ) = k γ ( 1 γ ) x 2 γ .
The degree of risk aversion diminishes with increasing x as
r ( x ) = 1 γ x > 0 .
One more example is the logarithmic utility function,
u ( x ) = c ln ( 1 + k x ) ( c > 0 , k > 0 ) ,
again, keeping in mind x 0 . Then,
u ( x ) = c k 1 + k x , u ( x ) = c k 2 ( 1 + k x ) 2 .
Hence, the degree of risk aversion diminishes with the increasing payoff x,
r ( x ) = c k 1 + k x .
Sometimes, one uses an exponential utility function,
u ( x ) = c 1 e k x ( c > 0 , k > 0 ) ,
for which
u ( x ) = c k e k x , u ( x ) = c k 2 e k x ,
and the degree of risk aversion is constant, r ( x ) = k .
In the above examples, the utility function is twice differentiable, such that u ( x ) and u ( x ) exist; it is non-decreasing and concave, hence risk-averse; it is also non-negative for non-negative x; and it is normalized to zero, so that u ( 0 ) = 0 , implying that the utility of nothing is zero.
Some of the utility functions exemplified above are defined only for positive payoffs x > 0 . In the case of losses, payoffs are negative. Then, one defines different utility functions for gains and losses, for example, u g ( x ) for gains and u l ( x ) for losses. Certainly, these should coincide at zero, so that u g ( 0 ) = u l ( 0 ) .
It is also useful to mention that there are two types of utility: cardinal and ordinal. Cardinal utility can be precisely measured, and the magnitude of the measurement is meaningful, similar to how distance is measured in meters, time in hours, or weight in kilograms. For ordinal utility, its precise magnitude is not important, but the magnitude of the ratios between different utilities only is meaningful.

4.3. Expected Utility

The notion of expected utility was introduced by Bernoulli [52], and axiomatic utility theory was formulated by von Neumann and Morgenstern [53]. The definitions below follow the classical exposition of von Neumann and Morgenstern [53].
One defines a probability measure over the set of payoffs,
{ p ( x i ) [ 0 , 1 ] : i = 1 , 2 , , N } ,
with the probabilities, p ( x i ) , normalized to one,
i = 1 N p ( x i ) = 1 .
The probabilities (203) can be objective, prescribed by a rule [53], or they can be subjective probabilities, evaluated by a decision-maker [54].
The probability distribution over a set of payoffs is termed a lottery,
L = { x i , p ( x i ) : i = 1 , 2 , , N } .
A compound lottery is a lottery whose outcomes are other lotteries. For two lotteries
L 1 = { x i , p 1 ( x i ) } , L 2 = { x i , p 2 ( x i ) } ,
the compound lottery is the linear combination,
α 1 L 1 + α 2 L 2 = { x i , α 1 p 1 ( x i ) + α 2 p 2 ( x i ) } ,
where
α 1 0 , α 2 0 , α 1 + α 2 = 1 .
The lottery mean is the lottery expected value,
E ( L ) i = 1 N x i p ( x i ) .
The lottery volatility, lottery spread, or lottery dispersion is the variance,
var ( L ) i = 1 N x i 2 p ( x i ) E 2 ( L ) .
The lottery volatility, or lottery dispersion, is a measure of lottery uncertainty.
The expected utility of a lottery is the function,
U ( L ) i = 1 N u ( x i ) p ( x i ) .
The expected utility (209) is proportional to the lottery mean in the case of a linear utility function. Lotteries are ordered, implying the following definitions.
Lotteries L 1 and L 2 are indifferent if, and only if,
U ( L 1 ) = U ( L 2 ) ( L 1 = L 2 ) .
A lottery L 1 is preferred to L 2 if, and only if,
U ( L 1 ) > U ( L 2 ) ( L 1 > L 2 ) .
A lottery L 1 is preferred or indifferent to L 2 if, and only if,
U ( L 1 ) U ( L 2 ) ( L 1 L 2 ) .
The expected utility satisfies the following properties.
(i) Completeness. For any two lotteries L 1 and L 2 , one of the conditions is valid:
L 1 = L 2 , L 1 < L 2 , L 1 > L 2 , L 1 L 2 , L 1 L 2 .
(ii) Transitivity. For any three lotteries, such that L 1 L 2 and L 2 L 3 , it follows that L 1 L 3 .
(iii) Continuity. For any three lotteries, ordered so that L 1 L 2 L 3 , there exists α [ 0 , 1 ] for which
L 2 = α L 1 + ( 1 α ) L 3 .
(iv) Independence. For any L 1 L 2 and arbitrary L 3 , there exists α [ 0 , 1 ] , such that
α L 1 + ( 1 α ) L 3 α L 2 + ( 1 α ) L 3 .
The properties (213)–(215) follow directly from the properties of the utility function described above.
The standard decision-making process proceeds through the following steps. For the same set of payoffs { x n : n = 1 , 2 , . . . , N } , there exist several lotteries,
L n = { x i , p n ( x i ) } ( n = 1 , 2 , ) .
Defining the utility function u ( x ) , one calculates the expected lottery utilities, U ( L n ) . Comparing the values U ( L n ) , one selects the largest among them.
A lottery L * is named optimal when it is preferred to all others from the given choice of lotteries. A lottery is optimal if, and only if, its expected utility is the largest:
U ( L * ) sup n U ( L n ) .
The aim of a decision-maker is assumed to choose an optimal lottery maximizing the related expected utility.

4.4. Time Preference

Betweenwhiles, people need to compare present goods that are available for use at the present time with future goods that are defined as present expectations of goods becoming available at some date in the future. Time preference is the insight that people prefer present goods to future goods [84,113].
A mathematical description of the time preference effect can be performed as follows. Let us consider at time t = 0 a lottery,
L = { x i , p ( x i ) } ( t = 0 ) .
With a utility function u ( x ) at time t = 0 , the expected utility of the lottery is
U ( L ) = i u ( x i ) p ( x i ) ( t = 0 ) .
The lottery, expected at time t > 0 , has the form,
L ( t ) = { x i ( t ) , p ( x i ( t ) ) } ( t > 0 ) .
Denoting the utility function at time t > 0 as u ( x ( t ) , t ) , we have the expected utility of L ( t ) as
U ( L ( t ) ) = i u ( x i ( t ) , t ) p ( x i ( t ) ) .
According to the meaning of time preference, the same goods at a future time are valued lower than at the present time just because any goods can be used during the interval of time [ 0 , t ] . In the case of money, its value increases with time since it can bring additional profit through an interest rate; hence, the utility of a fixed amount of money decreases with time. This can be formalized as the inequality,
u ( x i , t ) < u ( x i ) ( t > 0 ) .
One may introduce a discount function, D ( x , t ) , by the relation,
u ( x i , t ) = u ( x i ) D ( x i , t ) ,
with the apparent condition,
D ( x i , 0 ) = 1 .
From the time preference condition (222):
D ( x i , t ) < 1 ( t > 0 ) .
In this way, if the payoffs at the present time t = 0 and at a future time t > 0 are the same, then the present lottery is preferable over the future one,
U ( L ( t ) ) < U ( L ) ( x i ( t ) = x i ) .
In particular, if the discount function is uniform with respect to the payoffs,
D ( x i , t ) = D ( t ) ,
then
U ( L ( t ) ) = U ( L ) D ( t ) ,
where D ( 0 ) = 1 .
In decision-making, one uses the following discount functions.
Power-law discount function,
D ( t ) = 1 ( 1 + r ) t .
Exponential discount function,
D ( t ) = exp ( γ t ) .
Note that the power-law discount function (229) is equivalent to the exponential form (230), since they are related by the reparametrization,
1 + r = e γ , γ = ln ( 1 + r ) .
Hyperbolic discount function,
D ( t ) = 1 ( 1 + t / τ ) γ τ ,
with positive parameters γ and τ .
A detailed review on the effect of time preference is given in Ref. [84].

4.5. Stochastic Utility

There exists a number of factors that influence decision-making. These factors are random or stochastic, and because of this, the related approach to decision-making is named stochastic. These factors, for instance, are:
(i) external conditions under which the decision is made, such as weather, situation in the country, relations with people, opinions of other people, etc.;
(ii) internal physical state of the decision-maker, such as fatigue, illness, pain, influence of alcohol, etc.;
(iii) internal psychological state, including biases, emotions, mood, impatience, etc.;
(iv) previous experience, information, prejudices, religion, principles, character, etc.;
(v) occasional errors in decisions.
Stochastic decision theory assumes that all these interrelated factors can be characterized by some variables, for example, ξ , called states of nature. The total collection of the states of nature forms the nature set, { ξ } . The variables ξ are random and stochastic. It is supposed that there should be a probability measure, μ ( ξ ) , on the nature set, { ξ } . The utility function, u ( x , ξ ) , becomes a random variable. The probability of payoffs, p ( x , ξ ) , is also a random variable.
In this way, a lottery becomes a stochastic lottery
L ( ξ ) = { x i , p ( x i , ξ ) : i = 1 , 2 , , N } .
Respectively, one comes to a stochastic utility,
U ( L , ξ ) = i = 1 N u ( x i , ξ ) p ( x i , ξ ) .
As far as the states on nature are random, one needs to average over these states; thus, coming to the expected utility,
U ( L ) U ( L , ξ ) d μ ( ξ ) .
The process of stochastic decision-making consists in the choice between several stochastic lotteries,
L n ( ξ ) = { x i , p n ( x i , ξ ) : i = 1 , 2 , , N } .
Technically, this implies the choice between several expected utilities,
U ( L n ) = U ( L n , ξ ) d μ ( ξ ) .
One needs to choose a lottery with the largest expected utility U ( L * ) . This approach, however, confronts several quite serious difficulties:
(i) it is not clear how the random variables should be incorporated into lotteries;
(ii) calculations become rather complicated;
(iii) the nature set is not fixed, generally, depending on time;
(iv) the nature set also can depend on the set of payoffs;
(v) the explicit form of the nature states probability is not known and needs to be postulated.
More details on stochastic utility theory can be found, for example, in the books [114,115].

4.6. Affective Decisions

People make decisions based not merely on rational grounds by calculating utility, but also when affected by emotions, which are irrational. Several attempts have been made to take account of emotions in decision-making by modifying expected utility, which is equivalent to some variants of non-expected utility models [91,116,117]. Here, we present the main points of a probabilistic approach to taking emotions into account. First, this approach was formulated by resorting to techniques of quantum theory [103,104]; however, later, it was shown [105,106] that it can be reformulated into classical terms without invoking any quantum expressions. The basics of the probabilistic affective decision theory are as follows.
The aim of any decision-making process is to choose from a set of several alternatives. Let the set of alternatives be denoted as
A = { A n : n = 1 , 2 , , N A } .
Each alternative from this set is assumed to be equipped with a probability, p ( A n ) , with the normalization condition,
n p ( A n ) = 1 , 0 p ( A n ) 1 .
This probability shows how it is probable that the alternative A n can be chosen.
An alternative, A 1 , is said to be stochastically preferable to A 2 if, and only if,
p ( A 1 ) > p ( A 2 ) .
Two alternatives are called stochastically indifferent if, and only if,
p ( A 1 ) = p ( A 2 ) .
An alternative, A opt , is stochastically optimal if its probability is maximal:
p ( A opt ) = sup n p ( A n ) .
The usefulness of an alternative is characterized by a utility factor,  f ( A n ) , whose form is to be prescribed by normative rules. This factor shows the probability of choosing an alternative, A n , based on a rational understanding of its utility. The standard probability normalization is applied:
n f ( A n ) = 1 , 0 f ( A n ) 1 .
One says that an alternative A 1 is more useful than A 2 if, and only if,
f ( A 1 ) > f ( A 2 ) .
Two alternatives are equally useful if, and only if,
f ( A 1 ) = f ( A 2 ) .
Affective features of an alternative are represented by an attraction factor,  q ( A n ) , with the normalization,
n q ( A n ) = 0 , 1 q ( A n ) 1 .
Positive attraction factors imply attractive alternatives, while negative attraction factors mean that the corresponding alternatives are repulsive. An alternative A 1 is more attractive than A 2 if, and only if,
q ( A 1 ) > q ( A 2 ) .
Two alternatives are equally attractive if, and only if,
q ( A 1 ) = q ( A 2 ) .
The probability, p ( A n ) , of an alternative A n is a function of the related utility factor, f ( A n ) , and of attraction factor, q ( A n ) , such that the rational boundary condition is valid:
p ( A n ) = f ( A n ) , q ( A n ) = 0 ,
when, in the absence of emotions, the probability of an alternative coincides with the rational utility factor.
The simplest form of the probability function satisfying the rational boundary condition (248) is prescribed by the superposition axiom,
p ( A n ) = f ( A n ) + q ( A n ) .
An explicit expression for the utility factor can be derived from the principle of minimal information in Section 2. Assume that for each alternative, A n , there exists a utility functional, U ( A n ) , so that an average utility is given by the standard condition,
n f ( A n ) U ( A n ) = U .
Then, the information function, under conditions (242) and (250), reads:
I [ f ( A n ) ] = n f ( A n ) ln f ( A n ) f 0 ( A n ) + α 1 n f ( A n ) + β U n f ( A n ) U ( A n ) ,
where f 0 ( A n ) is a trial distribution. Minimizing the information function with respect to the utility factor f ( A n ) , yields
f ( A n ) = f 0 ( A n ) e β U ( A n ) n f 0 ( A n ) e β U ( A n ) .
The trial distribution, f 0 ( A n ) , can be taken following the Luce rule [118,119,120],
f 0 ( A n ) = a n n a n ,
where a n is the attribute of the alternative A n , having the form,
a n = U ( A n ) , U ( A n ) 0 ,
for semi-positive utility functions and
a n = 1 | U ( A n ) | , U ( A n ) < 0 ,
for negative utility functions [106,121,122].

4.7. Wisdom of Crowds

Wisdom of crowds is the notion that assumes that large groups of people are collectively smarter than individual members of the same groups. This concerns any kind of problem-solving and decision-making. The justification of this idea is based on the understanding that the viewpoint of an individual can inherently be biased, being influenced by the individual’s emotions and prejudices, whereas taking the average knowledge of a crowd results in the elimination of the noise of subjective biases and emotions; thus, producing a wiser aggregated result [123,124,125,126]. When discussing the wisdom of crowds, one keeps in mind the following crowd features: (i) The crowd should have a diversity of opinions; (ii) Each personal opinion should remain independent to those around them, not being influenced by anyone else; (ii) Each individual from the crowd should make their own decision based solely on their individual knowledge. These conditions exclude the situation where crowd members consult with each other and mimic the actions of their neighbors, which can lead to herding effects. The latter is considered in Section 4.8 below.
Let us enumerate the members of a crowd or a society by the index j = 1 , 2 , N . According to the conisderation in Section 4.6, each member of the considered group chooses an alternative A n with the probability,
p j ( A n ) = f j ( A n ) + q j ( A n ) ,
under the standard normalization condition,
n = 1 N A p j ( A n ) = 1 , 0 p j ( A n ) 1 .
The aggregate opinion implies the arithmetic averaging over the society members, which yields the average probability,
p ( A n ) 1 N j = 1 N p j ( A n ) ,
composed of the superposition of the average utility factor,
f ( A n ) 1 N j = 1 N f j ( A n ) ,
and the average attraction factor,
q ( A n ) 1 N j = 1 N q j ( A n ) .
Thus, coming to expression (249).
The utility factor is prescribed by a rational evaluation of the utility of the considered alternatives; hence, it is weakly dependent on subjective emotions. This means that f j ( A n ) is approximately the same for any group member, which is equivalent to the condition,
f j ( A n ) = f ( A n ) .
Notice that the normalization conditions,
n = 1 N A f j ( A n ) = 1 , 0 f j ( A n ) 1
remain valid.
The attraction factor, on the contrary, is subjective, being essentially influenced by the agent’s emotions. In that sense, the attraction factor is a random quantity. It is random because of several causes. First, its randomness is due to the variability of emotions experienced by rather different people. Second, emotions, even of the same person, vary at different times. Third, emotions randomly influence choice due to the generic variability and local instability of neural networks in the brain, as has been found in numerous psychological and neurophysiological studies [127,128,129,130,131,132,133,134,135].
Nevertheless, despite the intrinsic randomness of the attraction factor, some of its aggregate properties can be well defined. First of all, the alternation law is satisfied:
n = 1 N A q j ( A n ) = 0 .
Equation (262) follows straightforwardly from Equations (256) and (261). The expression (256) also tells us that the attraction factor for a j-th society member is in the range
f j ( A n ) q j ( A n ) 1 f j ( A n ) .
Respectively, the aggregate attraction factor varies in the interval
f ( A n ) q ( A n ) 1 f ( A n ) .
The set of alternatives can be separated into three classes, depending on the signs of the attraction factors, the class of positive attraction factors,
q + ( A n ) = q ( A n ) > 0 ,
negative attraction factors,
q ( A n ) = q ( A n ) < 0 ,
and zero attraction factors when q ( A n ) = 0 . According to the limits (264), positive attraction factors are in the interval
0 < q + ( A n ) < 1 f ( A n ) ,
while negative attraction factors are in the range
f ( A n ) < q ( A n ) < 0 .
Non-informative priors for the attraction factors can be estimated by means of the related arithmetic averages. Recall that, if a quantity y lies in an interval [ a , b ] , its arithmetic average is y ¯ = ( a + b ) / 2 . If the interval limits are in the ranges a 1 a a 2 and b 1 b b 2 , then they are expressed through their averages, so that
y ¯ = 1 2 a ¯ + b ¯ = 1 2 a 1 + a 2 2 + b 1 + b 2 2 .
Keeping in mind that 0 < q + ( A n ) < 1 f ( A n ) and f ( A n ) < q ( A n ) < 0 , while 0 f ( A n ) 1 , one obtains the quarter law,
q + ( A n ) = 1 4 , q ( A n ) = 1 4 .
Employing the non-informative priors for the aggregate attraction factors, one can estimate the probability of alternatives, averaged over the crowd, as
p ( A n ) = f ( A n ) ± 1 4 ,
depending on whether the alternatives on average are attractive or not. The quarter law has been found to be in quite a good agreement with empirical data [106,136].

4.8. Herding Effect

In Section 4.7, the process of decision-making by a crowd of independent agents was considered. However, generally, the members of a society interact with each other, which can result in drastic changes in the agents’ behavior, such as the occurrence of the herding effect [137,138,139,140,141,142,143]. There can exist two kinds of interactions between the society members, which can lead to collective effects such as herding. First, the members can mimic the actions of others, replicating their behavior. Second, the members communicate through information exchange. Strictly speaking, the correct description of collective effects arising in the process of agents’ interactions requires the study of temporal processes, just because collective effects need time to form and develop. The consideration of temporal collective effects is beyond the scope of the present article. However, due to their importance for social systems, we delineate the principal points of how collective interactions can be incorporated into affective decision theory. Further details can be found in Refs. [144,145,146].
Multistep decision theory deals with quantities depending on time. Then, the probability of choosing by an agent j an alternative A n at time t is p j ( A n , t ) , with the normalization,
n = 1 N A p j ( A n , t ) = 1 , 0 p j ( A n , t ) 1 .
The utility factor becomes f j ( A n , t ) , with the normalization,
n = 1 N A f j ( A n , t ) = 1 , 0 f j ( A n , t ) 1 ,
and the attraction factor reads as q j ( A n , t ) , satisfying the conditions,
n = 1 N A q j ( A n , t ) = 0 , 1 q j ( A n , t ) 1 .
Taking into account the tendency of society members to replicate the actions of others defines the probability dynamics,
p j ( A n , t + 1 ) = ( 1 ϵ j ) f j ( A n , t ) + q j ( A n , t ) + ϵ j N 1 i ( j ) N f i ( A n , t ) + q i ( A n , t ) ,
where ε j is a replication parameter satisfying the condition,
0 ϵ j 1 ( j = 1 , 2 , , N ) .
As is explained in Section 4.6 above, there are two types of agent interactions in a society: the replication of the actions of other members and the exchange of information. The latter is known to attenuate the influence of emotions, which results in the attenuation of the attraction factor. The attenuation of emotional influence due to agent interactions is well confirmed by empirical observations [147,148,149,150,151,152,153,154,155,156,157,158]. The attenuation of the attraction factor is described [144,145,146,158] by the form,
q j ( A n , t ) = q j ( A n ) exp { M j ( t ) } ,
where q j ( A n ) is the attraction factor of an agent j in the absence of social interactions, and M j ( t ) is the amount of information received at the moment of time t by an agent j. The quantity M j ( t ) that, for brevity, can be called the memory, is expressed as
M j ( t ) = t = 0 t i = 1 N J j i ( t , t ) μ j i ( t ) .
Here, J j i ( t , t ) is the information transfer function from agent i to agent j in the interval of time from t to t, and μ j i ( t ) is the Kullback–Leibler information gain (13) received by agent j from agent i at time t,
μ j i ( t ) = n = 1 N A p j ( A n , t ) ln p j ( A n , t ) p i ( A n , t ) .
The interaction of the society members through replication and information exchange results in a rich variety of behavioral types, including herding, periodic cycles, and chaotic fluctuations, which depend on the society’s parameters. A detailed analysis of multistep decision-making by intelligent members of a society is given in Refs. [144,145,146].

5. Conclusions

In this review, the principle of minimal information has been formulated, which gives the key for constructing probability distributions for equilibrium and quasi-equilibrium social systems. Several simple enough examples, based on the yes–no model, have been considered. Despite their seeming simplicity, the models allow one to describe rather nontrivial effects, including the role of regulation costs and the existence of fluctuating groups inside a society. Quite surprisingly, it turns out that the occurrence of self-organized disorder can make a society more stable. The peculiarity of coexisting populations explains when these populations can mix and live peacefully in the same country and when this coexistence becomes unstable, so that the country separates with different populations in different locations. Since, behind all actions of any society, there are decisions of its members, the basics of the probabilistic affective decision theory are delineated.
The author hopes that the material in the above survey has given the reader a feeling of the wide possibilities of applying mathematical models of physics to describe social systems, even being limited to equilibrium systems. The other part [7] of the review will be devoted to nonequilibrium systems studying various evolution equations and dynamical effects.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Acknowledgments

The author very much appreciates the hospitality of and numerous discussions with Didier Sornette, without whose influence this work would not have been done. I am grateful for the discussions and help of Elizaveta P. Yukalova.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Waldrop, M.M. Complexity: The Emerging Science at the Edge of Order and Chaos; Touchstone/Simon & Schuster: New York, NY, USA, 1993; Available online: https://uberty.org/wp-content/uploads/2017/04/Waldrop-M.-Mitchell-Complexity-The-Emerging-Science-at-Edge-of-Order-and-Chaos.pdf (accessed on 12 May 2023).
  2. Bar-Yam, Y. Dynamics of Complex Systems; Addison-Wesley: Reading, MA, USA, 1997; Available online: https://necsi.edu/dynamics-of-complex-systems (accessed on 12 May 2023).
  3. Mitchell, M. Complexity: A Guided Tour; Oxford University Press: New York, NY, USA, 2009; Available online: https://dokumen.tips/documents/complexity-a-guided-tour-56b13a81a2c4f.html (accessed on 12 May 2023).
  4. Thurner, S.; Hanel, R.; Klimekl, P. Introduction to the Theory of Complex Systems; Oxford University Press: New York, NY, USA, 2018. [Google Scholar] [CrossRef]
  5. Ladyman, J.; Wiesner, K. What Is a Complex System; Yale University Press: New Haven, CT, USA, 2020. [Google Scholar] [CrossRef]
  6. Browning, R. Browning’s Complete Poetical Works (The Complete Poetic and Dramatic Works of Robert Browning. Cambridge Edition); Houghton, Mifflin and Company: Boston, MA, USA, 1895. [Google Scholar]
  7. Yukalov, V.I. Selected topics of social physics: Nonequilibrium systems. Physics, 2023; 5, to be published. [Google Scholar]
  8. Loewenstein, G.; Rick, S.; Cohen, J.D. Neuroeconomics. Annu. Rev. Psychol. 2008, 59, 647–672. [Google Scholar] [CrossRef] [PubMed]
  9. Simon, H.A. A behavioral model of rational choice. Quart. J. Econ. 1955, 69, 99–118. [Google Scholar] [CrossRef]
  10. Shannon, C.E. A mathematical theory of communication. Bell Syst. Techn. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  11. Gibbs, J.W. The Collected Works; Longmans, Green and Co.: New York, NY, USA, 1928; Volume 1, Available online: https://gallica.bnf.fr/ark:/12148/bpt6k95192s.image (accessed on 12 May 2023).
  12. Gibbs, J.W. The Collected Works; Yale University Press: New Haven, CT, USA, 1948; Volume 2, Available online: https://archive.org/details/collectedworksof02gibb/page/n5/mode/2up (accessed on 12 May 2023).
  13. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  14. Kullback, S. Information Theory and Statistics; Peter Smith: Gloucester, MA, USA, 1978; Available online: https://doc.lagout.org/Others/Information%20Theory/Information%20Theory/Information%20theory%20and%20statistics%20-%20Solomon%20Kullback.pdf (accessed on 12 May 2023).
  15. Shore, J.E.; Johnson, R.W. Axiomatic derivation of the principle of maximum entropy and the principle of minimum cross-entropy. IEEE Trans. Inf. Theory 1980, 26, 26–37. [Google Scholar] [CrossRef] [Green Version]
  16. Janes, E.T. Information theory and statistical mechanics. Phys. Rev. 1957, 106, 620–630. [Google Scholar] [CrossRef]
  17. Jaynes, E.T. Probability Theory: The Logic of Science; Cambridge University Press: New York, NY, USA, 2003; Available online: http://www.med.mcgill.ca/epidemiology/hanley/bios601/GaussianModel/JaynesProbabilityTheory.pdf (accessed on 12 May 2023).
  18. Yukalov, V.I. Nonequilibrium representative ensembles for isolated quantum systems. Phys. Lett. A 2011, 375, 2797–2801. [Google Scholar] [CrossRef] [Green Version]
  19. Ter Haar, D. Elements of Statistical Mechanics; Reed Elsevier plc group/Butterworth-Heinemann Ltd: Oxford, UK, 1995. [Google Scholar] [CrossRef]
  20. Ter Haar, D. Theory and applications of the density matrix. Rep. Prog. Phys. 1961, 24, 304–362. [Google Scholar] [CrossRef]
  21. Yukalov, V.I. Phase transitions and heterophase fluctuations. Phys. Rep. 1991, 208, 395–492. [Google Scholar] [CrossRef]
  22. Yukalov, V.I. Fluctuations of composite observables and stability of statistical systems. Phys. Rev. E 2005, 72, 066119. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Yukalov, V.I. Representative ensembles in statistical mechanics. Int. J. Mod. Phys. B 2007, 21, 69–86. [Google Scholar] [CrossRef] [Green Version]
  24. Yukalov, V.I. Representative statistical ensembles for Bose systems with broken gauge symmetry. Ann. Phys. 2008, 323, 461–499. [Google Scholar] [CrossRef] [Green Version]
  25. Kubo, R. Thermodynamics; North-Holland Publishing Company: Amsterdam, The Netherlands, 1968; Available online: https://www.scribd.com/doc/279300920/Kubo-R-Thermodynamics-NH-1968-T-310s-pdf (accessed on 12 May 2023).
  26. Yukalov, V.I. Theory of cold atoms: Basics of quantum statistics. Laser Phys. 2013, 23, 062001. [Google Scholar] [CrossRef] [Green Version]
  27. Yukalov, V.I. Stochastic instability of quasi-isolated systems. Phys. Rev. E 2002, 65, 056118. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Yukalov, V.I. Irreversibility of time for quasi-isolated systems. Phys. Lett. A 2003, 308, 313–318. [Google Scholar] [CrossRef] [Green Version]
  29. Yukalov, V.I. Expansion exponents for nonequilibrium systems. Phys. A 2003, 320, 149–168. [Google Scholar] [CrossRef] [Green Version]
  30. Yukalov, V.I. Equilibration of quasi-isolated quantum systems. Phys. Lett. A 2012, 376, 550–554. [Google Scholar] [CrossRef] [Green Version]
  31. Yukalov, V.I. Decoherence and equilibration under nondestructive measurements. Ann. Phys. 2012, 327, 253–263. [Google Scholar] [CrossRef] [Green Version]
  32. Dembo, A.; Cover, T.M.; Thomas, J.A. Information theoretic inequalities. IEEE Trans. Inform. Theory 1991, 37, 1501–1517. [Google Scholar] [CrossRef] [Green Version]
  33. Weidlich, W. Physics and social science: The approach of synergetics. Phys. Rep. 1991, 204, 1–163. [Google Scholar] [CrossRef]
  34. Parsons, T. The Social System; Taylor & Francis Group/Routledge: London, UK, 2005; Available online: https://voidnetwork.gr/wp-content/uploads/2016/10/The-Social-System-by-Talcott-Parsons.pdf (accessed on 12 May 2023).
  35. Galam, S. Sociophysics: A Physicist’s Modeling of Psycho-Political Phenomena; Springer Science+Business Media, LLC: New York, NY, USA, 2012. [Google Scholar] [CrossRef]
  36. Perc, M.; Gomez-Gardenes, J.; Szolnoki, A.; Floria, L.M.; Moreno, Y. Evolutionary dynamics of group interactions on structured populations: A review. J. Roy. Soc. Interface 2013, 10, 20120997. [Google Scholar] [CrossRef] [Green Version]
  37. Perc, M.; Jordan, J.J.; Rand, D.G.; Wang, Z.; Boccaletti, S.; Szolnoki, A. Statistical physics of human cooperation. Phys. Rep. 2017, 687, 1–51. [Google Scholar] [CrossRef] [Green Version]
  38. Jusup, M.; Holme, P.; Kanazawa, K.; Takayasu, M.; Romic, I.; Wang, Z.; Gecek, S.; Lipic, T.; Podobnik, B.; Wang, L.; et al. Social physics. Phys. Rep. 2022, 948, 1–148. [Google Scholar] [CrossRef]
  39. Yukalov, V.I.; Yukalova, E.P. Zeroth-order nucleation transition under nanoscale phase separation. Symmetry 2021, 13, 2379. [Google Scholar] [CrossRef]
  40. Huang, K. Statistical Mechanics; John Wiley & Sons, Inc.: New York, NY, USA, 1987; Available online: https://physicsgg.files.wordpress.com/2016/09/huang-kerson-1987-statistical-mechanics-2ed-wileyt506s.pdf (accessed on 12 May 2023).
  41. Kubo, R. Statistical Mechanics; North-Holland Publishing Company: Amsterdam, The Netherlands, 1965; Available online: https://emineter.files.wordpress.com/2018/08/kubo_zbirka.pdf (accessed on 12 May 2023).
  42. Isihara, A. Statistical Physics; Academic Press, Inc.: New York, NY, USA, 1971. [Google Scholar] [CrossRef]
  43. Kadanoff, L.P. Statistical Physics. Statics, Dynamics and Renormalization; World Scientific Co. Ltd: Singapore, 2000. [Google Scholar] [CrossRef]
  44. Sornette, D. Physics and financial economics (1776–2014): Puzzles, Ising and agent-based models. Rep. Prog. Phys. 2014, 77, 062001. [Google Scholar] [CrossRef] [PubMed]
  45. Brush, S.G. History of the Lenz-Ising model. Rev. Mod. Phys. 1967, 39, 883–893. [Google Scholar] [CrossRef]
  46. Hayek, F. The Road to Serfdom; Taylor & Francis Group/Routledge: New York, NY, USA, 2006; Reprint of 1944 edition; Available online: https://ctheory.sitehost.iu.edu/img/Hayek_The_Road_to_Serfdom.pdf (accessed on 12 May 2023).
  47. Yukalov, V.I. Mesoscopic phase fluctuations: General phenomenon in condensed matter. Int. J. Mod. Phys. B 2003, 17, 2333–2358. [Google Scholar] [CrossRef] [Green Version]
  48. Yukalov, V.I. Systems with symmetry breaking and restoration. Symmetry 2010, 2, 40–68. [Google Scholar] [CrossRef] [Green Version]
  49. Yukalov, V.I. Renormalization of quasi-Hamiltonians under heterophase averaging. Phys. Lett. A 1987, 125, 95–100. [Google Scholar] [CrossRef]
  50. Yukalov, V.I. Procedure of quasi-averaging for heterophase mixtures. Physica A 1987, 141, 352–374. [Google Scholar] [CrossRef]
  51. Yukalov, V.I. Lattice mixtures of fluctuating phases. Physica A 1987, 144, 369–389. [Google Scholar] [CrossRef]
  52. Bernoulli, D. Specimen Theoriae Novae de Mensura Sortis. Comment. Academ. Scient. Imper. Petropolit. [Proc. Imper. Acad. Sci. St. Petersburg] 1738, V, 175–192. English translation: Exposition of a new theory on the measurement of risk. In The Kelly Capital Growth Investment Criterion. Theory and Practice; MacLean, L.C., Thorp, E.O., Ziemba, W.T., Eds.; World Scientific Co. Ltd.: Singapore, 2011; pp. 11–24. [Google Scholar] [CrossRef]
  53. von Neumann, J.; Morgenstern, O. Theory of Games and Economic Behavior; Princeton University Press: Princeton, NJ, USA, 1953. [Google Scholar] [CrossRef]
  54. Savage, L.J. The Foundations of Statistics; Dover Publications, Inc.: New York, NY, USA, 1972; Available online: https://gwern.net/doc/statistics/decision/1972-savage-foundationsofstatistics.pdf (accessed on 12 May 2023).
  55. Arrow, K.J. Essays in the Theory of Risk Bearing; North Holland Publishing Company: Amsterdam, The Netherlands, 1974; Available online: https://archive.org/details/essaysintheoryof0000arro (accessed on 12 May 2023).
  56. Pratt, J.W. Risk aversion in the small and in the large. Econometrica 1964, 32, 122–136. [Google Scholar] [CrossRef]
  57. Rothschild, M.; Stiglitz, J. Increasing risk: A definition. J. Econ. Theory 1970, 2, 225–243. [Google Scholar] [CrossRef]
  58. Rothschild, M.; Stiglitz, J. Increasing risk: Its economic consequences. J. Econ. Theory 1971, 3, 66–84. [Google Scholar] [CrossRef]
  59. Friedman, M.; Savage, L. The utility analysis of choices involving risk. J. Polit. Econ. 1948, 56, 279–304. [Google Scholar] [CrossRef]
  60. Markovitz, H. The utility of wealth. J. Polit. Econ. 1952, 60, 151–158. [Google Scholar] [CrossRef]
  61. Lindgren, B.W. Elements of Decision Theory; Macmillan: New York, NY, USA, 1971. [Google Scholar]
  62. White, D.I. Decision Theory; Taylor & Francis Group/Routledge: New York, NY, USA, 2006. [Google Scholar] [CrossRef]
  63. Rivett, P. Model Building for Decision Analysis; John Wiley & Sons: Chichester, UK, 1980. [Google Scholar]
  64. Berger, J.O. Statistical Decision Theory and Bayesian Analysis; Springer Science+Business Media New York: New York, NY, USA, 1985. [Google Scholar] [CrossRef]
  65. Marshall, K.T.; Oliver, R.M. Decision Making and Forecasting; McGraw-Hill, Inc.: New York, NY, USA, 1995. [Google Scholar]
  66. Bather, J. Decision Theory: An Introduction to Dynamic Programming and Sequential Decisions; John Wiley & Sons, Ltd.: Chichester, UK, 2000; Available online: https://archive.org/details/decisiontheoryin0000bath (accessed on 12 May 2023).
  67. French, S.; Rios Insua, D. Statistical Decision Theory. Kendall’s Library of Statistics 9; John Wiley & Sons Ltd/Hodder Arnold: London, UK, 2000. [Google Scholar]
  68. Raiffa, H.; Schlaifer, R. Applied Statistical Decision Theory; John Wiley & Sons: New York, NY, USA, 2000. [Google Scholar]
  69. Weirich, P. Decision Space; Cambridge University Press: Cambridge, UK, 2001. [Google Scholar] [CrossRef] [Green Version]
  70. Gollier, C. Economics of Risk and Time; The MIT Press: Cambridge, MA, USA, 2001. [Google Scholar] [CrossRef] [Green Version]
  71. Ariely, D. Predictably Irrational; Harper Collins Publishers: New York, NY, USA, 2008; Available online: https://radio.shabanali.com/predictable.pdf (accessed on 12 May 2023).
  72. Allais, M. Le comportement de l’homme rationnel devant le risque: Critique des postulats et axiomes de l’ecole Americaine. Econometrica 1953, 21, 503–546. [Google Scholar] [CrossRef] [Green Version]
  73. Edwards, W. The prediction of decision among bets. J. Experim. Psychol. 1955, 50, 200–204. [Google Scholar] [CrossRef]
  74. Edwards, W. Subjective probabilities inferred from decisions. Psychol. Rev. 1962, 69, 109–135. [Google Scholar] [CrossRef]
  75. Ellsberg, D. Risk, ambiguity, and the Savage axioms. Quart. J. Econ. 1961, 75, 643–669. [Google Scholar] [CrossRef] [Green Version]
  76. Kahneman, D.; Tversky, A. Prospect theory: An analysis of decision under risk. Econometrica 1979, 47, 263–291. [Google Scholar] [CrossRef] [Green Version]
  77. Rabin, M. Risk aversion and expected-utility theory: A calibration theorem. Econometrica 2000, 68, 1281–1292. [Google Scholar] [CrossRef] [Green Version]
  78. Tversky, A.; Shafir, E. The disjunction effect in choice under uncertainty. Psychol. Sci. 1992, 3, 305–309. [Google Scholar] [CrossRef]
  79. Tversky, A.; Kahneman, D. Extensional versus intuitive reasoning: The conjunction fallacy in probability judgement. Psychol. Rev. 1983, 90, 293–315. [Google Scholar] [CrossRef]
  80. Shafir, E.B.; Smith, E.E.; Osherson, D.N. Typicality and reasoning fallacies. Mem. Cognit. 1990, 18, 229–239. [Google Scholar] [CrossRef] [Green Version]
  81. McCaffery, E.J.; Baron, J. Isolation effects and the neglect of indirect effects of fiscal policies. J. Behav. Decis. Mak. 2006, 19, 289–302. [Google Scholar] [CrossRef]
  82. Kydland, F.E.; Prescott, E.C. Rules rather than discretion: The inconsistency of optimal plans. J. Polit. Econ. 1977, 85, 473–492. [Google Scholar] [CrossRef] [Green Version]
  83. Strotz, R.H. Myopia and inconsistency in dynamic utility maximization. Rev. Econ. Stud. 1955, 23, 165–180. [Google Scholar] [CrossRef]
  84. Frederick, S.; Loewenstein, G.; O’Donoghue, T. Time discounting and time preference: A critical review. J. Econ. Literat. 2002, 40, 351–401. [Google Scholar] [CrossRef]
  85. Tversky, A.; Kahneman, D. Advances in prospect theory: Cumulative representations of uncertainty. J. Risk Uncert. 1992, 5, 297–323. [Google Scholar] [CrossRef]
  86. Karmarkar, U. Subjectively weighted utility: A descriptive extension of the expected utility model. Org. Behav. Hum. Perform. 1978, 21, 61–72. [Google Scholar] [CrossRef]
  87. Karmarkar, U. Subjectively weighted utility and the Allais paradox. Org. Behav. Hum. Perform. 1979, 24, 67–72. [Google Scholar] [CrossRef]
  88. Chew, S.H. A generalization of the quasilinear mean with applications to the measurement of income inequality and decision theory resolving the Allais paradox. Econometrica 1983, 51, 1065–1092. [Google Scholar] [CrossRef]
  89. Loomes, G.; Sugden, R. Regret theory: An alternative theory of rational choice under uncertainty. Econ. J. 1982, 92, 805–824. [Google Scholar] [CrossRef]
  90. Hey, J. The economics of optimism and pessimism: A definition and some applications. Kyklos 1984, 37, 181–205. [Google Scholar] [CrossRef]
  91. Yaari, M. The dual theory of choice under risk. Econometrica 1987, 55, 95–115. [Google Scholar] [CrossRef] [Green Version]
  92. Green, J.; Jullien, B. Ordinal independence in nonlinear utility theory. J. Risk Uncert. 1988, 1, 355–387. [Google Scholar] [CrossRef]
  93. Chew, S.; Epstein, L.; Segal, U. Mixture symmetry and quadratic utility. Econometrica 1991, 59, 139–163. [Google Scholar] [CrossRef]
  94. Camerer, C.F.; Loewenstein, G.; Rabin, R. Advances in Behavioral Economics; Princeton University Press: Princeton, NJ, USA, 2004. [Google Scholar] [CrossRef]
  95. Machina, M.J. Non-expected utility theory. In The New Palgrave Dictionary of Economics. Second Edition; Durlauf, S.N., Blume, L.E., Eds.; Macmillan Publishers Ltd/Palgrave Macmillan: Hampshire, UK, 2008; pp. 4586–4596. [Google Scholar] [CrossRef]
  96. Safra, Z.; Segal, U. Calibration results for non-expected utility theories. Econometrica 2008, 76, 1143–1166. [Google Scholar] [CrossRef]
  97. Al-Najjar, N.I.; Weinstein, J. The ambiguity aversion literature: A critical assessment. Econ. Philos. 2009, 25, 249–284. [Google Scholar] [CrossRef] [Green Version]
  98. Al-Najjar, N.I.; Weinstein, J. Rejoinder: The ambiguity aversion literature: A critical assessment. Econ. Philos. 2009, 25, 357–369. [Google Scholar] [CrossRef]
  99. Bohr, N. Light and life. Nature 1933, 131, 421–423. [Google Scholar] [CrossRef] [Green Version]
  100. Bohr, N. Light and life. Nature 1933, 131, 457–459. [Google Scholar] [CrossRef] [Green Version]
  101. Bohr, N. Atomic Physics and Human Knowledge; John Wiley & Sons, Inc.: New York, NY, USA, 1958; Available online: https://www.holybooks.com/wp-content/uploads/Atomic-Physics-and-Human-Knowledge.pdf (accessed on 12 May 2023).
  102. von Neumann, J. Mathematical Foundations of Quantum Mechanics, New Edition; Wheeler, N.A., Ed.; Princeton University Press: Princeton, NJ, USA, 2018. [Google Scholar] [CrossRef]
  103. Yukalov, V.I.; Sornette, D. Quantum decision theory as quantum theory of measurement. Phys. Lett. A 2008, 372, 6867–6871. [Google Scholar] [CrossRef]
  104. Yukalov, V.I.; Sornette, D. Mathematical structure of quantum decision theory. Adv. Compl. Syst. 2010, 13, 659–698. [Google Scholar] [CrossRef] [Green Version]
  105. Yukalov, V.I. A resolution of St. Petersburg paradox. J. Math. Econ. 2021, 97, 102537. [Google Scholar] [CrossRef]
  106. Yukalov, V.I. Quantification of emotions in decision making. Soft Comput. 2022, 26, 2419–2436. [Google Scholar] [CrossRef]
  107. Krause, J.; Ruxton, G.D. Living in Groups; Oxford University Press: Oxford, UK, 2002. [Google Scholar]
  108. Seeley, T.D. Honeybee Democracy; Princeton University Press: Princeton, NJ, USA, 2010; Available online: https://hadinur1969.files.wordpress.com/2017/10/thomas_d_seeley-honeybee_democracy_-princeton_univ.pdf (accessed on 12 May 2023).
  109. Easley, D.; Kleinberg, J. Networks, Crowds, and Markets: Reasoning about a Highly Connected World; Cambridge University Press: New York, NY, USA, 2010. [Google Scholar] [CrossRef]
  110. Bourke, A.F.G. Principles of Social Evolution; Oxford University Press Inc.: New York, NY, USA, 2011. [Google Scholar] [CrossRef]
  111. Marshall, J.A.R. Social Evolution and Inclusive Fitness Theory: An Introduction; Princeton University Press: Princeton, NJ, USA, 2015. [Google Scholar]
  112. Hamann, H. Swarm Robotics: A Formal Approach; Springer International Publishing AG: Cham, Switzerland, 2018. [Google Scholar] [CrossRef]
  113. Samuelson, P. A note on measurement of utility. Rev. Econ. Stud. 1937, 4, 155–161. [Google Scholar] [CrossRef]
  114. Pratt, J.W.; Raiffa, H.; Schlaifer, R. Introduction to Statistical Decision Theory; The MIT Press: Cambridge, MA, USA, 2008. [Google Scholar]
  115. Liese, F.; Miescke, K.-J. Statistical Decision Theory: Estimation, Testing, and Selection; Springer: New York, NY, USA, 2008. [Google Scholar] [CrossRef] [Green Version]
  116. Reynaa, V.F.; Brainer, C.J. Dual processes in decision making and developmental neuroscience: A fuzzy-trace model. Developm. Rev. 2011, 31, 180–206. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  117. Bracha, A.; Brown, D.J. Affective Decision Making: A Theory of Optimism Bias. Games Econ. Behav. 2012, 1, 67–80. [Google Scholar] [CrossRef] [Green Version]
  118. Luce, R.D. Individual Choice Behavior: A Theoretical Analysis; Wiley: New York, NY, USA, 1959; Available online: https://www.scribd.com/book/271527880/Individual-Choice-Behavior-A-Theoretical-Analysis (accessed on 12 May 2023).
  119. Luce, R.D.; Raiffa, R. Games and Decisions: Introduction and Critical Survey; Dover Publications, Inc.: New York, NY, USA, 1989; Available online: https://www.scribd.com/book/271636770/Games-and-Decisions-Introduction-and-Critical-Survey (accessed on 12 May 2023).
  120. Gul, F.; Natenzon, P.; Pesendorfer, W. Random choice as behavioral optimization. Econometrica 2014, 82, 1873–1912. [Google Scholar] [CrossRef] [Green Version]
  121. Yukalov, V.I.; Sornette, D. Manupulating decision making of typical agents. IEEE Trans. Syst. Man Cybern. Syst. 2014, 44, 1155–1168. [Google Scholar] [CrossRef] [Green Version]
  122. Yukalov, V.I.; Sornette, D. Quantitative predictions in quantum decision theory. IEEE Trans. Syst. Man Cybern. Syst. 2018, 48, 366–381. [Google Scholar] [CrossRef] [Green Version]
  123. Surowiecki, J. The Wisdom of Crowds; Random House, Inc./Anchor Books: New York, NY, USA, 2005. [Google Scholar]
  124. Sunstein, C. Infotopia: How Many Minds Produce Knowledge; Oxford University Press: Oxford, UK, 2006. [Google Scholar]
  125. Page, S.E. The Difference: How the Power of Diversity Creates Better Groups, Firms, Schools, and Societies; Princeton University Press: Princeton, NJ, USA, 2007; Available online: https://www.scribd.com/book/232949830/The-Difference-How-the-Power-of-Diversity-Creates-Better-Groups-Firms-Schools-and-Societies-New-Edition (accessed on 12 May 2023).
  126. Fiechter, J.; Kornell, N. How the wisdom of crowds, and of the crowd within, are affected by expertise. Cogn. Res. 2021, 6, 5. [Google Scholar] [CrossRef]
  127. Werner, G.; Mountcastle, V.B. The variability of central neural activity in a sensory system and its implications for the central reflection of sensory events. J. Neurophysiol. 1963, 26, 958–977. [Google Scholar] [CrossRef]
  128. Arieli, A.; Sterkin, A.; Grinvald, A.; Aertsen, A. Dynamics of ongoing activity: Explanation of the large variability in evoked cortical responses. Science 1986, 273, 1868–1871. [Google Scholar] [CrossRef] [Green Version]
  129. Gold, J.I.; Shadlen, M.N. Neural computations that underlie decisions about sensory stimuli. Trends Cognit. Sci. 2001, 5, 10–16. [Google Scholar] [CrossRef]
  130. Glimcher, P.W. Indeterminacy in brain and behaviour. Annu. Rev. Psychol. 2005, 56, 25–56. [Google Scholar] [CrossRef] [Green Version]
  131. Schumacher, J.F.; Thompson, S.K.; Olman, C.A. Contrast response functions for single Gabor patches: ROI-based analysis over-represents low-contrast patches for GE BOLD. Front. Syst. Neurosci. 2011, 5, 19. [Google Scholar] [CrossRef] [Green Version]
  132. Shadlen, M.N.; Shohamy, D. Perspective decision making and sequential sampling from memory. Neuron 2016, 90, 927–939. [Google Scholar] [CrossRef]
  133. Webb, R. The neural dynamics of stochastic choice. Manag. Sci. 2019, 65, 230–255. [Google Scholar] [CrossRef] [Green Version]
  134. Kurtz-David, V.; Persitz, D.; Webb, R.; Levy, D.J. The neural computation of inconsistent choice behaviour. Nature Commun. 2019, 10, 1583. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  135. Woodford, M. Modeling imprecision in perception, valuation, and choice. Annu. Rev. Econ. 2020, 12, 579–601. [Google Scholar] [CrossRef]
  136. Yukalov, V.I.; Sornette, D. Preference reversal in quantum decision theory. Front. Psychol. 2015, 6, 1538. [Google Scholar] [CrossRef] [Green Version]
  137. Martin, E.D. The Behavior of Crowds: A Psychological Study; Harper & Brothers: New York, NY, USA, 1920; Available online: https://www.gutenberg.org/files/40914/40914-h/40914-h.htm (accessed on 12 May 2023).
  138. Sherif, M. The Psychology of Social Norms; Harper & Brothers Publishers: New York, NY, USA, 1936; Available online: https://archive.org/details/in.ernet.dli.2015.264611 (accessed on 12 May 2023).
  139. Smelser, N.J. Theory of Collective Behavior; Macmillan Company/The Free Press: New York, NY, USA, 1965; Available online: https://archive.org/details/theoryofcollecti00smel/mode/2up (accessed on 12 May 2023).
  140. Merton, R.K. Social Theory and Social Structure; Macmillan Publishing Co., Inc./The Free Press: New York, NY, USA, 1968; Available online: https://edisciplinas.usp.br/pluginfile.php/4250035/mod_folder/content/0/Textos/Merton%2C%20Social%20Theory%20and%20Social%20Structure.pdf (accessed on 12 May 2023).
  141. Turner, R.H.; Killian, L.M. Collective Behavior; Prentice-Hall: Englewood Cliffs, NJ, USA, 1993. [Google Scholar]
  142. Hatfield, E.; Cacioppo, J.T.; Rapson, R.L. Emotional Contagion; Cambridge University Press: New York, NY, USA, 1993. [Google Scholar] [CrossRef]
  143. Brunnermeier, M.K. Asset Pricing under Asymmetric Information: Bubbles, Crashes, Technical Analysis, and Herding; Oxford University Press: New York, NY, USA, 2001. [Google Scholar] [CrossRef]
  144. Yukalov, V.I.; Yukalova, E.P.; Sornette, D. Information processing by networks of quantum decision makers. Physica A 2018, 492, 747–766. [Google Scholar] [CrossRef] [Green Version]
  145. Yukalov, V.I.; Yukalova, E.P.; Sornette, D. Role of collective information in networks of quantum operating agents. Physica A 2022, 598, 127365. [Google Scholar] [CrossRef]
  146. Yukalov, V.I.; Yukalova, E.P. Self-excited waves in complex social systems. Physica D 2022, 433, 133188. [Google Scholar] [CrossRef]
  147. Kühberger, A.; Komunska, D.; Perner, J. The disjunction effect: Does it exist for two-step gambles? Org. Behav. Hum. Decis. Process. 2001, 85, 250–264. [Google Scholar] [CrossRef] [PubMed]
  148. Charness, G.; Rabin, M. Understanding social preferences with simple tests. Quart. J. Econ. 2002, 117, 817–869. [Google Scholar] [CrossRef] [Green Version]
  149. Blinder, A.; Morgan, J. Are two heads better than one? An experimental analysis of group versus individual decision-making. J. Money Credit Bank. 2005, 37, 789–811. [Google Scholar]
  150. Cooper, D.; Kagel, J. Are two heads better than one? Team versus individual play in signaling games. Am. Econ. Rev. 2005, 95, 477–509. [Google Scholar] [CrossRef] [Green Version]
  151. Sutter, M. Are four heads better than two? An experimental beauty-contest game with teams of different size. Econ. Lett. 2005, 88, 41–46. [Google Scholar] [CrossRef] [Green Version]
  152. Tsiporkova, E.; Boeva, V. Multi-step ranking of alternatives in a multi-criteria and multi-expert decision making environment. Inform. Sci. 2006, 176, 2673–2697. [Google Scholar] [CrossRef]
  153. Charness, G.; Karni, E.; Levin, D. Individual and group decision making under risk: An experimental study of bayesian updating and violations of first-order stochastic dominance. J. Risk Uncert. 2007, 35, 129–148. [Google Scholar] [CrossRef] [Green Version]
  154. Charness, G.; Rigotti, L.; Rustichini, A. Individual behavior and group membership. Am. Econ. Rev. 2007, 97, 1340–1352. [Google Scholar] [CrossRef] [Green Version]
  155. Chen, Y.; Li, S. Group identity and social preferences. Am. Econ. Rev. 2009, 99, 431–457. [Google Scholar] [CrossRef] [Green Version]
  156. Liu, H.H.; Colman, A.M. Ambiguity aversion in the long run: Repeated decisions under risk and uncertainty. J. Econ. Psychol. 2009, 30, 277–284. [Google Scholar] [CrossRef] [Green Version]
  157. Charness, G.; Karni, E.; Levin, D. On the conjunction fallacy in probability judgement: New experimental evidence regarding Linda. Games Econ. Behav. 2010, 68, 551–556. [Google Scholar] [CrossRef] [Green Version]
  158. Yukalov, V.I.; Sornette, D. Role of information in decision making of social agents. Int. J. Inf. Technol. Decis. Mak. 2015, 14, 1129–1166. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Yes–no model with regulation cost. Behavior of the free energy, F, as a function of the regulation force, h, for different temperatures (noise intensity): (a) T < 1 and (b) T > 1 .
Figure 1. Yes–no model with regulation cost. Behavior of the free energy, F, as a function of the regulation force, h, for different temperatures (noise intensity): (a) T < 1 and (b) T > 1 .
Physics 05 00041 g001
Figure 2. Yes–no model with regulation cost. The order parameter, s, as a function of the regulation force, h, for (a) low ( T < 1 ) and (b) high ( T > 1 ) temperatures.
Figure 2. Yes–no model with regulation cost. The order parameter, s, as a function of the regulation force, h, for (a) low ( T < 1 ) and (b) high ( T > 1 ) temperatures.
Physics 05 00041 g002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yukalov, V.I. Selected Topics of Social Physics: Equilibrium Systems. Physics 2023, 5, 590-635. https://doi.org/10.3390/physics5020041

AMA Style

Yukalov VI. Selected Topics of Social Physics: Equilibrium Systems. Physics. 2023; 5(2):590-635. https://doi.org/10.3390/physics5020041

Chicago/Turabian Style

Yukalov, Vyacheslav I. 2023. "Selected Topics of Social Physics: Equilibrium Systems" Physics 5, no. 2: 590-635. https://doi.org/10.3390/physics5020041

APA Style

Yukalov, V. I. (2023). Selected Topics of Social Physics: Equilibrium Systems. Physics, 5(2), 590-635. https://doi.org/10.3390/physics5020041

Article Metrics

Back to TopTop