1. Introduction
By studying certain properties of the surrounding world, a physicist would usually formulate some tentative theories. Such theories are usually composed of three parts: (1) Basic (initial) concepts, (2) laws (or principles), and (3) a number of assumptions and idealizations (models). In the case of mechanics (in what follows, we shall base our reasoning on mechanics as the most developed and fundamental branch of physics), the first part may consist of concepts of time, space, mass, etc.; the second may be represented by the Newton laws, including the law of gravity and the principle of least action. In mechanics, there are many assumptions and idealizations. For example, these include point mass, free body, closed system, perfectly elastic collision, Euclidean (as well as isotropic, homogeneous) space, etc. When corollaries of the tentative theory are analyzed, especially when compared with experiment, a number of problems arise which, in turn, lead to refinements and modifications of the tentative theory (
Figure 1a). Thanks to such refinements, contradictions between the components forming the three parts of the theory are eliminated or smoothed out. It should be noted that experimental data are interpreted within the framework of the tentative theory in its current, provisional form. After a number of iterations, the theory looks more or less complete (to the majority of scientists) for a rather long period until some new, “unexpected” problem comes up. This results in new and sometimes substantial adjustments in the tentative theory causing the so-called scientific revolution. This is a well-known process described by historians and philosophers of science long ago (specifically by Popper, Kuhn, and Lakatos) (
Figure 1a). In the history of physics and technology, this method proves to be very effective and productive. However, such a development of theory has methodological disadvantages. Let us list them.
1. Basic concepts (the “language” of a theory) are usually based neither on measurements nor on logical operations. They are introduced by definition (axiomatically), based on conscious and/or subconscious experience gained through interactions with the surrounding world (observations and experiments). The best example of this is the introduction of time in mechanics (see, for example, the well-known and respected textbook on theoretical physics [
1]). The conventional description fails to give any idea of how time should be measured or introduced (without being based on laws of mechanics obtained only on the assumption that time exists and has some properties). Why can time be considered homogeneous, regular, expressed in real numbers, etc.? These questions are not new and were raised many times, e.g., by Mach, Poincaré, and Bridgman.
2. Theories as a whole (basic concepts and laws/principles, as well as assumptions and idealizations) adapt to experimental data. In the case of mechanics, this led to assertions, for example, that if all coordinates and velocities are specified, it allows one to fully determine the state of a mechanical system and predict its motion, or, for instance, that gravitational force is exactly inversely proportional to the square of the distance between bodies [
1]. All three parts of a theory (basic concepts, laws, and models) effectively have one origin and validation tool: observation and experiment. In the light of the above, dividing a theory into three parts is, in many respects, a very approximate procedure. In essence, these parts are identical and they form a
single interrelated system that evolves historically, following the diagram in
Figure 1a, to describe some natural phenomena. Due to the absence of strong logical connections between the parts of a theory, its iterative adaptations to experimental data enable the change of any of its elements rather freely (a typical example here is the well-known “transformation” of classical mechanics in quantum or relativistic cases when problems with experimental results were found). Such an optimization is accompanied by an unlimited number of degrees of freedom. As a result, such theories can no longer be trusted as unique and fundamental. In spite of this conclusion, the history of science shows that basic concepts and laws (principles) are often assigned meanings that essentially go beyond the framework of their original tentative theories and are regarded as universally valid. For example, the laws of classical thermodynamics are freely applied to cosmology. Obviously, given the origin of a tentative theory, such an “extrapolation” can prove to be either very bold and fruitful, or risky and erroneous.
3. Different tentative theories exploring different, sometimes unrelated areas of science (not necessarily only natural science) can have similar basic concepts. However, every tentative theory develops
independently (
Figure 1b). Iterations and optimizations in individual theories can lead to basic concepts acquiring increasingly different properties. As a result, science is divided into weakly interacting parts, scientific knowledge loses its universality, and the gap of understanding between scientists of various specializations widens. Undoubtedly, all this has a negative effect on scientific progress as a whole. For example, the concept of time in science now has a different meaning not only for biologists, psychologists, and historians but also for physicists in different fields: thermodynamics, classical and quantum mechanics, or cosmology.
A possible methodological way out of this situation was previously suggested in [
2]. It is schematically shown in
Figure 2a. Let us comment on this diagram. Basic concepts (for example, time) should be formulated in such a way that they can be defined (measured or calculated) in a uniform manner for every branch of the modern science (liberal arts included). Then, by studying some phenomenon of the surrounding world, a model (a set of axioms, idealizations, and assumptions) should be made. Thereafter, using the introduced basic concepts and model, laws (corollaries) should be
deductively obtained and then compared with experiments and analyzed. In case contradictions or problems are found, the model should be adjusted with the cycle repeating. Further, the basic concepts should be treated in an extremely conservative way. These concepts concern and bind together the entire scientific framework and its multiple theories, thus forming the basis of our scientific worldview (
Figure 2b). So, in comparison with the old method of theory formulation (
Figure 1), the new one has the following main features:
- (1)
Basic concepts are introduced explicitly, operationally (constructively), and universally. On these specific grounds, the concepts may be called basic.
- (2)
Laws are a logical corollary of the deductive development of basic concepts with the help of models. So, the strict hierarchy is established: Basic concept + model → law.
- (3)
During the analysis and comparison with experiments, only elements of the model are optimized. As a result, there is a drastic reduction in the number of degrees of freedom when bridging the gap between theoretical predictions and experimental data.
This method is unusual for physics. However, similar ideas have been expressed before. Notably, Milne established a number of basic concepts and a model of the expanding universe in order to logically derive a number of conclusions in mechanics and electromagnetism, which he later compared with traditionally-established laws [
3]. This is a very interesting, original, and comprehensive study. In his approach, time (clock) is a primary basic concept which is related to the properties of light rays emitted and received by various observers. In Milne’s theory, spatial interval is a secondary concept. In our opinion, the lack of universality in the introduced concept of time is a significant drawback of this method. Indeed, the concept is introduced using light rays, the physical properties of which (including constant velocity in a vacuum) determine the properties of time. Such “time” can hardly be introduced and used in other scientific disciplines.
In the paper [
2] (developing ideas mentioned in [
4,
5]), a measure of time related to the concept of informational entropy was introduced. The essential properties of this measure are its universality (i.e., applicability to both natural sciences and liberal arts) and its constructive potential. The latter means that, on the basis of this measure, one could deduce dynamic laws, if required. In comparison with the traditional approach, such a method of developing a physical theory (e.g., mechanics), is rather new and promising. As a consequence, the purpose of this paper is to obtain, based on such time measures (and using a simple model), a number of laws of mechanics which could be subsequently compared with experiments.
Here, one should stress the following. There exists a large range of options for constructing physical theories on the basis of informational entropy. The best known is the classical approach of Jaynes, which obtains statistical mechanics. Derivations of the quantum theory (including its generalizations and classical mechanics limit), using maximum entropy production rate and entropic inference, are presented, for example, in the papers by Beretta, Gheorghiu-Svirschevski, and Caticha [
6,
7,
8,
9,
10]. Caticha’s approach deserves special mention. According to him [
10], "Entropic dynamics is a framework in which dynamical laws such as those that arise in physics are derived as an application of entropic methods of inference." A theory is constructed by maximizing the relative entropy subject to constraints reflecting the information that is relevant to the problem at hand. Caticha introduces "entropic" time as a book-keeping device designed to keep track of the accumulation of change [
10]. Caticha defines time in a special way using transition probability, thus this is an important independent assumption with regard to the inference algorithm used by him. In the approach presented in this article time is defined through entropy directly, and in order to construct a physical theory it is not necessary to maximize this entropy (entropic measure of time) at all times.
2. Measure of Time and Model
Let us assume according to [
2] that time,
τ, up to a dimensionless constant, is equal to a logarithm of the number of possible microstates (Γ) representing some state (macrostate) at a given moment of observation.
where
ε is a number and
S is entropy (in the considered model, entropy turns out to be a familiar configurational (Boltzmann’s) entropy). It is note that all the quantities assumed herein, particularly time and size, are dimensionless. This does not impose any fundamental restrictions but only simplifies notation.
Let us accept two postulates with regard to time [
2]: (1) Time exists (is defined) only if changes in the observer–observed system occur; (2) its value may only increase. These postulates restrict potential changes in the system (specifically, they rule out changes that result in Γ decreasing or being constant).
Let us discuss a crucial issue arising when complex systems (i.e., systems consisting of multiple coexisting subsystems) are considered using the introduced measure of time. An individual time scale can be assigned to each subsystem using entropy and this poses a question: Are these time scales uniform or not in relation to each other? In other words, are elementary intervals of the change of time
for two different scales non-linearly connected? The answer is no, because, as is known from the maximum entropy production principle [
11,
12], coexisting subsystems have the same change of entropy (if the coexisting subsystems have different entropy production (change of entropy), then according to the maximum entropy production principle, the subsystem with the maximum entropy production will be selected during the evolution of a complex system. As a result, such subsystems cannot coexist). According to Equation (1), the change of time is directly related to the change of entropy and as a result the time scales are linearly connected. As a consequence, for such subsystems, laws derived from these time measures using the same model are similar. In this regard, it can be stated that laws for different coexisting subsystems are the same. This statement can be considered as an analog of the Galileo–Einstein relativity principle.
Let us define the measure of length in a traditional way using some accepted standard (e.g., a perfectly rigid rod). Using this measure of length, we shall represent the available space as the number, G, of identical cells. The number of cells, G, thus represents the size of the system under consideration.
Let us consider the following simplest model. There are two types of identical non-interacting particles, b and f, in the space of size G, the number of which are and , respectively (where N is the total number of particles in the system and is the fraction of particles b, ). Particles b and f differ in terms of their possible distribution in the cells of the space. Let us assume that one cell of the space can contain any number of particles b (up to N if α = 1) and not more than one particle f. Particles with such properties resemble the well-known particles bosons and fermions. The principal difference is that, in the case of particles b and f, cells have no associated energy values (the notion of energy will be defined later through the kinetic energy of the expansion).
By the number of microstates we shall mean the number of possible distributions of particles b and f among G cells. A macrostate shall mean a system with some G, α, and N values.
Thus, our model contains four quantities (τ, G, α, and N). Of these, τ and G can be regarded as basic, whilst the other two in the proposed model are auxiliary. Let us make three assumptions: (1) Particles b and f are distributed in cells independently; (2) Time is defined by a specific (per the number of particles) entropy, i.e., ; (3) The fraction of particles α in the system shall be such that the system’s specific entropy reaches its maximum value.
3. Corollaries of the Measure of Time and the Model: A Law
It is well known [
13] that the number (Γ
1) of possible microstates for particles
b corresponding to a given macrostate is:
the number (Γ
2) of possible microstates for particles
f is:
Using the first assumption of the model, we have:
for the specific (per
N) entropy,
s, based on the Stirling’s formula and assuming that
G ≫ 1,
N ≫ 1 and
G ≫ (1 –
α)
N, it can be obtained that:
Using the third assumption of the model, let us find an optimal value of
α for which the system’s specific entropy would be at its maximum. Differentiating (5) with respect to
α and setting it equal to zero, we obtain:
Subject to conditions
,
,
, we get the unique solution of the last equation:
Using (5) and (8), and applying the second assumption of the model, we get:
Expression (law) (9) interrelates all the model quantities. We shall call it the kinematic relationship. It is valid for times larger than the time of maximization of the specific entropy due to the redistribution of particles
b and
f in cells. Using (8) and (9), we can plot the variation of fractions of particles
b and
f with time (
Figure 3). It can be seen in
Figure 3 that with the passage of time the fraction of particles
b decreases from 1 in the superdense state (
) to 0.5 in the low dense state (
).
Figure 4, based on Equation (9), shows how the size of the system and the number of particles depend on time. It can be concluded from
Figure 4 that for any values of
G and
N, time is greater than zero.
According to the second postulate, time in the system can only increase. This allows one to determine the restrictions on possible variations of the parameters. Indeed,
According to (9):
where
. We can show that
.
As a result, (10) is reduced to the form:
So, in order for time to increase, a relative change of the system’s size must be greater than a relative change of the number of particles in the system. In the particular case when the number of particles remains constant, according to (13), the system’s size must inevitably increase, whereas when the system evolves maintaining its constant size, the number of particles must decrease at all times.
Figure 4 illustrates exactly this behavior.
The kinematic relation Equation (9) is rather complex. However, it can be considerably simplified for two extreme cases.
4. Corollaries from the Derived Law
Let us discuss some results directly following from the kinematic law derived above.
1. We define the rate of the system’s size change as
υ = (∂G/∂
τ)
N. According to Equation (12):
The rate of change of the size of the system as a function of its size, number of particles, and time, as calculated on the basis of Equations (9) and (18), is shown in
Figure 6. It can be seen from
Figure 6a that
υ grows with the increase of
G and
N and this change in
G occurs much faster. With time,
υ increases too (
Figure 6b).
For the two extreme cases, dependences of velocity on time can be obtained explicitly. For the superdense state, using (15) and the property of the Lambert function [
14]:
we obtain:
For the state of low density, according to (17), we have:
Interestingly, for the state of low density, according to (17) and (21), the rate at which the system grows (expands) is identically equal to its size.
For the two considered particular cases, the rate of change of the system’s size increases with time (
Figure 7). Additionally, in the superdense state this increase occurs with decreasing acceleration and in the state of low density with increasing acceleration.
2. Let us define the energy of the system as
(where
ρ = N/G is an average density of particles in the system). By definition, this energy is related exclusively to the rate of change of the system’s size and is analogous to the kinetic energy. Based on Equation (18), the system’s energy has the general form (
Figure 8):
For the two extreme cases, we obtain energy explicitly dependent on time. For the superdense state, based on (20) and (15), the system’s energy is:
Based on (17) and (21), the system’s energy for the state of low density has the form:
As can be seen from the above, the system’s energy with time reaches its minimum; it decreases in a dense state but grows exponentially during further evolution and transition to a state of low density. This minimum has no special properties (not an attractor) because we consider an evolving system where time cannot “stop” according to the introduced postulate; it always increases.
3. As mentioned above, the system’s kinetic energy changes and therefore some forces do work on the system. Let us define force as
F = −(∂
E/∂
G)
N, then, using Equation (22), we obtain:
As can be seen, the force has two parts, one of which is positive and the other negative (it can be easily proved that Ω is always positive at any
N and
G). By analogy with the classical representations [
1], let us consider that the first term
F1 in Equation (25) corresponds to “attraction”:
and the second term
F2 represents “repulsion” of particles in the system:
It is seen from
Figure 9 that the attraction between particles decreases and the repulsion increases as the system grows in size. Additionally, at larger sizes both contributions reach constant values. Let us find from (26), (27) an asymptotic behavior of the forces. For the superdense state (
or
N ≫
G ≫ 1), the attractive and the repulsive forces are inversely proportional to the square of the system size. They are:
We note that, at a sufficiently large
N, the repulsive force can be neglected:
F1 ≫
F2. It is important to note here that the system size,
G, is defined as the number of “space boxes” (or “volumetric size”). In a 3D world, the linear size should scale as power
1/
3 of volumetric size. So, in a 3D world, the forces (28) are inversely proportional to the power
2/
3 of the system linear size.
For the state of low density (
or
G ≫
N ≫ 1), the forces of attraction and repulsion are comparable:
Obviously, the forces arising in the considered model have a purely kinematic origin, i.e., the forces are “born” when the measure of time is introduced and then, on this basis, particles in the growing volume of space are considered.