1. Introduction
The history of multiple time boundary conditions goes back—as far as I know—to Schottky [
1,
2], who, in 1921, considered a single slice of time inadequate for prediction or retrodiction (see
Appendix A). There was later work of Watanabe [
3] concerned with prediction and retrodiction. Then, Schulman [
4] uses this as a conceptual way to eliminate “initial conditions” prejudice from Gold’s [
5] rationale for the arrow of time and Wheeler [
6,
7] discusses two time boundary conditions. Gell–Mann and Hartle also contributed to this subject [
8] and include a review of some previous work. Finally, Aharonov et al. [
9,
10] proposed that this could solve the measurement problem of quantum mechanics, although this is disputed [
11]. See also
Appendix B.
This formalism has also allowed me to come to a conclusion: effect follows cause in the direction of entropy increase. This result is not unanticipated; most likely everything having to do with arrows of time has been anticipated. What is unusual, however, is the ability to prove the thesis mathematically.
The proof given [
12] only applies to classical mechanics. The present work extends it to quantum mechanics. Moreover, since the previous paper was buried in conference proceedings, and the present work uses similar arguments, I will repeat some parts of that proof. Note though that the present work is extremely limited: it applies to particles in a gas with Gaussian wave functions. I think it should apply more generally, but that is not what I prove.
A remark is in order on the arrow of time. Often the sequence of cause and effect is taken as primary [
13], but in the present context this is shown not to hold. Our argument depends on two conditions: the nature of perturbation and two-time boundary conditions. Both will be elaborated on. A future condition on quantum problems may make measurement deterministic but is hollow in that it does not lead (to my knowledge) to testable claims. There is the possibility of advanced effects ([
6,
14]) but so far, this has not proved measurable.
3. Review of Previous Work
The argument for causality to follow the direction of entropy increase depends on the dynamics. The principle assumption on the dynamics is that there is a relaxation time. Call it
(this is aside from the cosmological assumptions, discussed below). Let the dynamics be
, so that from time
t to
a subset of
of the phase space goes from
to
. Then, by giving two times separated by more than
and low entropy in the beginning and end, one gets increase, then constant, then decline of the entropy. Thus, the set of points satisfying both boundary conditions (
at 0 and
at
T) at time 0 is
Note that both beginning and end are forced to have low entropy (by definition, assuming
and
are small). This is illustrated for the “cat map” in the first graph of
Figure 1. See
Appendix C for information on the “cat map” and on our procedure.
Then, we do something different! Two times for perturbation are contemplated: the first during the rising of the entropy, the second during its fall. By detailed calculation (not reproduced here, but see below,
Section 5) it is found that causality, the macroscopic change in behavior, follows the direction of entropy increase, that is in the falling entropy case, the macroscopic behavior changes earlier in the time parameter (which is neutral, i.e., has two different arrows and in between doesn’t have an arrow).
Intuitively, this is simple. Let the perturbation be at and . Between 0 and (before the perturbation) both perturbed and unperturbed have the same boundary conditions, hence their macroscopic behavior is similar (but the microscopic behavior is, in general, different). After the perturbation there is no constraint, i.e., no effective boundary condition: both systems go to equilibrium, from which there is adequate time (recall ) to reach (whether perturbed or not) a specific region within the unit square. There was however a perturbation and there are different microscopic and macroscopic paths. Similarly, for the boundary conditions at and (just after the perturbation in the neutral parameter t) are the same for perturbed and unperturbed systems, hence the macroscopic behavior is the same. However, they both go (in ) to equilibrium; hence, the two systems (perturbed and unperturbed) have different macroscopic paths (N.b., distinguish carefully the words “macroscopic” and “microscopic.”).
This is illustrated in the case of the cat map, both for
and
. See
Figure 1, second and third images.
4. Quantum Version
The first step is to show that with low entropy conditions at both (distantly separated) times, the entropy increases in between. To calculate entropy in subspaces of Hilbert space presence or absence in that subspace should be defined and numbers of states counted. This leads to having regions of 6-dimensional
-
space—to mimic the classical space—and that is possible. What we do is take a region of phase space and select a basis of states whose maximum value is in this region. Coherent states will do the job. For convenience, the value of the spread used to define those states might be the fixed point of [
15], but that is not necessary. Finally the dimension of the Hilbert subspace will be the number of states in that region. What this means is that one can go back to the classical method of using (the logarithm of) volume in phase space as a measure of entropy.
Therefore, as was done previously, we take coarse grains that are regions of phase space—coordinates and momentum. The density matrix involves mainly the diagonal elements. Thus even for identical particles the separation of locale as well as separation of momenta makes the density matrix nearly diagonal. For more on this theme, see [
16].
Even for identical particles this leads to cancellation. In one dimension the wave function for a pair of Gaussians is (not normalized)
(The variables are
and
; all the others are constants.) The diagonal elements of the density matrix (calculated from Equation (
2)) already show signs of cancellation, as follows:
As is evident, if is significantly different from or from , then there is already cancellation or rapid oscillation. With more particles the effect is stronger. This, by the way, is the reason that isolated systems can be analyzed without paying attention to symmetrization with respect to all identical particles in the universe.
To get low entropy at the beginning and the end confine the wave functions—at the beginning and end—to particular coarse grains. In between the two times considered, wave functions will spread. This sounds like the classical definitions, but once you have coarse grains the definitions are not all that different. However, to imitate [
12] it is necessary that
, where
is the propagator from time-0 (zero is the initial time) to time-
T (
T is the final time). This imposes a significant constraint on the wave function, in particular, the wave function, under
, should not spread. If spreading were to happen, the entire space available, not just the target in phase space, would be occupied and the entropy would not drop.
In a recent article [
15], we found that Gaussians that scatter do not spread. (This was explored further in [
17], but the principal application to Gaussians was done in [
15].) The idea is that scattering provides localization. In [
17], it is argued that wave functions become Gaussian (often) but that involves decoherence, which is human perception. Is everything a Gaussian? Obviously not; atomic wave functions can be complicated and even the hydrogen atom is a different function. Nevertheless, for the purpose of this demonstration a Gaussian will be adequate, at least for showing that
sometimes causality is an effect.
The requirement that
and that both be confined to a (small) region of phase space (at
and
) is severe. However, based on the results of [
15], it can be done. It is possible that the entropy would not be strictly zero (due to the Gaussian’s never vanishing) but it can be made small. The same holds in momentum space.
At this point we are back in classical mechanics and the proof is straightforward. Since non-standard definitions are used in [
12] we repeat the proof (now using standard definitions).
5. Classical Proof
We present a précis of what has been done in our previous work ([
12]).
On the phase space
let
be a measure and
. Let the dynamics be a measure-preserving map
on
, with
the time-
t image of an initial point
. The coarse graining of
, providing a notion of “macroscopic,” are sets with the following properties:
,
, with
,
for
. Let
be the characteristic function of
and let
(
). If
f is a function on
, its coarse graining is defined as
Let the system’s distribution in
be described by a density function
. The entropy to be used for studying irreversibility involves coarse graining and is defined as
with
formed from
as in Equation (
4). (In other notation,
.) The
relative entropy (Kullback–Leibler divergence), to which
is related, was given in [
12] with a different sign from the usual. Moreover, the illustration given in
Figure 1 uses a different definition of entropy.
Turning to the system at hand, it is required to start (
) in a subset
and end (
) in a subset
. (The fact that Gaussian wave functions necessarily do not vanish anywhere may lead to a small correction.) The points of
satisfying this two-time boundary condition are
In [
12,
18,
19] it is argued that for chaotic dynamics and for sufficiently long times
T,
. (Whether this carries over to quantum mechanics will be dealt with later.) We assume that there is a relaxation time
, and that
. As a consequence
(Recall that for mixing dynamics,
satisfies
. This is true for the
limit, but we assume that there is a time (
) such that the decorrelation condition holds. Moreover,
is measure preserving.) Under
,
becomes
To calculate the entropy, the density, which was
at time-0, must be coarse grained. The important quantity for the entropy calculation is
If
then the following will hold
Using the measure-preserving property of
, the factors
in both numerator and denominator of
cancel, leading to
This is precisely what one gets
without future conditioning, so that all macroscopic quantities, and in particular the entropy, are indistinguishable from their unconditioned values.
Working backward from time-
T one obtains an analogous result. Define a variable
and set
. Then
If
s satisfies
, then when the density associated with
is calculated, its dependence on
will drop out. It follows that
For a time-reversal invariant dynamics this will give the entropy the same time dependence coming back from
T as going forward from 0. Even in the absence of T invariance, there should be roughly the same behavior because of the absence of any dissipative dynamics.
Now we turn to perturbations. Call the unperturbed system A. The microstates are in the set
(formerly called
). System B, the perturbed case, has an instantaneous transformation act on it at time-
. Call this transformation
. (This transformation is called
in [
12]. The letter
is chosen to avoid confusion with the wave function.) It is not dissipative—the arrow does not arise from such an asymmetry.
is invertible and measure preserving. Successful solutions must go from
to
under the transformation
. The microstates for system B are therefore
Again, there is the assumption that
is not empty, which will be taken up for quantum mechanics in
Section 6. Clearly,
and
are different—at
all times. But as will now be shown, for mixing dynamics and for sufficiently large
T, the following hold: (1) for
close to 0, the only differences in
macroscopic behavior between A and B are for
; (2) for
close to
T, the only differences in
macroscopic behavior between A and B are for
.
The direction of causality follows the direction of entropy increase.The proof is nearly the same as before. Again we use a time
such that the mixing decorrelation holds for time intervals longer than
. First consider
close to 0. The observable macroscopic quantities are the densities in grain-
, which are, for
,
As before, the mixing property, for
, yields
, which is the initial-value-only macroscopic time evolution. For
, the only difference is to add a step,
. But this step is as measure preserving as
itself, and therefore, as before,
. Thus A and B have the same macrostates before
.
For
,
continues its behavior as before. For
things are different:
Now, I require
. If this is satisfied the
dependence drops out and
This shows that the effect of
is the usual initial-conditions-only phenomenon.
If we repeat these arguments for
t such that
is small, then just as we showed in
Section 3, the effect of
will only be at times
t less than .
7. Conclusions
There are many defects to this quantum version of entropy increase determining the direction of causality. What is not clear is whether a quantum theory is at all needed. Causality as usually understood is a macroscopic phenomenon and not quantum mechanical. Nevertheless, we point out the deficiencies of the present work.
There is uncertainty over black holes. If one considers two time boundary conditions purely as a way to get rid of “initial conditions” prejudice, OK, but if there is any thought that the world is actually symmetric (approximately) then that too must be dealt with. Do black holes evaporate? Do they become white holes? Are there black holes at all? (See the blog by Cramer [
20].) We will not deal with any of these questions and allow our results to depend on the answers.
Our arguments are phrased in terms of chaotic dynamics. Chaos is problematic for quantum mechanics. Our previous paper ([
12]) deals also with harmonic oscillators (which are not chaotic) but has other limitations. For the argument above to be meaningful all that is really required is a relaxation time. It is plausible that this exists in quantum theory.
Although we have briefly mentioned identity of particles, that has not been seriously dealt with.
Finally, there is the requirement that
(where
is the evolution operator for time
T) and that both be low entropy, confined to a region of phase space. In some versions of quantum mechanics this is a natural requirement, but if that is wrong, there are severe limitations on the applicability of these results (see
Section 6).
We mention in passing that our understanding of the “destiny” of Aharonov et al. does not fix the results of measurements (the “destiny” being a future value of the wave function). This is because general wave functions allow many outcomes to a given experiment.