quantum entanglement

Answering Mermin’s Challenge with the Relativity Principle

Estimated Read Time: 19 minute(s)
Common Topics: quantum, outcomes, principle, mermin, spin

Note: This Insight was previously titled, “Answering Mermin’s Challenge with Wilczek’s Challenge.” While that version of this Insight did not involve any particular interpretation of quantum mechanics, it did involve the block universe interpretation of special relativity. I have updated this Insight to remove the block universe interpretation, so that it now answers Mermin’s challenge in “principle” fashion alone, as in this Insight.

Nearly four decades ago, Mermin revealed the conundrum of quantum entanglement for a general audience [1] using his “simple device,” which I will refer to as the “Mermin device” (Figure 1). To understand the conundrum of the device required no knowledge of physics, just some simple probability theory, which made the presentation all the more remarkable. Concerning this paper Feynman wrote to Mermin, “One of the most beautiful papers in physics that I know of is yours in the American Journal of Physics” [2, p. 366-7]. In subsequent publications, he “revisited” [3] and “refined” [4] the mystery of quantum entanglement with similarly simple devices. In this Insight, I will focus on the original Mermin device as it relates to the mystery of entanglement via the Bell spin states.

Figure 1. The Mermin Device

The Mermin device functions according to two facts that are seemingly contradictory, thus the mystery. Mermin simply supplies these facts and shows the contradiction, which the “general reader” can easily understand. He then challenges the “physicist reader” to resolve the mystery in an equally accessible fashion for the “general reader.” Here is how the Mermin device works.

The Mermin device is based on the measurement of spin angular momentum (Figure 2). The spin measurements are carried out with Stern-Gerlach (SG) magnets and detectors (Figure 3). The Mermin device contains a source (middle box in Figure 1) that emits a pair of spin-entangled particles towards two detectors (boxes on the left and right in Figure 1) in each trial of the experiment. The settings (1, 2, or 3) on the left and right detectors are controlled randomly by Alice and Bob, respectively, and each measurement at each detector produces either a result of R or G. The following two facts obtain:

  1. When Alice and Bob’s settings are the same in a given trial (“case (a)”), their outcomes are always the same, ##\frac{1}{2}## of the time RR (Alice’s outcome is R and Bob’s outcome is R) and ##\frac{1}{2}## of the time GG (Alice’s outcome is G and Bob’s outcome is G).
  2. When Alice and Bob’s settings are different (“case (b)”), the outcomes are the same ##\frac{1}{4}## of the time, ##\frac{1}{8}## RR and ##\frac{1}{8}## GG.

The two possible Mermin device outcomes R and G represent two possible spin measurement outcomes “up” and “down,” respectively (Figure 2), and the three possible Mermin device settings represent three different orientations of the SG magnets (Figures 3 & 4).

Figure 2. A pair of Stern-Gerlach (SG) magnets showing the two possible outcomes, up (##+\frac{\hbar}{2}##) and down (##-\frac{\hbar}{2}##) or ##+1## and ##-1##, for short. The important point to note here is that the classical analysis predicts all possible deflections, not just the two that are observed. This difference uniquely distinguishes the quantum joint distribution from the classical joint distribution for spin entangled pairs [5].

 

Figure 3. Alice and Bob making spin measurements on a pair of spin-entangled particles with their SG magnets and detectors. In this particular case, the plane of conserved spin angular momentum is the xz plane.
Figure 4. Three orientations of SG magnets in the plane of symmetry for Alice and Bob’s spin measurements corresponding to the three settings on the Mermin device.

Mermin writes, “Why do the detectors always flash the same colors when the switches are in the same positions? Since the two detectors are unconnected there is no way for one to ‘know’ that the switch on the other is set in the same position as its own.” This leads him to introduce “instruction sets” to account for the behavior of the device when the detectors have the same settings. He writes, “It cannot be proved that there is no other way, but I challenge the reader to suggest any.” Now look at all trials when Alice’s particle has instruction set RRG and Bob’s has instruction set RRG, for example.

That means Alice and Bob’s outcomes in setting 1 will both be R, in setting 2 they will both be R, and in setting 3 they will both be G. That is, the particles will produce an RR result when Alice and Bob both choose setting 1 (referred to as “11”), an RR result when both choose setting 2 (referred to as “22”), and a GG result when both choose setting 3 (referred to as “33”). That is how instruction sets guarantee Fact 1. For different settings, Alice and Bob will obtain the same outcomes when Alice chooses setting 1 and Bob chooses setting 2 (referred to as “12”), which gives an RR outcome. And, they will obtain the same outcomes when Alice chooses setting 2 and Bob chooses setting 1 (referred to as “21”), which also gives an RR outcome. That means we have the same outcomes for different settings in 2 of the 6 possible case (b) situations, i.e., in ##\frac{1}{3}## of case (b) trials for this instruction set. This ##\frac{1}{3}## ratio holds for any instruction set with two R(G) and one G(R).

The only other possible instruction sets are RRR or GGG where Alice and Bob’s outcomes will agree in ##\frac{9}{9}## of all trials. Thus, the “Bell inequality” for the Mermin device says that instruction sets must produce the same outcomes in more than ##\frac{1}{3}## of all case (b) trials. But, Fact 2 for the Mermin device says you only get the same outcomes in ##\frac{1}{4}## of all case (b) trials, thereby violating the Bell inequality. Thus, the conundrum of Mermin’s device is that the instruction sets needed for Fact 1 fail to yield the proper outcomes for Fact 2.

Concerning his device Mermin wrote, “Although this device has not been built, there is no reason in principle why it could not be, and probably no insurmountable practical difficulties.” Sure enough, the experimental confirmation of the violation of Bell’s inequality per quantum entanglement is so common that it can now be carried out in the undergraduate physics laboratory [6]. Thus, there is no disputing that the conundrum of the Mermin device has been experimentally well verified, vindicating its prediction by quantum mechanics.

While the conundrum of the Mermin device is now a well-established fact, Mermin’s “challenging exercise to the physicist reader to translate the elementary quantum-mechanical reconciliation of cases (a) and (b) into terms meaningful to a general reader struggling with the dilemma raised by the device” arguably remains unanswered. To answer this challenge, it is generally acknowledged that one needs a compelling causal mechanism or a compelling physical principle by which the conundrum of the Mermin device is resolved. Such a model needs to do more than the “Copenhagen interpretation” [7], which Mermin characterized as “shut up and calculate” [8]. In other words, while the formalism of quantum mechanics accurately predicts the conundrum, quantum mechanics does not provide a model of physical reality or underlying physical principle to resolve the conundrum. While there are many interpretations of quantum mechanics, even one published by Mermin [9], there is no consensus among physicists on any given interpretation.

Rather than offer yet another uncompelling interpretation of quantum mechanics, I will share and expand on an underlying physical principle [10,11] that explains the quantum correlations responsible for the conundrum of the Mermin device. In other words, I will provide a “principle account” of quantum entanglement. Here I’m making specific reference to Einstein’s notion of a “principle theory” as explained in this Insight. While this explanation, conservation per no preferred reference frame (NPRF), may not be “in terms meaningful to a general reader,” it is pretty close. That is, all one needs to appreciate the explanation is a course in introductory physics, which probably represents the “general reader” interested in this topic.

That quantum mechanics accurately predicts the observed phenomenon without spelling out any means a la “instruction sets” for how it works prompted Smolin to write [12, p. xvii]:

I hope to convince you that the conceptual problems and raging disagreements that have bedeviled quantum mechanics since its inception are unsolved and unsolvable, for the simple reason that the theory is wrong. It is highly successful, but incomplete.

Of course, this is precisely the complaint leveled by Einstein, Podolsky and Rosen (EPR) in their famous 1935 paper [13], “Can Quantum-Mechanical Description of Physical Reality Be Considered Complete?” Contrary to this belief, I will show that quantum mechanics is actually as complete as possible, given Einstein’s own relativity principle (NPRF). Indeed, Einstein missed a chance to rid us of his “spooky actions at a distance.” All he would have had to do is extend his relativity principle to include the measurement of Planck’s constant h, just as he had done by extending the relativity principle from mechanics to include the measurement of the speed of light c per electromagnetism.

That is, the relativity principle (NPRF) entails the light postulate of special relativity, i.e., that everyone measure the same speed of light c, regardless of their motion relative to the source. If there was only one reference frame for a source in which the speed of light equalled the prediction from Maxwell’s equations (##c = \frac{1}{\sqrt{\mu_o\epsilon_o}}##), then that would certainly constitute a preferred reference frame. The light postulate then leads to time dilation, length contraction, and the relativity of simultaneity per the Lorentz transformations of special relativity. Indeed, this is the way special relativity is introduced by Serway & Jewett [14] and Knight [15] for introductory physics students. Let me show you how further extending NPRF to the measurement of Planck’s constant leads to quantum entanglement per the qubit Hilbert space structure (probability structure) of quantum mechanics.

 


Figure 5. In this set up, the first SG magnets (oriented at ##\hat{z}##) are being used to produce an initial state ##|\psi\rangle = |u\rangle## for measurement by the second SG magnets (oriented at ##\hat{b}##).

As Weinberg points out [16], measuring an electron’s spin via SG magnets constitutes the measurement of “a universal constant of nature, Planck’s constant” (Figure 2). So if NPRF applies equally here, then everyone must measure the same value for Planck’s constant h, regardless of their SG magnet orientations relative to the source, which like the light postulate is an “empirically discovered” fact. By “relative to the source,” I might mean relative “to the vertical in the plane perpendicular to the line of flight of the particles” [1], ##\hat{z}## in Figure 5 for example. Here the possible spin outcomes ##\pm\frac{\hbar}{2}## represent a fundamental (indivisible) unit of information per Dakic & Brukner’s first axiom in their information-theoretic reconstruction of quantum theory [17], “An elementary system has the information carrying capacity of at most one bit.” Thus, different SG magnet orientations relative to the source constitute different “reference frames” in quantum mechanics just as different velocities relative to the source constitute different “reference frames” in special relativity.

To make the analogy more explicit, one could have employed NPRF to predict the light postulate as soon as Maxwell showed electromagnetic radiation propagates at ##c = \frac{1}{\sqrt{\mu_o\epsilon_o}}##. All they would have had to do is extend the relativity principle from mechanics to electromagnetism. However, given the understanding of waves at the time, everyone rather began searching for a propagation medium, i.e., the luminiferous ether. Likewise, one could have employed NPRF to predict spin angular momentum as soon as Planck published his wavelength distribution function for blackbody radiation. All they would have had to do is extend the relativity principle from mechanics and electromagnetism to quantum physics. However, given the understanding of angular momentum and magnetic moments at the time, Stern & Gerlach rather expected to see their silver atoms deflected in a continuum distribution after passing through their magnets (Figure 2). In other words, they discovered spin angular momentum when they were simply looking for angular momentum. But, had they noticed that their measurement constituted a measurement of Planck’s constant (with its dimension of angular momentum), they could have employed NPRF to predict the spin outcome with its qubit Hilbert space structure (Figures 2 & 5) and its ineluctably probabilistic nature, as I will now explain.

If we create a preparation state oriented along the positive ##z## axis as in Figure 5, i.e., ##|\psi\rangle = |u\rangle## in the Dirac notation [18], our spin angular momentum is ##\vec{S} = +1\hat{z}## (in units of ##\frac{\hbar}{2} = 1##). Now proceed to make a measurement with the SG magnets oriented at ##\hat{b}## making an angle ##\beta## with respect to ##\hat{z}## (Figure 5). According to classical physics, we expect to measure ##\vec{S}\cdot\hat{b} = \cos{(\beta)}## (Figure 6), but we cannot measure anything other than ##\pm 1## due to NPRF (contra the prediction by classical physics), so we see that NPRF answers Wheeler’s “Really Big Question,” “Why the quantum?” in “one clear, simple sentence” to convey “the central point and its necessity in the construction of the world” [19,20].

 

 

Figure 6. The spin angular momentum of Bob’s particle ##\vec{S}## projected along his measurement direction ##\hat{b}##. This does not happen with spin angular momentum due to NPRF.

As a consequence, we can only recover ##\cos{(\beta)}## on average (Figure 7), i.e., NPRF dictates “average-only” projection
\begin{equation}
(+1) P(+1 \mid \beta) + (-1) P(-1 \mid \beta) = \cos (\beta) \label{AvgProjection}
\end{equation}
Solving simultaneously with our normalization condition ##P(+1 \mid \beta) + P(-1 \mid \beta) = 1##, we find that
\begin{equation}
P(+1 \mid \beta) = \mbox{cos}^2 \left(\frac{\beta}{2} \right) \label{UPprobability}
\end{equation}
and
\begin{equation}
P(-1 \mid \beta) = \mbox{sin}^2 \left(\frac{\beta}{2} \right) \label{DOWNprobability}
\end{equation}

Figure 7. An ensemble of 4 SG measurement trials with ##\beta = 60^{\circ}##. The tilted blue arrow depicts an SG measurement orientation and the vertical arrow represents our preparation state ##|\psi\rangle = |u\rangle## (Figure 5). The yellow dots represent the two possible measurement outcomes for each trial, up (located at arrow tip) or down (located at bottom of arrow). The expected projection result of ##\cos{(\beta)}## cannot be realized because the measurement outcomes are binary (quantum) with values of ##+1## (up) or ##-1## (down) per NPRF. Thus, we have “average-only” projection for all 4 trials (three up outcomes and one down outcome for ##\beta = 60^\circ## average to ##\cos{(60^\circ)}=\frac{1}{2}##).

This explains the ineluctably probabilistic nature of QM, as pointed out by Mermin [21]:

Quantum mechanics is, after all, the first physical theory in which probability is explicitly not a way of dealing with ignorance of the precise values of existing quantities.

That is, quantum mechanics is as complete as possible, given the relativity principle. Of course, these “average-only” results due to “no fractional outcomes per NPRF” hold precisely for the qubit Hilbert space structure of quantum mechanics [11]. Thus, we see that NPRF provides a principle explanation of the kinematic/probability structure of quantum mechanics, just as it provides a principle explanation of the kinematic/Minkowski spacetime structure of special relativity. In fact, this follows from Information Invariance & Continuity at the basis of axiomatic reconstructions of QM per information-theoretic principles (see No Preferred Reference Frame at the Foundation of Quantum Mechanics). Now let’s expand this idea to the situation when we have two entangled particles, as in the Mermin device. The concept we need to understand now is the “correlation function.”

The correlation function between two outcomes over many trials is the average of the two values multiplied together. In this case, there are only two possible outcomes for any setting, +1 (up or R) or –1 (down or G), so the largest average possible is +1 (total correlation, RR or GG, as when the settings are the same) and the smallest average possible is –1 (total anti-correlation, RG or GR). One way to write the equation for the correlation function is
\begin{equation}\langle \alpha,\beta \rangle = \sum (i \cdot j) \cdot p(i,j \mid \alpha,\beta) \label{average}\end{equation}
where ##p(i,j \mid \alpha,\beta)## is the probability that Alice measures ##i## and Bob measures ##j## when Alice’s SG magnet is at angle ##\alpha## and Bob’s SG magnet is at angle ##\beta##, and ##(i \cdot j)## is just the product of the outcomes ##i## and ##j##. The correlation function for instruction sets for case (a) is the same as that of the Mermin device for case (a), i.e., they’re both 1. Thus, we must explore the difference between the correlation function for instruction sets and the Mermin device for case (b).

To get the correlation function for instruction sets for different settings, we need the probabilities of measuring the same outcomes and different outcomes for case (b), so we can use Eq. (\ref{average}). We saw that when we had two R(G) and one G(R), the probability of getting the same outcomes for different settings was ##\frac{1}{3}## (this would break down to ##\frac{1}{6}## for each of RR and GG overall). Thus, the probability of getting different outcomes would be ##\frac{2}{3}## for these types of instruction sets (##\frac{1}{3}## for each of RG and GR). That gives a correlation function of

\begin{equation}\langle \alpha,\beta \rangle = \left(+1\right)\left(+1\right)\left(\frac{1}{6}\right) + \left(-1\right)\left(-1\right)\left(\frac{1}{6}\right) +  \left(+1\right)\left(-1\right)\left(\frac{2}{6}\right) + \left(-1\right)\left(+1\right)\left(\frac{2}{6}\right)= -\frac{1}{3}
\end{equation}
For the other type of instruction sets, RRR and GGG, we would have a correlation function of ##+1## for different settings, so overall the correlation function for instruction sets for different settings has to be larger than ##-\frac{1}{3}##. In fact, if all eight possible instruction sets are produced with equal frequency, then for any given pair of case (b) settings, e.g., 12 or 13 or 23, you will obtain RR, GG, RG, and GR in equal numbers giving a correlation function of zero. That means the results are uncorrelated as one would expect given that all possible instruction sets are produced randomly, i.e., with equal frequency. From this we would typically infer that there is nothing that needs to be explained.

Fact 2 for the Mermin device says the probability of getting the same results (RR or GG) for different settings is ##\frac{1}{4}## (##\frac{1}{8}## for each of RR and GG). Thus, the probability of getting different outcomes for case (b) must be ##\frac{3}{4}## (##\frac{3}{8}## for each of RG and GR). That gives a correlation function of
\begin{equation}\langle \alpha,\beta \rangle = \left(+1\right)\left(+1\right)\left(\frac{1}{8}\right) + \left(-1\right)\left(-1\right)\left(\frac{1}{8}\right) +  \left(+1\right)\left(-1\right)\left(\frac{3}{8}\right) + \left(-1\right)\left(+1\right)\left(\frac{3}{8}\right)= -\frac{1}{2}
\end{equation}
That means the Mermin device is more strongly anti-correlated for different settings than instruction sets. Indeed, if all possible instruction sets are produced with equal frequency, the Mermin device evidences something to explain (anti-correlated results) where instruction sets suggest there is nothing to explain (uncorrelated results). Thus, quantum mechanics predicts and we observe anti-correlated outcomes for different settings in need of explanation while its classical counterpart suggests there is nothing in need of explanation at all. Mermin’s challenge then amounts to explaining why that is true for the “general reader.”

At this point read my Insight Exploring Bell States and Conservation of Spin Angular Momentum.

Now you understand how the correlation function for the Bell spin states results from “average-only” conservation (as a mathematical fact, Figure 8) resulting from the fact that Alice and Bob both always measure ##\pm 1 \left(\frac{\hbar}{2}\right)## (quantum), never a fraction of that amount (classical), as shown in Figure 2 (empirical fact).

There are two important points to be made here. First, NPRF is just the statement of an “empirically discovered” fact, i.e., Alice and Bob both always measure ##\pm 1##. Second, it is simply a mathematical fact that the “average-only” conservation yields the quantum correlation functions. In other words, to paraphrase Einstein, “we have an empirically discovered principle that gives rise to mathematically formulated criteria which the separate processes or the theoretical representations of them have to satisfy.” That is why this principle account of quantum entanglement provides “logical perfection and security of the foundations.” Thus, we see how quantum entanglement follows from NPRF applied to the measurement of h in precisely the same manner that time dilation and length contraction follow from NPRF applied to the measurement of c. And, just like in special relativity, Bob could partition the data according to his equivalence relation (per his reference frame) and claim that it is Alice who must average her results (obtained in her reference frame) to conserve spin angular momentum (Figure 9).

 

Figure 8. An ensemble of 8 experimental trials for the Bell spin states showing Bob’s outcomes corresponding to Alice‘s ##+1## outcomes when ##\theta = 60^\circ##. Angular momentum is not conserved in any given trial, because there are two different measurements being made, i.e., outcomes are in two different reference frames, but it is conserved on average for all 8 trials (six up outcomes and two down outcomes average to ##\cos{60^\circ}=\frac{1}{2}##). It is impossible for angular momentum to be conserved explicitly in each trial since the measurement outcomes are binary (quantum) with values of ##+1## (up) or ##-1## (down) per NPRF. The conservation principle at work here assumes Alice and Bob’s measured values of spin angular momentum are not mere components of some hidden angular momentum with variable magnitude. That is, the measured values of angular momentum are the angular momenta contributing to this conservation, as I explained in my Insight Bell States and Conservation of Spin Angular Momentum.

Figure 9. Comparing special relativity with quantum mechanics according to no preferred reference frame (NPRF).

Of course, all of this does not provide any relief for those who still require explanation via “constructive efforts.” As Lorentz complained [22]:

Einstein simply postulates what we have deduced, with some difficulty and not altogether satisfactorily, from the fundamental equations of the electromagnetic field.

And, Albert Michelson said [23]:

It must be admitted, these experiments are not sufficient to justify the hypothesis of an ether. But then, how can the negative result be explained?

In other words, neither was convinced that NPRF was sufficient to explain time dilation and length contraction. Apparently for them, such a principle must be accounted for by some causal mechanism, e.g., the luminiferous ether. Likewise, if one requires “constructive efforts” to account for “conservation per NPRF” responsible for “average-only” conservation, then they will certainly want to continue the search for a causal mechanism responsible for quantum entanglement.

But after 115 years, physicists have largely abandoned theories of the luminiferous ether, having grown comfortable with the longstanding and empirically sound light postulate based on NPRF. Even Lorentz seemed to acknowledge the value of this principle explanation when he wrote [22]:

By doing so, [Einstein] may certainly take credit for making us see in the negative result of experiments like those of Michelson, Rayleigh, and Brace, not a fortuitous compensation of opposing effects but the manifestation of a general and fundamental principle.

Therefore, 85 years after publication of the EPR paper, perhaps we should consider the possibility that quantum entanglement will likewise ultimately yield to principle explanation. After all, we now know that our time-honored relativity principle is precisely the principle that resolves the mystery of “spooky actions at a distance.” As John Bell said in 1993 [24, p. 85]:

I think the problems and puzzles we are dealing with here will be cleared up, and … our descendants will look back on us with the same kind of superiority as we now are tempted to feel when we look at people in the late nineteenth century who worried about the ether. And Michelson-Morley .., the puzzles seemed insoluble to them. And came Einstein in nineteen five, and now every schoolboy learns it and feels .. superior to those old guys. Now, it’s my feeling that all this action at a distance and no action at a distance business will go the same way. But someone will come up with the answer, with a reasonable way of looking at these things. If we are lucky it will be to some big new development like the theory of relativity.

Perhaps causal accounts of quantum entanglement are destined to share the same fate as theories of the luminiferous ether. Regardless, we have certainly answered Mermin’s challenge, since conservation per NPRF is very accessible to the “general reader.”

References
  1. Mermin, N.D.: Bringing home the atomic world: Quantum mysteries for anybody. American Journal of Physics 49, 940-943 (1981).
  2. Feynman, M.: Perfectly Reasonable Deviations from the Beaten Track. Basic Books, New York (2005).
  3. Mermin, N.D.: Quantum mysteries revisited. American Journal of Physics 58, 731-734 (Aug 1990).
  4. Mermin, N.D.: Quantum mysteries refined. American Journal of Physics 62, 880-887 (Aug 1994).
  5. Garg, A., and Mermin, N.D.: Bell Inequalities with a Range of Violation that Does Not Diminish as the Spin Becomes Arbitrarily Large. Physical Review Letters 49(13), 901–904 (1982).
  6. Dehlinger, D., and Mitchell, M.W.: Entangled photons, nonlocality, and Bell inequalities in the undergraduate laboratory. American Journal of Physics 70(9), 903–910 (2002).
  7. Becker, A.: What is Real? The Unfinished Quest for the Meaning of Quantum Physics. Basic Books, New York (2018).
  8. Mermin, N.D.: Could Feynman Have Said This? Physics Today 57(5), 10 (Apr 2004).
  9. Mermin, N.D.: What is quantum mechanics trying to tell us? American Journal of Physics 66(9), 753-767 (1998).
  10. Stuckey, W.M., Silberstein, M., and McDevitt, T., and Le, T.D.: Answering Mermin’s challenge with conservation per no preferred reference frame. Scientific Reports 10, 15771 (2020).
  11. Silberstein, M., Stuckey, W.M., and McDevitt, T.: Beyond Causal Explanation: Einstein’s Principle Not Reichenbach’s. Entropy 23(1), 114 (2021).
  12. Smolin, L.: Einstein’s Unfinished Revolution: The Search for What Lies Beyond the Quantum. Penguin Press, New York (2019).
  13. Einstein, A., Podolsky, B., and Rosen, N.: Can Quantum-Mechanical Description of Physical Reality Be Considered Complete? Physical Review 47(10), 777–780 (1935).
  14. Serway, R., and Jewett, J.: Physics for Scientists and Engineers with Modern Physics. Cengage, Boston (2019).
  15. Knight, R.: Physics for Scientists and Engineers with Modern Physics. Pearson, San Francisco (2008).
  16. Weinberg, S.: The Trouble with Quantum Mechanics. (2017).
  17. Dakic, B., and Brukner, C.: Quantum Theory and Beyond: Is Entanglement Special? In: Deep Beauty: Understanding the Quantum World through Mathematical Innovation. Halvorson, H. (ed.). Cambridge University Press, New York (2009), 365–393.
  18. Ross, R.: Computer simulation of Mermin’s quantum device. American Journal of Physics 88(6), 483–489 (2020).
  19. Barrow, J.D., Davies, P.C.W., and Harper, C.: Science and Ultimate Reality: Quantum Theory, Cosmology, and Complexity. Cambridge University Press, New York (2004).
  20. Wheeler, J.: How Come the Quantum? Annals of the New York Academy of Sciences: New Techniques and Ideas in Quantum Measurement Theory 480(1), 304–316 (1986).
  21. Mermin, N.D.: Making better sense of quantum mechanics. Reports on Progress in Physics 82, 012002 (2019).
  22. Lorentz, H.A.: The Theory of Electrons and Its Applications to the Phenomena of Light and Radiant Heat. G.E. Stechert and Co., New York (1916).
  23. A. Michelson quote from 1931 in Episode 41 “The Michelson-Morley Experiment” in the series “The Mechanical Universe,” written by Don Bane (1985).
  24. Bell, J.S.: Indeterminism and Nonlocality. In: Mathematical Undecidability, Quantum Nonlocality and the Question of the Existence of God. Driessen, A., and Suarez, A. (eds.). Springer, Netherlands (1997), 78–89.

Comment Thread

89 replies
« Older Comments
  1. RUTA says:

    bhobba said

    For those interested here is Mermin's original paper:
    https://pdfs.semanticscholar.org/76f3/9c8a412b47b839ba764d379f88adde5bccfd.pdf
    Feynman in a letter to Mermin said 'One of the most beautiful papers in physics that I know of is yours in the American Journal of Physics.'

    I personally am finding my view of QM evolving a bit. Feynman said the essential mystery of QM was in the double slit experiment. I never actually thought so myself, but was impressed with it as an introduction to the mysteries of QM at the beginning level. I am now starting to think entanglement may be the essential mystery.

    Thanks
    Bill

    I ordered the book ILL and copied the letter from Feynman to Mermin and Mermin's response. I'll attach those copies here.

  2. PeterDonis says:

    RUTA said

    You can model Schwarzschild surrounding matter solutions and the Schwarzschild solution holds all the way inside the horizon to that matter. So, as you go smaller and smaller you approach ##\rho = \infty##.

    You can do this in a model such as the 1939 Oppenheimer-Snyder model, sure. That's not what I was using "the Schwarzschild solution" to refer to, but yes, it's a valid model, and is subject to the same limitation that Wald describes, assuming Wald's viewpoint that classical GR will break down at Planck scale curvatures is correct.

  3. RUTA says:

    PeterDonis said

    There is no "where M is" in the Schwarzschild solution; it's a vacuum solution with zero stress-energy everywhere. Also ##r = 0## is not even part of the manifold; it's a limit point that is approached but never reached, so it can't be "where" anything is.

    M in the Schwarzschild solution is a global property of the spacetime; there is no place "where it is".

    You can model Schwarzschild surrounding matter solutions and the Schwarzschild solution holds all the way inside the horizon to that matter. So, as you go smaller and smaller you approach ##\rho = \infty##.

  4. PeterDonis says:

    RUTA said

    it's difficult to say how "big" ##a = 0## is

    But however "big" it is, it does not extend to any value of ##t## greater than zero in the standard FRW models. That's the point I was making.

  5. PeterDonis says:

    RUTA said

    don't dismiss the model because you believe an empirically unverifiable mathematical extrapolation leads to "pathologies."

    Are you claiming that Wald is "dismissing" the Einstein-de Sitter model or similar models on these grounds? I don't see him dismissing it at all. I just see him (and MTW, and every other GR textbook I've read that discusses this issue) saying that any such model will have a limited domain of validity; we should expect it to break down in regions where spacetime curvature at or greater than the Planck scale is predicted.

  6. PeterDonis says:

    RUTA said

    The ρ=∞ρ=∞ is also a problem for Schwarzschild at r=0r=0 because that's where M is.

    There is no "where M is" in the Schwarzschild solution; it's a vacuum solution with zero stress-energy everywhere and no ##\rho = \infty## (and for that matter no ##\rho \neq 0##) anywhere. Also ##r=0## is not even part of the manifold; it's a limit point that is approached but never reached, so it can't be "where" anything is.

    M in the Schwarzschild solution is a global property of the spacetime; there is no place "where it is".

  7. RUTA says:

    PeterDonis said

    Then the Einstein-de Sitter model would be valid for any ##t > 0##, since it predicts finite and positive density and scale factor. So are you saying I could use any value of ##B## I like in your modified version of the model, putting the cutoff wherever I want, as long as it doesn't include the ##\rho = \infty## point?

    Yes, and you could even use ##\rho = \infty## if you could produce empirical verification. Use whatever you need, just don't dismiss the model because you believe an empirically unverifiable mathematical extrapolation leads to "pathologies."

  8. RUTA says:

    PeterDonis said

    But ##\rho = \infty## isn't a "region", it's a point. And that point is not even included in the manifold; as I said, it's a limit point that's approached but never reached. So again, I don't see what is wrong with the standard Einstein-de Sitter model, where ##B = 0## in your modified formula, if any finite value of ##\rho## is ok.

    Well, it's difficult to say how "big" ##a = 0## is because the spatial hyper surfaces to that point are ##\infty##. Its "size" is undefined so I was being careful with my language.

  9. RUTA says:

    PeterDonis said

    No, that's not what they say. What they say (Wald, for example) is that spacetime curvatures which are finite but larger than the Planck scale are problematic for a classical theory of gravity like GR, because we expect quantum gravity effects to become important at that scale. In the Einstein-de Sitter model, for example, the curvature becomes infinite at the point you have been labeling ##\rho = \infty##, not just the density. And in Schwarzschild spacetime, the curvature is the only thing that becomes infinite at the singularity at ##r = 0##, because it's a vacuum solution and the stress-energy tensor is zero everywhere. But that singularity is just as problematic on a viewpoint like Wald's.

    Yes, the curvature is also problematic as Wald points out in Chapter 9. The ##\rho = \infty## is also a problem for Schwarzschild at ##r = 0## because that's where M is. Am I missing something there?

  10. PeterDonis says:

    RUTA said

    the textbooks say GR cosmology models are problematic because they're "singular" in the sense that they entail ##\rho = \infty##.

    No, that's not what they say. What they say (Wald, for example) is that spacetime curvatures which are finite but larger than the Planck scale are problematic for a classical theory of gravity like GR, because we expect quantum gravity effects to become important at that scale. In the Einstein-de Sitter model, for example, the curvature becomes infinite at the point you have been labeling ##\rho = \infty##, not just the density. And in Schwarzschild spacetime, the curvature is the only thing that becomes infinite at the singularity at ##r = 0##, because it's a vacuum solution and the stress-energy tensor is zero everywhere. But that singularity is just as problematic on a viewpoint like Wald's.

  11. PeterDonis says:

    RUTA said

    the cutoff is not at all arbitrary, you keep whatever you can argue generates conceivable empirical results

    Then the Einstein-de Sitter model would be valid for any ##t > 0##, since it predicts finite and positive density and scale factor. So are you saying I could use any value of ##B## I like in your modified version of the model, putting the cutoff wherever I want, as long as it doesn't include the ##\rho = \infty## point?

    RUTA said

    I didn't say you can't keep inflationary models.

    Ok, that helps to clarify your viewpoint.

    RUTA said

    Adynamical explanation allows you to simply omit the problematic region if it is beyond empirical confirmation

    But ##\rho = \infty## isn't a "region", it's a point. And that point is not even included in the manifold; as I said, it's a limit point that's approached but never reached. So again, I don't see what is wrong with the standard Einstein-de Sitter model, where ##B = 0## in your modified formula, if any finite value of ##\rho## is ok.

  12. RUTA says:
    Again, the only difference between what I'm claiming and what exists in the common textbook explanations of GR cosmology is that the textbooks say GR cosmology models are problematic because they're "singular" in the sense that they entail ##\rho = \infty##. And I'm saying adynamical explanation does not force those solutions to entail that region. Adynamical explanation allows you to simply omit the problematic region if it is beyond empirical confirmation. So far that's true, the model is working great despite the fact that a purely mathematical extrapolation produces ##\rho = \infty##. So why worry about that region?
  13. RUTA says:
    I answered your question many times, the cutoff is not at all arbitrary, you keep whatever you can argue generates conceivable empirical results. I didn't say you can't keep inflationary models. I personally think they're not interesting as articulated by Paul Steinhardt, but others may want to pursue them.
  14. PeterDonis says:

    RUTA said

    Every part of M fits self-consistently with every other part of M via EEs.

    In your version of the Einstein-de Sitter model, there is only one part of M, the Einstein-de Sitter region with your arbitrary cutoff. So in your model, there is nothing to fit self-consistently with. But there certainly could be: you could, for example, fit your Einstein-de Sitter region self-consistently via EEs with an inflationary region, just as inflationary models do. Why didn't you?

  15. PeterDonis says:

    RUTA said

    That's where you're coming from

    No, you don't understand where I'm coming from. Let me try to get at the issue I see another way.

    RUTA said

    I've repeated many times that you only need to keep that part of M that you believe can or can conceivably produce empirically verifiable results.

    In the particular case of the Einstein-de Sitter model, as far as I can tell, to you this means: cut off the model at some spacelike hypersurface before it reaches ##\rho = \infty##. But how close to ##\rho = \infty## can I get before I cut the model off? Your cutoff procedure left a finite range of time (from ##t = 0## to ##t = – B## in your modified model) between the edge of the model and the problematic ##\rho = \infty## point. Could I make an equally viable model by taking, say, ##B / 2## instead of ##B## as the constant in the model?

    If your answer is yes, your procedure does not lead to a unique model; taken to its logical conclusion, it ends up being the same as the standard Einstein-de Sitter model, since that model does not consider the ##\rho = \infty## point to be part of the manifold in any case, it's just a limit point that is approached but never reached.

    If your answer is no, then you need to give a reason for picking the particular value ##B## as the constant in your model, instead of something else. So far I have not seen you give one.

    By contrast, my response to the fact that the Einstein-de Sitter model predicts ##\rho = \infty## at some particular limit point is to look for an alternate model that does not have that property, by taking an Einstein-de Sitter region, just like the one in your model, and joining it to another region, such as an inflationary region, that does not predict ##\rho = \infty## anywhere. You appear to think that any such extension is driven by a "dynamical" viewpoint, but I don't see why that must be the case. I think the desire to have a model that has no arbitrary "edges" where spacetime just stops for no reason, is a valid adynamical desire. You appear to disagree, but I can see no reason why you should, and you have not given any reason for why you do.

    RUTA said

    Only a dynamical thinker believes some part of M needs to be explained independently from it fitting coherently into the whole of M.

    This has nothing to do with my issue. I am not asking you to explain the Einstein-de Sitter region in your model independently from fitting it into a larger model. I am asking why you have no larger model: why you just have the Einstein-de Sitter region and nothing else, when that region is not fitted coherently into any larger model, it's just sitting there with an obvious edge that, as far as I can see, has no reason for being there. If you think that region all by itself, with its edge, is a coherent whole adynamical model, I would like you to explain why. Just saying "oh, you're thinking dynamically so you just don't understand" doesn't cut it.

  16. RUTA says:

    Elias1960 said

    No, I don't. I continue to think that your point is refuted. Maybe you formulate your point in a different way which makes it possible to understand that it is not refuted?

    Your objections have not in any way refuted my claim as I've stated it many times as clearly as I know how. Sorry, I can't help you further.

    Elias1960 said

    But it makes it unnecessary. You claim to provide an explanation for the violation of the Bell inequality where no causal explanation is possible, as proven in a theorem, not?

    Read very carefully what I claimed in the Insight. I warn the reader that if they are unwilling or unable to accept the adynamical constraints as explanatory without a corresponding dynamical counterpart, then they will not believe I have explained the violation of Bell's inequality. That is the case for you. But, if you do accept the premise, then the conclusion (Bell's inequality has been explained) follows as a matter of deductive logic.

    Elias1960 said

    So, the double standard does not disappear if you only allow causal explanations. Causal explanations are required in science.

    That is your belief. I'm saying, "look at what you get if you accept that the constraints we have in physics are explanatory even in the absence of causal mechanisms." You don't believe they are, fine. But, that does not refute my claim. Again, I don't know how to state my point any more clearly than that. Sorry.

  17. RUTA says:

    PeterDonis said

    Because otherwise our model would predict that spacetime just ends for no reason. Unless you can give a reason, an adynamical reason, as I have asked you to do several times now, and you haven't.

    I've repeated many times that you only need to keep that part of M that you believe can or can conceivably produce empirically verifiable results. Every part of M fits self-consistently with every other part of M via EEs. Only a dynamical thinker believes some part of M needs to be explained independently from it fitting coherently into the whole of M. That's where you're coming from and that's why you keep believing I haven't answered your question. You're thinking dynamically.

  18. PeterDonis says:

    RUTA said

    Why do you believe that?

    Because otherwise our model would predict that spacetime just ends for no reason. Unless you can give a reason, an adynamical reason, as I have asked you to do several times now, and you haven't.

    RUTA said

    What is wrong with that?

    That there's no reason for it. Unless you can give a reason. But you haven't.

    RUTA said

    I suspect we're getting to your dynamical bias.

    I have made no dynamical claims whatever. As I have repeatedly said, I am taking your blockworld viewpoint in which spacetime is a 4-D geometry that doesn't change or evolve, it just is. Asking for a reason does not mean asking for a dynamical reason. An adynamical reason would be fine. But you have given no reason.

  19. RUTA says:

    PeterDonis said

    So, in other words, you think ##\rho = \infty## is unreasonable (and I agree), but you also think it's perfectly OK for a model to predict that some timelike observer's worldline can just suddenly cease to exist in the past, because it hits an "edge" of spacetime?

    Absolutely. What is wrong with that? I suspect we're getting to your dynamical bias.

  20. RUTA says:

    PeterDonis said

    But I also don't think that just arbitrarily cutting off a 4-D spacetime geometry is reasonable; I think a reasonable model has to include everything that can be included up to the maximal analytic extension.

    Why do you believe that?

  21. PeterDonis says:

    RUTA said

    That the existence of past inextendable timelike or null geodesics is "pathological."

    So, in other words, you think ##\rho = \infty## is unreasonable (and I agree), but you also think it's perfectly OK for a model to predict that some timelike observer's worldline can just suddenly cease to exist in the past, because it hits an "edge" of spacetime?

  22. PeterDonis says:

    RUTA said

    I said IF you believe …, then why?

    You didn't come across to me as saying "IF", but fine. I don't think ##\rho = \infty## is reasonable. But I also don't think that just arbitrarily cutting off a 4-D spacetime geometry is reasonable; I think a reasonable model has to include everything that can be included up to the maximal analytic extension. If the maximal analytic extension of a particular idealized model leads to ##\rho = \infty## somewhere, to me that's a reason for adjusting the model. Inflationary cosmology adjusts the model by changing the stress-energy tensor prior to the end of inflation to one that violates the energy conditions and therefore does not require ##\rho = \infty## anywhere.

    RUTA said

    Then we're in agreement that the flat, matter-dominated cosmology model does not have to include ##a = 0## with ##\rho = \infty##.

    I would say that a model which fixes the ##\rho = \infty## problem, by adjusting the stress-energy tensor prior to some spacelike hypersurface, is no longer a simple "flat, matter-dominated cosmology model"; it includes a region that is flat and matter-dominated, but that is not the entire model. (Note that in our best current model of our universe, the flat, matter-dominated region ends a few billion years before the present; our universe at present in our best current model is dark energy dominated, not matter dominated. So even the flat, matter-dominated region itself is an extrapolation; it's not what we currently observe.)

  23. RUTA says:

    PeterDonis said

    I'm a little confused about what you think Wald's position is. Wald describes the singularity theorems and what they show in Chapter 9, yes. And you have already agreed that cutting off a solution that, when maximally extended, has a singularity, before the singularity is reached, as you did with your version of the Einstein-de Sitter model, does not contradict the singularity theorems. So what, exactly, do you disagree with Wald about?

    That the existence of past inextendable timelike or null geodesics is "pathological."

  24. PeterDonis says:

    RUTA said

    Wald

    I'm a little confused about what you think Wald's position is. Wald describes the singularity theorems and what they show in Chapter 9, yes. And you have already agreed that cutting off a solution that, when maximally extended, has a singularity, before the singularity is reached, as you did with your version of the Einstein-de Sitter model, does not contradict the singularity theorems. So what, exactly, do you disagree with Wald about?

  25. RUTA says:

    PeterDonis said

    And if you want to ask @Elias1960 to justify his viewpoint, or email Wald to ask him to justify his, that's fine. But that doesn't explain why you keep asking me to justify a claim that I have never made.

    I never claimed you did say that. I said IF you believe …, then why? Your response should have been simply, "I don't believe … ." Then we're in agreement that the flat, matter-dominated cosmology model does not have to include ##a = 0## with ##\rho = \infty##.

  26. PeterDonis says:

    RUTA said

    from an adynamical perspective it's perfectly reasonable to include only that region of M that you believe can or can conceivably render empirical results

    Why? Why should spacetime just suddenly end at the point where our ability to observe stops?

    For example, consider Schwarzschild spacetime at and inside the horizon. This region is in principle unobservable from outside the horizon. Are you saying we should arbitrarily cut off our models of black holes just a smidgen above the horizon?

    Note that I am not saying that "dynamics" requires us to continue spacetime. I am considering spacetime just like you are in your blockworld viewpoint, as a 4-D geometry that doesn't change or evolve, it just is. I'm asking for an adynamical reason why 4-D spacetime should just suddenly end, and "because that's all we can observe" doesn't seem like a valid one to me.

  27. PeterDonis says:

    RUTA said

    I never said you shouldn't explore the observable consequences of pushing the model back in time. That's exactly what was done to make the predictions of anisotropies in the CMB power spectrum many years before we made the observations. But, and this IS what I'm saying, you let the physics dictate that extrapolation, not the math.

    To me, "let the physics dictate the extrapolation" means "explore the observable consequences of pushing the model back in time". So I don't see the distinction you are making here.

    RUTA said

    the only problematic region is ##\rho = \infty##, so that's why I'm asking what physics you believe justifies me keeping that region

    And, as I have said repeatedly now, I have never made any claim that there is justification for keeping ##\rho = \infty##. So I don't see why you keep asking me to justify a claim that I have never made.

    RUTA said

    you keep agreeing with me that we should push back farther into time

    But I have never claimed that "push farther back into time" requires including ##\rho = \infty##. In fact I have explicitly said the opposite, when I pointed out that inflation models do not require ##\rho = \infty## anywhere and that eternal inflation models do not have it anywhere.

    RUTA said

    Wald and Elias1960 are clear

    And if you want to ask @Elias1960 to justify his viewpoint, or email Wald to ask him to justify his, that's fine. But that doesn't explain why you keep asking me to justify a claim that I have never made.

  28. RUTA says:

    Elias1960 said

    No, I do not plan to accept that constraints are explanatory, I'm sure they are not and have arguments for this.

    Fine, but that does not in any way refute my point. Do you understand that fact?

    Elias1960 said

    It is this typical appearance of double standards that I try to attack with my tobacco industry, astrology, and creationist analogies.

    There is no double standard here. I told you constraint-based explanation does not rule out causal explanation in my view. Indeed, the vast majority of our experience can be easily explained via causal mechanisms. I never said otherwise, I'm simply showing how everything can be explained self-consistently by assuming adynamical constraints are fundamental. If you have a dynamical counterpart, do what I'm doing, i.e., publish papers explaining the idea and use it to explain experimental results (which involves fitting data and comparing to other fitting techniques), present at conferences, write a book with a legit academic press, etc. That's the academic game.

  29. RUTA says:

    PeterDonis said

    I don't see the difference. In both cases you have a model and an obvious way to extend it. The only difference is that the ball is not the entire universe, but if that actually made a difference it would mean, by your logic, that we can never extrapolate anything for the entire universe beyond what we have already observed. Which, as I have said, is not how progress has been made in science.

    You're not even reading what I'm saying. I have never said we have to do that. You are talking as if this is the only possible extension of any cosmological model beyond what we have already observed. It isn't.

    We are also looking at extending ##\Lambda\text{CDM}##, for example with inflation models. By your logic, nobody should be bothering to do that unless and until we get some actual direct observations from an inflationary epoch.

    I have never made any such claim. I don't know who you think you are responding to with these repeated references to ##\rho = \infty##, but it isn't me. You need to read what I'm actually saying instead of putting words in my mouth.

    You're not responding to what I'm saying. I never said you shouldn't explore the observable consequences of pushing the model back in time. That's exactly what was done to make the predictions of anisotropies in the CMB power spectrum many years before we made the observations. But, and this IS what I'm saying, you let the physics dictate that extrapolation, not the math. Again, the only problematic region is ##\rho = \infty##, so that's why I'm asking what physics you believe justifies me keeping that region. And you keep agreeing with me that we should push back farther into time, which does not answer my question about the only problematic region. Wald and Elias1960 are clear about why they believe we are forced to include ##a = 0## with ##\rho = \infty## in M — it's pathological from a dynamical perspective not to do so. But, from an adynamical perspective it's perfectly reasonable to include only that region of M that you believe can or can conceivably render empirical results. This doesn't rule out the exploration of theories like inflation at all, regardless of what motivates them. They constitute exploration of alternate cosmology models, which is of course a perfectly reasonable thing to do. If the new model makes a prediction that disagrees with the current best cosmology model in some respect while agreeing with all currently available data that the current model gets right, and that prediction vindicates the new model, then the new model wins. There is nothing in this process that says we have to accept empirically unmotivated mathematical extrapolations, i.e., those that cannot or cannot conceivably render empirical results. So, do you believe such empirically unmotivated extrapolations are required? If so, why?

  30. PeterDonis says:

    RUTA said

    The difference here is that there is an external context for the ball's trajectory where there is no such external context for cosmology.

    I don't see the difference. In both cases you have a model and an obvious way to extend it. The only difference is that the ball is not the entire universe, but if that actually made a difference it would mean, by your logic, that we can never extrapolate anything for the entire universe beyond what we have already observed. Which, as I have said, is not how progress has been made in science.

    RUTA said

    without an external context and a physical motivation otherwise, what would motivate you to include ##a = 0## with ##\rho = \infty## in your model?

    You're not even reading what I'm saying. I have never said we have to do that. You are talking as if this is the only possible extension of any cosmological model beyond what we have already observed. It isn't.

    RUTA said

    That is precisely what we're doing now with ##\Lambda\text{CDM}##, i.e., we're using it where it can account for observations.

    We are also looking at extending ##\Lambda\text{CDM}##, for example with inflation models. By your logic, nobody should be bothering to do that unless and until we get some actual direct observations from an inflationary epoch.

    RUTA said

    There is something that is driving you to believe ##a = 0## with ##\rho = \infty## should be included

    I have never made any such claim. I don't know who you think you are responding to with these repeated references to ##\rho = \infty##, but it isn't me. You need to read what I'm actually saying instead of putting words in my mouth.

  31. RUTA says:

    Elias1960 said

    What means "to account for our observations"? This is clear and obvious for me, given my "dynamical thinking", and my insistence on Reichenbach's common cause principle which defines what is a reasonable explanation. You reject both and rely on consistency only.

    To account for the Planck distribution of the CMB or the anisotropies in its power spectrum, for example. The self-consistency I'm talking about is in EEs. Did you read my GR Insight on that? And I do not reject dynamical explanation. I use it all the time.

    My claim is that if you view adynamical constraints as fundamental to dynamical laws, then many mysteries of modern physics, such as entanglement per the Bell states, disappear. You have said nothing to refute that point. All you have done is espouse your dynamical bias in response. If I had claimed that you MUST use constraints to dispel the mysteries of modern physics, then your replies would be relevant. But, I never made that claim. To refute my claim, you would have to accept my premise that constraints are explanatory and show how they fail to explain something that I claim they explain. So, for example, show how my constraint, conservation per NPRF, cannot explain conservation per the Bell states, conceding first that constraints are explanatory. I don't see how that's possible, but I'll let you try. You haven't even made an effort wrt cosmology, all you've done is espouse your dynamical bias there.

  32. RUTA says:

    PeterDonis said

    Because we already know there is a constraint: the ball wasn't freely flying at negative times, it was sitting on the ground. So we don't extend the parabolic trajectory to negative times because we know it doesn't apply. Instead, we join that trajectory to a different trajectory for negative times.

    Now, suppose that we weren't watching the ball at all at negative times and had no empirical evidence whatever of its trajectory then. But we do know that the surface of the Earth is there and that the parabolic trajectory intersects that surface at ##t = 0##. How would we model the ball? Would we just throw up our hands and say, well, we don't have any evidence at negative times so we'll just cut off our model at ##t = 0## and stop there? Or would we exercise common sense and predict that, at negative times, the ball is sitting on the surface of the Earth, and someone threw it upwards at time ##t = 0##, and extend our model accordingly?

    As far as I can tell, you prefer the first alternative and I (and others, it appears) prefer the second. Can I give you a logical proof that you must use the second alternative? No. But I can tell you that the first alternative makes no sense to me, and I suspect it makes no sense to a lot of other people.

    The difference here is that there is an external context for the ball's trajectory where there is no such external context for cosmology. You're tacitly using that external context to infer empirical results. Again, in physics there must be some empirical rational for using the mathematics. So, yes, without an external context and a physical motivation otherwise, what would motivate you to include ##a = 0## with ##\rho = \infty## in your model? The burden is on you to motivate the use of the math. That is precisely what we're doing now with ##\Lambda\text{CDM}##, i.e., we're using it where it can account for observations. If someone used ##a = 0## with ##\rho = \infty## to make a testable empirical prediction and that prediction was verified, then we would include it in our model. It's that simple. You're just not giving me any empirical reason to include that region, so why would I? There is something that is driving you to believe ##a = 0## with ##\rho = \infty## should be included despite the lack of empirical motivation. Can you articulate that motive?

  33. PeterDonis says:

    RUTA said

    We don’t include the mathematical extension into negative times demanding therefore we must include ##y = 0##. Why?

    Because we already know there is a constraint: the ball wasn't freely flying at negative times, it was sitting on the ground. So we don't extend the parabolic trajectory to negative times because we know it doesn't apply. Instead, we join that trajectory to a different trajectory for negative times.

    Now, suppose that we weren't watching the ball at all at negative times and had no empirical evidence whatever of its trajectory then. But we do know that the surface of the Earth is there and that the parabolic trajectory intersects that surface at ##t = 0##. How would we model the ball? Would we just throw up our hands and say, well, we don't have any evidence at negative times so we'll just cut off our model at ##t = 0## and stop there? Or would we exercise common sense and predict that, at negative times, the ball is sitting on the surface of the Earth, and someone threw it upwards at time ##t = 0##, and extend our model accordingly?

    As far as I can tell, you prefer the first alternative and I (and others, it appears) prefer the second. Can I give you a logical proof that you must use the second alternative? No. But I can tell you that the first alternative makes no sense to me, and I suspect it makes no sense to a lot of other people.

    RUTA said

    in adynamical thinking the onus is on you to produce a prediction with empirical evidence showing you need to include ##a = 0## with ##\rho = \infty##.

    I haven't made any such prediction. I don't have a problem with looking for a solution that does not have ##\rho = \infty## at ##t = 0##. And we have such solutions: inflationary models do not require ##\rho = \infty## at ##t = 0##. Eternal inflation is a possibility. Other possibilities have been suggested as well. If your position is that everybody except you is stuck in a rut thinking we have to have ##\rho = \infty## at ##t = 0##, then I think you are ignoring a lot of work being done in cosmology.

    OTOH, what I do have a problem with is saying, oh, well, we don't have any empirical evidence for times before the hot, dense, rapidly expanding state that in inflationary models occurs at the end of inflation, so we'll just cut off the model there and pretend nothing existed before that at all, it just suddenly popped into existence for no reason. That, to me, is not valid adynamical thinking. Valid adynamical thinking, to me, would be that the 4-D spacetime geometry, which does not "evolve" but just "is", should extend to wherever its "natural" endpoint is. The most natural thing would be for it to have no boundary at all, which means that if your model has a boundary in it, which it certainly does if you arbitrarily cut off the model the way you are describing, your model is obviously incomplete. Unless you can show some valid adynamical constraint that requires there to be a boundary at that particular place in the 4-D geometry. I have not seen any such argument from you.

    RUTA said

    I'm saying we should take the model as far back as necessary to account for our observations.

    But why should we stop there? Why should our observations be the criterion for where the 4-D spacetime geometry of the universe has a boundary?

    RUTA said

    There is no reason to include mathematics in physics unless that mathematics leads to empirically verifiable predictions.

    Inflationary models, which carry the 4-D spacetime geometry of the universe back past the earliest point we can currently observe directly, do make empirically verifiable predictions. But those models were developed before anyone knew that they would be able to make such predictions. You seem to be saying nobody should bother working on any model unless it covers a domain we already have empirical data from. That doesn't make sense to me; if we did that we would never make any predictions about observations we haven't made yet. But science progresses by making predictions about observations we haven't made yet.

    RUTA said

    No one is saying, “Well, if you extrapolate that cosmology model backwards in time far enough, you get ##\rho = \infty##, so I guess we have to stop using it otherwise.”

    You're right that no one is saying that. But that's because no one is extrapolating the model backwards in time to ##\rho = \infty## in the first place. Everyone appears to me to be looking at how to extend our best current model in ways that don't require ##\rho = \infty## anywhere. Nobody appears to me to be saying, "oh, well, we'll just have to arbitrarily cut off the model at the earliest point where we can make observations, and say that adynamical thinking prevents us from going further until we have more evidence".

  34. RUTA says:
    It's always true, even with dynamical thinking, that reality could have begun 10 min ago with all the signs of a past beyond that. You can choose to model reality that way, but I'm not suggesting we do so. I'm saying we should take the model as far back as necessary to account for our observations. With anisotropies in the CMB that's before decoupling. @Elias1960, you have a strong dynamical bias, so you should pursue a model consistent with that. Again, you're not in any way refuting the point I'm making by simply espousing your dynamical bias.
  35. RUTA says:

    martinbn said

    … that is what is done in GR anyway.

    Exactly, we’re using ##\Lambda\text{CDM}## successfully to make predictions relying on conditions even before decoupling (anisotropies in CMB power spectrum depend on pre-decoupling oscillations). No one is saying, “Well, if you extrapolate that cosmology model backwards in time far enough, you get ##\rho = \infty##, so I guess we have to stop using it otherwise.” That’s silly. Again and again, as I keep showing, adynamical thinking vindicates what we’re doing in modern physics, revealing its coherence and integrity. The Insight here is refuting Smolin et al. who believe “quantum mechanics is wrong because it’s incomplete.” Modern physics isn’t wrong or incomplete, it’s true it isn’t finished (we need to connect quantum and GR), but what we have is totally right. All its mysteries can be attributed to our dynamical bias (you don’t have to attribute them to that, but you can).

  36. martinbn says:

    RUTA said

    The key phrase there is "if I need the value of ##y = 0##" (the origin of the time parameterization is irrelevant of course). So, what dictates your need? Empirical results, not math results. Same with ##a = 0## with ##\rho = \infty##. Do you have an empirically verifiable prediction requiring we keep ##a = 0## with ##\rho = \infty##? If so, we'll check it and if you're right, we'll need to keep that region. Otherwise, why would we keep it?

    I suppose I missunderstood. I thought you were claiming that at ##t=0##, the quantity ##a## has to have a value, and since the value zero is problematic you don't use that value, but you use a different value. Of course not using any value and saying that the solution is valid only vor ##t>0## is fine, and that is what is done in GR anyway.

  37. RUTA says:
    The key phrase there is "if I need the value of ##y = 0##" (the origin of the time parameterization is irrelevant of course). So, what dictates your need? Empirical results, not math results. Same with ##a = 0## with ##\rho = \infty##. Do you have an empirically verifiable prediction requiring we keep ##a = 0## with ##\rho = \infty##? If so, we'll check it and if you're right, we'll need to keep that region. Otherwise, why would we keep it?
  38. martinbn says:
    My point was, that in the given example if I need the value ##y(0)##, I don't have the freedom the choose it, it is a consequence of the equation and the other values. It seemed to me that you were saying that in the cosmological model you can just addjast that value?
« Older Comments

Leave a Reply

Want to join the discussion?
Feel free to contribute!

Leave a Reply