Sunday, December 18, 2016

Is the Decoherent Histories Approach Consistent?


One particular approach of interpreting quantum mechanics is Decoherent Histories. All major non-Copenhagen approaches have serious issues: 
- MWI has the issue of the very meaning of probability and without a non-circular derivation (impossible in my opinion) of Born rule does not qualify for anything but a "work in progress" status.
-GRW-type theories make different predictions than quantum mechanics which soon will be confirmed or rejected by ongoing experiments. (My bet is on rejection since later GRW versions tuned their free parameters to avoid collision with known experimental facts instead of making a falsifiable prediction) 
-Bohmian approach has issues with "surreal trajectories" which invalidates their only hard ontic claim: the position of the particle.

Now onto Decoherent Histories. I did not closely follow this approach and I cannot state for sure if there are genuine issues here, but I can present the debate. On one hand, Robert Griffiths states: 

"What is different is that by employing suitable families of histories one can show that measurement actually measure something that is there, rather than producing a mysterious collapse of the wave function"

On the other hand he states:

"Any description of the properties of an isolated physical system must consists of propositions belonging together to a common consistent logic" - in other words he introduces contextuality.

Critics of decoherent (or consistent) histories use examples which are locally consistent but globally inconsistent to criticize the interpretation.

Here is an example by Goldstein (other examples are known). The example can be found in Bricmont's recent book: Making Sense of Quantum Mechanics on page 231. Consider two particles and two basis for a two-dimensional spin base \((|e_1\rangle\, |e_2\rangle), (|f_1\rangle, |f_2\rangle))\) and consider the following state:

\(|\Psi\rangle = a |e_1\rangle|f_2\rangle + a|e_2\rangle|f_1\rangle - b |e_1\rangle|f_1\rangle\)

Then consider four measurements A, B, C, D corresponding to projectors on four vectors, respectively: \(|h\rangle, |g\rangle, |e_2\rangle, |f_2\rangle\) where:

\(|g\rangle = c|e_1\rangle + d|e_2\rangle\)
\(|h\rangle = c|f_1\rangle + d|f_2\rangle\)

Then we have the following properties:

(1) A and C can be measured simultaneously, and if A=1 then C=1 
(2) B and D can be measured simultaneously, and if B=1 then D=1
(3) C and D can be measured simultaneously, but we never get both C and D = 1
(4) A and B can be measured simultaneously, and sometimes we get both A and B = 1

However all 4 statements cannot be true at the same time: when A=B=1 as in (4) then by (1) and (2) C=D=1 and this contradicts (3).

So what is going on here? The mathematical formalism of decoherent histories is correct as they predict nothing different than standard quantum mechanics. The interpretation assigns probabilities to events weather we observe them or not, but does it only after taking into account contextuality. Is this a mortal sin of the approach? Nature is contextual and I don't get the point of the criticism. The interpretation would be incorrect if it does not take into account contextuality. Again, I am not an expert of this approach and I cannot offer a definite conclusion, but to state my bias I like the approach and my gut feeling is that the criticism is without merit.
 
PS: I'll be going on vacation soon and my next post will be delayed: I will skip a week.

Sunday, December 11, 2016

Measurement = collapse + irreversibility


I got a lot of feedback from last 2 posts and I need to continue the discussion. Even Lubos with his closed mind unable to comprehend anything different than the textbooks from 50 years ago and his combative style said something worth discussing.

But first let me thank Cristi for the picture below which will help clarify what I am trying to state. Let me quickly explain it: the interferometer arms are like the two sides of Einstein's box and once the particle was launched -for the duration of the flight- you can close the input and output of the interferometer, open the exit just in time and still have the interference. So this seems to contradict my prediction. But does it?




This time I do need to dig a bit deeper into the mathematical formalism. First, the role of the observer is paramount: no observer=no measurement. Second, the observer is described by quantum mechanics as well: there is the wavefunction of the quantum system, and there is the wavefunction of the observer. Now here is the new part: while we can combine the quantum system and the observer by tensor product and do the usual discussion of how unitary evolution does not predict an unique outcome, we need to combine the quantum system and the observer using the Cartesian product. This is something new, not present in standard quantum mechanics textbooks. However this follows naturally from the category theory derivation of quantum mechanics from first principles. There are equivalent Cartesian products corresponding to potential measurement outcomes:

\((|collapsed ~1 \rangle, | observer~ see~1\rangle ) \equiv ( |collapsed~2\rangle, | observer~see~2\rangle)\) 

This equivalence exists in a precise mathematical sense and respects the three properties of the equivalence relationship: reflexivity, symmetry, and transitivity. Break the Cartesian pair equivalence by any mechanism possible and you get the collapse of the wavefunction. 

Closing the interferometer, or cutting Einstein's box in half kills the equivalence and the wavefunction collapses. However while the particle is still in flight the process is reversible!!! Open the interferometer's exits in time and you restore the equivalence, undo the collapse still get the interference (Han Solo kills the stormtrooper 100% of the time). 

However there is a way to make the collapse permanent: just wait enough time with the ends closed such that the energy-time uncertainty relation allows you to reduce the energy uncertainty to the point that you can detect the particle inside by weighing the half-boxes or the arms of the interferometer. Suppose the ends of the interferometer is made out of perfect mirrors. If you wait long enough (even though you are not physically weighing anything) and then reopen the exits will result in loss of interference: this is my prediction.

But what happens if you only wait a little bit of time and you are in between full interference and no interference? You get a weak measurement.

Now let me discuss Lubos objection and then come back to the nonlocality point Bricmont was making.

First Lubos was stating: "If you place some objects (a wall) at places where a particle is certain not to be located, the effect on the particle's future behavior is obviously non-existent". The objection is vacuous. Obviously I don't disagree with the statement, but his entire line of argument is incorrect because collapse is not a dynamic process. If collapse would have had a dynamic origin then we would have had a unitary explanation for it we would have had to talk about the "propagation" of collapse. What the Cartesian pair mathematical framework does is first getting the rid of the consciousness factor, and second clarifying the precise mathematical framework of how the observer should be treated within the formalism. Contextuality is paramount in quantum mechanics and cutting the box changes the context of the experiment.

Now onto Bricmont argument. Andrei stated in his comments: " I still do not see the relevance of all this in regards to the locality dilemma.". It has deep relevance as I will explain. And by the way, the rest of Andrei's comments were simply not worth answering-nothing personal: I don't have the luxury of enough free time to answer each and every comment.

Bricmont's point on Einstein's boxes was this: "either there is action at a distance in nature (opening B1 changes the situation at B2), or the particle was in B2 all along and quantum mechanics is incomplete "

Let's discuss the two options:
  1. opening B1 changes the situation at B2
  2. or the particle was in B2 all along and quantum mechanics is incomplete
Option one is clearly not the case. Wait long enough and the interference will no longer happen. At that point the particle IS in either B1 or B2 and shipping one box far away changes nothing. But how about option 2? Is quantum mechanics incomplete? Bohmian supporters think so because they augment the wavefunction with a hidden variable: the particle's initial position. Do we actually need this initial condition to make predictions? Not at all. Last thing to consider: was the particle in say B2 all along? If yes, there is no interference because the which way information. What about weak measurements? This is a case of even more examples of "surrealistic trajectories": combine two interferometers and you can obtain disjoint paths!!! The only thing which makes sense is that the particle does not have a well defined trajectory.

My question to Bohmian interpretation supporters is as follows: In the above picture close the arms long enough. What happens to the quantum potential? Does it dissipate? If yes how? If no, do you always get interference after opening the stormtrooper end regardless of the wait time? 

Finally back to measurement. There is no (strong) measurement without collapse. Collapse happens when a particular equivalence relationship no longer holds. Mathematically it can be proven that the wavefunction is projected on a subspace of the original Hilbert space. Moreover uniqueness can be proven as well: that this is the only mathematically valid mechanism in which projection can occur. Interaction between the quantum system and the measurement device can break the equivalence, but changing the experimental context can achieve the same thing as well. A measurement does not happen however until irreversibility occurs: there could be amplification effects, or as above enough time passes such that the energy uncertainty is low enough and the "which way" information becomes available regardless if we seek this information or not.

Sunday, December 4, 2016

A measurement can be more than an observer learning the value of a physical observable


Last post created quite a stir and I want to expand on the ideas from it. This will also help me get out of an somewhat embarrassing situation. For months now Lubos Motl tried to get revenge on his bruised ego after a well deserved April Fool's joke and became a pest at this blog. The problem is that although I have yet to see a physics post at his blog that is 100% correct, we share roughly the same intuition about quantum mechanics: I agree more much more with his position than say with the Bohmian, GRW, or MWI approaches. The differences are on the finer points and I found his in depth knowledge rusty and outdated. For his purpose: to discredit the opposite points of view at all costs this is enough, but it does not work if you are a genuine seeker of truth.

So last time he commented here: "A measurement is a process when an observer actually learns the value of a physical observable" which from 10,000 feet is enough. However this is not precise enough, and now I do have a fundamental disagreement with Lubos which hopefully will put enough distance between him and me. 

More important than my little feud with Lubos, I can now propose an experiment which will either validate or reject my proposed solution to the measurement problem. I do have a novel proposal on how to solve the measurement problem and this is distinct from all other approaches. I was searching for months for a case of a novel experimental prediction, but when I applied it to many problems I was getting the same predictions as standard quantum mechanics. Here is however a case where my predictions are distinct. I will not work out the math and instead let me simply present the experiment and make my experimental claim.


Have a box with a single particle inside. The box has a middle separator and also two slits A and B which can be placed next to a two-slit screen. We can then carry two kinds of experiments:

  1. open the two slits A and B without dropping the separator allowing the particle to escape the box and hit a detector screen after the two-slit screen.
  2. drop the separator and then open the two slits A and B allowing the particle to escape the box and hit a detector screen after the two-slit screen.
Next we repeat the experiments 1 or 2 enough times to see the pattern emerge on the final screen. Which pattern would we observe?

For experiment 1 we already know the answer: if we repeat it many times we obtain the interference pattern, but what will we get in the case of experiment number 2?

If dropping the separator constitutes a measurement, the wavefunction would collapse and we get two spots on the detector screen corresponding to two single slit experiments. If however dropping the separator does not constitute a measurement, then we would get the same interference pattern as in experiment 1.

My prediction (distinct from textbook quantum mechanics) is that there will be no interference pattern.


Monday, November 28, 2016

Are Einstein's Boxes an argument for nonlocality?

(an experimental proposal)


Today I want to discuss a topic from an excellent book by Jean Bricmont: Making Sense of Quantum Mechanics which presents the best arguments for the Bohmian interpretation. Although I do not agree with this approach I appreciate the clarity of the arguments and I want to present my counter argument.

On page 112 there is the following statement: "... the conclusion of his [Bell] argument, combined with the EPR argument is rather that there are nonlocal physical effects (and not just correlations between distant events) in Nature". 

To simplify the argument to its bare essentials, a thought experiment is presented in section 4.2: Einstein's boxes. Here is how the argument goes: start with a box B and a particle in the box, then cut the box into two half-boxes B1 and B2. If the original state is \(|B\rangle\), after cutting the state it becomes:

\(\frac{1}{\sqrt{2}}(|B_1\rangle+|B_2\rangle)\) 

Then the two halves are spatially separated and one box is opened. Of course the expected thing happens: the particle is always found in one of the half-boxes. Now suppose we find the particle in B2. Here is the dilemma: either there is action at a distance in nature (opening B1 changes the situation at B2), or the particle was in B2 all along and quantum mechanics is incomplete because \(\frac{1}{\sqrt{2}}(|B_1\rangle+|B_2\rangle)\) does not describe what is going on. My take on this is that the dilemma is incorrect. Splitting the box amounts to a measurement regardless if you look inside the boxes or not and the particle will be in either B1 or B2.  

Here is an experimental proposal to prove that after cutting the box the state is not \(\frac{1}{\sqrt{2}}(|B_1\rangle+|B_2\rangle)\):

split the box and connect the two halves to two arms of a Mach-Zehnder interferometer (bypassing the first beam splitter). Do you get interference or not? I say you will not get any interference because by weighing the boxes before releasing the particle inside the interferometer gives you the which way information.

If we do not physically split the box, then indeed \(|B\rangle = \frac{1}{\sqrt{2}}(|B_1\rangle+|B_2\rangle)\), but if we do physically split it \(|B\rangle \neq \frac{1}{\sqrt{2}}(|B_1\rangle+|B_2\rangle)\). There is a hidden assumption in Einstein's boxes argument: realism which demands non-contextuality. Nature and quantum mechanics is contextual: when we do introduce the divider the experimental context changes. 

Bohmian's supporters will argue that always \(|B\rangle = \frac{1}{\sqrt{2}}(|B_1\rangle+|B_2\rangle)\). There is a simple way to convince me I am wrong: do the experiment above and show you can tune the M-Z interferometer in such a way that there is destructive interference preventing the particle to exit at one detector.

Sunday, November 20, 2016

Gleason's Theorem


It feels good to be back to physics, and as a side note going forward I will do the weekly posts on Sunday. Today I want to talk about Gleason's theorem. But what is Gleason's theorem?

If you want to assign a non-negative real valued function \(p(v)\) to every vector v of a Hilbert space H of dimension greater than two, then subject to some natural conditions the only possible choice is \(p(v) = {|\langle v|w \rangle |}^{2}\) for all vectors v and an arbitrary but fixed vector w.

Therefore there is no alternative in quantum mechanics to compute the average value of an observable A the standard way by using:

\(\langle A \rangle = Tr (\rho A)\)

where \(\rho\) is the density matrix which depends only on the preparation process. 

Gleason's theorem is rather abstract and we need to unpack its physical intuition and the mathematical gist of the argument. Physically, Gleason's theorem comes from three axioms:

  • Projectors are interpreted as quantum propositions
  • Compatible experiments correspond to commuting projectors
  • KEY REQUIREMENT: For any two orthogonal projectors P, Q, the sum of their expectation values is the expectation value of P+Q: \(\langle P \rangle + \langle Q\rangle = \langle P+Q\rangle\)
In an earlier post I showed how violating the last axiom (which is the nontrivial one), in the case of spin one particles, can be used to send signals faster than the speed of light and violate causality. But how does Gleason arrives at his result?

Let's return at the original problem: to obtain a real non-negative function p. Now add the key requirement and demand that for any complete orthonormal basis \(e_m\) we have:

\(\sum_m p(e_m) = 1\)

For example in two dimensions on a unit circle we must have:

\(p (\theta) + p(\theta + \pi/2) = 1\)

which constrain the Fourier expansion of \(p (\theta)\) such that only components 2, 6, 10, etc can be non zero. In three dimensions the constraints are much more severe and this involves rotations under SO(3) and spherical harmonics. I'll skip the tedious math, but it is not terribly difficult to show that the only allowed spherical harmonics must be of order 0 and 2 which yields: \(p(v) = {|\langle v|w \rangle |}^{2}\).

The real math heavy lifting is on dimensions larger than three and to prove it Gleason first generalizes  \(\sum_m p(e_m) = 1\) to \(\sum_m f(e_m) = k\) where k is any positive value. He names this "f" a "frame function". Then he proceeds to show that dimensions larger than three do not add anything new.

If you are satisfied with the Hilbert spaces of dimension 3, the proof of the theorem is not above undergrad level, and I hope it is clear what the argument is. But what about Many Worlds Interpretation? Can we use Gleason's theorem there to prove Born rule? Nope. The very notion of probabilities is undefined in MWI, and I am yet to see a non-circular derivation of Born rule in MWI. I contend it can't be done because it is a mathematical impossibility and I blogged about it in the past. 

Sunday, October 30, 2016

Einstein's reality criterion


Sorry for the delay, this week I want to have a short post continuing to showcase the results of late Asher Peres which unfortunately are not well known and this is a shame.



In the famous EPR paper, Einstein introduced a reality criterion:

"If, without in any way disturbing a system, we can predict with certainty ... the value of a physical quantity, then there exists an element of physical reality corresponding to this physical quantity"

Now it is generally accepted that the difference between classical and quantum mechanics is noncomutativity. While there are some subtle points to be made about this assertion (by the mathematical community), from the physics point of view the distinction is rock solid and we can build in confidence upon it.

Now consider again the  EPR-B experiment with its singlet state. Suppose the x and y components of the spin exists independent of measurement and let's call the measurement values: \(m_{1x}, m_{1y}, m_{2x}, m_{2y} \). From experimental results we know:

\(m_{1x} = - m_{2x}\)
and
\(m_{1y} = - m_{2y}\)

And now for the singlet state \(\psi\) let's compute:

\((\sigma_{1x} \sigma_{2y} + \sigma_{1y} \sigma_{2x})\psi\)

which turns out to be zero. The beauty of this is that \(\sigma_{1x} \sigma_{2y} \) commutes with \(\sigma_{1y} \sigma_{2x}\) and by Einstein's reality criterion extended to commuting operators it implies that \(m_{1x} m_{2y} = - m_{1y} m_{2x}\) which contradicts \(m_{1x} = - m_{2x}\) and \(m_{1y} = - m_{2y}\)

This contradiction is in the same vein as the GHZ result, but it is not well known. The catch of this result is that measuring S1xS2y cannot be done at the same time with a measurement of  S1yS2x and so we are reasoning counterfactually. However, counterfactual reasoning is allowed in a noncontextual setting (in classical mechanics and in quantum mechanics for commuting operators) and the result is valid. 


Friday, October 21, 2016

von Neumann  and Gleason vs. Bell


Returning to physics topics today I want to talk about an important contention point between von Neuman and Gleason on one had, and Bell on the other. I had a series of posts about Bell in which I discussed his major achievement. However I do not subscribe to his ontic point of view and today I will attempt to explain why and perhaps persuade the reader with what what I consider to be a solid argument. 

Before Bell wrote his famous paper he had another one in which he criticized von Neumann, Jauch and Piron, and Gleason. The key of the criticism was that additivity of orthogonal projection operators not necessarily implies the additivity of expectation values:

\(\langle P_u + P_v \rangle = \langle P_{u}\rangle + \langle P_{v}\rangle \)

The actual technical requirements in von Neumann and Gleason case were slightly different, but they can be reduced to the statement above and more importantly this requirement is the nontrivial one in a particular proof of Gleason's theorem

Andrew Gleason

To Bell, additivity of expectation values is a non-natural requirement because he was able to construct hidden variable models violating this requirement. And this was the basis for his criticism of von Neumann and his theorem of the impossibility of hidden variables. But is this additivity requirement unnatural? What can happen when it is violated? I will show that violation on additivity of expectation values can allow instantaneous communication at a distance.

The experimental setting is simple and involves spin 1 particles. The example which I will present is given in late Asher Peres book: Quantum Theory: Concepts and Methods at page 191. (This book is one of my main sources of inspiration for how we should understand and interpret quantum mechanics. )

The mathematical identity we need is:

\(J_{z}^{2} = {(J_{x}^{2} - J_{y}^{2})}^2\)

and the experiment is as follows: a beam of spin 1 particles is sent through a beam splitter which sends to the left particles of eigenvalue zero for \(J_{z}^{2}\) and to the right particles of eigenvalue one for \(J_{z}^{2}\).

Now a lab on the right decides to measure either if \(J_z = 1\) or if \(J_{x}^{2} - J_{y}^{2} = 1\)

For the laboratory on the right let's call the projectors in the first case \(P_u\) and \(P_v\) and in the second case \(P_x\) and \(P_y\)

For the lab on the left let's call the projectors in the first case \(P_{w1}\) and in the second case\(P_{w2}\). 

Because of the mathematical identity: \(P_u + P_v = P_x +P_y\) the issues becomes: should the expectation value requirement hold as well?

\(\langle P_{u}\rangle + \langle P_{v}\rangle  = \langle P_{x}\rangle + \langle P_{y}\rangle \)

For the punch line we have the following identities:

\(\langle P_{w1}\rangle = 1 - \langle P_{u}\rangle - \langle P_{v}\rangle\) 
and
\(\langle P_{w2}\rangle = 1 - \langle P_{x}\rangle - \langle P_{y}\rangle\) 

and as such if the additivity requirement is violated we have:

\(\langle P_{w1}\rangle \neq \langle P_{w2}\rangle\)

Therefore regardless of the actual spatial separation, the lab on the left can figure out which experiment the lab on the right decided to perform!!!

With this experimental setup, if additivity of expectation values is false, you can even violate causality!!!

Back to Bell: just because von Neumann and Gleason did not provide a justification for their requirements, this does not invalidate their arguments. The justification was found at a later time. 

But what about the Bohmian interpretation of quantum mechanics? Although there are superluminal speeds in the theory, superluminal signaling is not possible in it. This is because Bohmian interpretation respects Born rule which is a consequence of Gleason't theorem and it respects the additivity of  expectation values as well. Bohmian interpretation suffers from other issues however.  

Saturday, October 8, 2016

Local Causality in a Friedmann-Robertson-Walker Spacetime


A few days ago I learned about a controversy regarding Joy Christian's paper: 
Local Causality in a Friedmann-Robertson-Walker Spacetime which got published in Annals of Physics and was recently withdrawn: http://retractionwatch.com/2016/09/30/physicist-threatens-legal-action-after-journal-mysteriously-removed-study/


The paper repeats the same mathematically incorrect arguments of Joy Christian against Bell's theorem and has nothing to do with Friedmann-Robertson-Walker spacetime. The FRW space was only used as a trick to get the wrong referees which are not experts on Bell theorem. In particular the argument is the same as in this incorrect Joy's one-pager preprint

The mistake happens in two steps: 
  • a unification of two algebras into the same equation
  • a subtle transition from a variable to an index in a computation mixing apples with oranges
I will run the explanation in parallel between the one-pager and the withdrawn paper because it is easier to see the mistake in the one-pager.

Step 1: One-pager Eq. 3 is the same as FRW paper Eq. 49:

\(\beta_j \beta_k = -\delta_{jk} - \epsilon_{jkl} \beta_l\)
\(L(a, \lambda) L(b, \lambda) = - a\cdot b - L(a\times b, \lambda)\)

In the FRW paper \(L(a, \lambda) = \lambda I\cdot  a\) while in the 1-pager: \(\beta_j (\lambda) = \lambda \beta_j\) where \(\lambda\) is a choice of orientation. This make look as an innocuous unification but in fact it describes two distinct algebras with distinct representations.

This means that Eqs. 3/49 describe two multiplication rules (and let's call them A for apples and O for oranges). Unpacked, the multiplication rules are:

\(A_i A_j = -\delta_{jk} + \epsilon_{jkl} A_l\)
\(O_i O_j = -\delta_{jk} - \epsilon_{jkl} O_l\)

The matrix representations are:

\( A_1 = \left( \begin{array}{cc} i & 0 \\ 0 &  -i \end{array}\right) = i\sigma_3\)
\( A_2 = \left( \begin{array}{cc} 0 & -1 \\ 1 &  0 \end{array}\right) = -i \sigma_2\)
\( A_3 = \left( \begin{array}{cc} 0 & -i \\ -i &  0 \end{array}\right)= -i \sigma_1\)

and \(O_i = - A_i = {A_i}^{\dagger}\)

Try multiplying the above matrices to convince yourself that they are indeed a valid representation of the multiplication rule.

There is even a ket and bra or column and row vector representation of the two distinct algebras, but I won't go into details since it requires a math detour which will takes the focus away from Joy's mistake.

Step 2: summing apples with oranges (or column vectors with row vectors)

The summation is done in steps 5-7 and 67-75. The problem is that the sum from 1 to n contains two kinds of objects apples and oranges and should be in fact broken up in two sums. If this needs to be combined into a single sum then we need to convert apples and oranges to orientation independent objects. Since \(L(a, \lambda) = \lambda I\cdot  a\) and  \(\beta_j (\lambda) = \lambda \beta_j\) with \(I \cdot a\) and \(\beta_j\) orientation independent objects, when we convert the two kinds of objects to a single unified kind there is an additional missing factor of lambda.

Since \(O_j=\beta_j (\lambda^k) = \lambda^k \beta_j\) with \(\lambda^k = +1\) and \(A_j=-\beta_j (\lambda^k) = \lambda^k \beta_j\) with \(\lambda^k = -1\) where \(\lambda^k\) is the orientation of the k-th pair of particles, in the transition from 6 to 7 and 72 to 73  in an unified sum we are missing a \(\lambda^k \)  factor.

Again, either break up the sum into apples and oranges (where the index k tells you which kinds of objects you are dealing with) or unify the sum and adjust it by converting it into orientation-free objects and this is done by multiplication by \(\lambda^k\). If we separate the sums, they will not cancel each other out because there is -1 a conversion factor from apples to oranges  \(O = - A\), and if we unify the sum as Joy does in Eq. 74 the sum is not of \(\lambda^k\) but of \({(\lambda^k)}^2\) which does not vanish.

As it happens Joy's research program is plagued by this -1 (or missing lambda) mistake in his attempt to vanquish a cross product term. But even if his proposal were mathematically valid it would not represent a genuine challenge to Bell's theorem. Inspired by Joy's program, James Weatherall found a mathematically valid example very similar with Joy's proposal but one which does not use quaternions/Clifford algebras.

The lesson of Weatherall is that correlations must be computed using actual experimental results and the computation (like the one Joy is doing at steps 67-75) must not be made in a hypothetical space of "beables". 

Now back to the paper withdrawal, the journal did not acted properly: it should have notified Joy before taking action. However Joy did not act in good faith by masquerading the title to sneak it past imperfect peer review and his attempt at victimization in the comments section has no merit. In the end the paper is mathematically incorrect, has nothing to do with FRW spacetime, and (as proven by Weatherall) Joy's program is fatally flawed and cannot get off the ground even if there were no mathematical mistakes in it. 

Saturday, October 1, 2016

The whole is greater than the sum of its parts


The tile of today's post is a quote from Aristotle, but I want to illustrate this in the quantum formalism. Here I will refer to a famous Hardy paper: Quantum Theory From Five Reasonable Axioms. In there one finds the following definitions:

  • The number of degrees of freedom, K, is defined as the minimum number of probability measurements needed to determine the state, or, more roughly, as the number of real parameters required to specify the state. 
  • The dimension, N, is defined as the maximum number of states that can be reliably distinguished from one another in a single shot measurement.  
Quantum mechanics obeys \(K=N^2\) while classical physics obeys \(K=N\).

Now suppose nature is realistic and the electron spin does exist independent of measurement. From Stern-Gerlach experiments we know what happens when we pass a beam of electrons through two such devices rotates by an angle \(\alpha\): suppose we pick only the spin up electrons, on the second device the electrons are still deflected up \(\cos^2 (\alpha /2)\) percent of time and are deflected down \(\sin^2 (\alpha /2)\) percent of time . This is an experimental fact.

Now suppose we have a source of electron pairs prepared in a singlet state. This means that the total spin of the system is zero. There is no reason to distinguish a particular direction in the universe and with the assumption of the existence of the spin independent of measurement we can very naturally assume that our singlet state electron source produces an isotropic distribution of particles with opposite spins. Now we ask: in an EPR-B experiment, what kind of correlation would Alice and Bob get under the above assumptions?

We can go about finding the answer to this in three ways. First we can cheat and look the answer up in a 1957 paper by Bohm and Aharonov who first made the computation, This paper (and the answer) is cited by Bell in his famous "On the Einstein-Podolsky-Rosen paradox". But we can do better than that. We can play with the simulation software from last time. Here is what you need to do:

-replace the generating functions with:

function GenerateAliceOutputFromSharedRandomness(direction, sharedRandomness3DVector) {
    var cosAngle= Dot(direction, sharedRandomness3DVector);
    var cosHalfAngleSquared = (1+cosAngle)/2;
    if (Math.random() < cosHalfAngleSquared )
        return +1;
    else
        return -1;
};

function GenerateBobOutputFromSharedRandomness(direction, sharedRandomness3DVector) {
    var cosAngle= Dot(direction, sharedRandomness3DVector);
    var cosHalfAngleSquared = (1+cosAngle)/2;
    if (Math.random() < cosHalfAngleSquared )
        return -1;
    else
        return +1;
};

-replace the -cosine curve drawing with  a -0.3333333 cosine curve: 

boardCorrelations.create('functiongraph', [function(t){ return -0.3333333*Math.cos(t); }, -Math.PI*10, Math.PI*10],{strokeColor:  "#66ff66", strokeWidth:2,highlightStrokeColor: "#66ff66", highlightStrokeWidth:2});

replace the fit test for the cosine curve with one for with 0.3333333 cosine curve: 

var diffCosine = epsilon + 0.3333333*Math.cos(angle); 

and the result of the program (for 1000 directions and 1000 experiments) is:



So how does the program work? The sharedRandomness3DVector is the direction on which the spins are randomly generated. The dot product compute the cosine of the angle between the measurement direction and the spin, and from it we can compute the cosine of the half angle. The square of the cosine of the half angle is used to determine the random outcome. The resulting curve is 1/3 of the experimental correlation curve. Notice that the output generation for Alice and Bob are completely independent (locality).

But the actual analytical computation is not that hard to do either. We proceed in two steps.

Step 1: Let \(\beta\) be the angle between one spin \(x\) and a measurement device direction \(a\). We have: \(\cos (\beta) = a\cdot x\) and:

\({(\cos \frac{\beta}{2})}^2 = \frac{1+\cos\beta}{2} = \frac{1+a\cdot x}{2}\)

Keeping the direction \(x\) constant, the measurement outcomes for Alice and Bob measuring on the directions \(a\) and \(b\) respectively are:

++ \(\frac{1+a\cdot x}{2} \frac{1+b\cdot (-x)}{2}\) percent of the time
-- \(\frac{1-a\cdot x}{2} \frac{1-b\cdot (-x)}{2}\) percent of the time
+-\(\frac{1+a\cdot x}{2} \frac{1-b\cdot (-x)}{2}\) percent of the time 
-+\(\frac{1-a\cdot x}{2} \frac{1+b\cdot (-x)}{2}\) percent of the time 

which yields the correlation: \(-(a\cdot x) (b \cdot x)\)

Step 2: integrate \(-(a\cdot x) (b \cdot x)\) for all directions \(x\). To this aim align \(a\) on the z axis and have \(b\) in the y-z plane:

\(a=(0,0,a)\)
\(b=(0, b_y , b_z)\)

then go to spherical coordinates integrating using:

\(\frac{1}{4\pi}\int_{0}^{2\pi} d\theta \int_{0}^{\pi} \sin\phi d\phi\)

\(a\cdot x = \cos\phi\)
\(b\cdot x = b(0, \sin\alpha, -\cos\alpha)\cdot(\sin\phi \cos\theta, \sin\phi\sin\theta, \cos\phi)\)

where \(\alpha\) is the angle between \(a\) and \(b\).

Plugging all back in and doing the trivial integration yields: \(-\frac{\cos\alpha}{3}\)

So now for the moral of the story. the quantum mechanics prediction and the experimentally observed  correlation is  \(-\cos\alpha\) and not \(-\frac{1}{3} \cos\alpha\)

The 1/3 incorrect correlation factor comes from demanding (1) the experimentally proven behavior from two consecutive S-G device measurements, (2) the hypothesis that the electron spins exist before measurement, and (3) and isotropic distribution of spins originating from a total spin zero state.

(1) and (3) cannot be discarded because (1) is an experimental behavior, and (3) is a very natural demand of isotropy. It is (2) which is the faulty assumption.

If (2) is true then circling back on Hardy's result, we are under the classical physics condition: \(K=N\) which means that the whole is the sum of the parts. 

Bell considered both the 1/3 result and the one from his inequality and decided to showcase his inequality for experimental purposes reasons: "It is probably less easy, experimentally, to distinguish (10) from (3), then (11) from (3).". Both hidden variable models:

    if (Dot(direction, sharedRandomness3DVector) < 0)
        return +1;
    else
        return -1;

and 

    var cosAngle= Dot(direction, sharedRandomness3DVector);
    var cosHalfAngleSquared = (1+cosAngle)/2;
    if (Math.random() < cosHalfAngleSquared )
        return -1;
    else
        return +1; 

are at odds with quantum mechanics and experimental results. The difference between them is on the correlation behavior for 0 and 180 degrees. If we allow information transfer between Alice generating function and Bob generating function (nonlocality) then it is easy to generate whatever correlation curve we want under both scenarios (play with the computer model to see how it can be done).

So from realism point of view, which hidden variable model is better? Should we insist on perfect anti-correlations at 0 degrees, or should we demand the two consecutive S-G results along with realism? It does not matter since both are wrong. In the end local realism is dead.

Friday, September 23, 2016

Explanation for Bell's theorem modeling program


Today I will explain in detail the code from last time and show how can you change it to experiment with Bell's theorem. The code below needs only a text editor to make modifications and requires only a web browser to run. In other words, it is trivial to play with provided you understand the basics of HTML and Java Script. For elementary introductions to those topics see here and here.

In a standard HTML page we start in the body section with 3 entrees responsible to plot the graph in the end.

<body>
<link href="http://jsxgraph.uni-bayreuth.de/distrib/jsxgraph.css" rel="stylesheet" type="text/css"></link>
<script src="http://jsxgraph.uni-bayreuth.de/distrib/jsxgraphcore.js" type="text/javascript"></script>
<script src="http://jsxgraph.uni-bayreuth.de/distrib/GeonextReader.js" type="text/javascript"></script>


Then we have the following HTML table

<table border="4" style="width: 50%px;">
<tr><td style="width: 25%;">
<br />
Number of experiments: <input id="totAngMom" type="text" value="100" />
<br />
Number of directions: <input id="totTestDir" type="text" value="100" />
<br />
  
<input onclick="clearInput();" type="button" value="Clear Data" />

<input onclick="generateRandomData();" type="button" value="Generate Shared Random Data" />
<br />

<textarea cols="65" id="in_data" rows="7">
</textarea>
<br />

<input onclick="clearTestDir();" type="button" value="Clear data" />

<input onclick="generateTestDir();" type="button" value="Generate Random Alice Bob directions (x,y,z,x,y,z)" />
<textarea cols="65" id="in_test" rows="4">
</textarea>
<br />
<input onclick="clearOutput();" type="button" value="Clear Data" />

<input onclick="generateData();" type="button" value="Generate Data from shared randomness" />
<br />
Legend: Direction index|Data index|Measurement Alice|Measurement Bob
<textarea cols="65" id="out_measurements" rows="4">
</textarea>
<input onclick="clearBoard();" type="button" value="Clear Graph" />
<input onclick="plotData();" type="button" value="Plot Data" />

</td>
</tr>
<tr>
<td>
<div class="jxgbox" id="jxgboxCorrelations" style="height: 400px; width: 550px;">
</div>

</td></tr>
</table>

and we close the body:

</body>

The brain of the page is encapsulated by script tags:

<script type="text/javascript">
</script>

which can be placed anywhere inside the HTML page. Here are the functions which are declared inside the script tags:

//Dot is the scalar product of 2 3D vectors
function Dot(a, b)
{
 return a[0]*b[0] + a[1]*b[1] + a[2]*b[2];
};

This simply computes the dot product of two vectors in ordinary 3D Euclidean space. As a Java Script reminder, the arrays start at index zero and go to N-1. Also in Java Script comments start with two double slash // and lines end in semicolon ;

Next there is a little utility function which computes the magnitude of a vector:

//Norm computes the norm of a 3D vector
function GetNorm(vect)
{
 return Math.sqrt(Dot(vect, vect));
};

This is followed by another utility function which normalizes a vector:

//Normalize generates a unit vector out of a vector
function Normalize(vect)
{
 //declares the variable
 var ret = new Array(3);
 //computes the norm
 var norm = GetNorm(vect);

 //scales the vector
 ret[0] = vect[0]/norm;
 ret[1] = vect[1]/norm;
 ret[2] = vect[2]/norm;
 return ret;
};

To create an random oriented vector we use the function below which first randomly generates a point in a cube of side 2, eliminated the points outside a unit sphere, and then normalizes the vector:

//RandomDirection create a 3D unit vector of random direction
function RandomDirection()
{
 //declares the variable
 var ret = new Array(3);

 //fills a 3D cube with coordinates from -1 to 1 on each direction
 ret[0] = 2*(Math.random()-0.5);
 ret[1] = 2*(Math.random()-0.5);
 ret[2] = 2*(Math.random()-0.5);

 //excludes the points outside of a unit sphere (tries again)
 if(GetNorm(ret) > 1)
  return RandomDirection();
 return Normalize(ret);
};

The rest of the code is this:

var generateData = function()
{
   clearBoard();
 clearOutput();
 //gets the data
 var angMom = new Array();
 var t = document.getElementById('in_data').value;
 var data = t.split('\n');
 for (var i=0;i<data.length;i++)
 {
    var vect = data[i].split(',');
    if(vect.length == 3)
   angMom[i] = data[i].split(',');
   }
  
   var newTotAngMom = angMom.length;
   clearBoard();
 var varianceLinear = 0;
 var varianceCosine = 0;
 var totTestDirs = document.getElementById('totTestDir').value;


 var abDirections = new Array();
 var AliceDirections = new Array();
 var BobDirections = new Array();
 var t2 = document.getElementById('in_test').value;
 var data2 = t2.split('\n');
 for (var k = 0; k < data2.length; k++) 
 {
         var vect2 = data2[k].split(',');
         if (vect2.length == 6)
  {
              abDirections[k] = data2[k].split(',');
              AliceDirections[k] = data2[k].split(',');
   BobDirections[k] = data2[k].split(',');

       AliceDirections[k][0] = abDirections[k][0];
       AliceDirections[k][1] = abDirections[k][1];
       AliceDirections[k][2] = abDirections[k][2];
       BobDirections[k][0]   = abDirections[k][3];
       BobDirections[k][1]   = abDirections[k][4];
       BobDirections[k][2]   = abDirections[k][5];
  }
 }

 var TempOutput = "";

 //computes the output
 for(var j=0; j<totTestDirs; j++)
 {
         var a = AliceDirections[j];
         var b = BobDirections[j];
  for(var i=0; i<newTotAngMom; i++)
  {
       TempOutput = TempOutput + (j+1);
       TempOutput = TempOutput + ",";
       TempOutput = TempOutput + (i+1);
       TempOutput = TempOutput + ",";
       TempOutput = TempOutput + (GenerateAliceOutputFromSharedRandomness(a, angMom[i]));
       TempOutput = TempOutput + ",";
       TempOutput = TempOutput + (GenerateBobOutputFromSharedRandomness(b, angMom[i]));
   if(i != newTotAngMom-1 || j != totTestDirs-1)
        TempOutput = TempOutput + " \n";
  }
 }

 apendResults(TempOutput);
};

var plotData = function()
{
   clearBoard();
  boardCorrelations.suspendUpdate();
 //gets the data
 var angMom = new Array();
 var t = document.getElementById('in_data').value;
 var data = t.split('\n');
 for (var i=0;i<data.length;i++)
 {
    var vect = data[i].split(',');
    if(vect.length == 3)
   angMom[i] = data[i].split(',');
   }
  
   var newTotAngMom = angMom.length;
 var varianceLinear = 0;
 var varianceCosine = 0;
 var totTestDirs = document.getElementById('totTestDir').value;

 //extract directions
 var abDirections = new Array();
 var AliceDirections = new Array();
 var BobDirections = new Array();
 var t2 = document.getElementById('in_test').value;
 var data2 = t2.split('\n');
 for (var k = 0; k < data2.length; k++) 
 {
         var vect2 = data2[k].split(',');
         if (vect2.length == 6)
  {
              abDirections[k] = data2[k].split(',');
              AliceDirections[k] = data2[k].split(',');
   BobDirections[k] = data2[k].split(',');

       AliceDirections[k][0] = abDirections[k][0];
       AliceDirections[k][1] = abDirections[k][1];
       AliceDirections[k][2] = abDirections[k][2];
       BobDirections[k][0]   = abDirections[k][3];
       BobDirections[k][1]   = abDirections[k][4];
       BobDirections[k][2]   = abDirections[k][5];
  }
 }


 var tempLine = new Array();
 var Data_Val = document.getElementById('out_measurements').value;
 var data_rows = Data_Val.split('\n');

 var directionIndex = 1;
 var beginNewDirection = false;

       var a = new Array(3);
 a[0] = AliceDirections[0][0];
 a[1] = AliceDirections[0][1];
 a[2] = AliceDirections[0][2];
        var b = new Array(3);
 b[0] = BobDirections[0][0];
 b[1] = BobDirections[0][1];
 b[2] = BobDirections[0][2];
 var sum = 0;


 for (var ii=0;ii<data_rows.length;ii++)
 {
  //parse the input line
    var vect = data_rows[ii].split(',');
    if(vect.length == 4)
   tempLine = data_rows[ii].split(',');

  //see if a new direction index is starting
  if (directionIndex != tempLine[0])
  {
   beginNewDirection = true;
  }

  if(!beginNewDirection)
  {
   var sharedRandomnessIndex = tempLine[1];
   var sharedRandomness = angMom[sharedRandomnessIndex];
   var aliceOutcome = tempLine[2];
   var bobOutcome = tempLine[3];
   sum = sum + aliceOutcome*bobOutcome;
  }


  if (beginNewDirection)
  {
   //finish computation
   var epsilon = sum/newTotAngMom;
   var angle = Math.acos(Dot(a, b));
   boardCorrelations.createElement('point', [angle,epsilon],{size:0.1,withLabel:false});

   var diffLinear = epsilon - (-1+2/Math.PI*angle);
   varianceLinear = varianceLinear + diffLinear*diffLinear;
   var diffCosine = epsilon + Math.cos(angle); 
   varianceCosine = varianceCosine + diffCosine*diffCosine;

   //reset and start a new cycle 
   directionIndex = tempLine[0];
   a[0] = AliceDirections[directionIndex-1][0];
   a[1] = AliceDirections[directionIndex-1][1];
   a[2] = AliceDirections[directionIndex-1][2];
   b[0] = BobDirections[directionIndex-1][0];
   b[1] = BobDirections[directionIndex-1][1];
   b[2] = BobDirections[directionIndex-1][2];
   sum = 0;
   var sharedRandomnessIndex = tempLine[1];
   var sharedRandomness = angMom[sharedRandomnessIndex];
   var aliceOutcome = tempLine[2];
   var bobOutcome = tempLine[3];
   sum = sum + aliceOutcome*bobOutcome;
   beginNewDirection = false;
  }

   }
 //finish computation for last element of the loop above
 var epsilon = sum/newTotAngMom;
 var angle = Math.acos(Dot(a, b));
 boardCorrelations.createElement('point', [angle,epsilon],{size:0.1,withLabel:false});
 var diffLinear = epsilon - (-1+2/Math.PI*angle);
 varianceLinear = varianceLinear + diffLinear*diffLinear;
 var diffCosine = epsilon + Math.cos(angle); 
 varianceCosine = varianceCosine + diffCosine*diffCosine;
 //display total fit
 boardCorrelations.createElement('text',[2.0, -0.7, 'Linear Fitting: ' + varianceLinear],{});
 boardCorrelations.createElement('text',[2.0, -0.8, 'Cosine Fitting: ' + varianceCosine],{});
 boardCorrelations.createElement('text',[2.0, -0.9, 'Cosine/Linear: ' + varianceCosine/varianceLinear],{});
 boardCorrelations.unsuspendUpdate();
};


var clearBoard = function()
{
 JXG.JSXGraph.freeBoard(boardCorrelations); 
 boardCorrelations = JXG.JSXGraph.initBoard('jxgboxCorrelations',{boundingbox:[-0.20, 1.25, 3.4, -1.25],axis:true, 

showCopyright:false});
 boardCorrelations.create('functiongraph', [function(t){ return -Math.cos(t); }, -Math.PI*10, Math.PI*10],{strokeColor: 

"#66ff66", strokeWidth:2,highlightStrokeColor: "#66ff66", highlightStrokeWidth:2});
 boardCorrelations.create('functiongraph', [function(t){ return -1+2/Math.PI*t; }, 0, Math.PI],{strokeColor: "#6666ff", 

strokeWidth:2,highlightStrokeColor: "#6666ff", highlightStrokeWidth:2});
};

var clearInput = function()
{
 document.getElementById('in_data').value = '';
};

var clearTestDir = function()
{
 document.getElementById('in_test').value = '';
};

var clearOutput = function()
{
 document.getElementById('out_measurements').value = '';
};

var generateTestDir = function()
{
   clearBoard();
 var totTestDir = document.getElementById('totTestDir').value;
 var testDir = new Array(totTestDir);
 var strData = "";
 for(var i=0; i<totTestDir; i++)
 {
  //first is Alice, second is Bob
  testDir[i] = RandomDirection();
  strData = strData + testDir[i][0] + ", " + testDir[i][1] + ", " + testDir[i][2]+ ", " ;
  testDir[i] = RandomDirection();
  strData = strData + testDir[i][0] + ", " + testDir[i][1] + ", " + testDir[i][2] + '\n';
 }

 document.getElementById('in_test').value = strData;
};

var generateRandomData = function()
{
   clearBoard();
 var totAngMoms = document.getElementById('totAngMom').value;
 var angMom = new Array(totAngMoms);
 var strData = "";
 for(var i=0; i<totAngMoms; i++)
 {
  angMom[i] = RandomDirection();
  strData = strData + angMom[i][0] + ", " + angMom[i][1] + ", " + angMom[i][2] + '\n';
 }

 document.getElementById('in_data').value = strData;
};

var apendResults= function(newData)
{
 var existingData = document.getElementById('out_measurements').value;
 existingData = existingData + newData;
 document.getElementById('out_measurements').value = existingData;
};


function GenerateAliceOutputFromSharedRandomness(direction, sharedRandomness3DVector) {
    //replace this with your own function returning +1 or -1
    if (Dot(direction, sharedRandomness3DVector) > 0)
        return +1;
    else
        return -1;
};

function GenerateBobOutputFromSharedRandomness(direction, sharedRandomness3DVector) {
    //replace this with your own function returning +1 or -1
    if (Dot(direction, sharedRandomness3DVector) < 0)
        return +1;
    else
        return -1;
};

var boardCorrelations = JXG.JSXGraph.initBoard('jxgboxCorrelations', {axis:true, boundingbox: [-0.25, 1.25, 3.4, -1.25], showCopyright:false});

clearBoard();
generateRandomData();
generateTestDir();
generateData();
plotData();

At loading time the page executes:

clearBoard();
generateRandomData();
generateTestDir();
generateData();
plotData();

The key to the whole exercise are the following two functions:

function GenerateAliceOutputFromSharedRandomness(direction, sharedRandomness3DVector) {
    //replace this with your own function returning +1 or -1
    if (Dot(direction, sharedRandomness3DVector) > 0)
        return +1;
    else
        return -1;
};

function GenerateBobOutputFromSharedRandomness(direction, sharedRandomness3DVector) {
    //replace this with your own function returning +1 or -1
    if (Dot(direction, sharedRandomness3DVector) < 0)
        return +1;
    else
        return -1;
};

To experiment with various hidden variable models all you have to do is replace the two functions above with your own concoction of hidden variable which uses the shared variable "sharedRandomness3DVector". 

There are certain models for which if we return zero (which in the correlation computation is equivalent with discarding the data since the correlations are computed by this line in the code: sum = sum + aliceOutcome*bobOutcome;) a certain number of times as a function of the angle between direction and sharedRandomness3DVector, then one can obtain the quantum mechanics correlation curve. This is the famous detection loophole (or (un)fair sampling) for Bell's theorem.

If we talk about the detection loophole the paper to read is an old one by Philip Pearle: http://journals.aps.org/prd/abstract/10.1103/PhysRevD.2.1418 In there Pearle found an entire class of solutions able to generate the quantum correlations. The original paper is hard to double check (it took me more than a week and I was still not done completely), but Richard Gill did manage to extract a useful workable detection loophole model out of it: https://arxiv.org/pdf/1505.04431.pdf

Manipulating the generating functions above one can easily test various ideas about hidden variable models. For example an isotropic model of opposite spins generates -1/3 a.b correlations. It is not that hard to double check the math in this case: a simple integrals will do the trick. in particular this shows that the spins do not exist independent of measurement. 

More manipulations using the detection loophole are even able to generate super-quantum Popescu-Rohrlich box correlations, but I let the user to experiment with this and discover how to do it for themselves. Happy computer modeling!

Thursday, September 15, 2016

Playing with Bell's theorem

In this post I'll write just a little text because editing is done straight in the HTML view which is very tedious. Below I have a Java script program which illustrates Bell's theorem. If you want to play with this code just right click on the page to view the source and extract it from there. If you do not know how to do that then you are not going to understand it in a few sentences. Next time I'll describe the code and how to experiment with various hidden variable models.
This is about an EPR-B Alice-Bob experiment where each ("measurement") generate a deterministic +1 or -1 outcome for a particular measurement direction using a shared piece of information: a random vector. Then the correlations are computed and plotted. No matter what deterministic model you try the correlation near the origin you generate a straight line vs. a curve of zero slope in the case of quantum mechanics. For this particular program, given a measurement direction specified as a unite vector in Cartesian coordinates I am computing the scalar product and I return +1 if positive I and -1 if negative. The experiment is repeated a number of times on various random measurement directions.
If you do not trust the randomly generated data, you can enter you own random Alice-Bob shared secret and your own measurement directions. Part of the credit for this program goes to Ovidiu Stoica.

Number of experiments:
Number of directions:




Legend: Direction index|Data index|Measurement Alice|Measurement Bob

Saturday, September 10, 2016

A sinister mystification


Once in a while, events in the society at large overshadow all other considerations. I will put on hold the series about Bell's theorem for this week because, such an event occurred: Mother Teresa was proclaimed to be a saint. So what? What is the big deal?

Growing up in Romania, all I heard about her was that she was the symbol of selfless devotion to the poor, a truly remarkable person symbolizing all that it is good in mankind. Coming to US, the public perception was on similar lines and her 1979 Nobel piece prize seemed well deserved. Her recent sainthood was only the realization of a natural public expectation. 

However things are not always what they appear and in this case the truth it is complete opposite with the perception. The sainthood outcome is the result of public gullibility masterly exploited by a morally bankrupt Catholic Church in collusion with dirty politicians, media, corrupt businessman, a dictatorship monarchy, and at the center of it all a pure evil person advancing a religious fanatic agenda for the benefit of the Catholic Church and her own perverted pleasure: Mother Teresa. 

The person who blew the whistle on Mother Teresa's con artist mystification was a remarkable person: Christopher Hitchens with his book: The Missionary Position. I never heard of Mr. Hitchens until a year ago when I discovered by accident his anti theistic stance. Coming from a country which suffered under communism for decades, I was turned off by his hints of admiration for Marxist ideas. It took me some time to properly asses his integrity and the value of his arguments. In the end I found him a very sharp clear thinker with a courageous attitude. I was surprised to discover he was a mini celebrity in the left political circles in US who alienated part of that audience due to his hawkish attitude and support for the Iraq war and who was also a personal friend of late Justice Scalia-the most conservative member of the US Supreme Court.

Now I don't think I will change the minds of the devout Catholics about Mother Theresa, so if you are such a person either take the blue pill and stop reading the rest of this post or take the red pill and keep reading to be shaken from your intellectual complacency and maybe stop buying the bridge the church keeps selling you.

To start I encourage you to watch the following videos:


It's too long to explain the whole mystification story but here is the gist:

Mother Teresa was not a friend of the poor but of poverty and suffering. She derived a perverted gratification from witnessing and encouraging suffering because she thought this would bring her closer to her salvation. This is the mark of a psychopath which derives meaning and pleasure from other's suffering. The places she established were not designed to alleviate suffering but decrepit places of abject poverty and suffering were people were simply brought to die. Young people were denied simple medical care which could have easily saved their lives because their suffering was sanctioned by a fanatical religious agenda. 

So maybe Mother Teresa applied the same principles on herself. Not at all. When she was ill no expenses were spared and she took advantage of the latest medical advances. What a cosmic hypocrite. But where did the money come from to establish her places of suffering? Among other sources from the brutal Duvalier dictatorship family in Haiti responsible for the deaths of tens of thousands of people over decades, and from a corrupt business person convicted of stealing the life savings of thousands of people. But perhaps the Catholic Church upon learning about this returned the blood and dirty money to the victims? After all Mother Teresa acted with the full blessing of the church. Think again.

Hitchens has a simple but true position: religion poisons everything. It takes time to evaluate his claims and try to refute it if you can. It is much more convenient to ignore it, but then I assume that if you reached this paragraph you took the red pill. Many religious figures and "religious scholars" tried debating with  Hitchens only to be shamefully debunked. None won the debates. Hitchslap is now an urban dictionary verb.

On Mother Teresa Hitchens put it like this: "It is a certainty that millions of people died because of her work. And millions more were made poorer, or stupider, more sick, more diseased, more fearful, and more ignorant."

Her sainthood is a scandal due to a sinister and cynical mystification perpetuated by many people over decades for their own benefits. The shame list includes the Catholic Church with all its popes who sanctioned Mother Teresa fanatical religious agenda, politicians like Ronald Reagan, public figures like Princess Dianna, the Nobel Peace prize committee, media like CNN, all who exploited public opinions for their own agendas, as well as corrupt businessman and dictators who provided the money in exchange of whitewashing their public image.  

Friday, September 2, 2016

I was wrong and Lubos Motl was right 

but quantum correlations are not like Bertlmann's socks


It finally happened. I was too careless and I stupidly challenged Lubos to "Show me a non-factorizable state in CM for spatially separated physical systems!!!" 

And he indeed presented one:

"Bertlemann's socks? The state is described by the probability distribution that has 0%,50%,50%,0% for left-and-right-green, left-green right-red, left-red right-green, left-and-right-read. The state, P(colorLEFT, colorRIGHT) isn't factorized. "

This is obscurely written but it is ultimately correct. Glup, glup, glup, he sank my battleship, bruised my ego, won the battle, but not the war :) I will attempt to show that quantum correlations are not like Bertlmann's socks correlations because quantum correlations depend in an essential way on the observer. First let me clarify Lubos example and why he is right in his example.

First what is this business of Bertlmann's socks? This came from a famous Bell paper:


Mr Bertlmann is a real person who uses to wear socks of different colors. The moment you see that one of his socks is pink you know that his other sock is not pink. (Lubos is using red and green) and there is a correlation between the sock colors. In general, you learn from statistic 101 that a joint probability \(P(A, B)\) factorizes: \(P(A,B)=P(A)P(B)\) if the two probabilities are independent. Therefore it was blatantly stupid on my part to ask Lubos for a non-factorizable example. But I had something else in mind for which I carelessly skipped the essential details and now I will explain it properly.

The Bell theorem factorization condition is not on independent probabilities but on residual probabilities:

\(P(A,B, \lambda) = P(A, \lambda) P(B, \lambda)\)

In the case of Bertlmann's socks there are no residual probabilities! But what does this mean?

Correlations can be generated due to physical interaction between two systems or because they share a common cause. One way people explain this is by inventing silly games where two players agree in advance on a strategy but they are not allowed to communicate during the actual game. For example consider this: Alice and Bob each have a coin and they flip it heads or tails. Then they both have to guess the result of each other and they win the game when successful. If they guess randomly the best odds of wining the game is 25% of the time. However there are two strategies they can agree beforehand which increases their odds of winning to 50%. (Can you guess what those strategies might be?). Strategies, common causes and interactions, outcome filtering, all generate correlations and ruin the factorization condition. 

But quantum mechanics is probabilistic and even if you account for all interactions the outcome is still random. After accounting for all those factors in the form of a generic variable \(\lambda\), the remaining probability is called a residual probability. Lambda can be a variable, or a set of variables of an unspecified format. The point is that after accounting for all common causes, if the physical systems A and B are spatially separated, they cannot communicate and in the quantum case it seems reasonable to demand \(P(A,B, \lambda) = P(A, \lambda) P(B, \lambda)\).

But such a factorization is at odds with both quantum mechanics predictions, and experimental results. This is the basis on which people in foundations of quantum mechanics call nature nonlocal.

I think I know how Lubos may attack this. I bet he will say that \(\lambda\) is basically a hidden variable and quantum mechanics does not admit hidden variables. This line of argument is faulty. Carefully read (several times) Bell's paper from the link above and you see that \(\lambda\) represents the coding of the usual way correlations are  accounted by this common \(\lambda\) which is present in all 3 factors.

Let me spell out better Bell's argument following a well written paper by Bernard d'Espagnat. Bell considers the singlet state and in there you have Alice and Bob in two spatially separated labs measuring the spins on directions a, and b and obtaining the outcomes A, and B respectively, Let \(\lambda\) represent a common source of the correlation between A and B. Then one can write the standard rule of statistics: \(P(M,N) = P(M|N)P(N)\) like this:

\(P(A,B|a,b,\lambda) = P(A|a,b,B,\lambda)P(B|a,b,\lambda)\)

then because what happens at Alice's side does not depend on what happens on Bob side and the other way around:

\(P(A|a,b,B,\lambda) = P(A|a, \lambda)\)
\(P(B|a,b,\lambda) = P(B|b, \lambda)\)

yields:

\(P(A,B|a,b,\lambda) = P(A|a, \lambda) P(B|b,\lambda)\)

From this the usual Bell theorem follows and disagreement with experiment is used to point out that what happens at Alice's side does depend on what happens on Bob side. In other words, nonlocality.

I disagree (with good arguments) with several points of view:

  • I disagree with Lubos that nature does not follow the logic of projectors and follows the Boolean logic instead. (measurements project on a Hilbert subspace and quantum OR and quantum NOT are different than their classical counterparts)
  • I disagree with Tim Maudlin who best defends the point that Bell proved nonlocality based on the argument above (Quantum mechanics is contextual and because of that: \(P(A|a,b,B,\lambda) \ne P(A|a, \lambda)\) . As such Bell's factorization condition is not justified. Only if you think in terms of classical Boolean logic avoiding contextuality the nonlocality conclusion is inescapable. )
  • I disagree with Lubos that quantum correlations are like Bertlmann's socks. To eliminate thinking about lambda as a hidden variable, picture it as fixing all conceivable sources of correlations (Bertlmann's socks type or any other type). Now add locality independence, Boolean logic, and you get correlations at odds with experiments. Pick your poison: give up locality or give up Boolean logic. The one to give up is Boolean logic. Unlike Bertlmann's socks correlations, quantum correlations depend in an essential way on the observer.
  • I disagree with giving up on locality and that the Bohmian position represents a valid description of nature (what happens with he quantum potential of a particle after the particle encounters its antiparticle? Both vanish or not vanish result in predictions incompatible with observations)

Bell himself provided 4 possible explanations of quantum mechanics' correlations:
-quantum mechanics is wrong sometimes
-superdeterminism (lack of free will)
-faster than light causal influences
-non-realism

My take on this is that quantum mechanics is the complete and correct description of nature, there is free will, there are no faster than light causal influences, realism is incorrect, and what people call nonlocality is actually a manifestation of contextuality because the observer (but not consciousness) does play an active role in generating the experimental outcome. This active role happens even when parts of the composed system are out of causal reach because quantum mechanics is blind to space-time separations.