Wednesday, February 16, 2011

"Bit from It" vs. "It from Bit"

Julian Barbour presented his essay "Bit from It" at the FQXi essay contest Is Reality Digital or Analog?.

The essay is beautiful and I agree with the conclusion "Bit from It", in a way I will try to make clear. But I disagree with the way the conclusion was reached - it seems to me that the central part of Wheeler's 'ontology' "It from Bit" was overlooked, and this makes it look naive, while it is in fact very profound.

In a classical world, Wheeler's "It from Bit" would be obviously silly. When we measure something, we can write down the outcome as a string of digits, and by collecting all these digits we can determine the state. In such a world, "bit" would indeed originate from "it".

But Wheeler is discussing the quantum world. And for Wheeler, the quantum world is not just "classical world" plus "probability". Julian Barbour said: "Crucially, even if individual quantum outcomes are unpredictable, the probabilities for them are beautifully determined by a theory based on 'its'", but this is not the whole story. If this would be all, then he would indeed be right to say "I see nothing in Wheeler's arguments to suggest that we should reverse the mode of explanation that has so far served science so well". Julian Barbour tries to understand how Wheeler could do so trivial mistakes: "Wheeler's thesis mistakes abstraction for reality", and "A 'bit' has no meaning except in the context of the universe". Yet, there is no such a gross mistake.

Wheeler's "It from Bit" can be understood in the context of the "delayed choice experiment". He realizes that it is not enough to specify the outcome, but also what we measure - for example "which way" or "both ways" in the Mach-Zehnder experiment. But he realizes that our choice of what to measure determines how the state was (yes, in the past). This is the key problem of quantum mechanics, and this is the fundamental obstacle of all realistic interpretations of quantum mechanics: we choose "now" what to measure, and our present choice dictates how the state was, long time before we made our choice. We can think that there is an ontology behind the outcomes of our measurements, as in the classical world. But the "delayed choice experiment" shows that the "elements of reality" depend of the future choice of our measurements. And the outcomes depend of these choices too. So, it is in fact "the choice of what to measure" (Hermitian operator) plus "the outcome" (eigenvalue) that forms the "Bit" from Wheeler's "It from Bit". And the "It" is in fact the eigenstate corresponding to the obtained eigenvalue, given that the observable was that particular Hermitian operator. Wheeler was not that naive to think that eigenvalues determine eigenstates by themselves, without considering the Hermitian operator, so he accounted well for the prescription "A 'bit' has no meaning except in the context of the universe".

The central point of Wheeler's "It from Bit" is that the reality of today depends on the choices we make tomorrow, when we decide what to observe, and of the outcomes of the observations. He compares this with the game of 20 questions, when we try to guess a word by asking 20 yes/no questions, under the prescription that the choice of the word is not done at the beginning. The person who "knows" the word changes it by wish, so long as it remains consistent with the answer she already gave to our question. Wheeler wants to emphasize by this the similarity with the quantum state we try to determine, but which depends on what we choose to observe. This is why he was led to the idea that the state of the universe (it) results from the observations (bit).

I give more credit than Julian Barbour to the "It from Bit" philosophy - I view it as a way to present a central problem of quantum mechanics. I think, nevertheless, that it is exaggerated to conclude from this, as many do, that the world is digital. It may be or it may be not, but we should not force the conclusion. After all, the "It from Bit" philosophy is intended to clarify some points of a theory based on continuum - Quantum Mechanics.

My viewpoint on "It from Bit" is that we should regard the outcomes of measurements as "delayed initial conditions" for the Schrödinger's equation. I presented my view in this article, and this video. A solution of a partial differential equation like Schrödinger's is determined by a set of initial conditions. Classically, the initial conditions can be determined from future observations. In Quantum Mechanics, the future observations determine the state in the two meanings of the word "determine": passive - "find out what it is" (by the selection of an eigenvalue of the observable), and active - "choose what it is" (by the choice of that observable). Another central problem is that two consecutive observations of the same quantum system are incompatible, if the observables do not commute. That is, they impose incompatible initial conditions to the wavefunction. But, the second measurement is not, in fact, a measurement of the same system. The system interacted with the first measurement device, and this measurement device has many degrees of freedom which are not determined yet. So, the second observation measures in fact the composed system - the observed system plus the apparatuses used for the previous observations, and all the past interactions of the observed system. This may offer enough degrees of freedom to maintain the unitary evolution and to avoid a discontinuous collapse of the wavefunction.

My interpretation comes with a realistic wavefunction, which is not yet determined among the possible wavefunctions, but whose "delayed initial conditions" are determined by all future and past observations. I think that we cannot avoid the idea of "delayed initial conditions", no matter what "It" we choose to consider as the underlying ontology.

My view is therefore that "It from Bit" and "Bit from It" are reciprocal: a set of possible "It"s (solutions to the Schrödinger's equation), a set of possible "Bit"s (observations, delayed initial conditions) and the Universe is a pair (It, Bit), so that the "It" and the "Bit"s are compatible.

On the other hand, the "Bit" itself is part of the solution of the Schrödinger's equation, that is, of the "It". This is why I said at the beginning that I agree with "Bit from It". But if we have some "delayed initial conditions" - the "Bit"s - the "It" that satisfies to them is not necessarily unique. So, in fact, what we have is not a pair (It, Bit), but a pair ("It"s that satisfy to the observed "Bit", the observed "Bit"). There is a relation "one-to-many" between the "Bit" and the "It"s. The "Bit" appears to be discrete, but the "It" may very well be continuous. So, although "It from Bit" reflects an important aspect of Quantum Mechanics, it should not be taken too far.

Friday, February 11, 2011

Heisenberg's Relations and Uncertainty

Quantum Mechanics, in particular the Uncertainty Relations, need indeed a good interpretation. Well, I think that it is more than a matter of interpretation. If its internal logic is self-consistent, then there would not be needed an interpretation. The long discussions about interpretations actually reveal the existence of internal inconsistencies in the formalism of Quantum Mechanics. The "no interpretation" alternative, the "operational interpretation", tries to ignore the inconsistencies by avoiding discussing about reality, focusing only on the operations we perform when making experiments of Quantum Mechanics. I think that what really is needed is to resolve the internal conflicts of Quantum Mechanics. Actually, I think that the expression "interpretation of Quantum Mechanics" is used in fact for alternative theories, which propose mechanisms by which QM is implemented. Because what we can observe is described already by QM, such mechanisms are usually hidden, practically impossible to observe. So, in my opinion, they are named "interpretations" and not "theories" because of the exigencies of modern science to name them "theories" only if they are testable. We may call them "hypotheses", because they are not interpretations - they actually propose new mechanisms, but they cannot be tested, so they don't qualify to the modern definition of the word "theory". Of course, it can be argued that the assumption (superstition?) that Nature really gave us access to all its mechanisms, as if She had the purpose to allow us to test every statement we can make about them, should be kept open to debate.

Seeing the Uncertainty Relations as fundamental is indeed problematic for several reasons. First, they are in fact the mix of two principles. The second of these principles is the Born rule, giving the probability to obtain a given state as outcome of an observation of a quantum state. The Born rule, by specifying the probability, provides the probabilistic interpretation of a wavefunction. If the Born rule already contains the probabilities, I think it would be better if we could see the Heisenberg Relations separated of the probabilities.

If we take the solutions of the Schrödinger's equation - that is, the wavefunctions - as fundamental, then the basic Heisenberg relations appear from their very properties. We just take the relations between the size of the interval of the time (position) and the size of the interval of the frequency (wave vector), known from Fourier analysis. These relations are much more general: if we represent the same wavefunction in two different bases in the space of all possible wavefunctions, there is always such a relation between the corresponding intervals. Of course, an observable (Hermitian operator) comes with its own set of eigenfunctions, which are orthogonal, so it is naturally to obtain similar relations if we refer only to the observables and their commutation relations.

Therefore, the Uncertainty Relations come directly from the wave nature of the solutions to Schrödinger's equation, combined with the Born rule. By "Heisenberg Relations", I will refer to the relations as they appear from the wave nature of the wavefunction, reserving the names "Heisenberg Uncertainty Relations" or "Uncertainty Relations" for their probabilistic interpretation.

In a similar way, the entanglement between two or more particles is in fact a property of the tensor products between wavefunctions representing single particles. When the total state cannot be represented as a pure tensor product (which can be a combination of symmetric and antisymmetric products), but only as a superposition, we have entanglement. When we appeal to the Born rule, the entanglement manifests as correlations between the possible outcomes of the observation of the particles.

The Born rule has been thus tested by all experiments in QM, involving entanglement or not. Being probabilistic, they are tested only statistical, but this doesn't mean that they reveal an intrinsic probabilistic reality.

One central problem of Quantum Mechanics is to accommodate the unitary evolution described by the Schrödinger's equation, and the apparent collapse of the wavefunction due to the observation. There is clearly a contradiction here. If we introduce an internal mechanism to explain this collapse, then we have to make this mechanism able to explain both the unitary evolution and the collapse. This is difficult, because both processes are very simple. In a vector space, what can be simpler than unitary transformations and projections? Any hidden mechanism would have to compete with them. This is why it is so difficult to explain QM in terms of hidden variables, of multiverse, of nonlinear collapse and spontaneous diagonalization of the density matrix caused by the environment.

On the other hand, there are already enough unknown factors even if we consider the wavefunction as the only real element. The Schrödinger's equation gives us the evolution, it doesn't give us the initial conditions. The initial conditions can be partially obtained from observation. Due to the particular nature of quantum observation, our choice of what to observe also is a choice of what the initial conditions were (yes, in the past). This is why the initial conditions are delayed until the measurement is taken. To this, let us add that we do not observe the initial conditions of just a particle, but of that particle and every system with which it interacted in the past - such as the preparation device, which ensures the state of that particle at a previous time. Since such a device is large and complex, we don't really know its initial conditions, so when we observe the particle, we also observe the preparation device, and everything with which they interacted. Therefore, there are much more factors to introduce in the Schrödinger's equation. These factors are complex enough to make the conclusion that the wavefunction collapse is discontinuous not so necessary as it initially seemed. It is possible to have a unitary evolution leading from the state before the preparation to that after the measurement, given that we need to account for the interaction with the preparation device, which also have much freedom in its initial conditions. I described these ideas here, and there is also a video. In this view, the wavefunctions are real, therefore the Heisenberg Relations are real too. By applying to them the Born rule, it follows their probabilistic meaning, the Heisenberg Uncertainty Relations. It would be nice to have an explanation for the Born rule as well, because it is very plausible that it just follows somehow from a measure defined over the space of all possible wavefunctions.

"Explanation" between concrete and abstract

I realized that an apparently well-understood word, "explanation", may lead to controversies in discussions about the foundations of physics. The foundations are already controversial enough, but this adds even more to the confusion. It gives you a double featured feeling: on the one hand, of being misunderstood, and on the other hand, that you don't understand where the interlocutor is going on.

What is an "explanation"? Probably the most usual meaning is that explanation is to reduce the unknown to the known, the unfamiliar to the familiar. When this happens, we get the sense of understanding.

Even since childhood, we had so many questions, and the grown ups explained them - reduced the unfamiliar to more familiar notions. In school, the teachers continued to provide us explanations, and we appreciated most the teachers who managed to make the unclear things more intuitive for us. When reading about the foundations of physics, we usually start with popular physics books. The most recommended such books are those providing the feeling of understanding, appealing to our intuition. When we try to read something more advanced, even if it is recommended by our favorite pop-sci books, we find ourselves in a totally different situation. Instead of finding the deeper explanations we are looking for, we find ourselves thrown in the turbulent torrents of the abstract mathematics, drifting without an apparent purpose. And what is most annoying, these textbooks and articles full of equations actually claim to explain things!

Why is this happening? I think that they are guided by another meaning of the term "explanation": "to give an explanation to a phenomenon is to deduce the existence of that phenomenon from hypotheses considered more fundamental. For example, when from the principles of General Relativity was deduced the correct value from the perihelion precession of Mercury, it was considered that GR explained this precession. On the other hand, the deflection of light by the Sun was considered a prediction. After the full experimental confirmation, it became an explanation. I consider that "prediction" is just a temporary status of a scientific explanation, and that the fact that many explanations are first predictions is a historical accident.

There seem to be a similarity between principles/phenomena and axioms/theorems. This similarity suggests the reason why mathematics plays such an important role in the explanation of phenomena. To deduce more from less, complicated from simple, diverse from universal, this means to use logic and mathematics. And there is no limit of the difficulty of the needed mathematics, even if the principles are not that difficult.

This notion of explanation, I understand now, it is not shared by all of us. The reason is simple: because "explanation" usually means to reduce the unfamiliar to familiar. When somebody claimed to explain a phenomenon, we expect him to show how this strange phenomenon can be described in more familiar, concrete terms. Instead, we find that he or she starts describing it in more abstract terms. How come that such more and more abstract terms are shamelessly named "more basic principles", "more elementary principles" and so on? Isn't this a lie?

Maybe the explanation by "reducing to concrete things" has pedagogical reasons, and the explanation by "reducing to universal principles" is in fact foundational research. But does this means that the gap between pedagogical and scientific explanation should grow as it does nowadays? Wouldn't be much, much better to have a mechanistic explanation? After all, Maxwell sought for such an explanation of the electromagnetic waves, even though he had the equations! The ether theorists of the XIXth century tried to reduce electromagnetism to vibrations in a medium. This tradition still continues, and we encounter on a daily basis renowned scientists trying to explain things which other renowned scientists consider to be already explained: electromagnetism, wave-particle duality, gravity, entropy, the Unruh effect, spacetime, time, black holes and so on.

Probably it would be better to have a mechanistic explanation of everything. This would definitely help the public outreach of physics, and will help physics to advance faster. This may have a huge impact on technology, and on our lives. But who can bet that God, when created the world, bothered about our need to reduce the things to what we know? Why would the universe care about our limited understanding, when decided what principles to follow? Who are we, why would we be so important? I think that, although it would be desirable to find concrete, familiar universal principles behind this complex and diverse world, we have no guarantee that this will ever happen. "You shall not make for yourself a carved image, or any likeness of anything that is in heaven above, or that is in the earth beneath, or that is in the water under the earth."

The definition of "explanation" as a reduction to universal principles has its own advantages, given that we do not take these principles as ultimate truths, but just as hypotheses. One of these advantages is that it allows us to equally appreciate theories which seem to contradict each other. We can appreciate its explanatory power in the sense stated above: as its efficient encapsulation of a wide variety of phenomena in fewer, simpler, and more general principles. This doesn't mean that we should consider these principles as being "true". It is not about being "true", just about encapsulating as much phenomena as possible in as few principles as possible, even if these principles are more abstract. If we insist to become fans of one theory or another as the ultimate "truth", we may reduce our capacity to grasp other explanations. This would not be a problem, if we could prove our theories beyond any doubt, but the truth is that we cannot, no matter how convincing they may look to us.

Sunday, April 25, 2010

The Essence of Quantum Theory

The purpose of this short post is to provide a very brief presentation of Quantum Theory.

Short:

In quantum theory, particles are waves of various shapes. You cannot directly observe the waves, only some of their properties. Each property is well defined only for some of the possible shapes. There is no shape for which the properties "position" and "momentum" are simultaneously well defined (Heisenberg's principle). When you observe a property, you find the wave in a shape corresponding to that property (like magic!), without regard of its previous shape. Entanglement: n particles are a single wave in a space with n x 3 dimensions, they don't have individual shapes.

Details:

In classical physics, particles are points moving on well-defined trajectories. This picture turned out to be an approximation: a particle is in fact a wave (although there is no waving medium for this wave). We know it is a wave, because it interferes, it can be diffracted, its allowed states in an atom are those corresponding to an integral number of wavelengths, and it is governed by a wave equation. As a wave, it has no definite trajectory, and insisting in discussing in terms of position and momentum as for point particles leads to problems.

But you can't observe the wave directly, only classical properties, like position or momentum. Each property you observe is well defined only for a particular set of possible shapes of the wave. When you observe its position, the wave appears to be concentrated at a point, but it has an undefined momentum. Conversely, the possible shapes that have well defined momentum have no well defined position – they are spread in all the space. Similar things happen when you want to observe any other classical property.

The first strangest thing about quanta is that when you look at them, they take precisely one of those shapes corresponding to the property you observe, without regard of their previously known shape. If further you try to observe another property, which is not well defined for the previously observed shape, you will find the new kind of shape, allowed by the new property. Knowing its shape before an observation, you can not predict which of the allowed shapes you obtain, but only the probability for each allowed shape.

The second strangest thing is the entanglement. When dealing with more particles, say n, they are not described by individual waves, but by a single wave on a space obtained by multiplying the usual three-dimensional space with itself n times. This means that after two particles interact, they have no individual shape, but a common shape on this space with 6 dimensions. We can still observe one of the particles, and obtain a particular 3-dimensional shape for it, but if we try to observe both particles, the shape of one is dependent on the shape of the other. The strangest part is that their shapes are correlated even if the particles are separated by very large distances.

Thursday, April 22, 2010

Are vector bundles fundamental in Physics?

Vector bundles and gauge theory

The idea in Gauge Theory is that the fields of the known forces can be expressed starting with some principal bundles and their associated vector bundles. To be more precise, let's consider Maxwell's electromagnetic field $F_{ab}$. It can be represented with the help of a principal bundle of group $U(1)$, and a connection on this bundle. The connection corresponds to the electromagnetic potential, and the curvature to the electromagnetic field. It is known that we can modify the potential to $A_a(t,x)\mapsto A_a(t,x) + \partial_a \theta(t,x)$, and obtain the same $F_{ab}$. In terms of bundles, this transformation corresponds to a gauge transformation of $\mathbb C$ by the action of $e^{i\theta(t,x)}$. The connection will appear to depend on the gauge, but the curvature is gauge invariant.


A bundle is just another manifold

Both principal bundles and vector bundles are differential manifolds (that is, topological spaces which looks locally, from topological viewpoint, like a vector space with with a fixed number of dimensions, and on which we can define partial derivatives). A fiber bundle over spacetime looks locally like the cartesian product between the spacetime and a fixed manifold named fiber. For the vector bundles the fiber is a vector space, for the principal bundle it is a Lie group. The $U(1)$ bundle looks locally like a cartesian product between the spacetime and a circle. This space is 5-dimensional, and it was used by Kaluza and Klein in their attempt to unify electromagnetism with gravity by using a 5-dimensional version of general relativity.

After the electromagnetic force was understood as a gauge field, Yang and Mills provided a generalization which allowed us to see as gauge fields also the strong and electroweak forces. It seemed as easy as replacing the $U(1)$ group with a non-abelian group like $U(2)$ for the electroweak force, and $SU(3)$ for the color force. New bundles resulted, and they can be viewed as well as spacetimes with more dimensions, from which some are compactified.

The obvious problem with these extra dimensions is that we cannot "see" them. What explanation is that we cannot test? To avoid this questions, these dimensions are referred as corresponding to "internal spaces", and the Kaluza-Klein interpretation is in general avoided, being preferred that in terms of bundles.


What is more fundamental, the field or the connection?

It was believed that the potential is only a mathematical trick to simplify Maxwell's equations, and that it has no correspondent in reality. There are some reasons to change this view.

One is, as I explained here, in chapter III., the following. Maxwell's equations contain constraints imposed on the field for equal time, that is, between the values of the field at spacelike separated points (Gauss' law). This may seem a little bit acausal, because requires the initial conditions at two spacelike separated points to be related. Of course, the separation between the two points is infinitesimal, but it still exists, and has non-local consequences. In terms of the potential, these constraints are no longer needed. If we consider the connection as fundamental, then the curvature will be a derived field. It will still obey Gauss' law, but this time just as a consequence of being associated to the connection, which is the true fundamental field. And the connection is not constrained.

Taking a charged field, such as the Dirac electron field, under a gauge transformation it is multiplied by $e^{i\theta(t,x)}$. The Dirac-Maxwell equations maintain their form, if we apply the corresponding gauge transformation to the potential. This allows us to perform an experiment to see whether the potential is a real field, or just a mathematical trick. This experiment was imagined by Werner Ehrenberg and Raymond E. Siday, and Aharonov and Bohm, a decade later. It was verified experimentally by S. Olariu and I. Iovitzu Popescu, and confirmed two years later by Osakabe et. al..

Basically, this effect shows that the electromagnetic potential has a fundamental nature. But how can a potential be the fundamental quantity? Which potential, considering that there can be an infinity such choices, related by $A_a(t,x)\mapsto A_a(t,x) + \partial_a \theta(t,x)$? The only way known for this is if it represents a connection on a $U(1)$-bundle. This way, the potential is just the expression of the connection, in a particular frame on the bundle. Gauge transformations are just changes of that frame.

The Aharonov-Bohm effect is interpreted topologically as an effect of the holonomy of a connection on this bundle (which is the electromagnetic potential). These properties are captured by Wilson’s loops.


Are those "internal spaces" real?

It is easy to check the number of dimensions of our space: it is the number of coordinates required to indicate the position of a point, that is, 3. The number of numbers needed to express a rotation, 3(3-1)/2=3, indicates also that we live in a 3-dimensional space. How can we check the extra, "internal" dimensions? We just count the numbers needed to represent them. Since the electromagnetic potential can be changed in a way indicating a rotation of a circle, we conclude that the internal space has one dimension. It is the same as in the case of the 3-dimensional space. The only difference is that we can actually move in this space, and this is why we consider it real. We cannot move in the internal dimensions. But can we, at least, send particles to move in those dimensions?

In fact we can. The Aharonov-Bohm effect shows that we can rotate the wavefunction of an electron. We can compare the rotation of a part of the wavefunction of an electron with that of another part. To do this, we just make them interfere, and see the relative rotation between them. Isn't this remind us of comparing the speed of light in two arms of the Michaelson-Morley interferometer? Only that the Aharonov-Bohm effect succeded, and showed that there is an "internal rotation".

Now, it is time to remember the notion of existence as it is used by mathematicians. Something exists from a mathematical viewpoint if it is logically consistent. The 5-dimensional spacetime (3+1+1) of the electromagnetism exists, in this respect. Did the rotation verified by the Aharonov-Bohm effect confirm its physical existence? In fact, we can take for the internal space, instead of a circle, the complex space $\mathbb C$. The group $U(1)$ acts as well on this space, and we can think that the physical spacetime is in fact 6-dimensional (3+1+2). What is the true number of dimensions? I would say that this number is given by the number of dimensions of the $U(1)$-bundle, that is, 3+1+1. And the internal space happens to be a circle because the $U(1)$ group itself is, topologically, a circle. It has one dimension too. And both the circle bundle and the $\mathbb C$ bundle are associated to this principal bundle, that is, they are obtained from representations of the $U(1)$ group.

OK, so the space dimensions are more real for us, because we can move almost freely in these dimensions. Time is the fourth, at least mathematically, and some people can accept that it is the fourth physically too. They think that this is true, because of the great beauty and symmetry of the Lorentz group. But the internal dimensions, have they more than a mathematical existence? We can ask as well whether the three space dimensions are true or not. What if the real number of space dimensions is two, as the holographic principle suggests?

Do we have a criterion to distinguish between real dimensions and simple mathematical constructions in physics? Can this criterion be the experiment?