Sunday, November 1, 2009
Maxwell's Demon: Computation & Physical Entropy
The second law of thermodynamics states that we cannot decrease the entropy of a closed system on average. This is important because if it would not hold, one could in principle build a "perpetuum mobile" (a machine that generates energy out of thin air). However, this is not a trivial law as we will see in the following. Physicist have repeatedly tried to construct machines (on paper mostly) that would violate the second law. The result of these thought experiments was that information (in the form of abstract sequences of bits) is intricately tied to the notion of physical entropy.
Let's first define entropy of a physical system. The usual definition counts the number of microstates contained in a macrostate. This means that we first have to define the macrostate and this usually involves coarse-graining of the micro-world.
The easiest example is a sequence of T coin-flips. Every sequence has an equal probably to occur, however if we study the space of the number of times heads comes up we can easily see that the entropy of the state N=T/2 is much larger than that of N=T. This is because there is only one unique sequence with N=T but NchooseT sequences of N heads (and this quantity peaks at N=T/2).
For a more realistic physical system let's imagine that we have contained 2N molecules in some volume, N of them are colored white, the remaining N are colored black. Let's divide the space in two imaginary halves and also imagine that the molecules randomly move around (as is realistic). Macrostates are now defined as counting the number of black balls in the left half and the number of white balls in the right half, which we will denote with B(l) and W(r). Again, B(l)=N, W(r)=N has a much lower entropy than B(l=N/2), W(r)=N/2, simply because there are many more ways to accomplish the second macrostate in terms of microstates.
If we use macrostates as the actual microstates themselves, then the entropy will always stay constant because there is exactly 1 microstate state per macrostate. However, if we use coarse-grained macrostates, then the entropy will typically increase, because you will tend to observe macrostates states that are more likely to occur. Thus in the case of the molecules, you will approximately find an equal number of white and black balls in the left and right halves. The second law says that you will not move back from this state to the ordered state of all black balls to the left and all white balls to the right. (There are fluctuations around the average, but with many molecules they are very small and the odds against fluctuating into the completely ordered state or enormous.)
All good so far. But Maxwell invented a "demon" that would build a separating wall between the two halves and control a gate. If a white ball approaches the gate from the left, the demon quickly opens the gate and lets the white ball through. Similarly for a black ball coming from the right (see Figure). The end result is that all black balls end up to the left and all white balls to the right.
So did we break the second law? To start with, the system of the molecules is no longer closed because it now includes the demon. But how will we treat the demon thermodynamically? Our first guess is that the act of moving the gate up and down and measuring the white and black balls as they approach the gate increase the entropy by a significant amount (more than can the demon can decrease the entropy of the environment). Although this is what will usually happen in practice, it turns out that the demon's trick can actually be achieved by a computer with vanishing increase of entropy (of course the construction is subtle, but you will have to take it on faith).
However, the demon will have to obtain information from the system and record it (say on a tape) in order to be able to operate the gate. As it turns out, one can actually decrease the entropy of the molecules in a vat only at the expense of recording the information about the system. Note that we are talking about an abstract sequence of 0's and 1's here. One can show that the act of erasing this information again will necessarily increase the entropy by an amount more then was extracted from the environment in the first place. Again, this is a very subtle argument, because the information needs to be erased irreversibly and that may not be as easy as it looks. Anyway, we will take it for granted now.
This then restores the second law before and after we recorded and erased the information, but what happened in the middle? Did we beat the second law for an arbitrary long time (the time information was recorded on a tape)? In a way yes, and physicist have been forced to change the definition of the second law to include the abstract (Shannon type) information of the sequence of 0's and 1's living on the tape. The new law says that the total entropy of the physical system plus that of the bit-string will not decrease over time. Hence, we have moved some of the entropy of the physical system onto a computer memory and froze it there. The boarders between the real physical world and the abstract world of computer science have been blurred.
This story re-enforces the opinion that physics can be understood as computation ("it from bit" according to Wheeler). But it also points to a potentially deep connection between intelligence (learning systems) and physics. The demon had to study the system and record the information in his "brain" before he could lower the entropy of his environment. But according to a Bayesian, the amount of information obtained by the demon has only meaning relative to his own subjective prior about the world. Is entropy then in the end a subjective quantity? There are many more indications that at least certain aspects about the world are rather subjectively defined (for instance the collapse of the wave function in quantum mechanics leading to the Schrodinger's cat paradox). Is there even an objective world out there? That will have to wait for another blog (;-).
Friday, October 23, 2009
Leaf Blowers (Suck)
Saturday morning 8:00am. The leaf blowers come. Now, we sleep with our windows wide open. So we wake up. Because leaf blowers make a lot of noise. We call that "noise pollution". Especially Saturday morning at 8:00am.
Now let's review the leaf blower. It moves leafs from A to B. In fact it moves leafs away from places where they belong, namely on the ground between your plants. Once, blown on a pile, they can be shipped off. Now, leafs make humus (degraded organic material in soil). Yes, they rot, but the rotting doesn't smell bad. It's natural (really). And as an additional bonus: it makes your soil fertile.
Leaf blowing removes leafs. So now we have to ship-in degraded, other organic material (that actually does smell strongly and is probably made of your leafs in the first place). We call that mulch. In summary: we use a very noisy machine (at 8am) that uses gas to then remove organic material which is then turned into mulch that needs to be shipped back in from far away, once again, using gas. How stupid is that, given that nature has figured out ways to do that all by itself.
However, what I experienced recently went even further. The leaf blowing humanoid was directing his all-powerful blowing machine into a tree (he was a tree blower). The reason was presumably that he wanted the leafs that had turned a little brown to fall off the tree early. As it turns out, trees have evolved to figure this out all by themselves; they do not need help with that (certainly not at 8am).
One more example of how we have alienated ourselves from nature. Half a year ago or so it actually rained in Southern California. I picked up the following phrase when entering a restaurant: "I am going to move back to the desert". Rain is good for plants. In fact, rain is good for people too. So don't complain if it rains; you only have to endure it about twice per year. Try living in Sudan (no rain) or Ireland (always rain).
I worry: when my oldest child got (consciously) confronted with rain for the first time, and I explained that rain made plants grow she said in surprise: "no daddy, sprinklers make plants grow". Yes, that was actually correct. 60% of our water usage is on watering our lawns. That means that we have to drain the Colorado river to the last drop before it hits Mexico, just to water our lawns (and wash our SUVs). Actually, we also drain the west side of the Sierra's Nevada's resulting in "salt storms" in those regions. So let's get rid of those lawns now and replace them with native plants (almost don't need any watering). Additional advantage: lot's of hummingbirds in your garden!
Sunday, October 18, 2009
What Language Does Our Universe "Speak"?
Many profound physicist have come to think of physics at the very tiniest length scale (the Planck length) as a computer or information processor. To name a few: John A. Wheeler, Richard Feynman, Roger Penrose, Gerard 't Hooft. Wheeler expressed this view as "it from bit".
One of the main reasons for this view is the realization that physics at that scale will have to be discrete. If not, it becomes very hard to reconcile relativity and quantum mechanics into one theory. In the continuous domain calculation simply blow up: they cannot be re-normalized. In addition to that, the uncertainty principle of quantum mechanics demands that we can not even pinpoint things down to such precision without creating a black hole which would immediately render any measurement at a scale smaller than its horizon impossible....So these physicist think that physics at that scale is some sort of cellular automaton.
Around the end of every century some people seem the need to make the rather absurd claim that science is coming to an end (I believe we have barely started, but anyway). This century this view is expressed in the book: "The End Of Science: Facing The Limits Of Knowledge In The Twilight Of The Scientific Age" by John Horgan. He argues that there are four recent theories that have shown the fundamental limitations of science:
1. Relativity: anything inside the horizon of a black hole will never get out. So we cannot study the inside of a black hole.
2. Quantum Mechanics: the world is irreducibly random.
3. Chaos: The dynamics of many real physical phenomenon displays extreme sensitivity to initial conditions.
4. Complexity Theory: Godel's theorem of incompleteness of formal systems.
Let's see how these theories would fare in the face of a fundamental theory of "Physics as Computation" (PAC). I think the black hole issue is already close to being resolved. A quantum mechanical treatment of BHs will involve BH-radiation (or Hawking radiation). As such, in-falling matter will cause disturbances on the surface of the BH-horizon that encodes the information of the in-falling matter and which will eventually be radiated out again. No information is lost in the process. (Every BH will eventually die in an explosion that is more violent than the most energetic supernova, but it takes a while..) For the observer that stays outside the BH, the BH horizon is the edge of the universe in a very real sense. It will see his colleague that falls into the BH freeze onto the horizon, get disintegrated and eventually be radiated out again in bits and pieces. For the in-falling observer the edge of the universe is not the BH horizon, but a singularity at the center of the BH. In this case we have to deal with a singularity but it seems evident to me that the final PAC theory will describe that singularity not as an infinitely dense point but rather a sensible finite object.
How the irreducibility of quantum mechanics may be resolved in terms of a cellular automaton was described in my previous blog on "Quantum Mechanics is not the Final Theory".
The phenomenon of chaos in nonlinear dynamical systems makes claims on unpredictability of a more every day nature: for instance the weather patterns are unpredictable because a small error in the initial conditions may result in large differences a few days later (except in California where we don't need weather forecasting). The canonical example is this: x[t+1]=2*x[t] mod 1. This means that at every iteration we move all digits one decimal place to the left and set the number to the left of the dot to 0: 0.12345...
Finally Godel's theorem. It says that within any sufficiently complex formal system there will be true theorems that cannot be proved. I am still thinking about these issues, but I seem to have an issue with the notion of "a true theorem". True can only acquire meaning as an interpretation of the formal system (say mapping sequences to mathematical or physical "truths"). But mathematics is itself a formal system. Truth does not exist outside any axiomatic system and the interpretation that Godel's theorem shows that truth is bigger than formal reasoning just doesn't sit well with me. Anyway, some future blogs will unquestionably be devoted to these deep issues.
It will be very interesting to be able to answer the question: "what is the complexity class of the sequences generated by the cellular automaton that governs our universe". Or phrased more informally: "What language does our universe speak". Here is my prediction: Dutch ;-) (or maybe a language of the same complexity). It seems that Dutch is more complex than context-free languages due to cross-referencing but still decidable in polynomial time. It represents a possible level of complexity where things are not too regular but also not too unwieldy. Anyway, my prediction here should be taken with a huge grain of salt of course.
Soooo, the universe is a huge computer that is computing "something". It is our task as scientists to figure what and how it is computing. Actually, we already know the answer: 42 ;-). But what was the original question? Let's leave that to religion.
Wednesday, October 7, 2009
Complexity
In the previous blog I argued that a truly random sequence can (should) be defined as an incompressible one. A sequence from which nothing can be learned. Moreover, it can be shown (mathematically) that certain deterministic chaotic dynamical systems can indeed generate these random sequences.
If randomness is a slippery subject, complexity is worse. Complexity has turned into an entire research field of its own but is being criticized for not even having a good definition for the term itself. In fact, there seem to be 31 definitions floating around. Complexity is one of those things for which we seem to have a clear intuition but is hard to capture in math.
Let's try to build some intuition first. Consider the tree-image above and compare that to say a completely white image and an image of white noise (your television screen if it receives no signal). We feel that the white image is boring because it is utterly predictable but the noise image is also boring because there is nothing to predict about it. The tree-image on the contrary seems to interesting. There is lots of structure. We want our complexity measure to peak at the tree-image but vanish at boring images and random images. Therefore, complexity should not be equated to randomness.
An attractive definition was given by Murray Gell-Mann (Nobel laureate and inventor of the quark). I will discuss a slightly adapted version of his story. We return to the MDL principle (minimum description length). Given a string of symbols we wish to optimally compress it. That means, we want to build a model with a bunch of parameters, P, and use that model to predict most of the symbols in the sequence correctly. The symbols that were wrongly predicted will have to be stored separately. One can imagine a communication game where I want to send the string to you. Instead of sending the string, I send the model and only those symbols which where predicted wrongly by the model. If you use my model on your end to generate the predictable symbols you have all you need to generate the original string (use the model at the locations where predictions are correct but use the separate list of corrections where the model was wrong). The trick is now to find the model that has the optimal complexity in the sense that the total information necessary to encode the model and the unpredictable symbols is minimized. In other words, I want the procedure that maximally compresses the original string.
There is something fundamentally important about this procedure because it guarantees that the model will be optimally predictive of future data. If your model was too small, there was more structure that you could have predicted in the test string. If on the other hand you use a more complex model, you will have "modeled" the randomness of the input string (called overfitting) and this is obviously not helpful either. Living things survive because they are able to predict the world (and leverage it). It is our business to compress the input stream of our senses!
We are now ready to define complexity: it is the information necessary to encode the model (parameters), excluding the residual random bits of the string that remained unpredictable. A white image has very little information to start with and even though it can all be predicted by our model (no random bits) so the total is still very small. On the other, a random string is completely unpredictable and even though it carries a lot of information (a la Shannon) the model part is zero. Both cases have vanishing complexity as we wanted to, and the tree-image will have lots of complexity.
Complexity is a relative quantity. For instance, it depends on factors such as the length of the string. For small strings it doesn't pay off to use a complicated model. This is not so bad in my opinion, because imagine a primordial world where all that has ever been created is the string "GOD". The string are just three symbols embedded in nothingness. It will only acquire its structure (meaning) after we read the bible.
Another more serious objection is perhaps the fact that the true complexity is in-computable because it can be shown that one can not decide whether the model you have is really the optimal one in general. Imagine the fractal image below.
The image looks wonderfully complex. Who would have thought that there is a really simple algorithm to produce it? So, practically speaking, complexity is defined relative to how good of a modeler you are.
Then there is the issue of scale. If we view the world at a large scale, it may look very predictable. However, the world is composed of quadrizillions of elementary particles that are moving in chaotic motion. At the small scale the world is random, but at a larger scale order emerges through averaging. This may be best illustrated by going back to the coin-toss. If we are asked to predict the next outcome of a coin-flip, we will be at a loss. However, if we are asked to predict the total number of heads over the next 100 coin tosses we can make a good guess (50 for a fair coin). This is because even though every sequence of 0's and 1's is equally likely, there are many more sequences with 50 heads and 50 tails than that there are sequences with only heads for instance.
Finally there is the issue of the time that it needs to compress and decompress the string. We may have found a great model that does a wonderful job at compressing the input string, but that requires a very long computation time to use for encoding and decoding. This seems important at least biologically. If I have a model of lions that takes 5 hours to use (i.e. make predictions with) then it is not very useful in terms of my chances of survival. As another example, consider the Mandelbrot fractal again. There is very small prescription to generate it but one needs considerable time to actually compute that image (in fact, this connects to the previous point because you would have to first specify the scale at which you want to stop otherwise it would take infinitely long).
In machine learning researchers have also started to realize the importance of computation in addition to prediction quality. They acknowledge that modern day problems are so large that a really good model may require too long to train up and use for predictions. So, the computational budget is factored into the total equation, favoring faster algorithms and models over slower ones even though the latter may be more accurate (it is in this sense that machine learning is really different from statistics and in fact the reason why machine learning is in the computer science department and not the mathematics department).
Should we adapt our definition of complexity to accommodate running time? Gell-Mann thinks so and he argues for a bound on the computation time. There are indeed many definitions of complexity, such as "logical depth" that do just that. But if we include computational complexity, shouldn't we also include the amount of memory we need during computation (since they can be traded off)? Uh-oh we are descending in the bottomless pit call complexity research with its 31 definitions. This may be a good point to stop.:-)
Sunday, October 4, 2009
Is A Coin Toss Random?
Randomness is a slippery concept. If you teach it in an AI class you are quick to point out that what we call random is really a matter of ignorance: because we don't know and/or model all the details of the world we lump together everything that is outside the scope of our models and call it a random effect. Under this reasoning the toss of a coin is *not* really random in a philosophical sense.
Unless of course, quantum mechanics has anything to do with it. Imagine we toss a coin through a quantum measurement. Now we are confident to claim true randomness. The quantum world is with a nice world "irreducibly random", meaning that there is no deeper theory from which quantum mechanics would emerge and where the coin toss is once again the result of our ignorance. However, as I reported in an earlier blog, there may even be ways out of the seemingly "irreducibly random" nature of quantum mechanics. In this view we are back to square one: what then is randomness?
There is a beautiful definition of randomness in a mathematical discipline called algebraic complexity. To understand this definition let's consider a very long sequence of coin flips. Now let's try to compress this sequence of "HTTTHHHTTHTTTTHHHTHTHTHHH" into a smaller sequence by building a model. If the sequence where just "HTHTHTHTH..." things would be easy: "I will simply state that head and tails alternate and the sequence starts with H (I will have to count the symbols in this sentence as the cost of encoding it). The sequence is not random because I could compress it. The definition of randomness is thus: a sequence is random if it cannot be described (encoded) using a smaller sequence. In terms of learning theory: one cannot learn anything from a random sequences.
Using this definition, a sequence can be random yet and at the same time be generated by a fully deterministic process. This process will have to be "chaotic" which we will not define here but has nothing to do with quantum mechanics. In fact, it snugly fits the definition of a random coin toss. For instance, if you toss a coin in a deep tank of moving water and record whether it landed head or tails on the bottom, then I am pretty sure that the details of the trajectory of the coin are chaotic. Hence, there is no model that can predict anything about the final outcome. This is not because we don't know how to model moving water in a tank, but because trajectories of coins in these tanks have something inherently unpredictable about them.
I therefore conclude that the flip of a coin toss can be truly random and deterministic at the same time. We don't need quantum mechanics for true randomness, we have something far better: chaotic dynamics. In fact, I predict that the so called "irreducible randomness" of quantum mechanics will in the end be reduced to randomness derived from chaotic dynamics.
Sunday, September 20, 2009
On Intelligent Design
Consider the following reasoning I witnessed a while back at a parent meeting. Some trees are so tall that there is no scientific explanation for the fact that these trees can pump water from its roots all the way up to its top (I actually seriously doubts that there is no explanation but let's ignore that for now.) Therefore there exists a upward force that pushes things in the opposite direction as gravity...
What is the problem here. I claim there are two fallacies in the logic. Firstly, the fact that we currently don't have an explanation for some phenomenon in terms of ordinary physics, doesn't mean there is none. A scientist should readily admit that many measurable phenomenon remain unexplained today. It doesn't mean that we have to introduce a new mysterious force. Secondly, the "upward force" feels like a valid explanation, however we have just given the unexplained phenomenon a new name. Instead of saying "I don't know the answer to this question" we say "I know the answer, it is X". It turns out this is just another way of saying the same thing. Giving names to things generates the illusion that we understand it. And people really don't like to not understand things. Our survival depends on it, so it to give us peace of mind to simply delude ourselves.
Isn't this is even true in real science? Apparently, when Feynman asked his father about the reason why objects tend to persist in their straight motion, his father said something of the sort. That's because momentum is conserved. But he warned the young Feynman that giving something a distinguished name isn't the same as explaining it.
When does something become an explanation then? Isn't all we are doing just giving names to things? Well no, the distinguishing factor is prediction. When you can genuinely predict a new phenomenon, then you have found useful structure in the world. Whatever your naming conventions, that predictive power is useful and real.
Now to intelligent design (ID). In ID, the most scientific concept seems to be "irreducible complexity (IC)". This idea should be given a chance. If someone can compute the IC of some complex system as the gap in complexity between the system under study the next simplest version of the system that still performs some useful function, than that seems like a genuine advance. Note though that this is extremely difficult because one would have to search over all possible complex systems that can lead to the system under study and one would have to have a notion of what it means to be useful. I believe these problems basically make the concept of irreducible complexity a non-starter, but we want to be extremely open minded here.
Where things go desperately wrong is when we think we found a system that has a high irreducible complexity and then exclaim that therefore it must be designed by an intelligent designer (aka God). This is like the tree example: we have shown that contemporary science cannot explain the phenomenon and therefore we give the problem another name, namely God. However, stating that God designed certain things in the world does not make us understand and thus predict the phenomenon any better. Our attitude should be: 1) maybe Darwin's version of evolution needs to be improved to properly explain structures with high irreducible complexity, or 2) perhaps our estimate of the irreducible complexity was wrong and we need to look harder for simpler functional structures that could evolve into the "problematic" structure.
Personally, I am very tolerant to religion. It brings some good and some bad things to people, but everyone should make up their own mind on these issues. It would be arrogant to claim there is no God. My hope is that we can teach religion in schools on an objective level. I favor teaching all religions to our children, not just one. However, I do not believe science and religion should be mixed. They live on different planes. Trying to prove the existence of God is a lost battle. Trying to prove that someone performed three miracles and then declaring that person saint is simply silly (at the level of Santa Claus). Declaring the existence of God based on the fact that you (supposedly) cannot explain some natural phenomenon using modern science is equally backward.
Saturday, September 19, 2009
Jared Diamond's Psychohistory of the Human Race
In his Foundation Trilogy, Isaac Asimov describes a new kind of science: "psychohistory". The idea of this discipline is to predict the course of human history through mathematical analysis. The fundamental assumption is that the impact of individuals is "washed out" due to the law of large numbers: if you have enough elements it is only their average behavior that counts. The same idea underlies thermodynamics and statistical mechanics: if we have a *very* large number of individual atoms behaving chaotically, we will have no chance in predicting properties of individual atoms. However, new emergent properties such as temperature and pressure are predictable.
Humans are no atoms of course, and although in Asimov's universe there are many more humans to average over than the 6 billion that live today, the appearance of "The Mule" does cause a breakdown of the predictions. Interestingly, one can relate this idea to recent insights in physics and mathematics that societies are in a state of "self criticality". At its core this means that small disturbances can have large consequences that propagate through the entire system. To give an example from human history: the invention of the atomic bomb by a few scientists changed history radically.
Despite these potential objections, Jared Diamond's book "Guns, Germs and Steel" is the best psychohistory of the human race I have read so far. In fact, personally I find this book the best popular scientific read on my list, right next to the "Selfish Gene" by Richard Dawkins.
What are Diamonds claims? He claims that despite the sometimes large impact of individuals there are also very important and predictive regularities of human history that depend on geography. Why did the European colonist basically wipe out the native Indian population in America and not the reverse? The core reason, so he argues, lies in the fact that massive food production was first invented in the Fertile Crescent (the Iraque, Iran region). The climate was ideal for many species of plant and animal to become domesticated. Moreover, the East-West orientation of Eurasia made it easy for inventions to spread to Europe (and as far as China, although China seemed to have invented food production around the same time). Food production made it possible to switch from a hunter-gatherer life style to a farming lifestyle which in turn made it possible for many people to specialize in other things than farming. This way, larger cities and states started to emerge with a specialized fighting caste. These "successful" states then spread either by conquest or by simply producing more offspring.
But surprisingly, even more important than having powerful new weapons was the fact that cattle generates diseases and thus living among cattle causes a population to become resistant to lethal diseases such as small pox, measles and so on. (of course the price paid for this resistance was a high death toll because for a population to become resistant an aweful lot of cruel "selection of the fittest" will have to take place first.) In the America's, food production was much tougher due to climate issues and due to the North-South axis which prevented effective spreading of inventions. Also, very few large animals were available for domestication (perhaps they were all killed when the first people started settling a long time ago). So when the Spanish arrived they did not only have superior weapons and administration skills, they carried many more lethal diseases with them as well.
Diamond goes on to tell the amazing story of many continents: how the Polynesians were former Chinese evicted from the main land, how the aboriginals from Australia were decimated by Europeans, how the Buntu people spread over much of Africa by using superior farming practices. If you need to remember one thing it is this: it is massive food production that has caused all the major migrations and colonizations of populations. And perhaps equally important, the fact that one civilization ended up dominating another has nothing to do with race, it has to do with geography.
So it seems some form of psychohistory is still possible, even though we are averaging over tiny numbers compared to the numbers Asimov had in mind. I warmly recommend this book to anyone who is even vaguely interested in how the world has become what it is today.
Friday, September 4, 2009
Quantum Mechanics is not the final answer.
The first thing I was told when starting my course on quantum mechanics (QM) was: "if you think you understand QM, you don't understand QM". It turned out to be true. I could master the rules of the game, but things always struck me as fundamentally weird. Information is carried by wave-functions that would collapse into definite states when you decide to look at it?? "Oh well, that is because we were not evolved to think about the very tiny" I kept telling myself.
For decades very few serious physicist would dare to propose something more interpretable and fundamental than QM. The main reason is the Bell inequalities that say that QM can not be explained by a deterministic hidden variable theory, i.e. a more fundamental, deterministic, causal and local theory at a smaller scale that would give rise to QM at a larger scale. It turned out that in order to do that one would need a non-local theory or a non-causal theory.
Now all of that seems to change. It takes a genius with a Nobel prize at the end of his career to stick out his neck. Gerard 't Hooft from the university Utrecht has started a lonely uphill battle to come up with, yes, a deterministic hidden variable theory of QM. So how does he propose to deal with the Bell paradox? The gist of the argument is that QM is an emergent theory, in a similar way as thermodynamics is an emergent theory of many particles that move chaotically. The concept of pressure, temperature, entropy etc. only make sense if we are talking about the average behavior of very many particles that move about in a way that is unpredictable at the level of an individual particle. However, treated as a group new structure emerges. That is what we call thermodynamics.
In a similar sense, QM can emerge from a more fundamental, deterministic and interpretable theory at a smaller scale. A first draft of such a theory is based on the idea of cellular automata. Basically, imagine one can split the world up in small boxes and every box can be in a certain state. Time evolution of a particular box is based on rules that generate a new state (deterministically!) as a function of the states in its direct neighborhood. However, there is a twist. New states are created on two-dimensional surfaces such as the horizon of a black hole. The theory of nonlinear dynamics states that "information" can be created using what is known as "chaos". This basically means that any information about a system, e.g. where the constituent particles are located, is lost very quickly under the evolution of the system. In 't Hooft's universe information is lost in the three dimensional interior. This happens because two different states can evolve into a single state. This idea, that the information of a universe is encoded at two-dimensional surfaces is known as the "holographic principle", another brain-child of 't Hooft.
So how then do we beat the Bell inequalities? Here is the intuition. The large scale theory known as QM defines variables that are functions of the more fundamental states. However, the way that this works out according to 't Hooft is that a QM state at time "t" is an aggregation of all the states that will eventually evolve into a single state. This definition however is non-causal because it involves knowing which states in the future will collapse into a single state. Hence while the fundamental theory is causal and local, the emerged theory at a larger scale can exhibit strange features because the variables defined by us mix up the present and the future.
I may have missed a point or two in trying to translate these ideas, but I certainly think this is a very exciting new development. Einstein may turn out to be right after all in saying that "God does not play dice". These bold ideas definitely give me a sense of relief that it was OK to feel dissatisfaction with the strangeness of QM. Between string theory and deterministic QM, I will bet on the latter. Has 't Hooft ever been wrong?
For decades very few serious physicist would dare to propose something more interpretable and fundamental than QM. The main reason is the Bell inequalities that say that QM can not be explained by a deterministic hidden variable theory, i.e. a more fundamental, deterministic, causal and local theory at a smaller scale that would give rise to QM at a larger scale. It turned out that in order to do that one would need a non-local theory or a non-causal theory.
Now all of that seems to change. It takes a genius with a Nobel prize at the end of his career to stick out his neck. Gerard 't Hooft from the university Utrecht has started a lonely uphill battle to come up with, yes, a deterministic hidden variable theory of QM. So how does he propose to deal with the Bell paradox? The gist of the argument is that QM is an emergent theory, in a similar way as thermodynamics is an emergent theory of many particles that move chaotically. The concept of pressure, temperature, entropy etc. only make sense if we are talking about the average behavior of very many particles that move about in a way that is unpredictable at the level of an individual particle. However, treated as a group new structure emerges. That is what we call thermodynamics.
In a similar sense, QM can emerge from a more fundamental, deterministic and interpretable theory at a smaller scale. A first draft of such a theory is based on the idea of cellular automata. Basically, imagine one can split the world up in small boxes and every box can be in a certain state. Time evolution of a particular box is based on rules that generate a new state (deterministically!) as a function of the states in its direct neighborhood. However, there is a twist. New states are created on two-dimensional surfaces such as the horizon of a black hole. The theory of nonlinear dynamics states that "information" can be created using what is known as "chaos". This basically means that any information about a system, e.g. where the constituent particles are located, is lost very quickly under the evolution of the system. In 't Hooft's universe information is lost in the three dimensional interior. This happens because two different states can evolve into a single state. This idea, that the information of a universe is encoded at two-dimensional surfaces is known as the "holographic principle", another brain-child of 't Hooft.
So how then do we beat the Bell inequalities? Here is the intuition. The large scale theory known as QM defines variables that are functions of the more fundamental states. However, the way that this works out according to 't Hooft is that a QM state at time "t" is an aggregation of all the states that will eventually evolve into a single state. This definition however is non-causal because it involves knowing which states in the future will collapse into a single state. Hence while the fundamental theory is causal and local, the emerged theory at a larger scale can exhibit strange features because the variables defined by us mix up the present and the future.
I may have missed a point or two in trying to translate these ideas, but I certainly think this is a very exciting new development. Einstein may turn out to be right after all in saying that "God does not play dice". These bold ideas definitely give me a sense of relief that it was OK to feel dissatisfaction with the strangeness of QM. Between string theory and deterministic QM, I will bet on the latter. Has 't Hooft ever been wrong?
Sunday, April 5, 2009
His Holyness the Pope
I cannot say that I am a very religious man, although under the influence of my wife I do start to appreciate the teachings of Buddha. I am much less inspired by the current pope though. In fact the Vatican's history is one littered with crime and murder: witch burnings and torture by its inquisition. On a lesser but still significant scale: I found it "curious" to see President Mugabe being invited at the funeral of Pope John Paul II and to see the Vatican in a state of denial when it concerns its cowardice actions in the second world war. How does anyone embedded in such a history claim any moral leadership?
Let's have a look at today's pope Benedictus XVI. Recently he declared the use of a birth-control pill immoral apparently because it pollutes the environment. More damaging, he reiterated the churches condemnation of the use of condoms and he flatly denies that it helps in the fight against AIDS. This was contradicted by the Lancet recently. Then there is his unfortunate decision to reinstall bishop Richard Williamsen who thinks the holocaust is being highly exaggerated. This is now reverted after much critique. He also seems to lobby for sainthood of pope Pius XII (who had a very doubtful role in dealing with the nazi's in the second world war). The current pontiff also believes that homosexuals should be "cured" and protected against self-destruction.
What was the church's attitude when cases of child abuse in the US started to surface? A comprehensive study finds that child abuse of some form happens in as much as 4% of priests. I would have expected that the only morally justifiable action is to immediately remove these priests from the church and let the law bring them to justice. No,in many cases the priests were allowed to move to another parish and keep practicing there. In fact, pope Benedictus seemed to be involved in a secret document that instruct how to cover up child abuse, putting the churches' interest always ahead of child safety. I cite: "The document recommended that rather than reporting sexual abuse to the relevant legal authorities, bishops should encourage the victim, witnesses and perpetrator not to talk about it. And, to keep victims quiet, it threatened that if they repeat the allegations they would be excommunicated."
For more reading see this Wikipedia article.
The Vatican has enormous power over people, and its not helping to solve the challenges of this world.
I like to end with a positive note. At least the Roman church inspired (well, perhaps just pay) artists such as Michelangelo to make beautiful art...
Monday, March 16, 2009
The Infinite Universe that Started as a Point
The beginning of our universe is shrouded behind a huge paradox. Recent observations seem to indicate our universe is open. Which basically means that it has infinite volume. Yet, and here comes the paradox, it started in a single point: the Big Bang.
Sure enough there are alternative solutions to the equations. It could have a closed universe. It's much easier to convince oneself that a closed universe can start at a single point. Imagine you have lost 1 dimension, so you live on a sphere. (there is no such thing as space outside of this sphere: you have become 2-D yourself.) At the moment of the big bang the sphere come into existance and expands fast. whereever you are on the sphere, if you look around you you see other galaxies move away from you. But, since the volume of this space is finite, it's not hard to imagine that at some point in the past it was zero, which was the time of the Big Bang.
Alas, observations say we are not in a closed universe, but in an open one. Logic seems to demand that if space is infinite now, it must have been infinite when the universe got created (unless you believe there was a time after the Big Bang that it suddenly became infinite). It was filled with matter everywhere, and moreover, this matter was expanding outward. In fact, it doesn't matter who you would ask, everyone would tell the same story: the universe is exanding outwards and galaxies that are far away are expanding faster than the galaxies that are close by.
I find that paradoxical. Yet there is a way out. It is described very well here. The basic idea is this. You can switch to another frame of reference. In this view the universe has finite volume and you can visualize it as an expanding sphere again. However, galaxies are packed into this sphere in a peculiar way. Due to the so called Lorentz contraction, distances become contracted for galaxies flying away at very speeds. Since the galaxies at the outer edge of the expanding sphere fly at lightspeed, radial distances are infinitely contracted there. And so we can pack infinitely many galaxies in an infinitely thin slice of space.. In fact, because there is infinitely mass at that rim, it forms a sort of wall (a singularity) beyond which there is simply nothing. So it is not useful to think of this sphere as expanding inside something else. Spacetime is not defined outside this wall. Btw note that you can not visit this wall to peek over it, because it is receding at the speed of light. Also, there is noone really at this edge because the picture looks the same for everyone!
Ok, we have now packed infinitely many galaxies inside a finite volume. Why is space then infinite? For that we need to realize that not only does space contract at high speeds, it also slows down time. So watches on those distant galaxies are moving at a slower rate relative to your watch. Again, this situation is perfectly symmetric: people on those distant galaxies don't notice their watches going any slower (that's because their brains go slower too, so we reason) and moreover, they observe our watches going slower too. Now, we introduce a new cosmic time. This is the time where every observer in the galaxy sets their watch at 0 at the Big Bang and takes their watch with them when they fly outwards. If we now measure the volume of space defined when all these watches display some constant time space becomes infinite. This is because time is standing still at the rim (according to us) and so the infinite space contraction and the infinite time dilatation cancel.
It's head-spinning, but the conclusion is that space was created infinitely large in one point.
Sunday, March 8, 2009
Supersize Me
Just saw "Supersize Me". Didn't know it was that bad. I mean, I didn't know it was that unhealthy. What I did know and what shocked me was that you can find (all too easily) junk food in hospitals. Sometimes it is the only option. In hospitals! Then there is of course the schools. The perfect place to get our kids into unhealthy eating habits that will last them a lifetime. It makes me really angry that food companies have become so powerful that they can prevent changing the status quo through lobbying in Washington. (This whole lobbying thing is deeply disturbing and corrupt.)
Fast-food has become so much part of American culture that many people have forgotten how to cook! They eat out or shove pre-made food into microwaves. Every day. Cooking dinner and enjoying dinner with your family (not in front of the tv watching commercials for McDonalds) is becoming something people did in the old days.
So, congratulation and thanks Morgan Spurlock for exposing this worrying development at the cost of your own health. By the time my children are old enough, I will show them your movie. Hopefully it will prevent them from dumping their veggie sandwich in the trashbin and bying the big Mac equivalent in their school's cafetaria.
Saturday, March 7, 2009
How Long until "1984"
There are currently 4.2 millions surveillance cameras operational in the UK alone (thanks to Pietro Perona for pointing this out to me). That is one camera for every 14 people.... The data on your cellphone, such as your social network, your current location and of course your phone calls, can easily be accessed by phone companies. Your credit-card history, your bank transactions, your email, they all present easily accessible information to intelligence agencies. Add to this the immense growth in computer power and the improvement of artificial intelligence systems and it may not be hard to imagine that very soon it may be "1984".
Already, companies such as Google, Yahoo and Microsoft are designing web-applications which are based on tracking large numbers of cellphones. When many tracks intersect in both space and time an "event" happens (for example a baseball game). People present at these events upload text and photos into websites such as "Twitter", creating some sort of self-organizing online news experience. One can imagine a new Googlemaps application where events are shown on a map. You may then click on the event and see the rich multi-media content for that event as it develops.
We are increasingly covering every patch of (interesting) space-time with our multi-media. We may ask the question if this trend is desirable or perhaps even dangerous. Imagine a government having that much information at her fingertips. It may be awefully tempting to use it to blackmail political oponents etc. But, may be we will be forced into this direction, whether we want it or not.
I remember the 9/11 attacks were a shock to me because for the first time in my life I realized there are actually people out there who will go to any lengths to further their religious or/and political goals, even if it means killing millions of people. It is a matter of time before terrorist will get their hands on very small atomic bombs. I do not believe we will be able to prevent these devices from going off in one of our big cities. Perhaps even more dangerous is biological warfare. It's a matter of time before these technologies become so advanced and widespread that they may fall into the wrong hands.
In my opinion, we will see one of these extreme catastrophies in the somewhat near future (plus or minus 100 years that is). At that point, people will be willing to trade all their privacy for security. People will be tagged at birth with a chip and followed 24 hours a day, 7 days a week. A patchwork of cameras and smart software will detect individuals which are not tagged. They will be instantly arrested. The population of this future country will act as one organism that breaths through the internet. It will monitor its constituents and its environment. But who will monitor its government?
This scenario sounds like a nightmare now. But it may be a price people are willing to take in the future. If no explicit measures are taken to keep this information out of the hands of governments or other organisations, the sketched sceneario may simply be a byproduct of the technology that is rapidly becoming available. Of course these measures will presumably be taken in the short run, but my prediction is that in the long run it may prove to be the only weapon against an increasingly sophisticated terrorist kind of warfare.
Even though today this type monitoring and control seems to be taken directly from Orwell's science fiction story "1984", the world in 100 years may be so utterly different that people living then may find it "desirable". We may become "Borg" after all. Resistance is futile.
Sunday, March 1, 2009
National Healthcare
After seeing Sicko I was kind of shocked I must admit. I am sure it draws a highly exaggerated picture, but it started to make me think. Here in Orange County health-care seems to be excellent. Perhaps too good to be true. My wife has been offered an MRI scan for her knee after having been told the diagnosis with a confidence bordering certainty. After inquiring why an MRI was necessary in this case the answer was, well, "it's not necessary". So she refused and with no consequences to the remainder of the (successful) treatment.
Myself, I was also recently offered a (insert complicated name here) scan. After asking why I needed it, the response from this frank MD was: "because it is required if symptom X is present". Seems reasonable one could think. However, in my case the symptom was explained away by a very innocent issue. After confronting my (really very good) doctor with this fact, he said: "if we don't recommend it we can get sued". He offered in addition that "in our society we have chosen to over-treat one sub-population (read: the rich) at the expense of not treating another sub-population (read: the poor). I refused this treatment as well (not necessarily out of noble considerations: the scan is definitely uncomfortable and time consuming).
There is also another issue at stake. Hospitals are required to treat anyone who comes into the ER and many of those are not insured. This costs money, lots of money (ever seen an ER bill?) To make up for it, doctors send patients with insurance who don't need scans to have expensive scans anyway. This is their way of making money.
All of this on top of the message of Sicko, namely that it is a really bad idea to place health-care in the hands of companies who think about making profit rather than about making healthy.
So I say to president Obama this. There is a huge opportunity to save an enormous amount of money on health-care by making sure doctors don't over-treat patients (those with insurances that is). I am happy paying the enormous amounts of money I am paying to healthcare right now, but please use this money not to send me from one unnecessary scan to another. Use it to give every citizen in the US the RIGHT of health-care. Let's not be afraid to be socialistic about this (btw socialist is NOT communist). We are after all already quite socialistic about the right of protection through sponsoring the federal government to maintain an army.
In the long run, the increasing costs of new medical technology and drugs will force us to make very difficult decisions on who to treat and who not. A good start would be to assess whether the scan or treatment can detect and prevent a disease that has a probability of killing or disabling that is significantly higher than other risk factors for an average patient. We can be conservative about this and still save big time. In my case, the scan was meant to detect a form of cancer that I was so unlikely to have that my chances of dying in a car accident are many times higher.
Saturday, February 28, 2009
Fractals as Symmetries under Scaling
Symmetries are both beautiful and powerful tools to explain the physics of our world. One of the most salient symmetries are invariances of physical laws. For instance, laws remain constant 1) over time, implying (through Noether's theorem) that energy is conserved 2) throughout space, implying conservation of momentum, and 3) under rotations, implying conservation of angular momentum. The Dutch physicist Lorentz also introduced invariance under "Lorentz transformations" which involve both space and time and which led Einstein to develop his theory of special relativity. This in turn led to the equivalence of energy and mass and the constancy of the speed of light.
Going one step further, Einstein also postulated another symmetry, namely that "gravity = acceleration" leading to the general theory of relativity (see earlier blog). In solid-state physics discrete rotational symmetries play a crucial role in describing chrystals such as snowflakes (see Figure).
In quantum mechanics an entirely different kind of symmetry states that particles of the same kind are fundamentally indistinguishable. This leads to all sort of quantum mechanical effects such as Bose-Einstein condensation, entanglement and much more. And the list goes on to even more abstract symmetries that transform the elementary particles into each other. This means that instead of viewing them as different they are rather different sides of a die. These so called gauge symmetries form the fundament of the "standard model" of physics and help explain the existence of e.g. fotons (quanta of light).
And then there is invariance under changing scale. In statistical mechanics, physical laws are the same on all scales at phase transitions where materials radically change their organization from, say, solid to liquid. The really exciting discovery was that the physics at these critical points is universal, i.e. irrespective of the details of the material that undergoes the transition, the physics is the same.
But scale invariance has also led to an entirely new branch of mathematics, namely that of fractal geometry. Fractals are objects designed to be self-similar across all scales. This type of invariance leads to counter-intuitive notions such as objects having fractional dimensions, say between a line and a plane. But above all, fractals are beautiful. (see Figures).
Why are fractals beautiful? I believe they are beautiful because they look incredibly complex, yet are extremely structured. So structured in fact that they can often be described by a few lines of code. Now, our brain is in the business of searching for structure in the world. Understanding the world is understanding its structure and it allows one to make predictions and that will help us survive and produce more offspring (see earlier blog). It seems that we are not too good in discovering the underlying structure in fractals. Can you tell me the equation that generated the fractals in the figures? Our brain simply did not evolve to deal with these type of geometric ojects. yet, our brain does recognize the enormous opportunity to explain structure in fractals but it is tortured by its inability to do so. Beauty in this sense is fascination, it draws your attention because it never succeeds in its ultimate mission: discover the structure in the world.
Fractals also seem to appear in nature. Perhaps not surprisingly, because simple genetic code can generate intricate geometric shapes that can be functional. A famous examples is your vascular system. Another beautiful example is a type of broccoli shown below. It looks genetically engineered, but I believe it is natural.
Sunday, February 22, 2009
Fright in Disneyland
On Febuary 20, birthday of my oldest daughter, we went to Disneyland. For my children the first time in their life. Every day, 9.30 pm, we heard the fireworks from where we live in Irvine (yes, past tense, because the economic downturn has caused Disneyland to temporarily abondon this feature), so we figured it was time for us to experience the ``happiest place on earth''.
Our first attraction was absolutely wonderful: the all new ``it's a small world after all''. The Netherlands had a display the same size as all of Africa which was of course nice (for us...). But let's not be cynical here, this attraction has a message with educational value, the puppets are simply beautiful and you leave the place with a positive feeling: the world is a benign, beautiful place.
Next on our list was "Thunder Mountain". This was a bit scarier, and my youngest daughter (7) asked me why everyone was raising their arms and screaming all the time. She concluded it must be because they want to show off that they are not afraid. She assured me that she was definitely afraid. All in all not much harm here.
But then it was time for "Priates of Caribbean". It came very highly recommended by one of the Disneyland employees and even the Pirates outside assured me that it was allright for small kids. They were debating the steepness of the drops the boat was making during the ride. However, that turned out to be beside the point. A brief excerpt:
-> A skeleton speared against a rock with a sword in his (her ?) chest.
-> Pirates selling a line of women to other pirates as their new brides (an euphemism for Pirates raping women).
-> Pirates drinking lots of alcohol (perhaps also drugs?).
-> Pirates burning down an entire village (in modern terms: a bombing raid).
-> Pirates torturing a man by submerging him in a waterwell (somehow the word Abu-Ghraib came to mind).
Now keep in mind, the pirates are the good guys here. This is not a realistic display of the aweful pirate terror of the old days. These are role models for our children. How appropriate do you think it would be if we would have a similar Disneyland attraction about the holocaust, or the genocide in Rwanda. If you think about this for a minute or two you quickly realize how bizarre this form of entertainment really is. To us, but certainly to children.
Somewhat shocked through this experience we went for Fantasy land and entered the "Snow White" ride. It can't get much more innocent than that you would tink. To my horror, it too consisted of scary witches and, yes, another corpse. Now we are talking a ride for the very smallest children. Is this really "the happiest place on earth"? Why this obsession with fear? In fact, almost all Disney movies are packed with fear. Why are we showing movies where the mother of main character (Bambi) gets brutally killed? What does this do to our kids?
I believe epic battles between good and evil have an enormous archetypical appeal to us. I am a huge fan of starwars myself. The symbolism can be found an all major religions in the form of the dualism God and Devil, Heaven and Hell, etc. That's allright, but not for the very young. As a society we are disgusted with pictures of nudity. Indeed, I have not seen a single naked breast in Disneyland. Yet, breasts will not shock young children, they grow up with it, its completely natural to them. Breasts are good: they are symbols of motherhood, protection, kindness. However, violence, skeletons, weapons, torture and fear are not natural to children. These things transmit the message that the world is an evil place. How twisted is our society really?
Let's end on a postive note. I am teaching an introduction to artificial intelligence class and for the first time I came face to face with Asimo on that Disneyland trip. Asimo is simply amazing. One constantly has the feeling that there is a little person inside that suit. And everytime you realize that this is not the case, it gets a little scary (a good kind of scary this time). What if Asimo decided not to listen to the boss, somewhat like a dog with big teeth. Asimo, walks, talks, runs and walks chairs, all very gracefully. It's truly amazing, a peek into the future. My youngest child is now Asimo fan. I am happy Asimo won over "Pirates of Caribbean" after all.
Our first attraction was absolutely wonderful: the all new ``it's a small world after all''. The Netherlands had a display the same size as all of Africa which was of course nice (for us...). But let's not be cynical here, this attraction has a message with educational value, the puppets are simply beautiful and you leave the place with a positive feeling: the world is a benign, beautiful place.
Next on our list was "Thunder Mountain". This was a bit scarier, and my youngest daughter (7) asked me why everyone was raising their arms and screaming all the time. She concluded it must be because they want to show off that they are not afraid. She assured me that she was definitely afraid. All in all not much harm here.
But then it was time for "Priates of Caribbean". It came very highly recommended by one of the Disneyland employees and even the Pirates outside assured me that it was allright for small kids. They were debating the steepness of the drops the boat was making during the ride. However, that turned out to be beside the point. A brief excerpt:
-> A skeleton speared against a rock with a sword in his (her ?) chest.
-> Pirates selling a line of women to other pirates as their new brides (an euphemism for Pirates raping women).
-> Pirates drinking lots of alcohol (perhaps also drugs?).
-> Pirates burning down an entire village (in modern terms: a bombing raid).
-> Pirates torturing a man by submerging him in a waterwell (somehow the word Abu-Ghraib came to mind).
Now keep in mind, the pirates are the good guys here. This is not a realistic display of the aweful pirate terror of the old days. These are role models for our children. How appropriate do you think it would be if we would have a similar Disneyland attraction about the holocaust, or the genocide in Rwanda. If you think about this for a minute or two you quickly realize how bizarre this form of entertainment really is. To us, but certainly to children.
Somewhat shocked through this experience we went for Fantasy land and entered the "Snow White" ride. It can't get much more innocent than that you would tink. To my horror, it too consisted of scary witches and, yes, another corpse. Now we are talking a ride for the very smallest children. Is this really "the happiest place on earth"? Why this obsession with fear? In fact, almost all Disney movies are packed with fear. Why are we showing movies where the mother of main character (Bambi) gets brutally killed? What does this do to our kids?
I believe epic battles between good and evil have an enormous archetypical appeal to us. I am a huge fan of starwars myself. The symbolism can be found an all major religions in the form of the dualism God and Devil, Heaven and Hell, etc. That's allright, but not for the very young. As a society we are disgusted with pictures of nudity. Indeed, I have not seen a single naked breast in Disneyland. Yet, breasts will not shock young children, they grow up with it, its completely natural to them. Breasts are good: they are symbols of motherhood, protection, kindness. However, violence, skeletons, weapons, torture and fear are not natural to children. These things transmit the message that the world is an evil place. How twisted is our society really?
Let's end on a postive note. I am teaching an introduction to artificial intelligence class and for the first time I came face to face with Asimo on that Disneyland trip. Asimo is simply amazing. One constantly has the feeling that there is a little person inside that suit. And everytime you realize that this is not the case, it gets a little scary (a good kind of scary this time). What if Asimo decided not to listen to the boss, somewhat like a dog with big teeth. Asimo, walks, talks, runs and walks chairs, all very gracefully. It's truly amazing, a peek into the future. My youngest child is now Asimo fan. I am happy Asimo won over "Pirates of Caribbean" after all.
Saturday, January 10, 2009
Gravitation
I remember reading "The First Man in the Moon" by H.G. Wells quite a long time ago. Wells describes a machine with which a scientist travels to the moon. In this machine the scientist was pulled towards the rear end of the rocket (facing the earth) for the first two-thirds of the trip. At two-thirds he was momentarily weightless but after that he was pulled towards the front of the rocket (the side facing the moon). Now in this story, the rocket wasn't propelled by any real laws of physics, but it makes you wonder, what happened to Armstrong, Aldrin and Collins when they flew to the moon?
Well, clearly, when they took off and the engines were buzzing they were pulled hard towards the earth side of the rocket. But once in space, they turned off their engines. Was there a slight residual pull towards the rear of the rocket because the earth was much closer than the moon? Take a minute to think about this before you read on.
Gravitation pulls exactly as hard on the rocket as on the astronauts and therefore both their speeds decrease, but by exactly the same amount. So inside the rocket you won't notice a thing until you turn on your engines again (or hit an object in space.) The same thing happens to astronauts in the international space station. In this case, the space station is in an eternal fall towards earth. However, through its forward speed it keeps missing the earth! Since space station and astronauts fall with the same speed they are not noticing their fall.
But the story gets better. The gravitational force between two objects is F = -G M * m / R^2 where G is some constant, M the mass of the earth and m the mass of the space station or astronaut and R the distance between them. Thus, the force is stronger for the space station than for the astronaut because its mass (m) is much bigger. Now, when we compute the motion for the station or astronaut through space we need to equate the force with "m * a" where "m" is again the mass and "a" the acceleration. If this is hard to follow forget about the equations and remember this: the mass "m" drops from the equation and the orbit that one can now compute is independent of the mass m!
This magic cancellation is kind of coincidental or suspicious if you want. Perhaps there is a simpler principle at work here. Einstein thought so too. He imagined a version of the following thought experiment. Imagine you are knocked unconscious for a full year and wake up in an elevator. You feel a downward pull on your body and your company tells you this is because you have been abducted and now live on a bigger planet with more mass and thus more gravitational pull. He assures you the elevator is not moving. You are not so sure, because isn't a simpler explanation that you are accelerating upwards in a building? The point is that there is no way to tell the difference between gravitation and accelaretaion and Einstein concluded that if there is no measurement to tell the difference, well then there might be no difference... It's the Turing test for gravity!
In Einstein's general relativity there is no gravitational force and no mysterious cancellations. Mass and energy (remember E=M*c^2) bends space and time in such a way that what used to be straight lines become curved trajectories. Every object travels through this space in exactly the same way without any forces acting on it. The only thing that has changed is that a straight freeway became a roller coaster, but this change is identical for all objects traveling on it, big or small. You may only experience a "force" when you accelerate out of these "free fall geodesics". For instance, simply standing on the surface of the earth blocks the natural geodesic that would move to you the center of the earth. In this way, the earth accelerates you upward, as if you were firing the engines of a rocket. And so what you would classically call gravitational pull becomes upward acceleration, just as the man in the elevator.
Well, clearly, when they took off and the engines were buzzing they were pulled hard towards the earth side of the rocket. But once in space, they turned off their engines. Was there a slight residual pull towards the rear of the rocket because the earth was much closer than the moon? Take a minute to think about this before you read on.
Gravitation pulls exactly as hard on the rocket as on the astronauts and therefore both their speeds decrease, but by exactly the same amount. So inside the rocket you won't notice a thing until you turn on your engines again (or hit an object in space.) The same thing happens to astronauts in the international space station. In this case, the space station is in an eternal fall towards earth. However, through its forward speed it keeps missing the earth! Since space station and astronauts fall with the same speed they are not noticing their fall.
But the story gets better. The gravitational force between two objects is F = -G M * m / R^2 where G is some constant, M the mass of the earth and m the mass of the space station or astronaut and R the distance between them. Thus, the force is stronger for the space station than for the astronaut because its mass (m) is much bigger. Now, when we compute the motion for the station or astronaut through space we need to equate the force with "m * a" where "m" is again the mass and "a" the acceleration. If this is hard to follow forget about the equations and remember this: the mass "m" drops from the equation and the orbit that one can now compute is independent of the mass m!
This magic cancellation is kind of coincidental or suspicious if you want. Perhaps there is a simpler principle at work here. Einstein thought so too. He imagined a version of the following thought experiment. Imagine you are knocked unconscious for a full year and wake up in an elevator. You feel a downward pull on your body and your company tells you this is because you have been abducted and now live on a bigger planet with more mass and thus more gravitational pull. He assures you the elevator is not moving. You are not so sure, because isn't a simpler explanation that you are accelerating upwards in a building? The point is that there is no way to tell the difference between gravitation and accelaretaion and Einstein concluded that if there is no measurement to tell the difference, well then there might be no difference... It's the Turing test for gravity!
In Einstein's general relativity there is no gravitational force and no mysterious cancellations. Mass and energy (remember E=M*c^2) bends space and time in such a way that what used to be straight lines become curved trajectories. Every object travels through this space in exactly the same way without any forces acting on it. The only thing that has changed is that a straight freeway became a roller coaster, but this change is identical for all objects traveling on it, big or small. You may only experience a "force" when you accelerate out of these "free fall geodesics". For instance, simply standing on the surface of the earth blocks the natural geodesic that would move to you the center of the earth. In this way, the earth accelerates you upward, as if you were firing the engines of a rocket. And so what you would classically call gravitational pull becomes upward acceleration, just as the man in the elevator.
Thursday, January 8, 2009
Intelligence
Alan Turing, the father of computer science, proposed a test to determine if an artificial system could be intelligent. He proposed to play a game where a human interrogator would ask questions to either another human or to a machine which would be hidden out of sight (behind a brick wall or on the other side of the earth). If the human interrogator would be fooled 30% of the time the system would pass the test and ought to be called intelligent.
Now, this is a somewhat arbitrary line in the sand. Why 30%? How many humans must it fool? Moreover, why would we be so concerned about human intelligence? Can't we imagine other types of intelligence that would not be able to fool us? I don't think we should take the claim too seriously that this ought to be the working definition of intelligence, but rather a very useful illustration of what may be needed to build robots that are behaving in ways similar to humans. Turing believed that by the year 2000 this machine would be reality. It didn't turn out that way.
The Turing test did spark a lot of heated debate about whether we could ever build machines that can think in the same way as humans do. Most importantly, can we ever build robots that have a conscious mind? John Searle was the most vocal opponent of this idea. He invented interesting thought experiments which, according to his reasoning, would make it clear that such a thing could not happen. He imagined a "Chinese Room" with a human in it that would receive questions in Chinese. S/He would also have access to an enormous library with a (Chinese) answer to every conceivable (Chinese) question. Mechanically, the person would look up the answer, perhaps randomly picking from a large set of possible answers, and return it to the interrogator. Clearly this person has no idea of Chinese even though s/he would pass this variant of the Turing test.
One objection could be that it is really somewhat unrealistic to ask for a library with answers to every possible question; it would require more atoms to store than the universe has to offer. So perhaps we should restrict ourselves to systems that are limited in their capabilities of storing data. This seems like a workable sophistication of the Turing test, but I am pretty sure someone else will come with a system that may beat the Turing test in another way which would then require addition changes because "surely that system could not be really intelligent"!
In another famous thought experiment (the "brain prosthesis experiment" by Clark Glymour) one imagines replacing the neurons in one's brain one by one with electrical circuits. Each circuit models the input output function of a neuron perfectly. By the time we are done we have replaced the entire brain by a computer. We can do it again in reverse order to rebuild the human. Assuming for a moment this can be technically achieved, would we then have altered the human mind in any way? Many people (among whom Searle) believe this is the case. During the process they believe that one will gradually loose any conscious experience of the world, yet one's behavior will remain unaltered.
How about this thought experiment then? In 3000 AC a smart scientist comes up with a really clever way to build an intelligent machine. S/He first assembles a long strand of amino-acids which s/he places in a small protein cover. The strand of amino-acids will copy itself and code for new proteins to be assembled. The proteins become a robotic body (made mostly of carbohydrates and water) with a brain. S/He then decides to raise this robot among her biological children. This robot has a normal life because it looks so much like a real human being. And some day, it starts referring to itself as "me". It looks in the mirror and understands who that person is. When published, people cry out: "this is not a real robot" it was just a copy of a human being grown in a test tube". So the scientist develops new techniques to copy the exact same process by now with new synthetic molecules that have the same properties as carbon, hydrogen, iron etc. People are still not satisfied until the same procedure is now repeated with electronic circuits. Now what? Intelligent machine or not?
For me these examples drive home the fact that a mind has nothing to do with the substrate it is made from. It's about computation and information. And therefore, if our goal is creating human intelligence, yes, we will be able to build intelligent machines that may have conscious minds, real emotions and all the other good and bad things that make us human. These are not magical, but rather functional properties. We evolved them for a purpose. They are fascinating, and utterly un-understood today, but perhaps not un-understandable or non-replicable. Our brain is extremely complex, propelled by principles that we don't understand today. It therefore seems far to premature to claim that a consciousness mind is something reserved uniquely for human beings.
Now, this is a somewhat arbitrary line in the sand. Why 30%? How many humans must it fool? Moreover, why would we be so concerned about human intelligence? Can't we imagine other types of intelligence that would not be able to fool us? I don't think we should take the claim too seriously that this ought to be the working definition of intelligence, but rather a very useful illustration of what may be needed to build robots that are behaving in ways similar to humans. Turing believed that by the year 2000 this machine would be reality. It didn't turn out that way.
The Turing test did spark a lot of heated debate about whether we could ever build machines that can think in the same way as humans do. Most importantly, can we ever build robots that have a conscious mind? John Searle was the most vocal opponent of this idea. He invented interesting thought experiments which, according to his reasoning, would make it clear that such a thing could not happen. He imagined a "Chinese Room" with a human in it that would receive questions in Chinese. S/He would also have access to an enormous library with a (Chinese) answer to every conceivable (Chinese) question. Mechanically, the person would look up the answer, perhaps randomly picking from a large set of possible answers, and return it to the interrogator. Clearly this person has no idea of Chinese even though s/he would pass this variant of the Turing test.
One objection could be that it is really somewhat unrealistic to ask for a library with answers to every possible question; it would require more atoms to store than the universe has to offer. So perhaps we should restrict ourselves to systems that are limited in their capabilities of storing data. This seems like a workable sophistication of the Turing test, but I am pretty sure someone else will come with a system that may beat the Turing test in another way which would then require addition changes because "surely that system could not be really intelligent"!
In another famous thought experiment (the "brain prosthesis experiment" by Clark Glymour) one imagines replacing the neurons in one's brain one by one with electrical circuits. Each circuit models the input output function of a neuron perfectly. By the time we are done we have replaced the entire brain by a computer. We can do it again in reverse order to rebuild the human. Assuming for a moment this can be technically achieved, would we then have altered the human mind in any way? Many people (among whom Searle) believe this is the case. During the process they believe that one will gradually loose any conscious experience of the world, yet one's behavior will remain unaltered.
How about this thought experiment then? In 3000 AC a smart scientist comes up with a really clever way to build an intelligent machine. S/He first assembles a long strand of amino-acids which s/he places in a small protein cover. The strand of amino-acids will copy itself and code for new proteins to be assembled. The proteins become a robotic body (made mostly of carbohydrates and water) with a brain. S/He then decides to raise this robot among her biological children. This robot has a normal life because it looks so much like a real human being. And some day, it starts referring to itself as "me". It looks in the mirror and understands who that person is. When published, people cry out: "this is not a real robot" it was just a copy of a human being grown in a test tube". So the scientist develops new techniques to copy the exact same process by now with new synthetic molecules that have the same properties as carbon, hydrogen, iron etc. People are still not satisfied until the same procedure is now repeated with electronic circuits. Now what? Intelligent machine or not?
For me these examples drive home the fact that a mind has nothing to do with the substrate it is made from. It's about computation and information. And therefore, if our goal is creating human intelligence, yes, we will be able to build intelligent machines that may have conscious minds, real emotions and all the other good and bad things that make us human. These are not magical, but rather functional properties. We evolved them for a purpose. They are fascinating, and utterly un-understood today, but perhaps not un-understandable or non-replicable. Our brain is extremely complex, propelled by principles that we don't understand today. It therefore seems far to premature to claim that a consciousness mind is something reserved uniquely for human beings.
Monday, January 5, 2009
Our Unconscious Mind
Everyone is faced with making difficult decisions, buying a house, changing your job, moving to a new place, another addition to the family perhaps. For a long time, the going theory was that we take these decisions consciously, in full control of our faculties. Logically weighting pros and cons. However, this "Mr. Spock view" of ourselves is shifting. Perhaps the most recent influential work in social psychology in this respect is a recent book by Timothy Wilson: Strangers to Ourselves where he argues that most decisions are actually taken unconsciously. In fact, the argument goes that our "gut feeling" is very good at taking these kinds of complex decisions involving many sources of information. A similar story can be read in Damasio's "Descarte's Error" where he describes a patient who has lost this ability with disastrous results.
In Wilson's book one can read a endless list examples of how we seem to perform much of our "reasoning" unconsiously. The one that stood out most to me was an experiment by Lewicki, Hill and Bizot. Here, subjects get to see the symbol "X" in one of four quadrants of a computer screen. Their task is report as quickly as they can where they spotted the X. The sequence of the X locations follows a complex but predictable pattern. Over time, these subjects learn these patterns and improve their response time. However, when they are asked why they improve their performance they have no clue. Better even, after a while the sequence reverts to a random sequence resulting in a fast drop in response time. The subjects noticed their drop in response time but can not figure out what caused it! They do not know that the sequence suddenly became unpredictable (or that there was a predictable sequence in the first place.)
What conclusion can be drawn from this? Wilson argues that the pattern was learned completely unconsciously. We constantly pick up on regularities in the world around us. As argued in an earlier post, prediction is crucial for our survival so we tend to be really good at it. As a child, we have learned thousands if not millions of regularities about our world: water makes me wet and moreover I cannot walk on it, fire is really hot and hurts my body, when someone frowns at me s/he is angry etc. Some of them have become conscious over time, but you just wonder how many of these rules remain unconscious for our entire lifetime.
I like the idea of our consciousness as a kind of "searchlight". We can monitor some decisions that are being taken unconsciously but have much less conscious influence over this process then we like to think. Our mind tricks us in thinking we are in control, but in reality we mostly observe. There are indeed also a number of experiments which confirm this point to view to some extend. One can measure skin conductance which apparently changes when a decision has been made. Skin conductance seemed to change before people reported to have taken a decision. I don't quite think our consciousness is useless as the searchlight hypothesis seems to suggest. Otherwise, why did it evolve? Perhaps it can be used to guide our long term goals in life.
The searchlight hypothesis may be disconcerning to some. If we take decisions unconsciously, can we still be held accountable for it? Emphatically: yes! Fear of punishment does change our behavior, conscious or unconscious. We don't need to understand the details of how our decision making process to understand that the Law works (to some extend) and will work even after we do understand it. Another disconcerning factor is the fact that we have become a little less "human". After the realization that God's creation ship earth is not in the middle of our universe (hell, not even in the middle of our solar system), and the fact that man evolved from apes only 1 million years ago, we now have to face the fact that our awake mind is mostly an observer with her/his hands tied down! Personally I don't feel that way. An experience is an experience and an emotion remains an emotions whatever the underlying mechanisms are that generate it.
In Wilson's book one can read a endless list examples of how we seem to perform much of our "reasoning" unconsiously. The one that stood out most to me was an experiment by Lewicki, Hill and Bizot. Here, subjects get to see the symbol "X" in one of four quadrants of a computer screen. Their task is report as quickly as they can where they spotted the X. The sequence of the X locations follows a complex but predictable pattern. Over time, these subjects learn these patterns and improve their response time. However, when they are asked why they improve their performance they have no clue. Better even, after a while the sequence reverts to a random sequence resulting in a fast drop in response time. The subjects noticed their drop in response time but can not figure out what caused it! They do not know that the sequence suddenly became unpredictable (or that there was a predictable sequence in the first place.)
What conclusion can be drawn from this? Wilson argues that the pattern was learned completely unconsciously. We constantly pick up on regularities in the world around us. As argued in an earlier post, prediction is crucial for our survival so we tend to be really good at it. As a child, we have learned thousands if not millions of regularities about our world: water makes me wet and moreover I cannot walk on it, fire is really hot and hurts my body, when someone frowns at me s/he is angry etc. Some of them have become conscious over time, but you just wonder how many of these rules remain unconscious for our entire lifetime.
I like the idea of our consciousness as a kind of "searchlight". We can monitor some decisions that are being taken unconsciously but have much less conscious influence over this process then we like to think. Our mind tricks us in thinking we are in control, but in reality we mostly observe. There are indeed also a number of experiments which confirm this point to view to some extend. One can measure skin conductance which apparently changes when a decision has been made. Skin conductance seemed to change before people reported to have taken a decision. I don't quite think our consciousness is useless as the searchlight hypothesis seems to suggest. Otherwise, why did it evolve? Perhaps it can be used to guide our long term goals in life.
The searchlight hypothesis may be disconcerning to some. If we take decisions unconsciously, can we still be held accountable for it? Emphatically: yes! Fear of punishment does change our behavior, conscious or unconscious. We don't need to understand the details of how our decision making process to understand that the Law works (to some extend) and will work even after we do understand it. Another disconcerning factor is the fact that we have become a little less "human". After the realization that God's creation ship earth is not in the middle of our universe (hell, not even in the middle of our solar system), and the fact that man evolved from apes only 1 million years ago, we now have to face the fact that our awake mind is mostly an observer with her/his hands tied down! Personally I don't feel that way. An experience is an experience and an emotion remains an emotions whatever the underlying mechanisms are that generate it.
Sunday, January 4, 2009
Adam and Eve
All humans descended from Adam and Eve. That's the biblical story at least. The scientific community has their own Adam and Eve: Y-Chromosome Adam and Mitochondrial Eve to be more precise. They did not live in the garden of Eden but probably in Africa. They didn't live 6000 years ago, but about 60,000 years ago (Adam) and 140,000 (Eve). Who are these people, or rather, how are they defined?
Let's focus on Eve. She is defined to be the most recent common female ancestor of all humans that live today if we only consider female descend. Ok, let's explain that a bit more. Consider all living humans today. Next consider all their mothers, then all their mothers etc. The crucial point is that we only consider women and moreover that everyone has precisely 1 mother so there can never be more mothers than children. However, mothers can have more than one child (and typically will) while other females will not have any children. So, we expect that the set of our grand-grand-etc. mothers will actually shrink when we extrapolate back in time, until there is only a single mother left.
This reasoning can easily lead to confusion. For instance, it doesn't mean that Eve lived alone, nor that she was the sole female in her times. In fact, many other women lived in her times. More importantly, they may have well been our ancestors through their son's lineages. Eve is defined throught her daughters alone: to be Eve, you need to be the only female from which a path of daughters exists all the way to modern times (in fact to all humans). In contrast, female paths from all other female contemporaries will end prematurely.
Mitochondrial DNA (mitochondria are the power plants of our cells) happens to be passed along the female line and so we can claim is that everyone today inherits the mitochondrial DNA from Eve (not so for the DNA residing in our genes.) A similar effect occurs for the Y-chromosome in a man which is passed along the male line.
Perhaps, more bizarrely, the honor of being Eve changes hand continuously. Imagine that Eve lived in 140,000 BC. In 130,000 BC there happened to be two female descendants of Eve that are the two mothers of all modern humans through female lineage. One day, the last descendant of mother A will die and at that point in time, mother B will take over the title "Eve" (she is then more recent than the former Eve living 140,000 BC). Thus, the honor is only bestowed retrospectively.
The mathematics of the proof depends on the fact that it is likely that some two mothers in one generation will be the daughters of a single mother in the next generation. It is not necessary. If two populations are well separated it will take many steps of going back in time to "coalesce" the two mothers to one mother. So the "mother-set" will remain stable at two for a long time. However, if we go far enough back in time, it seems very unlikely that two species of human evolved from fish! In the extreme case, to argue for two Eve's, we would have to argue that life has started twice on earth. Not impossible, but very unlikely if we consider how much we look alike.
Then finally, we should also not confuse Eve with the "most recent common ancestor" which has all humans as descendants. This person or persons (male or female) lived much later (perhaps even 3000 years ago.) because its descends are defined through both male and female lineages. Moreover, their contemporaries may also have many descendents today (just not everyone). Going back in time one can show that there must be a point in time (the "identical ancestors point") where each person who lived then will have either everyone or no-one as a descendent today.
Let's focus on Eve. She is defined to be the most recent common female ancestor of all humans that live today if we only consider female descend. Ok, let's explain that a bit more. Consider all living humans today. Next consider all their mothers, then all their mothers etc. The crucial point is that we only consider women and moreover that everyone has precisely 1 mother so there can never be more mothers than children. However, mothers can have more than one child (and typically will) while other females will not have any children. So, we expect that the set of our grand-grand-etc. mothers will actually shrink when we extrapolate back in time, until there is only a single mother left.
This reasoning can easily lead to confusion. For instance, it doesn't mean that Eve lived alone, nor that she was the sole female in her times. In fact, many other women lived in her times. More importantly, they may have well been our ancestors through their son's lineages. Eve is defined throught her daughters alone: to be Eve, you need to be the only female from which a path of daughters exists all the way to modern times (in fact to all humans). In contrast, female paths from all other female contemporaries will end prematurely.
Mitochondrial DNA (mitochondria are the power plants of our cells) happens to be passed along the female line and so we can claim is that everyone today inherits the mitochondrial DNA from Eve (not so for the DNA residing in our genes.) A similar effect occurs for the Y-chromosome in a man which is passed along the male line.
Perhaps, more bizarrely, the honor of being Eve changes hand continuously. Imagine that Eve lived in 140,000 BC. In 130,000 BC there happened to be two female descendants of Eve that are the two mothers of all modern humans through female lineage. One day, the last descendant of mother A will die and at that point in time, mother B will take over the title "Eve" (she is then more recent than the former Eve living 140,000 BC). Thus, the honor is only bestowed retrospectively.
The mathematics of the proof depends on the fact that it is likely that some two mothers in one generation will be the daughters of a single mother in the next generation. It is not necessary. If two populations are well separated it will take many steps of going back in time to "coalesce" the two mothers to one mother. So the "mother-set" will remain stable at two for a long time. However, if we go far enough back in time, it seems very unlikely that two species of human evolved from fish! In the extreme case, to argue for two Eve's, we would have to argue that life has started twice on earth. Not impossible, but very unlikely if we consider how much we look alike.
Then finally, we should also not confuse Eve with the "most recent common ancestor" which has all humans as descendants. This person or persons (male or female) lived much later (perhaps even 3000 years ago.) because its descends are defined through both male and female lineages. Moreover, their contemporaries may also have many descendents today (just not everyone). Going back in time one can show that there must be a point in time (the "identical ancestors point") where each person who lived then will have either everyone or no-one as a descendent today.
Friday, January 2, 2009
The Baldwin Effect in Evolution
Evolution is the search for a genetic code that produces a phenotype that can produce lots of offspring in a particular environment. The disadvantage is that it can be quite slow. Variation in genotype are generated through crossover in chromosomes and to a lesser extend, mutations of the generatic code. Problem is, to come up with a good solution to a problem there needs to exist a series of very small changes that are all advantageous. Can we go faster?
Lamarck proposed that experience could directly affect the genotype and as such guide the evolution. It is now widely assumed that this is does not happen. But there is another way to ameliorate the negative effects of a mutation, and that is through phenotype plastcity or learning. Imagine a new adaptation is desperately needed to deal with new theats (e.g. new preditors). If available suboptimal mutations can be more easily adapted to the specific problem under selective pressure, then it will help the survival of those individuals who use it. In turn, once these indivuals have started to utilize this adaptation to their advantage (and perhaps teach their children how to use it as well), evolution can further improve it by selecting individuals who have more advanced adaptions and know how to to use it.
More generally, evolution is more effective in individuals who can learn. Thus, if learning is not too costly (in terms of energy consumption) evolution will probably select for those genotypes that develop brains. Hence evolution improves learning and leaning improves evolution. This positive feedback is what I understand as the Baldwin effect.
There seems to be a second phase in the Baldwin effect. Namely, that over time, when the environment remains stable, the learning element will be largely removed from the equation. Where the early versions of an adaptation relied on some learning (either by parents or by inventing the wheel every time), the later version is hardwired into the genes. The idea being that this solution is more robust, albeit less flexible.
Lamarck proposed that experience could directly affect the genotype and as such guide the evolution. It is now widely assumed that this is does not happen. But there is another way to ameliorate the negative effects of a mutation, and that is through phenotype plastcity or learning. Imagine a new adaptation is desperately needed to deal with new theats (e.g. new preditors). If available suboptimal mutations can be more easily adapted to the specific problem under selective pressure, then it will help the survival of those individuals who use it. In turn, once these indivuals have started to utilize this adaptation to their advantage (and perhaps teach their children how to use it as well), evolution can further improve it by selecting individuals who have more advanced adaptions and know how to to use it.
More generally, evolution is more effective in individuals who can learn. Thus, if learning is not too costly (in terms of energy consumption) evolution will probably select for those genotypes that develop brains. Hence evolution improves learning and leaning improves evolution. This positive feedback is what I understand as the Baldwin effect.
There seems to be a second phase in the Baldwin effect. Namely, that over time, when the environment remains stable, the learning element will be largely removed from the equation. Where the early versions of an adaptation relied on some learning (either by parents or by inventing the wheel every time), the later version is hardwired into the genes. The idea being that this solution is more robust, albeit less flexible.
Thursday, January 1, 2009
Simpson's Paradox
As I argued in my previous blog on three principles to learning, paradoxes are a great way to sharpen one's intuition. The following paradox was brought to my attention by Hans Welling in Evora, Portugal (incidentially, my brother.) Here is the most intuitive version I could think of.
Imagine two driving schools advertising by publishing their success rates. School A reports a success rate of 65% (720 of their 1100 students passe their driving exam), while school B reports 35% (only 380 out of 1100 students passed.) Clear evidence school A is better than school B right? Or not?
Turns out school A had 1000 students under the age of 30 of which 700 passed (70%) and only 100 old students over the age of 50 of which 20 passed (20%). School B on the other hand had only 100 young students of which 80 passed (80%) and 1000 old students of which 300 passed (30%). Here is the surprise: school B was performing better for both the young students (80% versus 70%) and the old students (30% versus 20%). Which school would you pick now? The paradox is resolved by realizing school B had to deal with so many more old students who on average have much lower passing results.
This seems harmless as long as you know which type of subgroups you are dealing with. But now imagine doing a drug test: does drug A or drug B work better? How do you know you didn't accidentally have a high percentage of subjects in group A that have gene X that makes them react much better to a drug A? You don't, and there seem to be an enormous number of possible subgroups that may randomly appear in your sample.
The best one can do is to make sure the subjects were chosen from the same population with no hidden biases to select subjects for drug A or for drug B. For instance, combining two results from the literature is dangerous, because one group me be English while the other American, or one group may have lived 20 years earlier than the other. But even so, to claim statistical significance for drug A to be better than drug B (or vice versa) one would have to correct for the possibility of randomly selecting unbalanced subgroups that react differently to one of the drugs. Seems a pretty daunting task and I am not so sure this is routinely done.
Imagine two driving schools advertising by publishing their success rates. School A reports a success rate of 65% (720 of their 1100 students passe their driving exam), while school B reports 35% (only 380 out of 1100 students passed.) Clear evidence school A is better than school B right? Or not?
Turns out school A had 1000 students under the age of 30 of which 700 passed (70%) and only 100 old students over the age of 50 of which 20 passed (20%). School B on the other hand had only 100 young students of which 80 passed (80%) and 1000 old students of which 300 passed (30%). Here is the surprise: school B was performing better for both the young students (80% versus 70%) and the old students (30% versus 20%). Which school would you pick now? The paradox is resolved by realizing school B had to deal with so many more old students who on average have much lower passing results.
This seems harmless as long as you know which type of subgroups you are dealing with. But now imagine doing a drug test: does drug A or drug B work better? How do you know you didn't accidentally have a high percentage of subjects in group A that have gene X that makes them react much better to a drug A? You don't, and there seem to be an enormous number of possible subgroups that may randomly appear in your sample.
The best one can do is to make sure the subjects were chosen from the same population with no hidden biases to select subjects for drug A or for drug B. For instance, combining two results from the literature is dangerous, because one group me be English while the other American, or one group may have lived 20 years earlier than the other. But even so, to claim statistical significance for drug A to be better than drug B (or vice versa) one would have to correct for the possibility of randomly selecting unbalanced subgroups that react differently to one of the drugs. Seems a pretty daunting task and I am not so sure this is routinely done.
Subscribe to:
Posts (Atom)