Revealing Stereo And 3D

Double Slit Experiment Revisited
Home | Downloads | DCB | Store | Boot | Xprmnts | Intro | Bib | Motion | Focus | Sgmnt | Rcgntn | Depth | Build | Fail

        Now since I only deal with computer vision people might be wondering, what this 200 year old, two slit experiment by Thomas Young has to do with it. If you are bright enough you would have already got the answer to it. Even though this experiment was a great success at his time and remained to be for nearly 100 years, Plank’s quantum hypothesis reopened it. For nearly 100 years now there had been a lot of new explanations to it including few from Einstein, Bhor and Fenyman. Finally now at the time of writing this document there might be a satisfactory explanation to it, but I am not quite sure of it, since I am not physicist. Anyway, whatever the explanation the final results are going to be the same, since an explanation is not going to change the way a photon is going to behave (Photons do not have brains!).

Now, am I trying to giving a new explanation to it? No, not at all, I am just trying to fit the results of the two slit experiment somehow to see if it solves the correspondence or the stereo problem. The two slits in our case are our eyes. Suppose the surrounding contains a single point source, the photons from this point source can be made to interfere with each other to form an interference pattern. This interference pattern along with the angle of sight for one of the eyes can be used to find the angle of sight for the other eye. Not very clear! Here’s the detailed explanation.

Let us assume that before the process of correspondence our eyes are looking at only those photons that pass through the center of the lens. At any point on the lens there will be photons coming from all possible directions, the center of the lens is no different. Therefore at the center of the lens we can construct a spherical region, with the center as the center of the lens and radius epsilon (epsilon being some small value). Now any line that can be drawn from the surface of the sphere that lies outside the lens to the center is a possible path for a photon. Every time we see anything in our surrounding one of our eyes is going to take precedence over the other and focus on the point that lies exactly on the optical axis of it. The focusing method may be the QED technique discussed in the other article. It actually doesn’t matter which method we use to focus as long as it is independent of the other eye. Before crossing the next gate, let me give you a small introduction to holograms.

Holograms are a way to embed 3D information on a 2D surface. It touches the ideal case of the Shannon’s theorem. To get a better reproduction of the stored information monochromatic light or LASER is used to develop a hologram. A holographic sensor not only stores the intensity information but also the phase information of light. In order to get these parameters on to the sensor two beams of LASER are used. One act’s as a reference and the other is used to illuminate the objects. The reference beam comes straight to the sensor, and the beam used to illuminate the objects is made to interact with the reference beam after it is reflected from the objects to generate an interference pattern which is also stored. This interference pattern is unique for each of the objects. Once the photographic film is developed the 3D information can be reproduced by illuminating it with light at different angles. There are different ways in which a hologram is produced, but the basic concept remains the same. Now where does LASER come from in our vision?

As I told you earlier one eye is taken as a reference, say in our case it is the left eye (it need not be the same every time you see an object). So the left eye only admits the photon that is traveling along the optical axis. The right eye doesn’t know what distance the left eye is looking, but knows the angle at which the photon is coming to it. Now the question is, is it possible for the right eye to generate an interference pattern with the photons of the left eye? There will be an infinite number of interference patterns due to the infinitely many photons coming to the center of the right eye from all the angles, as described earlier. So the problem is to find the right interference pattern that corresponds to the photon that is coming from the object along the optical axis of the left eye. This will give the angle the right eye has to look to match the images stereoscopically.

One thing that is missing here is the LASER aspect of the photons that are entering the eyes. In order to take care of this condition we can admit only those photons that arrive at the center of the lens at a particular phase. These things are not tried out by me experimentally and are only proposals. They are not blind proposals though, because they are based on the explanations given by quantum physics on the various phenomenons used here. One doubt that still remains in my mind is whether this problem can be solved without giving the LASER aspect to the photons. The only way to answer these questions is to experiment.
       

The double slit experiment was discovered in the early 1800’s to prove the wave theory of light. It was revisited in the early 1900’s and verified with the particle theory as well. It produced bizarre consequences, because it seemed to verify even when single photons were fired with a certain time gap between each firing. Now in the early 2000’s it is revisited again but for a completely new purpose. It is not a new theory for light, or there is no new interpretation to it, only the results of this rigorously and repeatedly done experiment is put to test in a new application.

        We know that human beings and a lot of other animals have a two eyed vision called stereovision. Stereovision helps us to perceive the world in 3D, or in other words it allows us to perceive depth. In order to stimulate this kind of a vision in robots, scientists calculate what is called the disparity map from two images of a same space taken from two slightly different angles i.e. with a small separation between the cameras, similar to our eyes. But this method has a lot of limitations one of which is shown below.

fail.jpg

The two ellipses are the two eyes. A, B, C and D are 4 objects such that B and C are identical. Objects A and D are exactly along the line bisecting the horizontal line joining the two eyes. B and C are placed such that the images projected by them in the two eyes occupy the same place as that occupied by A. B is not seen by the left eye and C is not seen by the right eye, because D projects itself in the same space as B and C in the left and right eyes respectively. Let’s now consider two cases.

Case1: Suppose only D and A are present in the diagram, the right eye has the image of D to the left of A and the left eye has the image of D to the right of A. You get a certain disparity map for this arrangement and that would indicate the correct depth between the objects A and D.

Case2: Now suppose B and C are introduced in the diagram, A is no longer visible. In fact the two new objects are projecting themselves in the same place as that of A and therefore since the objects are identical it would produce the same disparity map as earlier. But now the disparity map has failed to produce the right result.

This means that there should be a new way to look at the problem itself! Trying to solve it after the image formation is of no use as we have no extra information to distinguish between the objects A and either B or C. Since this problem is an old one and well known among computer vision scientists, a lot of attempts were made to solve the problem by making use of the focus parameter i.e. depth from focus. If you know where one of your cameras is focused and the angle that its optical axis makes with the horizontal, then you could readily pin point the location of the point along the plane of the eye. With this information being passed on to the other eye even this eye could know where it is supposed to see. Then a matching can be done and the objects at a particular depth can be easily segmented. This method definitely eliminates the search for a stereo match, but in reality there isn’t a perfect focusing method. So you definitely do not know whether the information you are passing is right or wrong. Even if you find an accurate focusing technique, such as the one that I have described in the article on focus, what if your robot mistakenly wears spectacles? Even though the object is correctly focused, the focus value obtained now is different from the one that will be obtained with out the spectacle. You are again passing the wrong information to the other camera!

Here is a new method using the results of the double slit experiment. The two eyes are considered as the two slits of the experiment and the source of light is the point that is to be matched by the two eyes. This method is independent of whether the eye has focused or not and hence independent of the focusing technique used. I am not a physicist and therefore cannot describe this technique in full detail from the physics point of view, but can somehow manage to scan the entire process to give you a good idea.

In a two slit experiment the source of light is a LASER and it is a single source. In order to get these two into the vision scenario let me assume that, the eye is capable of allowing only that photon that enters it along the optical axis to participate in the game called interference. This holds good only for one eye, say in this case it is the left eye. One of the eyes has to take predominance over the other so that this happens. This is because you always view an object or a point that lies on the optical axis of the two eyes, or in case of binocular rivalry it will be one eye. This would happen after the stereo correspondence is over, initially only one eye will be in the right direction. Now the problem is to get the right eye to the an angle at which it would be able to correspond the two points exactly. At any point on the surface of the lens or eye we can say that the photons can come from any direction. For the right eye we will consider only those photons that arrive at the center of the lens (the first point on the lens that the optical axis meets), and in any direction. Let us assume that it is capable of sampling the photons in any direction at any time. The technology to achieve these things may not be present at the time of writing this document, but let’s just try putting it in theory. So the right eye is going to sample each direction and allow the photons coming in that particular direction to participate in the game. The interference pattern produced will be from the photons that are arriving from the two eyes similar to the two slits. We still need to embed the aspect of LASER into this? For this to happen we have to give some more super powers to our lens. Let us assume that at the point of entry, the photons are down converted to a particular single frequency or filtered. So each eye is going to allow only one particular frequency, while the frequency selected may be different in both the eyes. Since the frequencies in both the eyes are known, the interference pattern can be interpreted without any problems. Instead of confusing ourselves a lot let us assume that there are a common set of visible frequencies between the eyes, and if it is not there let’s skip that direction. Here the only aim is to get an idea of how the double slit experiment will be used to solve the stereo correspondence problem and so I will make it as simple as possible with more than required assumptions at times. Every direction is going to give one or the other pattern, and our problem is to select the right one.

        For one particular common frequency we have an interference pattern, and with the help of it we guess the location of the point on the plane of the eye. If the photons in the two eyes are not coming from a single point, this will be a wrong guess. So we go on to verify it with a next common frequency. If the guessed point were right we already know the pattern that should appear. If the expected pattern matches with the pattern that is formed now the guessed point should be right! The right eye need not sample or scan the directions 3 dimensionally, because the plane of the eyes is already known. It only traces a semicircle to find a perfect match.
 
Here are some links to my photography gallaries: