Abstract
The fundamental principles of optical holography for capturing the optical waves of physical objects, and its difference from photography, are described. A photograph can only record a single view of an object scene; a hologram is capable of capturing the entire optical wavefront that impinges on it. There should be little difference between observing a hologram and the physical object scene. Numerical generation of digital holograms, commonly known as computer-generated holography (CGH), is presented. Two important approaches in CGH, the point-based and the layer-based methods, are described. The point-based method is suitable for generating holograms of simple objects with a small number of object points; the layer-based method is preferred for an object scene with a large number of object points concentrated in a few depth planes. The method for recovering a 3-D scene image from a digital hologram is provided. Three different methods for capturing digital holograms of physical object scene are described. The first method is similar to the art of optical holography, but instead of a photographic film, a digital camera is used to record the holographic waves emitted from the scene. As a digital camera can only record intensity information, the method can only be employed to capture an off-axis, amplitude-only hologram. The other two methods, known as phase-shifting holography (PSH) and optical scanning holography (OSH), are capable of capturing both the magnitude and phase components of the holographic signals. PSH is faster in operation, while OSH can be used to capture holograms of large objects. A simplified version of OSH, known as non-diffractive optical scanning holography (ND-OSH), is presented. ND-OSH is similar in principle to OSH, but the complexity of the optical and electronic setups is reduced.
1.1 Basic Concept of Holography
When light impinges from one medium to another, part of it will be transmitted while the rest will bounce back to the environment or be absorbed into the medium. Figure 1.1(a) shows a ray of monochromatic light (i.e., single frequency or wavelength) L falling on a planar interface between two different homogeneous media M1 and M2.The beam is inclined at an angle θ1 with respect to the normal of the interface. Assuming that M2 is partially transmissive, and L is a plane wave, the direction of part of the light transmitting through M2 will be changed from θ1 to θ2 . This process is known as refraction, and the relation between the pair of angles is governed by Snell’s law.
If the interface between the two media is smooth, the light that is not refracted will be reflected back as specular (mirror-like) reflection, in which case the angle of incident θ1 will be identical to the angle of reflection θ2 . For an interface with a rough surface composed of microscopic irregularities, the incident beam will be scattered, resulting in diffusive reflection, as shown in Figure 1.1(b). Scattering is a generic term that refers to the change in the propagation path of an electromagnetic wave as it is intercepted by some form of non-uniformities in a medium. Optical scattering results when a beam of light hits one or more tiny particles.
The light impinging on a physical object undergoes one or more of the abovementioned modes of propagation. Light emitted from the illuminated object surface conveys information on the shape and color of the object. As the emitted light waves fall on the retina of human eyes, an image of the object will be perceived by the brain.
Photography and optical holography are methods for capturing the emitted light waves from an object scene onto a photographic film. When the photographic film carrying the recorded optical signals is presented to an observer, the light waves of the object scene will be partially or fully reproduced to the eyes of an observer. At present, there are two major approaches in capturing the light waves of an illuminated object: photography and holography. Photography is the art of capturing the part of the optical waves corresponding to a two-dimensional (2-D) projected view of an object, while holography captures the entire optical waves. A photograph only presents the intensity or color of a scene; it is a planar image and does not contain any disparity or depth information. With holography, all the light waves emitted by the object are captured, and hence are capable of reproducing a three-dimensional (3-D) view of the object. These two image-recording techniques are explained as follows.
Light, or optical signal, is a kind of electromagnetic wave in the form of a time-varying sinusoidal wave oscillating with a certain amplitude and frequency. When a person observes a scene, optical waves of different intensities and frequencies will be received by different groups of sensory cells in the retina, creating the impression of color and brightness. The instantaneous amplitude v(t) of a single beam of monochromatic light L is generally expressed as:
In Eq. (1.1), t represents the time variable, while the terms ω and θL denote the frequency and phase of the light beam, respectively. For a monochromatic light wave of constant frequency, the frequency term is sometimes left out, and the representation of the wave can be simplified as v=Aexp(iθL) , where i denotes the imaginary unit. In the physical world, an object reflects an infinite number of light beams from its surface, and the collective excitation of these optical signals on the retina leads to the formation of an image of the object to an observer.
Next, we shall explore how the optical waves of an object can be captured. Imagine the light wave of an object, which is sometimes referred to as the object wave, is intercepted by a transparent medium that is capable of recording the amplitude and phase of every optical beam that passes through it, as shown in Figure 1.2(a). For the time being, let’s ignore the details of the mechanism behind the recording process. As this medium is illuminated by a coherent plane optical wave (i.e., the amplitude and phase at every part of the wavefront is identical), the light beam that emerges from every point on it will be modulated (in amplitude and phase), by an amount that is identical to that which has been recorded. As illustrated in Figure 1.2, the observer will see the same set of optical waves from either the object or the recording media. Theoretically, the optical waves received by the retina in both cases should be identical, hence casting an impression of the same image.
Figure 1.2 (a) Capturing object waves of an object on a recording medium. (b) Reconstructing a virtual object image from the recording medium with a coherent plane optical wave.
The magical recording media is known as a hologram, a word that finds its origin from the Greek words ὅλος (whole) and γραφή (drawing). Literally, the word hologram carries the meaning of something that is capable of recording the complete information of a visible image. Theoretically, there is little difference between looking at the virtual image of a hologram and the actual object. Hence, from a hologram, an observer should be able to see not only the color and intensity of the object, but also all the depth cues and disparity information. Depth cues are the visual information in an image that instigates the perception of distances between the observer and different parts of the object. For example, the lens of our eye will change its focus when objects at different distances are observed. Another example of depth cues is when the viewpoint of an observer is panned (shifted along a certain path): an object that is closer will appear to translate more than one that is farther away, which again creates the sensation of difference in depth between the objects. Disparity, sometimes referred as “binocular disparity,” is the deviation in the location of an object as observed by the left and the right eyes. Through the disparity, our brain is able to interpret the depth of different objects in the scene.
The art of capturing a hologram of an object is commonly referred to as optical holography. The method was invented in late 1940 by Dennis Gabor, who was awarded the Nobel Prize in Physics in 1971 for his contribution to the invention and development of the holographic method. Works of Gabor in optical holography were reported in [1], and later by Bragg in [2]. There have been numerous advancements in optical holography since, and a comprehensive description on the developments can be found in [3,4]. For completion, a brief outline of holography is given here.
The concept of holography can be illustrated from a more analytical point of view. Figure 1.3 shows the side view of an object scene containing a single object point. The intensity of the object point is white, represented with a value of unity. A recording medium is positioned at the hologram plane to record the light wave of the object scene. As the white object point is illuminated with a coherent plane wave (i.e., a wavefront that is homogeneous in magnitude and phase), it will behave as an isotropic point light source that scatters an object wave O0(x,y) onto the hologram. Extending this concept, for a single object point with intensity A , the object wave O(x,y) that is projected on the hologram is given by
Figure 1.3 Object wave projected by an object point of unit intensity on the recording medium.
In Eq. (1.2), the magnitude and phase of the object wave of a point source are denoted by |O0(x,y) | and θ0(x,y) , respectively. The term ϕ0 is the phase of the light illuminating the object point, which will be added to the phase component of the object wave.
The above principles of hologram formation can be easily applied to a scene with multiple object points. An arbitrary object surface can be considered as comprising many isotropic point sources, each emitting its own object wave when illuminated. The optical signal recorded on the hologram is the superposition of the optical wave emitted from each object point. The mixing of these waves is a process that is commonly known as optical interference:
where P is the total number of object points, and subscript p is the index of the object points. The object wave recorded on the hologram is referred as the interference pattern. For a coherent plane wave propagating along the axial direction (i.e., normal to the hologram plane), the phase ϕp of the illuminating wave is constant for all the object points, which can be ignored or included as a constant phase term in the computation of the object wave. If the illumination light beam is incoherent, the value of ϕp is random for different object points, and the object wave can no longer represent the light field of the source object.
For the time being, assume there are some means that enables the full object wave to be recorded on a medium, resulting in a hologram (denoted by the function H(x,y) in Figure 1.3). The signals on the hologram are complex-valued (with magnitude and phase components), and in the form of high-frequency fringe patterns. Different names have been adopted in various literatures to describe hologram signals, such as – but not limited to – fringe patterns, hologram fringes, diffractive waves, or holographic signals. These terms will be used throughout this book.
As shown in Figure 1.2(b), a coherent plane wave passing through a hologram is modulated, both in amplitude and phase to give the same object wave O(x,y) . From the reconstructed object wave, the observer will be able to see a realistic 3-D image of the object from the hologram. The process of recovering a 3-D image from the hologram is known as “reconstruction.” The use of a coherent optical plane wave is mandatory as it will not change the amplitude and phase of the object waves – that is, Ap and θp(x,y) of each object point in both the recording and the reconstruction processes.
1.2 Optical Recording in Practice
A conceptual optical setup for recording the hologram of an object scene is shown in Figure 1.4. The actual recording system is slightly more complicated. A thin laser beam is converted into a plane wave with the use of a beam expander (BX), and illuminates the specimen after passing through the semi-transparent beam splitter (BS). A laser beam is used because it is coherent, so the phase distribution of its wavefront is homogeneous. As explained in Section 1.1, it is necessary that the phase ϕp of the illumination beam on every point on the object surface is constant. The beam expander magnifies the cross-sectional area of the thin laser beam so that the illumination light will be wide enough to cover the entire object of interest. The object wave emitted by the object of interest is partially reflected by the beam splitter to the recording media. If both the amplitude and phase of the object wave can be captured, the recording medium, will become the hologram of the specimen. Unfortunately, this ideal scenario is difficult, if not impossible, to achieve in practice.
Figure 1.4 Conceptual optical setup in hologram recording.
The reason is that optical recording materials developed to date, commonly known as photographic films, are usually implemented with a layer of gelatin emulsion coated onto a plastic substrate. The emulsion comprises microscopic grains of silver halide crystal, each of which will be partially converted into silver upon exposure to light. Upon exposing the film to an optical image, the chemical property of each silver halide crystal will change proportionally to the intensity of light falling on it, forming an invisible or latent image. Subsequently, the film can be chemically developed into a visible image. Since the silver halide crystal is only responsive to the intensity of light, a photographic film is only capable of recording the intensity, but not the phase, of a light field. Instead of the full holographic information, the signal recorded on the photographic film is only constituted by the magnitude component of the object wave, as given by
Apparently, the signal recorded by the photographic film in Eq. (1.4) is different from the object wave, as it does not carry any information on the phase component. Due to the absence of the phase terms, an image of the specimen cannot be reconstructed from the signal OI(x,y) .
1.3 Photography
The classical art of photography can be interpreted as a special case of optical holography, although it was invented at a much earlier date when the art of holography was unknown to the world. As seen in Eq. (1.4), it is not possible to record the phase component of the object wave with a photographic film. Photography is based on capturing a small part of the object waves that carries the optical image of a specific viewpoint of the object scene. The concept is to place a small aperture between the object scene and the photographic film (a.k.a. a photograph), so that there is a one-to-one mapping between the light wave emitted from each object point and a single point on the recording plane. In another words, the light waves of different object points will not interfere with one another as they fall onto the recording media. Such an implementation is known as a “pinhole camera.” Figure 1.5(a) shows the imaging of a pair of point light sources, A and B, with a pinhole camera. The tiny pinhole only allows a light wave of a unique orientation for each object point to pass through, and impinges on an image plane at which a photographic film is placed. Restricted by the pinhole, each location on a photograph can only receive the object wave of a single object point. Mathematically, the magnitude of the object wave received by a photograph from a single object point at location (xp,yp) is expressed as
where T(xp,yp) denotes a one-to-one transformation on the coordinates (xp,yp) . From Eq. (1.5), the function OP(x,y) is simply a transformed image of the object scene, with the object point recorded at the position T(xp,yp) . On the image plane, the image of an object point is flipped horizontally and vertically, and relocated to a new position according to the location of the pinhole. Note that the object wave contains only the magnitude component (i.e., the intensity) of the light wave of a unique point source on the photograph. As a result, the phase component of the light wave is not involved in the recording process, and the object can be illuminated with incoherent light (e.g., ambient lighting from the sun and fluorescent lamps), whereby the phase of the illumination beam on each object point can be totally random and uncorrelated with each other.
The photograph can only capture an image of the object scene at a particular viewpoint that is constrained by the position of the pinhole. With a single view of the object scene, the recorded image cannot convey any depth cue or disparity information. To capture another viewpoint of the object scene, the position of the pinhole has to be changed. Figure 1.5(b) shows the imaging of the pair of object points, A and B, with the pinhole shifted to a different location. From these two figures, the images of the two object points are relocated to different positions on the recording plane.
1.4 Recording Setup in Optical Holography
Although holographic technology can be traced back to 1940, hologram recording was not practiced until the development of the laser in the 1960s, which is capable of generating a coherent light wave. As mentioned before, coherent illumination is mandatory in holographic recording. The first hologram was demonstrated in 1962 by the Soviet scientist Yuri Denisyuk [5], as well as by Emmett Leith and Juris Upatnieks of the University of Michigan in the USA [6]. Different to photography, optical holography records a 3-D – instead of 2-D – image of the object scene onto a photographic film by preserving both the amplitude and phase components of the object wave. Intuitively, this may sound a bit contradictory, as a photographic film is only responsive to the intensity of the light wave. In holographic recording, this problem is overcome by converting the complex-valued holographic signal into intensity information prior to recording on a photographic film. The conversion is achieved by adding the complex-valued object wave with an inclined reference plane wave, which is also derived from the coherent beam that illuminates the object. The result of adding the pair of waves is an interference pattern (which is also commonly referred to as a beat pattern), encapsulating both the magnitude and phase components of the object waves. A typical optical setup for recording an off-axis hologram is given in Figure 1.6.
Figure 1.6 Practical optical setup for recording an off-axis hologram.
Similar to Figure 1.4, the source of illumination is a monochromatic laser beam LR which is expanded into a plane wave via a beam expander. The expanded beam is divided into two separate paths after passing through a beam splitter. The first beam is reflected by the beam splitter BS0, propagates through the beam splitter BS1, and illuminates the specimen. The object wave O(x,y) from the specimen is reflected by BS1, and impinges on the hologram. The second part of the expanded beam is reflected to the hologram via mirrors M0 and M1, projecting the beam onto the hologram at an angle of incidence θR along the vertical direction, resulting in an inclined reference plane wave
where λ is the wavelength of the laser beam. Note that R(y) is a pure phase function. If θR=0, the reference wave is simply the plane wave that is used to illuminate the specimen. When the object wave and the reference wave meet on the surface of the photographic film, they combine to form an interference pattern. The magnitude of the interference pattern is then recorded on the photographic film, resulting in a hologram with the intensity given by
where C1=[|O(x,y)|2+|R(y)|2] , C2= O(x,y)R*(y) , and C3=O*(x,y)R(y).
From Eq. (1.7), the hologram H(x,y) is a real signal that can be recorded onto the photographic film. Subsequently, the object wave can be optically reconstructed from the hologram by illuminating it with a plane wave, as shown in Figure 1.7. The light wave emitted by the hologram is contributed by three different components, each corresponding to a term in Eq. (1.7).
The first term C1 is the sum of the intensities of the object and the reference beams. Due to the absence of the phase component, this part of the light wave does not carry useful information on the 3-D image of the specimen, and appears as a patch of highlighted region. The second term C2 is identical to the object wave apart from the constant phase term R*(y) , while the last term C3 is the conjugate of the object wave multiplied with the reference wave R(y) . Being different from C1 , the complex object wave of the specimen are encapsulated in both C2 and C3 .
From the optical wave of the illuminated hologram, three images can be observed. The first one is a focused, virtual image of the object scene that is reconstructed from the component C2 . This component of the optical wave contains the object wave O(x,y) in the hologram H(x,y) , with the addition of the phase term R*(y) . The virtual image can be viewed from an oblique angle, as shown in Figure 1.7. The second image is a blurred, defocused twin image corresponding to C3 , which carries O*(x,y) , the conjugate of the object wave, and the phase term R(y) . Similar to the virtual image, the twin image can be observed from another oblique angle. The third image, usually referred to as the zero-order beam, is constituted from the remaining term C1 , and can be observed from the axial direction. Apart from the virtual image, the other pair of images are unwanted content as they do not form a correct reconstruction of the object scene.
From Figure 1.7, it can be seen that mixing of the object wave O(x,y) with the inclined reference wave R(y) has imposed an angular separation θR between each consecutive pair of images that are reconstructed from the hologram. It can be inferred that if θR is small, the virtual image could be partially overlapping with, and hence contaminated by, the twin image and the zero-order beam. As such, the angular separation θR must be large enough that the virtual image can be separated from the unwanted content, and optically reconstructed as a standalone visible image. Even if this is achieved, there is still a good chance that all three components are visible to the observer, although they are not spatially overlapping with one another. This is one of the major disadvantages of an off-axis hologram.
Theoretically, the virtual image can be fully separated from the unwanted content by increasing the value of θR . In practice, due to the limited resolution of the photographic film, a ceiling on the angle θR is imposed. As explained previously, a photographic film comprises individual microscopic grains of photosensitive materials, each of which constitute a pixel on the film. The resolution of the photographic film, therefore, is limited by the size of the pixel. Assuming that all the pixels are square in shape and have identical size of δd×δd , an image recorded on it is effectively discretized with a 2-D sampling lattice with a sampling interval δd along the horizontal and vertical directions. According to sampling theory, this will impose an upper limit to the spatial frequency of the recorded signal.
To determine the maximum angle of incidence, consider a simple case where only the reference wave is present. For a given θR , the spatial frequency of the hologram fringe pattern is the minimum, as compared to the case when the object wave is present. Figure 1.8 shows the side view of a reference wave which is inclined at an angle θR along the vertical axis. As the signal on the hologram is discretized into pixels when it is recorded on the photographic film, the vertical axis is denoted by the variable n , where n is an integer representing the row indices of the pixels. The physical location of the kth row along the vertical axis is given by y=kδd . A pair of light beams of the reference wave is shown in Figure 1.8, impinging on the hologram at n=0 and n=k. The relative phase difference between these two beams is given by
Figure 1.8 An inclined reference wave.
where ω=2πδdλ−1sin(θR) rad/s. From Eq. (1.8) it can be visualized that the fringe pattern on the hologram is a sinusoidal wave with frequency ω . As the maximum frequency that can be represented in the discrete signal is π rad/s,
For small value of θR , this expression can be approximated as
In Eq. (1.10), a hologram with pixel size of δd can only record an incline plane wave with a maximum angle of incidence θmax=0.5λδd−1 . This implies that the maximum viewing angle of this hologram is also limited to the angle θmax .
A plot of the angle θmax versus the pixel size δd of the hologram is shown in Figure 1.9. It can be seen that θmax is inversely proportional to the pixel size of the hologram. Hence, the higher the resolution of the hologram, the larger will be the angular separation between the virtual image, the zero-order beam, and the twin image.
Figure 1.9 Plot of θmax versus pixel size of the hologram.
In optical holography, the hologram of a physical 3-D object is recorded in a photographic film. With the advancement of computing and electronics technologies, a digital hologram can be generated numerically from a computer graphic model, or captured from a physical object scene with a digital camera. In the following part of this chapter, the principles of operations of these two methods are described.
1.5 Computer-Generated Holography
Computer-generated holography (CGH) can be regarded as a computer simulation on the optical hologram-capturing processes depicted in Figures 1.4 and 1.6. Pioneering works on CGH can be traced back to the theoretical methods of J. Waters [7], and later production of the first binary computer-generated Fourier hologram by B.R. Brown and A.W. Lohmann [8–10]. Being different from optical holography, all entities in CGH, such as light, object, and hologram, are virtual and encapsulated as numbers or mathematical functions. For example, reference wave R(y) is represented by the expression in Eq. (1.6), and its associated set of values λ and θR . As shown in Figure 1.10, the object scene is a 3-D computer graphic (CG) model of one or more physical objects. In general, the CG model can be defined by a set of object points that are located at certain positions in 3-D object space. It is often assumed that all the object points are illuminated with a uniform plane wave (which may be inclined at an angle θR ), so that each object point will scatter a spherical light wave with intensity governed by its material properties. Such object points are sometimes referred to as “self-illuminating,” and if an inline plane wave with θR=0 is assumed, the illuminating light beam can be discarded in the simulation.
Figure 1.10 Computer-generated holography.
The set of self-illuminating points in the CG model is input into a computer, in which a software program is applied to compute a digital hologram of the CG model, and the result is stored in a data file. At this point, the data file is merely a set of numbers that does not provide a lot of clues about the object scene. As such, the data file is converted into a physical hologram by either printing it on a static medium (e.g., a photographic film) or displaying it on an electronic accessible device such as a spatial light modulator (SLM). An SLM is a device that is capable of spatially modulating the optical properties of light falling on or reflecting from it. Display chips such as liquid crystal (LC) and liquid crystal on silicon (LCoS) are typical examples of SLMs. In any case, the 3-D object scene will be reconstructed as a visible image that can be observed by our eyes when the hologram is illuminated with a coherent beam. The process of converting a 3-D CG model into a digital hologram is referred to as computer-generated holography.
The fundamental principles of CGH are introduced in this section. Euclidean space, terminologies, and parameters, which will be referenced throughout the book, are defined. Euclidean space, as shown in Figure 1.11, defines a 3-D virtual environment that houses the object scene and the hologram. There are two major parts in the environment: the object space and the hologram plane. The object space contains the 3-D object scene (i.e., the CG model) which is composed of one or more objects, while the hologram plane contains the digital hologram to be generated. Major parameters that are frequently used throughout the book are defined in Table 1.1, and outlined as follows. Locations in Euclidean space are addressed with discrete rectangular Cartesian coordinates, with m and n (both of which are integers) denoting the discrete lateral (i.e., horizontal and vertical) coordinates. The axial direction, or the depth, that is orthogonal to the m and the n axes, is represented by the variable z. The hologram is assumed to be square in shape, with an identical number of rows and columns equal to H_DIM. The sampling intervals of the hologram and the lateral plane of the object space are identical and denoted by δd . Individual points in a hologram or an object scene are sometimes referred to as pixels. The physical location (x,y) of a discrete coordinate at (m,n) is related by (x,y)=(mδd,nδd) . The wavelength of light is denoted by λ .