Light field
The light field is a vector function that describes the amount of light flowing in every direction through every point in space. The space of all possible light rays is given by the five-dimensional plenoptic function, and the magnitude of each ray is given by the radiance. Michael Faraday was the first to propose that light should be interpreted as a field, much like the magnetic fields on which he had been working for several years. The phrase light field was coined by Andrey Gershun in a classic paper on the radiometric properties of light in three-dimensional space.
The 5D plenoptic function
If the concept is restricted to geometric optics—i.e., to incoherent light and to objects larger than the wavelength of light—then the fundamental carrier of light is a ray. The measure for the amount of light traveling along a ray is radiance, denoted by L and measured in watts ' per steradian ' per meter squared . The steradian is a measure of solid angle, and meters squared are used here as a measure of cross-sectional area, as shown at right.The radiance along all such rays in a region of three-dimensional space illuminated by an unchanging arrangement of lights is called the plenoptic function. The plenoptic illumination function is an idealized function used in computer vision and computer graphics to express the image of a scene from any possible viewing position at any viewing angle at any point in time. It is never actually used in practice computationally, but is conceptually useful in understanding other concepts in vision and graphics. Since rays in space can be parameterized by three coordinates, x, y, and z and two angles θ and ϕ, as shown at left, it is a five-dimensional function, that is, a function over a five-dimensional manifold equivalent to the product of 3D Euclidean space and the 2-sphere.
Like Adelson, Gershun defined the light field at each point in space as a 5D function. However, he treated it as an infinite collection of vectors, one per direction impinging on the point, with lengths proportional to their radiances.
Integrating these vectors over any collection of lights, or over the entire sphere of directions, produces a single scalar value—the total irradiance at that point, and a resultant direction. The figure at right, reproduced from Gershun's paper, shows this calculation for the case of two light sources. In computer graphics, this vector-valued function of 3D space is called the vector irradiance field. The vector direction at each point in the field can be interpreted as the orientation one would face a flat surface placed at that point to most brightly illuminate it.
Higher dimensionality
One can consider time, wavelength, and polarization angle as additional variables, yielding higher-dimensional functions.The 4D light field
In a plenoptic function, if the region of interest contains a concave object, then light leaving one point on the object may travel only a short distance before being blocked by another point on the object. No practical device could measure the function in such a region.However, if we restrict ourselves to locations outside the convex hull of the object, i.e. in free space, then we can measure the plenoptic function by taking many photos using a digital camera. Moreover, in this case the function contains redundant information, because the radiance along a ray remains constant from point to point along its length, as shown at left. In fact, the redundant information is exactly one dimension, leaving us with a four-dimensional function. Parry Moon dubbed this function the photic field, while researchers in computer graphics call it the 4D light field or Lumigraph. Formally, the 4D light field is defined as radiance along rays in empty space.
The set of rays in a light field can be parameterized in a variety of ways, a few of which are shown below. Of these, the most common is the two-plane parameterization shown at right. While this parameterization cannot represent all rays, for example rays parallel to the two planes if the planes are parallel to each other, it has the advantage of relating closely to the analytic geometry of perspective imaging. Indeed, a simple way to think about a two-plane light field is as a collection of perspective images of the st plane, each taken from an observer position on the uv plane. A light field parameterized this way is sometimes called a light slab.
Sound analog
The analog of the 4D light field for sound is the sound field or wave field, as in wave field synthesis, and the corresponding parametrization is the Kirchhoff-Helmholtz integral, which states that, in the absence of obstacles, a sound field over time is given by the pressure on a plane. Thus this is two dimensions of information at any point in time, and over time a 3D field.This two-dimensionality, compared with the apparent four-dimensionality of light, is because light travels in rays, while by Huygens–Fresnel principle, a sound wave front can be modeled as spherical waves : light moves in a single direction, while sound simply expands in every direction. However, light travelling in non-vacuous media may scatter in a similar fashion, and the irreversibility or information lost in the scattering is discernible in the apparent loss of a system dimension.
Ways to create light fields
Light fields are a fundamental representation for light. As such, there are as many ways of creating light fields as there are computer programs capable of creating images or instruments capable of capturing them.In computer graphics, light fields are typically produced either by rendering a 3D model or by photographing a real scene. In either case, to produce a light field views must be obtained for a large collection of viewpoints. Depending on the parameterization employed, this collection will typically span some portion of a line, circle, plane, sphere, or other shape, although unstructured collections of viewpoints are also possible.
Devices for capturing light fields photographically may include a moving handheld camera or a robotically controlled camera, an arc of cameras, a dense array of cameras, handheld cameras, microscopes, or other optical system.
How many images should be in a light field? The largest known light field contains 24,000 1.3-megapixel images. At a deeper level, the answer depends on the application. For light field rendering, if you want to walk completely around an opaque object, then of course you need to photograph its back side. Less obviously, if you want to walk close to the object, and the object lies astride the st plane, then you need images taken at finely spaced positions on the uv plane, which is now behind you, and these images need to have high spatial resolution.
The number and arrangement of images in a light field, and the resolution of each image, are together called the "sampling" of the 4D light field. Analyses of light field sampling have been undertaken by many researchers; a good starting point is Chai. Also of interest is Durand for the effects of occlusion, Ramamoorthi for the effects of lighting and reflection, and Ng and Zwicker for applications to plenoptic cameras and 3D displays, respectively.
Applications
Computational imaging refers to any image formation method that involves a digital computer. Many of these methods operate at visible wavelengths, and many of those produce light fields. As a result, listing all applications of light fields would require surveying all uses of computational imaging in art, science, engineering, and medicine. In computer graphics, some selected applications are:- Illumination engineering: Gershun's reason for studying the light field was to derive the illumination patterns that would be observed on surfaces due to light sources of various shapes positioned above these surface. An example is shown at right. A more modern study is.
- Light field rendering: By extracting appropriate 2D slices from the 4D light field of a scene, one can produce novel views of the scene. Depending on the parameterization of the light field and slices, these views might be perspective, orthographic, crossed-slit, general linear cameras, multi-perspective, or another type of projection. Light field rendering is one form of image-based rendering.
- Synthetic aperture photography: By integrating an appropriate 4D subset of the samples in a light field, one can approximate the view that would be captured by a camera having a finite aperture. Such a view has a finite depth of field. By shearing or warping the light field before performing this integration, one can focus on different fronto-parallel or oblique planes in the scene. If a digital camera was able to capture the light field, its photographs would allow being refocused after they are taken.
- 3D display: By presenting a light field using technology that maps each sample to the appropriate ray in physical space, one obtains an autostereoscopic visual effect akin to viewing the original scene. Non-digital technologies for doing this include integral photography, parallax panoramagrams, and holography; digital technologies include placing an array of lenslets over a high-resolution display screen, or projecting the imagery onto an array of lenslets using an array of video projectors. If the latter is combined with an array of video cameras, one can capture and display a time-varying light field. This essentially constitutes a 3D television system.
- Brain imaging: Neural activity can be recorded optically by genetically encoding neurons with reversible fluorescent markers e.g. GCaMP that indicate the presence of calcium ions in real time. Since Light field microscopy captures full volume information in a single frame, it is possible to monitor neural activity in many individual neurons randomly distributed in a large volume at video framerate. A quantitative measurement of neural activity can even be done despite optical aberrations in brain tissue and without reconstructing a volume image, and be used to monitor activity in thousands of neurons in a behaving mammal.
Modern approaches to light field display explore co-designs of optical elements and compressive computation to achieve higher resolutions, increased contrast, wider fields of view, and other benefits.
- Glare reduction: Glare arises due to multiple scattering of light inside the camera's body and lens optics and reduces image contrast. While glare has been analyzed in 2D image space, it is useful to identify it as a 4D ray-space phenomenon. By statistically analyzing the ray-space inside a camera, one can classify and remove glare artifacts. In ray-space, glare behaves as high frequency noise and can be reduced by outlier rejection. Such analysis can be performed by capturing the light field inside the camera, but it results in the loss of spatial resolution. Uniform and non-uniform ray sampling could be used to reduce glare without significantly compromising image resolution.
Theory
- Adelson, E.H., Bergen, J.R.. , In Computation Models of Visual Processing, M. Landy and J.A. Movshon, eds., MIT Press, Cambridge, 1991, pp. 3–20.
- Arvo, J.. , Proc. ACM SIGGRAPH, ACM Press, pp. 335–342.
- Bolles, R.C., Baker, H. H., Marimont, D.H.. , International Journal of Computer Vision, Vol. 1, No. 1, 1987, Kluwer Academic Publishers, pp 7–55.
- Faraday, M., , Philosophical Magazine, S.3, Vol XXVIII, N188, May 1846.
- Gershun, A.. "The Light Field", Moscow, 1936. Translated by P. Moon and G. Timoshenko in Journal of Mathematics and Physics, Vol. XVIII, MIT, 1939, pp. 51–151.
- Gortler, S.J., Grzeszczuk, R., Szeliski, R., Cohen, M.. , Proc. ACM SIGGRAPH, ACM Press, pp. 43–54.
- Levoy, M., Hanrahan, P.. , Proc. ACM SIGGRAPH, ACM Press, pp. 31–42.
- Moon, P., Spencer, D.E.. The Photic Field, MIT Press.
- Wong, T.T., Fu, C.W., Heng, P.A., Leung C.S.. , IEEE Trans. Multimedia, Vol. 4, No. 3, pp. 361-371.
Analysis
- G. Wetzstein, I. Ihrke, W. Heidrich , International Journal of Computer Vision , Volume 101, Issue 2, pp. 384–400.
- Ramamoorthi, R., Mahajan, D., Belhumeur, P.. , ACM TOG.
- Zwicker, M., Matusik, W., Durand, F., Pfister, H.. , Eurographics Symposium on Rendering, 2006.
- Ng, R.. , Proc. ACM SIGGRAPH, ACM Press, pp. 735–744.
- Durand, F., Holzschuch, N., Soler, C., Chan, E., Sillion, F. X.. , Proc. ACM SIGGRAPH, ACM Press, pp. 1115–1126.
- Chai, J.-X., Tong, X., Chan, S.-C., Shum, H.. , Proc. ACM SIGGRAPH, ACM Press, pp. 307–318.
- Halle, M. , in SPIE Proc. Vol. #2176: Practical Holography VIII, S.A. Benton, ed., pp. 73–84.
- Yu, J., McMillan, L.. , Proc. ECCV 2004, Lecture Notes in Computer Science, pp. 14–27.
Light field cameras
- Marwah, K., Wetzstein, G., Bando, Y., Raskar, R.. , ACM Transactions on Graphics .
- Liang, C.K., Lin, T.H., Wong, B.Y., Liu, C., Chen, H. H.. , Proc. ACM SIGGRAPH.
- Veeraraghavan, A., Raskar, R., Agrawal, A., Mohan, A., Tumblin, J.. , Proc. ACM SIGGRAPH.
- Georgiev, T., Zheng, C., Nayar, S., Curless, B., Salesin, D., Intwala, C.. , Proc. EGSR 2006.
- Kanade, T., Saito, H., Vedula, S.. , Tech report CMU-RI-TR-98-34, December 1998.
- Levoy, M.. .
- Levoy, M., Ng, R., Adams, A., Footer, M., Horowitz, M.. , ACM Transactions on Graphics, Vol. 25, No. 3.
- Ng, R., Levoy, M., Brédif, M., Duval, G., Horowitz, M., Hanrahan, P.. , Stanford Tech Report CTSR 2005-02, April, 2005.
- Wilburn, B., Joshi, N., Vaish, V., Talvala, E., Antunez, E., Barth, A., Adams, A., Levoy, M., Horowitz, M.. , ACM Transactions on Graphics, Vol. 24, No. 3, pp. 765–776.
- Yang, J.C., Everett, M., Buehler, C., McMillan, L.. , Proc. Eurographics Rendering Workshop 2002.
-
Light field displays
- Wetzstein, G., Lanman, D., Hirsch, M., Raskar, R.. , ACM Transactions on Graphics
- Wetzstein, G., Lanman, D., Heidrich, W., Raskar, R.. , ACM Transactions on Graphics
- Lanman, D., Wetzstein, G., Hirsch, M., Heidrich, W., Raskar, R.. , ACM Transactions on Graphics
- Lanman, D., Hirsch, M. Kim, Y., Raskar, R.. , ACM Transactions on Graphics
- Matusik, W., Pfister, H.. , Proc. ACM SIGGRAPH, ACM Press.
- Javidi, B., Okano, F., eds.. , Springer-Verlag.
- Klug, M., Burnett, T., Fancello, A., Heath, A., Gardner, K., O'Connell, S., Newswanger, C.. , SID Symposium Digest of Technical Papers
- Fattal, D., Peng, Z., Tran, T., Vo, S., Fiorentino, M., Brug, J., Beausoleil, R.. , Nature 495, 348–351
Light field archives
-
Applications
- Grosenick, L., Anderson, T., Smith S. J. From Nano to Macro, 6th IEEE International Symposium on Biomedical Imaging. 1263–1266.
- Grosenick, L., Broxton, M., Kim, C. K., Liston, C., Poole, B., Yang, S., Andalman, A., Scharff, E., Cohen, N., Yizhar, O., Ramakrishnan, C., Ganguli, S., Suppes, P., Levoy, M., Deisseroth, K. bioRxiv 132688; doi: https://doi.org/10.1101/132688.
- Heide, F., Wetzstein, G., Raskar, R., Heidrich, W.
- Wetzstein, G., Raskar, R., Heidrich, W. , IEEE International Conference on Computational Photography
- Pérez, F., Marichal, J. G., Rodriguez, J.M.. , Proc. EUSIPCO
- Raskar, R., Agrawal, A., Wilson, C., Veeraraghavan, A.. , Proc. ACM SIGGRAPH.
- Talvala, E-V., Adams, A., Horowitz, M., Levoy, M.. , Proc. ACM SIGGRAPH.
- Halle, M., Benton, S., Klug, M., Underkoffler, J.. , SPIE Vol. 1461, Practical Holography V, S.A. Benton, ed., pp. 142–155.
- Zomet, A., Feldman, D., Peleg, S., Weinshall, D.. , IEEE Transactions on Pattern Analysis and Machine Intelligence , Vol. 25, No. 6, June 2003, pp. 741–754.
- Vaish, V., Garg, G., Talvala, E., Antunez, E., Wilburn, B., Horowitz, M., Levoy, M.. , Proc. Workshop on Advanced 3D Imaging for Safety and Security, in conjunction with CVPR 2005.
- Bedard, N., Shope, T., Hoberman, A., Haralam, M. A., Shaikh, N., Kovačević, J., Balram, N., Tošić, I.. . Biomedical optics express, 8, pp. 260–272.
- Karygianni, S., Martinello, M., Spinoulas, L., Frossard, P., Tosic, I.. "". IEEE International Conference on Image Processing
- Rademacher, P., Bishop, G.. , Proc. ACM SIGGRAPH, ACM Press.
- Isaksen, A., McMillan, L., Gortler, S.J.. , Proc. ACM SIGGRAPH, ACM Press, pp. 297–306.
- Buehler, C., Bosse, M., McMillan, L., Gortler, S., Cohen, M.. , Proc. ACM SIGGRAPH, ACM Press.
- Ashdown, I.. , Journal of the Illuminating Engineering Society, Vol. 22, No. 1, Winter, 1993, pp. 163–180.
- Chaves, J. , CRC Press
- Winston, R., Miñano, J.C., Benitez, P.G., Shatz, N., Bortz, J.C., , Academic Press
- Pégard, N. C., Liu H.Y., Antipa, N., Gerlock M., Adesnik, H., and Waller, L.. Compressive light-field microscopy for 3D neural activity recording. Optica 3, no. 5, pp. 517–524.
- Perez, C. C., Lauri, A., Symvoulidis, P., Cappetta, M., Erdmann, A., & Westmeyer, G. G.. Calcium neuroimaging in behaving zebrafish larvae using a turn-key light field camera. Journal of Biomedical Optics, 20, 096009-096009.
- León, K., Galvis, L., and Arguello, H.. Revista Facultad de Ingeniería Universidad de Antioquia 80, pp. 131.