A light-field camera captures both the intensity and the direction of incoming light1,2,3,4,5. This enables a user to refocus pictures and afterwards reconstruct information on the depth of field. Research on light-field imaging can be divided into two components: acquisition and rendering. Microlens arrays have been used for acquisition, but obtaining broadband achromatic images with no spherical aberration remains challenging. Here, we describe a metalens array made of gallium nitride (GaN) nanoantennas6 that can be used to capture light-field information and demonstrate a full-colour light-field camera devoid of chromatic aberration. The metalens array contains an array of 60 × 60 metalenses with diameters of 21.65 μm. The camera has a diffraction-limited resolution of 1.95 μm under white light illumination. The depth of every object in the scene can be reconstructed slice by slice from a series of rendered images with different depths of focus. Full-colour, achromatic light-field cameras could find applications in a variety of fields such as robotic vision, self-driving vehicles and virtual and augmented reality.
Access optionsAccess options
Subscribe to Journal
Get full journal access for 1 year
only $15.58 per issue
All prices are NET prices.
VAT will be added later in the checkout.
Rent or Buy article
Get time limited or full article access on ReadCube.
All prices are NET prices.
The code used for analyses and figures is available from the corresponding author upon reasonable request.
The data that support the plots within this paper and other findings of this study are available from the corresponding authors upon reasonable request.
Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Lippmann, G. Epreuves reversibles donnant la sensation du relief. J. Phys. Theor. Appl. 7, 821–825 (1908).
Adelson, E. H. & Bergen, J. R. in Computational Models of Visual Processing (eds Landy, M. S. & Movshon, J. A.) 3–20 (MIT Press, Cambridge, 1991).
Adelson, E. H. & Wang, Y. A. Single lens stereo with a plenoptic camera. IEEE Trans. Pattern Anal. Mach. Intell. 14, 99–106 (1992).
Levoy, M. & Hanrahan, P. Light field rendering. In Proc. 23rd Annual Conference on Computer Graphics and Interactive Techniques (ed. Fujii, J.) 31–42 (ACM Press, 1996).
Wilburn, B. et al. High performance imaging using large camera arrays. ACM Trans. Graph. 24, 765–776 (2018).
Wang, S. M. et al. A broadband achromatic metalens in the visible. Nat. Nanotechnol. 13, 227–232 (2018).
Isaksen, A., McMillan, L. & Gortler, S. J. Dynamically reparameterized light fields. In Proc. 27th Annual Conference on Computer Graphics and Interactive Techniques (eds Brown, J. S. & Akeley, K.) 297–306 (ACM Press, 2000).
Levin, A., Fergus, R., Durand, F. & Freeman, W. T. Image and depth from a conventional camera with a coded aperture. ACM Trans. Graph. 26, 70 (2007).
Georgiev, T. G. & Lumsdaine, A. Focused plenoptic camera and rendering. J. Electron. Imaging 19, 021106 (2010).
Lumsdaine, A. & Georgiev, T. Full Resolution Lightfield Rendering (Adobe Systems, 2008).
Georgiev, T. & Lumsdaine, A. Reducing plenoptic camera artifacts. Comput. Graph. Forum 29, 1955–1968 (2010).
Zeller, N., Quint, F. & Stilla, U. Depth estimation and camera calibration of a focused plenoptic camera for visual odometry. ISPRS J. Photogramm. Remote Sens. 118, 83–100 (2016).
Lin, X., Wu, J. M., Zheng, G. A. & Dai, Q. H. Camera array based light field microscopy. Biomed. Opt. Express 6, 3179–3189 (2015).
Pegard, N. C. et al. Compressive light-field microscopy for 3D neural activity recording. Optica 3, 517–524 (2016).
Hallada, F. D., Franz, A. L. & Hawks, M. R. Fresnel zone plate light field spectral imaging. Opt. Eng. 56, 081811 (2017).
Liu, J. D. et al. Light field endoscopy and its parametric description. Opt. Lett. 42, 1804–1807 (2017).
Sahin, E., Katkovnik, V. & Gotchev, A. Super-resolution in a defocused plenoptic camera: a wave-optics-based approach. Opt. Lett. 41, 998–1001 (2016).
Wu, C. S., Ko, J. & Davis, C. C. Plenoptic mapping for imaging and retrieval of the complex field amplitude of a laser beam. Opt. Express 24, 29853–29872 (2016).
Bok, Y., Jeon, H.-G. & Kweon, I. S. Geometric calibration of micro-lens-based light field cameras using line features. IEEE Trans. Pattern Anal. Mach. Intell. 39, 287–300 (2017).
Jin, X., Liu, L., Chen, Y. Q. & Dai, Q. H. Point spread function and depth-invariant focal sweep point spread function for plenoptic camera 2.0. Opt. Express 25, 9947–9962 (2017).
Zhu, W. M. et al. A flat lens with tunable phase gradient by using random access reconfigurable metamaterial. Adv. Mater. 27, 4739–4743 (2015).
Khorasaninejad, M. et al. Metalenses at visible wavelengths: diffraction-limited focusing and subwavelength resolution imaging. Science 352, 1190–1194 (2016).
Tseng, M. L. et al. Metalenses: advances and applications. Adv. Opt. Mater. 6, 1800554 (2018).
Hsiao, H. H. et al. Integrated resonant unit of metasurfaces for broadband efficiency and phase manipulation. Adv. Opt. Mater. 6, 1800031 (2018).
Khorasaninejad, M., Chen, W. T., Oh, J. & Capasso, F. Super-dispersive off-axis meta-lenses for compact high resolution spectroscopy. Nano Lett. 16, 3732–3737 (2016).
Chen, B. H. et al. GaN metalens for pixel-level full-color routing at visible light. Nano Lett. 17, 6345–6352 (2017).
Lin, D., Brongersma, M. L., Kik, P. G. & Wetzstein, G. Light-field imaging using a gradient metasurface optical element. US patent 15/358,114 (2017).
Wang, S. M. et al. Broadband achromatic optical metasurface devices. Nat. Commun. 8, 187 (2017).
Goldys, E. M. et al. Analysis of the red optical emission in cubic GaN grown by molecular-beam epitaxy. Phys. Rev. B 60, 5464–5469 (1999).
The authors acknowledge financial support from the Ministry of Science and Technology, Taiwan (grant nos MOST-107-2112-M-001-042-MY3, MOST-107-2911-I-001-508, MOST-107-2911-I-001-510, MOST-107-2923-M-001-010-MY3) and Academia Sinica (grant nos AS-103-TP-A06, AS-TP-108-M12, AS-iMATE-108-41). The authors are also grateful for financial support from the National Key R&D Program of China (2017YFA0303700, 2016YFA0202103) and the National Natural Science Foundation of China (nos 11822406, 11834007, 11674167, 11621091, 11774164, 91850204). They are also grateful to the National Center for Theoretical Sciences, the NEMS Research Center of National Taiwan University, the National Center for High-Performance Computing, Taiwan, and the Research Center for Applied Sciences, Academia Sinica, Taiwan, for their support. T.L. is grateful for the support from Dengfeng Project B of Nanjing University.