A light-field camera captures both the intensity and the direction of incoming light1,2,3,4,5. This enables a user to refocus pictures and afterwards reconstruct information on the depth of field. Research on light-field imaging can be divided into two components: acquisition and rendering. Microlens arrays have been used for acquisition, but obtaining broadband achromatic images with no spherical aberration remains challenging. Here, we describe a metalens array made of gallium nitride (GaN) nanoantennas6 that can be used to capture light-field information and demonstrate a full-colour light-field camera devoid of chromatic aberration. The metalens array contains an array of 60 × 60 metalenses with diameters of 21.65 μm. The camera has a diffraction-limited resolution of 1.95 μm under white light illumination. The depth of every object in the scene can be reconstructed slice by slice from a series of rendered images with different depths of focus. Full-colour, achromatic light-field cameras could find applications in a variety of fields such as robotic vision, self-driving vehicles and virtual and augmented reality.

Access optionsAccess options

Rent or Buy article

Get time limited or full article access on ReadCube.


All prices are NET prices.

Code availability

The code used for analyses and figures is available from the corresponding author upon reasonable request.

Data availability

The data that support the plots within this paper and other findings of this study are available from the corresponding authors upon reasonable request.

Additional information

Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.


  1. 1.

    Lippmann, G. Epreuves reversibles donnant la sensation du relief. J. Phys. Theor. Appl. 7, 821–825 (1908).

  2. 2.

    Adelson, E. H. & Bergen, J. R. in Computational Models of Visual Processing (eds Landy, M. S. & Movshon, J. A.) 3–20 (MIT Press, Cambridge, 1991).

  3. 3.

    Adelson, E. H. & Wang, Y. A. Single lens stereo with a plenoptic camera. IEEE Trans. Pattern Anal. Mach. Intell. 14, 99–106 (1992).

  4. 4.

    Levoy, M. & Hanrahan, P. Light field rendering. In Proc. 23rd Annual Conference on Computer Graphics and Interactive Techniques (ed. Fujii, J.) 31–42 (ACM Press, 1996).

  5. 5.

    Wilburn, B. et al. High performance imaging using large camera arrays. ACM Trans. Graph. 24, 765–776 (2018).

  6. 6.

    Wang, S. M. et al. A broadband achromatic metalens in the visible. Nat. Nanotechnol. 13, 227–232 (2018).

  7. 7.

    Isaksen, A., McMillan, L. & Gortler, S. J. Dynamically reparameterized light fields. In Proc. 27th Annual Conference on Computer Graphics and Interactive Techniques (eds Brown, J. S. & Akeley, K.) 297–306 (ACM Press, 2000).

  8. 8.

    Levin, A., Fergus, R., Durand, F. & Freeman, W. T. Image and depth from a conventional camera with a coded aperture. ACM Trans. Graph. 26, 70 (2007).

  9. 9.

    Georgiev, T. G. & Lumsdaine, A. Focused plenoptic camera and rendering. J. Electron. Imaging 19, 021106 (2010).

  10. 10.

    Lumsdaine, A. & Georgiev, T. Full Resolution Lightfield Rendering (Adobe Systems, 2008).

  11. 11.

    Georgiev, T. & Lumsdaine, A. Reducing plenoptic camera artifacts. Comput. Graph. Forum 29, 1955–1968 (2010).

  12. 12.

    Zeller, N., Quint, F. & Stilla, U. Depth estimation and camera calibration of a focused plenoptic camera for visual odometry. ISPRS J. Photogramm. Remote Sens. 118, 83–100 (2016).

  13. 13.

    Lin, X., Wu, J. M., Zheng, G. A. & Dai, Q. H. Camera array based light field microscopy. Biomed. Opt. Express 6, 3179–3189 (2015).

  14. 14.

    Pegard, N. C. et al. Compressive light-field microscopy for 3D neural activity recording. Optica 3, 517–524 (2016).

  15. 15.

    Hallada, F. D., Franz, A. L. & Hawks, M. R. Fresnel zone plate light field spectral imaging. Opt. Eng. 56, 081811 (2017).

  16. 16.

    Liu, J. D. et al. Light field endoscopy and its parametric description. Opt. Lett. 42, 1804–1807 (2017).

  17. 17.

    Sahin, E., Katkovnik, V. & Gotchev, A. Super-resolution in a defocused plenoptic camera: a wave-optics-based approach. Opt. Lett. 41, 998–1001 (2016).

  18. 18.

    Wu, C. S., Ko, J. & Davis, C. C. Plenoptic mapping for imaging and retrieval of the complex field amplitude of a laser beam. Opt. Express 24, 29853–29872 (2016).

  19. 19.

    Bok, Y., Jeon, H.-G. & Kweon, I. S. Geometric calibration of micro-lens-based light field cameras using line features. ‎IEEE Trans. Pattern Anal. Mach. Intell. 39, 287–300 (2017).

  20. 20.

    Jin, X., Liu, L., Chen, Y. Q. & Dai, Q. H. Point spread function and depth-invariant focal sweep point spread function for plenoptic camera 2.0. Opt. Express 25, 9947–9962 (2017).

  21. 21.

    Zhu, W. M. et al. A flat lens with tunable phase gradient by using random access reconfigurable metamaterial. Adv. Mater. 27, 4739–4743 (2015).

  22. 22.

    Khorasaninejad, M. et al. Metalenses at visible wavelengths: diffraction-limited focusing and subwavelength resolution imaging. Science 352, 1190–1194 (2016).

  23. 23.

    Tseng, M. L. et al. Metalenses: advances and applications. ‎Adv. Opt. Mater. 6, 1800554 (2018).

  24. 24.

    Hsiao, H. H. et al. Integrated resonant unit of metasurfaces for broadband efficiency and phase manipulation. Adv. Opt. Mater. 6, 1800031 (2018).

  25. 25.

    Khorasaninejad, M., Chen, W. T., Oh, J. & Capasso, F. Super-dispersive off-axis meta-lenses for compact high resolution spectroscopy. Nano Lett. 16, 3732–3737 (2016).

  26. 26.

    Chen, B. H. et al. GaN metalens for pixel-level full-color routing at visible light. Nano Lett. 17, 6345–6352 (2017).

  27. 27.

    Lin, D., Brongersma, M. L., Kik, P. G. & Wetzstein, G. Light-field imaging using a gradient metasurface optical element. US patent 15/358,114 (2017).

  28. 28.

    Wang, S. M. et al. Broadband achromatic optical metasurface devices. Nat. Commun. 8, 187 (2017).

  29. 29.

    Goldys, E. M. et al. Analysis of the red optical emission in cubic GaN grown by molecular-beam epitaxy. Phys. Rev. B 60, 5464–5469 (1999).

Download references


The authors acknowledge financial support from the Ministry of Science and Technology, Taiwan (grant nos MOST-107-2112-M-001-042-MY3, MOST-107-2911-I-001-508, MOST-107-2911-I-001-510, MOST-107-2923-M-001-010-MY3) and Academia Sinica (grant nos AS-103-TP-A06, AS-TP-108-M12, AS-iMATE-108-41). The authors are also grateful for financial support from the National Key R&D Program of China (2017YFA0303700, 2016YFA0202103) and the National Natural Science Foundation of China (nos 11822406, 11834007, 11674167, 11621091, 11774164, 91850204). They are also grateful to the National Center for Theoretical Sciences, the NEMS Research Center of National Taiwan University, the National Center for High-Performance Computing, Taiwan, and the Research Center for Applied Sciences, Academia Sinica, Taiwan, for their support. T.L. is grateful for the support from Dengfeng Project B of Nanjing University.

Author information

Author notes

  1. These authors contributed equally: Ren Jie Lin, Vin-Cent Su, Shuming Wang, Mu Ku Chen.


  1. Department of Physics, National Taiwan University, Taipei, Taiwan

    • Ren Jie Lin
    • , Mu Ku Chen
    • , Tsung Lin Chung
    • , Yu Han Chen
    • , Hsin Yu Kuo
    • , Jia-Wern Chen
    • , Yi-Teng Huang
    •  & Din Ping Tsai
  2. Department of Electrical Engineering, National United University, Miaoli, Taiwan

    • Vin-Cent Su
  3. National Laboratory of Solid State Microstructures, School of Physics, College of Engineering and Applied Sciences, Nanjing University, Nanjing, China

    • Shuming Wang
    • , Ji Chen
    • , Tao Li
    • , Zhenlin Wang
    •  & Shining Zhu
  4. Key Laboratory of Intelligent Optical Sensing and Manipulation, Ministry of Education, Nanjing, China

    • Shuming Wang
    • , Ji Chen
    • , Tao Li
    •  & Shining Zhu
  5. Collaborative Innovation Center of Advanced Microstructures, Nanjing, China

    • Shuming Wang
    • , Ji Chen
    • , Tao Li
    • , Zhenlin Wang
    •  & Shining Zhu
  6. Department of Electrical Engineering and Graduate Institute of Electronics Engineering, National Taiwan University, Taipei, Taiwan

    • Jung-Hsi Wang
  7. Research Center for Applied Sciences, Academia Sinica, Taipei, Taiwan

    • Cheng Hung Chu
    • , Pin Chieh Wu
    •  & Din Ping Tsai
  8. College of Engineering, Chang Gung University, Taoyuan, Taiwan

    • Din Ping Tsai


  1. Search for Ren Jie Lin in:

  2. Search for Vin-Cent Su in:

  3. Search for Shuming Wang in:

  4. Search for Mu Ku Chen in:

  5. Search for Tsung Lin Chung in:

  6. Search for Yu Han Chen in:

  7. Search for Hsin Yu Kuo in:

  8. Search for Jia-Wern Chen in:

  9. Search for Ji Chen in:

  10. Search for Yi-Teng Huang in:

  11. Search for Jung-Hsi Wang in:

  12. Search for Cheng Hung Chu in:

  13. Search for Pin Chieh Wu in:

  14. Search for Tao Li in:

  15. Search for Zhenlin Wang in:

  16. Search for Shining Zhu in:

  17. Search for Din Ping Tsai in:


R.J.L., V.-C.S., S.W. and M.K.C. contributed equally to this work. R.J.L. and S.W. conceived the design, performed the numerical design, optical measurement and data analysis, and developed the algorithm. R.J.L., S.W., M.K.C. and C.H.C. co-wrote the manuscript. V.-C.S. and J.-W.C. performed the sample preparation. M.K.C., Y.-T.H., H.Y.K., T.L.C., Y.H.C. and C.H.C. built the optical system for measurement. J.C. and P.C.W. performed the numerical simulation and data analysis. J.-H.W. provided GaN film and performed the sample preparation. T.L., Z.W. and S.Z. organized the project, designed experiments, analysed the results and prepared the manuscript. D.P.T. organized the project, designed and developed the numerical design and optical measurement, analysed the results and prepared the manuscript. All authors discussed the results and commented on the manuscript.

Competing interests

The authors declare no competing interests.

Corresponding authors

Correspondence to Shining Zhu or Din Ping Tsai.

Supplementary information

  1. Supplementary information

    Supplementary Figs. 1–9, Supplementary Table 1

  2. Supplementary video

    A real-time depth map video of a scene

About this article

Publication history