Abstract
We introduce an imaging system that can simultaneously record complete Mueller polarization responses for a set of wavelength channels in a single image capture. The division-of-focal-plane concept combines a multiplexed illumination scheme based on Fourier optics together with an integrated telescopic light-field imaging system. Polarization-resolved imaging is achieved using broadband nanostructured plasmonic polarizers as functional pinhole apertures. The recording of polarization and wavelength information on the image sensor is highly interpretable. We also develop a calibration approach based on a customized neural network architecture that can produce calibrated measurements in real-time. As a proof-of-concept demonstration, we use our calibrated system to accurately reconstruct a thin film thickness map from a four-inch wafer. We anticipate that our concept will have utility in metrology, machine vision, computational imaging, and optical computing platforms.
Similar content being viewed by others
Introduction
The measurement of the spectral and polarization response of an object given an incident light source is foundational for a broad range of sensing, imaging, and metrology applications. In medicine, spectral- and polarization-based imaging can be used to quantify mechanical stresses and chemical compositions within cells and tissues1. In agriculture and the food industry, hyperspectral imaging can inform the health of plants and freshness of food2. Infrared polarimetry and spectroscopy are the basis for many remote sensing and machine vision tasks, from identifying the chemical makeup of an object3,4 to differentiating between natural and man-made features within a scene5. For metrology methods such as ellipsometry, the polarization and spectral analysis of thin film structures is the industry standard for quantifying refractive indices and geometric structure parameters6,7,8.
For many applications, fast acquisition times are required due to the need for high throughput imaging or because the object properties are time-varying. To address this need, a variety of snapshot optical imaging system configurations have been developed that capture the spectral and polarization properties of objects with a single image sensor exposure9,10. Snapshot hyperspectral systems have been realized with the incorporation of dispersive elements11, computational imaging apertures12, and wavelength filters13 in the optical system. Snapshot imaging polarimetry has been accomplished by the use of schemes based on division-of-amplitude14, division-of-aperture15, and division-of-focal-plane16 architectures. Snapshot hyperspectral polarimetry systems that combine concepts from hyperspectral and polarimetry imaging17,18,19 have been developed to capture full polarization-spectrum hypercube information.
The most comprehensive linear optical characterization of an object involves hyperspectral Mueller-polarimetry imaging, where the full polarization response of an objective is recorded as a function of illumination polarization and wavelength. Such an imaging capability represents the gold standard in advanced metrology applications and is the basis for instruments such as imaging ellipsometers20,21,22,23. It remains a challenge to realize these systems with snapshot functionality due to the large number of wavelength and polarization parameters that need to be simultaneously measured. The systems that come closest are non-imaging hyperspectral Mueller polarimeters, which use amplitude modulation in the spectral domain to encode generated polarization states and either spectral24 or spatial25 modulation to encode polarization and spectral analysis information.
We report a snapshot Mueller spectropolarimeter imaging system that can record the Mueller matrix response of an object for visible wavelengths at 1200 spatial positions. It is configured to illuminate and analyze flat objects supporting specular interactions, making it ideally suited for metrology applications with flat substrates. The system as demonstrated images four-inch wafers and uses nanostructured plasmonic polarizers to perform the generation and analysis of polarization states with a horizontal, vertical, diagonal, and right-handed circular polarization (RCP) basis (Fig. 1a). The plasmonic polarizers comprise single- and bi-layer nanoscale aluminum structures that can be tailored to filter linear and circular polarized light with high selectivity and wide bandwidth (Fig. 1b). As a proof-of-concept demonstration, we use our calibrated apparatus to characterize silicon dioxide thin films on a silicon wafer and measure thin film thicknesses to be within 2% of ellipsometry values.
In addition to featuring the unprecedented capability of snapshot Mueller spectropolarimetry imaging, our system features other noteworthy distinctions. First, data readout is highly interpretable, and in a perfectly aligned system, the Mueller matrix for a given wavelength and spatial position can be directly read off of the image sensor with no need for data reconstruction. Even in an imperfectly aligned system, the data remains sufficiently interpretable to enable data denoising and calibration to be performed using only matrix multiplication operations at high speeds. Second, wavelength and polarization information are independently recorded, and unlike methods that encode polarization information using spectral amplitude modulation, no assumptions about the spectral dispersion properties of the sample are required. Third, polarization control and analysis are enabled by nanostructured polarizers fabricated on a chip, enabling multi-aperture arrays capable of supporting customized polarization responses and operating wavelength ranges. Fourth, the light field camera system with analyzing polarization filters can itself serve as a compact hyperspectral polarimeter imaging system.
Results
System architecture
A schematic detailing the optical system architecture is shown in Fig. 1c. The schematic is illustrated for a transmissive sample for clarity but readily applies to reflective samples without loss of generality. Conceptually, our imaging Mueller spectropolarimeter comprises a superposition of three optical systems that work synergistically to enable the extreme multiplexing capabilities of our system: an illumination system based on an array of pinhole apertures coupled to a symmetric 4f system (L1 and L2 are identical), a telescopic imaging system for large area wafer imaging, and a light field camera for capturing multiple images of the wafer with different polarization responses. The sensor pixels under individual microlens, which we term a super-pixel, record the wavelength and polarization response from a specific point of the sample (Fig. 1c, right). A calibration algorithm, based on a customized neural network architecture trained with calibration data, accounts for misalignments and component imperfections. Optical component metrics used in the experimental system are listed in Table S1 and labeled in Fig. S1.
The 4f system for illumination maps fields from the front focal plane of the first lens (i.e., the illumination plane) to the back focal plane of the second lens (i.e., the analyzer plane). We consider an array of 16 pinhole apertures at the illumination plane, each of which serves as a distinct illumination source. Placement of broadband polarization filters with linear- or circular-polarized responses at each pinhole leads to a set of 16 polarized illumination sources. At the analyzer plane, a second set of 16 pinhole apertures, each containing unique broadband polarization filters, is placed such that fields from one pinhole at the illumination plane register to one pinhole at the analyzer plane. The light transmitted through an individual analyzer pinhole, which we term a polarization channel, therefore corresponds to a light source paired with a unique combination of illumination and analyzer polarization filters.
The sample is placed at the Fourier plane of the 4f system and is imaged using a telescopic system, which enables the imaging of large sample areas. In this scheme, the telescope objective is the second lens of the 4f system (L2) and the eyepiece is L3 and placed within the light field imager. Demagnification of the sample image at the sensor plane is set to f2/f3, where f2 is the focal distance of L2 and f3 is the focal distance of L3. At the sample position, light from each pinhole at the illumination plane collimates to a plane wave spanning the sample width with a distinctive incidence angle that corresponds to the pinhole position (Fig. S2). In this manner, the illumination sources at the illumination plane interact with the sample in an independent and angle-multiplexed manner.
Finally, light transmitting through the analyzer pinhole apertures is recorded using a modified light field camera, which incorporates a microlens array one microlens focal distance away from the sensor plane26,27. The microlenses serve as miniaturized imaging systems that map light field information from the aperture, which in our case is the set of analyzer pinholes, to unique pixels in the sensor array. They also serve to concentrate light spanning the microlens cross-sectional area and enhance the measured signal-to-noise by a factor of N, where N is the number of sensor pixels under each microlens. A diffraction grating placed just above the microlenses disperses light to the +1 diffraction order to enable hyperspectral imaging for each polarization channel. The light from higher orders carries much less energy, and its effect can be removed using our calibration algorithm. Our spectropolarimetry system therefore is a division-of-focal-plane system that leverages the synergistic coupling of sample illumination in the Fourier plane, in which polarization channels are multiplexed by wavevector angle, with light field imaging, in which waves with different wavevector angles are independently imaged. More details pertaining to the design and operation of the light field imager are in the Supplementary Section.
To optimize the spectral resolution and bandwidth of the system while maintaining minimal crosstalk between super-pixels and polarization channels, we derive basic design guidelines. These guidelines assume that the optical components operate in the scalar diffraction limit and that the diffraction grating transmits light to only the 0 and +1 diffraction orders. First, the microlens diameter (dm) is specified based on the desired spatial resolution of the system, which is set to dmf2/f3. The microlens focal length (fm) is then specified to minimize the focused spot size onto the image sensor. Spot size is determined by two factors, the diffraction limit and off-axis aberrations. Using diffraction theory to determine the diffraction-limited spot size and ray simulations to quantify spot sizes for off-axis illumination, optimal combinations of dm and fm that minimize spot size can be computed (Fig. S5d).
To maximize spectral bandwidth, the spatial dimensions of exposed image sensor pixels within an individual super-pixel should span dm, which maximizes the super-pixel area without introducing crosstalk between super-pixels. Given λmax, the longest wavelength processed in the system, and nm, the microlens refractive index, the grating pitch (Λ) is fixed using basic relations from diffraction theory and is λmaxfm/dmnm. With set grating and microlens parameters, the spectral resolution (δλ) at a given wavelength (λ) is δλ = Λλ/dm. This expression assumes the sensor pixel dimensions match the diffraction-limited microlens focusing spot size.
With the microlens and grating parameters determined, the pinhole analyzer aperture sizes can be specified. For light passing through a finite-sized aperture, L3 will produce a beam that increases in divergence with increased aperture size. This divergence will lead to an angular distribution of plane waves incident onto the grating and microlens, leading to blurring and reduced spectral imaging resolution. To minimize this effect, we specify the maximum pinhole aperture size such that beam divergence from L3 produces a blurring length scale at the imaging sensor that is less than the diffraction-limited spot size: da = λf3/dm. The spacing between the apertures should be sufficiently large to minimize crosstalk between neighboring polarization channels on the sensor, but sufficiently small so as to not waste pixels on the sensor. We separate neighboring polarization channels by one full row of sensor pixels in a diffraction-limited system by specifying the center-to-center spacing between apertures to be: 2λf3/dm.
Experimental implementation
The experimentally implemented system is shown in Fig. 2a and records a 40 × 30 super-pixel image of the sample. Each super-pixel corresponds to the integrated optical response of a 3.8 mm diameter area on the sample and includes the full Mueller response for 11 wavelength channels for wavelengths ranging from 550 to 800 nm. A 10-W Tungsten halogen white light source is used for illumination and the exposure time is 17 ms and is limited by the camera frame rate. Our ability to utilize such fast exposure times in spite of our use of pinhole apertures is due to the signal-to-noise enhancement from light concentration mediated by the microlenses. Telescope objective lenses are used to achieve illumination beam widths suitable for wafer-scale imaging. The angle of incidence on the sample is 45 degrees. We note that the system configuration is specified in large part to the availability of specific components and is not fully optimized using the design framework above.
A detailed view of the light field imager, which comprises the analyzer pinhole apertures (A2), collimating lens (L3), and functionalized image sensor, is presented in Fig. 2b. Modifications to a commercial image sensor are performed by first removing the protective cover glass over the sensor array followed by mounting of the microlens array and diffraction grating. The pinhole aperture array comprises four sets of four pinholes each for a total of 16 pinholes. Each pinhole is 200 microns in diameter, and each pinhole set contains polarization filters with broadband horizontal, vertical, diagonal, and RCP filtering functionality.
The broadband polarizers consist of nanoplasmonic aluminum structures featuring subwavelength-scale nanoridges. Metallic nanoridge arrays are well known to support the selective filtering of linearly polarized light28,29,30, and they can also be configured to serve as broadband quarter waveplates31. The horizontal, vertical, and diagonal polarizers consist of a single-layer linear polarizer structure rotated at different orientations. To enhance transmission and bandwidth, the polarizers are clad in silicon dioxide but contain air gaps between the ridges. The RCP filter comprises a quarter waveplate nanoridge device vertically stacked with a linear polarizer, separated by a dielectric layer (Fig. 2b). We consider plasmonic polarization optics, as opposed to those based on dielectrics32, due to ease of fabrication and their ability to support high polarization selectivity over relatively wide bandwidths and incidence angles.
Device fabrication is performed by first depositing an aluminum thin film onto a glass substrate, followed by electron beam lithography patterning and reactive ion etching. As the aperture dimensions are microscale, fabrication of the full set of analyzer filters involves a total patterned area of less than a square millimeter. The multi-layer RCP filter is created by first fabricating the quarter wave plate, planarizing the device with spin-on-glass, and then fabricating the linear polarizer. A cross-sectional image of a representative RCP filter combining the linear polarizer and quarter waveplate, without the top cladding layer, is in the inset of Fig. 2 and shows vertically stacked periodically spaced nanoscale aluminum ridges with vertical side walls. The experimental transmittance and polarization selectivity of the linear (Fig. S8g) and RCP (Fig. 2) filters at normal incident angles indicate high transmittance and selectivity across broadband. Deviations between experimental and theoretical performance are attributed to experimental variations in metal quality, sidewall roughness, and thin film thickness. The filters for the polarization state generator are fabricated in the same manner as those for the imaging aperture. Additional details pertaining to system assembly, filter design, and filter fabrication are provided in the Supplementary Section.
A representative raw spectropolarimetry image recorded on the sensor shows a two-dimensional array of super-pixels (Fig. 3), with each super-pixel corresponding to light processed from individual microlens. The image as shown is rotated by 90° relative to prior schematic illustrations for clarity. An individual super-pixel comprises sixteen columns that each contain information about a specific incidence and analyzed polarization state. The top row within the super-pixel corresponds to zeroth-order diffracted light while the lower part of the image contains 11 rows from first-order diffracted light spanning 550 to 800 nm. An individual row from the lower part of the image produces our measurement matrix for a given wavelength bin within our experimental polarization basis. These measurements can be readily translated to Mueller matrices using transformations discussed in the Supplementary Section.
System calibration
In an ideally implemented system, the measurement matrix can be directly read off and used to accurately compute Mueller matrices. However, in an experimental system, misalignments between system components and physical imperfections within the components themselves lead to noise in the super-pixel images. The wave nature of light also prevents the ideal readoff of independent wavelength and polarization information from an individual sensor pixel, as the center of the Airy disk from a focused beam contains 84% of the total power and the remainder of the power resides in concentric rings around the disk. Collectively, these factors produce systematic nonidealities that manifest as reduced signal-to-noise and crosstalk within and between polarization and wavelength channels (Fig. 4a). Additionally, misalignment between the microlens focal spots and the discrete sensor grid (i.e., sub-pixel misalignment) can lead to detection sampling errors. Due to the complexity of this error and its combination of systematic and random sources, conventional approaches to denoising, such as the fitting of a global point spread function, are not effective (Fig. S14).
To correct these noise sources, we consider an approach where a given sensor pixel within a super-pixel, which targets a specific wavelength band and polarization measurement state in an ideally aligned system, is corrected using a tailored denoising kernel, K. Given P, a pixel sensor image centered at the pixel of interest with the same dimensionality as K, the calibrated measurement value is M = K · P. This approach assumes that nonidealities in our experimental system produce local variations in the point spread function on the image sensor. It also builds on our observation that the optical system is linear and the true intensity values at a given pixel must be a linear combination of intensity values from surrounding pixels within the point spread function.
In a naive denoising approach, denoising kernels can be directly learned for every wavelength band and polarization measurement state within a super-pixel (11 × 16 = 176 kernels) for every super-pixel (30 × 40 super-pixels), using calibration data taken from measurements from a known reference wafer. Such a direct learning approach with limited training data is not effective and will lead to significant overfitting. We instead take a multi-step approach in which we first consolidate data from all super-pixels and use a deep network to learn first generation 5 × 5 kernels with limited accuracy, discussed in more detail below, which overcomes the overfitting issue and enables the learning of correlated noise between super-pixels (Fig 4b). Second-generation kernels, also with dimensions of 5 × 5, are tailored to of each wavelength band and polarization state of each super-pixel for a total of 11 × 16 × 1200 kernels. These kernels are computed by performing local gradient-based optimization on first generation kernels using the original calibration data, and intensity information from the first generation kernel learning process is incorporated to correct illumination heterogeneity in the system. The third generation kernels, which are the final kernels used for calibration, are computed by padding the second generation kernels to 7 × 7 dimensions and performing local gradient-based optimization with the original calibration data.
To produce first generation kernels, we utilize a customized neural network architecture in which the input is an individual 64 × 64 super-pixel image and the output is a set of 16 denoising kernels that are tailored for each polarization channel. This concept builds on the observation that many sources of noise, such as component misalignment, produce systematic errors in the recorded image that lead to shared noise characteristics between different super-pixels. The neural network architecture (Fig. S11a, b, d) bears unique properties that differ from typical convolutional neural networks used in classical computer vision tasks33,34,35. First, the network is specified to be relatively shallow to minimize overfitting and provide better stability with the use of a limited training dataset. Second, the network explicitly accounts for sub-pixel misalignments by upscaling the input image by 8× with bilinear sampling. Third, the network explicitly incorporates translation symmetry such that translational shifts of whole- and sub-pixel shifts within the input image lead to corresponding shifts in the outputted kernels. For the training process, a total of 42 images were taken of an aluminum reflection mirror with different combinations of polarizers and spectral filters placed at the light source and analyzers. 232 randomly selected super-pixels from each image were selected to formulate the training set, yielding a total of 42 × 232 = 9744 super-pixels for the training set. A representative training loss curve of first generation kernels is shown in Fig. S11e and indicates good convergence over the course of network training. Loss curves for second and third generation kernels averaged over all super-pixels are shown in Fig. 4c and also indicates good convergence.
Calibrated measurement
To demonstrate the capabilities of our system, we perform Mueller spectropolarimetry imaging on the 4” wafer shown in Fig. 1a. Image calibration involves dot product multiplication of third-generation kernels with the super-pixel images and takes 200 ms using standard CPU hardware (Intel core i7-8700K). As a point of comparison, the direct use of a neural network to perform super-pixel calibration over a wafer scale would take 15 min with the same hardware. Faster and parallelized matrix computations can be achieved using graphics processing unit (GPU) hardware. The calibrated measurement matrices at all points on the wafer are visualized in Fig. 5a, where the spectral response at a given super-pixel is encoded by a standard color vision response. The smooth and consistent color within each silicon dioxide band on the wafer, together with polarization features such as the lack of signal for cross-polarized input and analyzed polarization states for horizontal and vertical polarization, indicate the efficacy of our calibration method.
The measurement matrices can be used to compute the corresponding Mueller matrices, shown in Fig. 5b, using methods discussed in the Supplementary Section. The absolute error maps for each Mueller element in each wavelength band and super-pixel, compared to Mueller elements computed from film values and indices measured at individual regions on the wafer using ellipsometry (Table 1), are presented in Fig. S15. A map of the absolute error from the Mueller matrix elements in each super-pixel, averaged over all matrix elements for all wavelengths, is shown in Fig. 5d. These results display consistently low error with an overall average of 8%. The presence of speckle patterns suggests random super-pixel-specific manufacturing defects in the lens array or grating that cannot be corrected with our calibration method, and their elimination reduces the total average error to 6%.
These experimental data are used to calculate the thickness of silica at each super-pixel by minimizing the absolute error between the computed experimental Mueller matrices and those corresponding to a thin film fitting model. The resulting best-fit thin film model, visualized as Mueller matrix elements, is shown in Fig. 5c and agrees well with the experimental results. The resulting thin film thickness map for the wafer is presented in Fig. 5e and shows clear uniformity within each band of silicon dioxide. Deviations of the thicknesses are computed with respect to pixels that belong to the region in the thickness map. The averaged thicknesses within each silicon dioxide layer are, on average, within 5 nm of ground truth values (Table 1). We note that deviations within the thickness map arise at boundaries between silicon dioxide layers and along the edges of the wafer, due to the limited spatial resolution of our system. The uncalibrated thickness map (Fig. 5f) is inaccurate and indicates the need for our calibration method.
Discussion
In summary, we present an optical imaging system architecture that is capable of snapshot Mueller spectropolarimetry imaging for flat objects supporting specular interactions. Light from multiple illumination sources is simultaneously imaged by coupling a Fourier optic illumination scheme with a telescopic light field image sensor. Arrays of microscale nanoplasmonic polarization filters at the illumination and analyzer planes enable field-multiplexed imaging of unique polarization responses from the object. The design rules for the system are simple and straightforward, and the combination of a customized calibration neural network architecture together with an interpretable data readout scheme ensures accurate data capture. A proof-of-concept demonstration with a 4” calibration wafer indicates the potential of the system to accurately perform metrology tasks.
There exist multiple immediate opportunities for improving system performance. As discussed, the specifications used in our study are far from optimal, and further component customization can enable the maximization of spectral and spatial system bandwidth. The use of additive manufacturing to directly print a customized microlens array onto the image sensor facet can lead to ideal microlens alignment with the sensor, fully tailorable microlens parameters, and integration of the diffraction grating directly onto the microlens surface. Improvements in experimental filter efficiency can be achieved using optimized fabrication processes, device schemes that feature relatively low metal fill fraction, and device schemes that utilize lower loss metal materials such as silver31. Filter device fabrication can be scaled up using a variety of patterning techniques that exceed the throughput of electron beam lithography, including nanoimprint, laser interference, and deep UV lithography. As a metrology tool, incorporating a goniometer stage with our system can enable measurements at different light incidence angles, producing more data for more accurate analysis. Longer term, we envision that our concept can extend to the analysis of microscopic domains through the use of microscope objectives in the 4f system and that it can be implemented in other imaging and optical data processing modalities through the utilization of metasurface apertures with more customized optical responses36,37,38,39,40,41. With proper co-design of aperture responses with software, our imaging system can be tailored for tasks as diverse as optical computing and data compression, and it can combine with concepts in computational imaging to enable enhanced imaging capabilities42.
Materials and methods
Methods pertaining to optical system construction, nanopolarizer fabrication, neural network architecture and training for calibration, and additional experimental results are provided in a separate Supplementary Section file.
References
Ozeki, Y. et al. High-speed molecular spectral imaging of tissue with stimulated Raman scattering. Nat. Photonics 6, 844–850 (2012).
Honkavaara, E. et al. Processing and assessment of spectrometric, stereoscopic imagery collected using a lightweight UAV spectral camera for precision agriculture. Remote Sens. 5, 5006–5039 (2013).
Opilik, L., Schmid, T. & Zenobi, R. Modern Raman imaging: vibrational spectroscopy on the micrometer and nanometer scales. Annu. Rev. Anal. Chem. 6, 379–398 (2013).
Kalasinsky, K. S. et al. Raman chemical imaging spectroscopy reagentless detection and identification of pathogens: signature development and evaluation. Anal. Chem. 79, 2658–2673 (2007).
Ainsworth, T. L., Schuler, D. L. & Lee, J.-S. Polarimetric sar characterization of man-made structures in urban areas using normalized circular-pol correlation coefficients. Remote Sens. Environ. 112, 2876–2885 (2008).
Schubert, M. 9—Theory and application of generalized ellipsometry. In Handbook of Ellipsometry (eds Tompkins, H. G. & Irene, E. A.) 637–717 (William Andrew Publishing, Norwich, NY, 2005).
Yoon, C. et al. Toward realization of high-throughput hyperspectral imaging technique for semiconductor device metrology. J. Micro/Nanopatterning Mater. Metrol. 21, 021209 (2022).
Kwon, S., Park, J., Kim, K., Cho, Y. & Lee, M. Microsphere-assisted, nanospot, non-destructive metrology for semiconductor devices. Light: Sci. Appl. 11, 32 (2022).
Tyo, J. S., Goldstein, D. L., Chenault, D. B. & Shaw, J. A. Review of passive imaging polarimetry for remote sensing applications. Appl. Opt. 45, 5453–5469 (2006).
Qi, J. & Elson, D. S. Mueller polarimetric imaging for surgical and diagnostic applications: a review. J. Biophotonics 10, 950–982 (2017).
Baek, S.-H., Kim, I., Gutierrez, D. & Kim, M. H. Compact single-shot hyperspectral imaging using a prism. ACM Trans. Graph. 36, 1–12 (2017).
Monakhova, K., Yanny, K., Aggarwal, N. & Waller, L. Spectral diffusercam: lensless snapshot hyperspectral imaging with a spectral filter array. Optica 7, 1298–1307 (2020).
Lambrechts, A. et al. A CMOS-compatible, integrated approach to hyper and multispectral imaging. In 2014 IEEE International Electron Devices Meeting 10.5.1–10.5.4 (IEEE, 2014).
Mudge, J., Virgen, M., Dean, P. Near-infrared simultaneous Stokes imaging polarimeter. In Polarization Science and Remote Sensing IV Vol. 7461 (eds Shaw, J. A. & Scott Tyo, J.) 74610L (International Society for Optics and Photonics, SPIE, 2009).
Mu, T. et al. Optimized design, calibration, and validation of an achromatic snapshot full-stokes imaging polarimeter. Opt. Express 27, 23009–23028 (2019).
Garcia, M., Davis, T., Blair, S., Cui, N. & Gruev, V. Bioinspired polarization imager with high dynamic range. Optica 5, 1240–1246 (2018).
Meng, X., Li, J., Xu, T., Liu, D. & Zhu, R. High throughput full Stokes Fourier transform imaging spectropolarimetry. Opt. Express 21, 32071–32085 (2013).
Quan, N. et al. Linear stokes imaging spectropolarimeter based on the static polarization interference imaging spectrometer. Opt. Commun. 391, 30–36 (2017).
Mu, T. et al. Snapshot hyperspectral imaging polarimetry with full spectropolarimetric resolution. Opt. Lasers Eng. 148, 1–12 (2022).
Jin, G., Jansson, R. & Arwin, H. Imaging ellipsometry revisited: developments for visualization of thin transparent layers on silicon substrates. Rev. Sci. Instrum. 67, 2930–2936 (1996).
Liu, S., Du, W., Chen, X., Jiang, H. & Zhang, C. Mueller matrix imaging ellipsometry for nanostructure metrology. Opt. Express 23, 17316–17329 (2015).
Losurdo, M. et al. Spectroscopic ellipsometry and polarimetry for materials and systems analysis at the nanometer scale: state-of-the-art, potential, and perspectives. J. Nanopart. Res. 11, 1521–1554 (2009).
Garcia-Caurel, E., De Martino, A. & Dr´evillon, B. Spectroscopic Mueller polarimeter based on liquid crystal devices. Thin Solid Films 455-456, 120–123 (2004).
Hagen, N., Oka, K. & Dereniak, E. L. Snapshot Mueller matrix spectropolarimeter. Opt. Lett. 32, 2100–2102 (2007).
Oka, K., Haga, Y. & Michida, H. Snapshot Mueller—matrix spectropolarimeter using spectral and spatial carriers. In Polarization Science and Remote Sensing VII Vol. 9613 (eds Shaw, J. A. & Daniel A. LeMaster, D. A.) 96130E (International Society for Optics and Photonics, SPIE, 2015).
Ng, R. et al. Light Field Photography with a Hand-held Plenoptic Camera. Research Report CSTR 2005-02 (Stanford University, 2005).
Mart´ınez-Corral, M. & Javidi, B. Fundamentals of 3d imaging and displays: a tutorial on integral imaging, light-field, and plenoptic systems. Adv. Opt. Photon. 10, 512–566 (2018).
Kulkarni, M. & Gruev, V. Integrated spectral-polarization imaging sensor with aluminum nanowire polarization filters. Opt. Express 20, 22997–23012 (2012).
Inoue, D. et al. Polarization filters for visible light consisting of subwavelength slits in an aluminum film. J. Lightwave Technol. 30, 3463–3467 (2012).
Zheng, B. Y., Wang, Y., Nordlander, P. & Halas, N. J. Color-selective and cmos-compatible photodetection based on aluminum plasmonics. Adv. Mater. 26, 6318–6323 (2014).
Nyman, M., Maurya, S., Kaivola, M. & Shevchenko, A. Optical wave retarder based on metal-nanostripe metamaterial. Opt. Lett. 44, 3102–3105 (2019).
Basiri, A. et al. Nature-inspired chiral metasurfaces for circular polarization detection and full-stokes polarimetric measurements. Light: Sci. Appl. 8, 78 (2019).
Lecun, Y., Bottou, L., Bengio, Y. & Haffner, P. Gradient based learning applied to document recognition. Proc. IEEE 86, 2278–2324 (1998).
Long, J., Shelhamer, E. & Darrell, T. Fully convolutional networks for semantic segmentation. In 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 3431–3440 (IEEE, 2015).
He, K., Zhang, X., Ren, S. & Sun, J. Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 770–778 (IEEE, 2016).
Silva, A. et al. Performing mathematical operations with metamaterials. Science 343, 160–163 (2014).
Zhou, Y., Zheng, H., Kravchenko, I. I. & Valentine, J. Flat optics for image differentiation. Nat. Photonics 14, 316–323 (2020).
Sell, D., Yang, J., Doshay, S., Yang, R. & Fan, J. A. Large-angle, multifunctional metagratings based on freeform multimode geometries. Nano Lett. 17, 3752–3757 (2017).
Sell, D., Yang, J., Doshay, S. & Fan, J. A. Periodic dielectric metasurfaces with high-efficiency, multiwavelength functionalities. Adv. Opt. Mater. 5, 1700645 (2017).
Sell, D. et al. Ultra-high-efficiency anomalous refraction with dielectric metasurfaces. ACS Photonics 5, 2402–2407 (2018).
Chen, M., Jiang, J. & Fan, J. A. Algorithm-driven paradigms for freeform optical engineering. ACS Photonics 9, 2860–2871 (2022).
Wetzstein, G. et al. Inference in artificial intelligence with deep optics and photonics. Nature 588, 39–47 (2020).
Acknowledgements
This work was supported by the Samsung Global Outreach Program and the Office of Naval Research under Award Number N00014-16-1-2630. EW was supported by the Stanford Graduate Fellowship. Fabrication was performed in part at the Stanford Nanofabrication Facility (SNF) and the Stanford Nano Shared Facilities (SNSF), which are supported by the National Science Foundation as part of the National Nanotechnology Coordinated Infrastructure under award ECCS-1542152. We thank Y. Zhou and M. Chen for technical discussions.
Author information
Authors and Affiliations
Contributions
T.P. and J.F. conceived of the concept, T.P. and E.W. fabricated the polarizers, T.P. assembled the system, T.D. analyzed the data, T.D, S.K, J.S, and M.L. developed the calibration methods, and T.D. and J.F. wrote the manuscript.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no competing interests.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Dai, T., Phan, T., Wang, E.W. et al. Snapshot Mueller spectropolarimeter imager. Microsyst Nanoeng 9, 125 (2023). https://doi.org/10.1038/s41378-023-00588-y
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41378-023-00588-y