Abstract
Digital photonic sensors have greatly evolved to maximize sensitivity and spatial, spectral, and temporal imaging resolution. For low-energy photons, new designs have generated new types of noise that degrade the formed-image signal-to-noise ratio to values lower than 1. Fixed-pattern noise (FPN), which is produced by the non-uniform focal-plane-array optoelectronics response, is an ill-posed problem in infrared and hyperspectral imaging science. Here, we experimentally show that the FPN behaves as an object at a depth of infinity when a light field is captured by an imaging system. The proposed method is based on the capture of the light field of a scene and digital refocusing to any nearby objects in the scene. Unlike standard techniques for FPN reduction, our method does not require knowledge of the physical parameters of the optoelectronic transducer, the motion scene, or the presence of off-line blackbody sources. The ability of the proposed method to reduce FPN is measured by evaluating the structural similarity (SSIM) index employing a blackbody-based FPN reduction technique as a reference. This new interpretation of the FPN opens avenues to create new cameras for low-energy photons with the ability to perform denoising by digital refocusing.
Similar content being viewed by others
Introduction
Researchers in the field of digital imaging technology are constantly attempting to improve image resolution, whether in spatial, spectral or temporal quality, particularly because such capabilities are required to discover new information in fields such as astronomy, medicine, and biology, among many others. Imaging technologies present distinct types of noise characteristics, varying in intensity, features, and nature. For example, the noise characteristics of various infrared (IR) imaging systems vary in intensity and spatial and temporal features depending on the nature of the IR-radiation transducer and the electronics of the readout integrated circuit (ROIC). These noise characteristics become superimposed on the captured IR radiation, thus causing image degradation, which can be observed by naked-eye visual inspection when viewing a raw digital image, since the resulting images exhibit a signal-to-noise ratio lower than 1. The numerous varieties of noise that affect imaging systems1,2,3,4, can be classified based on their temporal characteristics. Some noise sources vary as rapidly in time as the imaging system integration time. In contrast, other noise sources remain fixed for many integration time periods, forming a static pattern over the scene, as if one were viewing the image through a dirty or streaked window. This latter type of noise is typically called fixed-pattern noise (FPN) and is attributed to manufacturing imperfections, inhomogeneous pixel responsivities, dark currents of detectors, and typical readout architectures, such as a passive-pixel sensor or active-pixel sensor, introducing both temporal and spatially FPN1,3. In general terms, non-uniformity correction techniques can be classified into two main families: calibration-based and scene-based techniques5. The calibration method uses an absolute temperature reference; for this purpose, we used a blackbody radiation source, and for quantitative estimation of quality assessment, we used the blackbody radiation two-point (BBRTP) calibration method. Scientific applications require the highest radiometric precision available; thus, bulky and expensive calibration sources must be employed frame-by-frame to reduce FPN. In contrast, military applications and certain industrial applications require on-line classification capabilities; in this case, transducer technology, ROIC technology, and application dependency model-based algorithms must be specially designed and tuned to filter out the FPN on a case-by-case basis.
Here, we present technology- and application-independent solutions to the problem of FPN in the context of wide-band IR imaging systems. However, the proposed solution can also be tested in visible-IR-hyperspectral and terahertz imaging systems. The key idea is to realize that FPN behaves as an optical object located at an image depth of infinity. When a collection of images in the same scene is captured with relative displacement between them, such as a scene image light field, such set can be used to achieve digital refocusing to any nearby objects in the scene. Naturally, this approach filters out the FPN as a consequence of the defocusing. This denoising is based on the most notable property of this type of noise, namely, that its location is fixed on the images over time.
To explain this new concept, we show in Fig. 1 a simplified scene where FPN is depicted as an optical object located at the depth infinity. The Z coordinate describes the scene depth. The baseline of the camera is at the position Z0, the plane of nearby objects is at the position Z1, intermediate objects are located at the position Z2, and the FPN at Z ∞ represents the optical infinity of the system.
This new way of treating the FPN as an optical object at a depth of infinity allows one to develop a global solution to the problem of FPN reduction. Specifically, in this work, we build a plenoptic system with an IR camera with high FPN content using scanning. We show that it is possible to perform denoising by defocusing the FPN. For this purpose, the light field L f of the scene (plenoptic collection of images that captures scene information as a 4-dimensional function) is processed using a digital refocused algorithm6,7,8,9,10,11. In this work, we decided to employ the seminal work developed by Ren Ng6 known as the Fourier slice photography theorem. As a result of performing digital refocus, the effect of the FPN on objects at planes close to the imaging system is significantly mitigated. The level of noise reduction is quantified by the structural similarity (SSIM) index12, showing that it is possible to obtain a high degree of FPN cancellation using as a reference a blackbody-based FPN calibration technique.
Results
The reconstruction of digitally refocused images at far object planes shows severe FPN. However, we shall show that FPN is continuously reduced when refocused images at closer object planes are reconstructed by applying the refocusing algorithm to the captured light field.
The image of Fig. 2a shows the scene that was captured by the plenoptic IR system (using a Cedip camera and scanning to produce a matrix of 16 × 16 images with a step of 15 mm), which corresponds to an industrial sector with slow temperature variations with high thermal contrast contents. The Cedip camera is a JADE-UC uncooled microbolometer focal-plane-array (FPA) model that operates in the long-wave IR range of 8 to 12 μm and is equipped with a 14-bits, 320 × 240 FPA.
The image of Fig. 2a was FPN corrected by using the two-point calibration method (TPC)13,14,15, that utilizes a first-order model for the camera response. This calibration technique was applied using two uniform radiation sources acquired from two high-accuracy blackbody radiators in the Mikron M300 series, at 20 °C and 30 °C, respectively.
The region of interest (RoI) in Fig. 2a is shown as a rectangle of segmented lines. This RoI corresponds to a gas cylinder located 3 m from the camera (closest object in the image scene) and is used to enable a comparison of results for this entire work. It is important to note that we have used the most complex camera pixel region to locate our object of study since there are two bright spot artifacts in the image (attributed to partially saturated pixels of the FPA).
Figure 2b shows the same image of Fig. 2a without FPN correction; this is a raw image, and corresponds to an image of the L f with severe FPN. From left to right, Fig. 2c shows a sequence of 5 RoI images corresponding to the refocusing planes, namely, c) N = 1, d) N = 70, e) N = 140, f) N = 174 (the exact refocused image), and g) N = 240, thereby suppressing FPN. The reduction of noise by defocusing is limited by the nearest refocused object plane. In other words, we must determine the N associated with the plane of the focused near object. For this case, the value is N = 174, and it is obtained using the level of focus algorithm based on the Brenner function16,17, which coincides with naked-eye visual observations.
The degree of denoising FPN was quantified by SSIM index12, which employs an image reference free of FPN and the corresponding image at a refocused image from the raw-data light field. The range of this index are values from 0 to 1, where the lowest value (SSIM = 0) corresponds to entirely different images and the highest value (SSIM = 1) corresponds to identical images. Thus, we used the BBRTP calibration method, which allows generating the reference sets of light fields when applying the method to the raw-data light field. When applying the plenoptic algorithm for quantitative comparison, the refocused images generated from the raw-data light field are always compared with the refocused images from the BBRTP light field; thus, the SSIM index is a distance measure between these refocused images.
We studied the level of FPN reduction on the RoI by digital refocusing at 300 different object planes. As a reference, we used the same RoI images corrected using the BBRTP method. Since the images being compared differ only in the amount of FPN, a high value of the SSIM index is directly associated with a high level of noise reduction. We emphasize based on our experience that values greater than 0.4 correspond to levels of noise reduction that make the residual FPN imperceptible in naked-eye inspection. Additionally, an SSIM index greater than 0.4 means that the FPN has been reduced to an extent to which objects can be distinguished in the image. Note that in IR imaging artefacts such as dead, saturated, and flickering pixels and any computational imaging processing may change the IR contrast of the raw scene imaging. Thus, the SSIM index primarily corresponds to the SSIM of the objects within the IR image.
The graph in Fig. 3 shows the dependence of the SSIM index versus the refocusing parameter, N, of the nearby object presented in the RoI. The first values of N correspond to distant planes, where there is a high presence of FPN, as evidenced by reaching SSIM values lower than 0.13. As the closest planes are refocused (corresponding to large values of N), the FPN begins to defocus. This result is most strongly pronounced in the first 100 refocused planes, where the SSIM index increases from 0.13 to 0.49. Although there is a tendency to continue improving the FPN reduction, we can use the focused image only up to the value of N = 174.
To experimentally demonstrate that the method works satisfactorily in different scenarios, we apply the proposed method to two additional scenes. The first scene, shown in Fig. 4, corresponds to a winter scene where only half of the scene is under direct sunlight illumination. The second scene, shown in Fig. 5, corresponds to a daylight city-landscape in the spring. In Fig. 4, we selected two regions of interest, a portion of the roof of a warehouse and a street light, which are shown in Fig. 4c,e at refocused object planes with N = 47 and N = 100, respectively. In both cases, the structures can be recognized within the RoI, which does not occur in the raw-data in Fig. 4a. Similarly, in Fig. 5 we selected two regions of interest, a portion of the tree and a shrubbery, which are shown in Fig. 5c,e at refocused object planes with N = 70 and N = 117, respectively. Again, in both cases, the structures can be recognized in the RoI, which does not occur in the raw-data in Fig. 5a. The entire datasets for the above described light fields are publicly available, both raw data and BBRTP calibrated data18.
The results show that the proposed noise cancellation method reaches significant levels of FPN reduction. Such FPN reductions are quantified by SSIM index reaching values corresponding to a human naked eye’s structure recognition. Of course, the SSIM index difference with respect to the blackbody FPN method is not attributable to the structural objects scene recovering process, but to the contrast modification introduced by both techniques to the raw-data in the process of the FPN correction. Notably, unlike standard techniques19,20,21,22,23, our method does not require knowing physical parameters associated with the optoelectronic transducer system with which the images are acquired, the motion scene, or the presence of off-line blackbodies for estimation of non-uniformity sensor parameters.
Evaluating the effectiveness of the proposed method for reducing the FPN on different IR transducer technologies, where each type of transducer presents FPN with distinctive features, is quite interesting. Using pre-recorded blackbody data and the previously used L f corrected by BBRTP, it is possible to artificially generate an L f containing real FPN from the blackbody data. For this experiment, three additional cameras from different technologies are considered. The first camera is a commercial Amber camera, model AE-4128, which operates in the mid-wave infrared (MWIR) (3–5 μm) region and is equipped with a cooled InSb FPA with 128 × 128 pixels and 16-bits resolution. The second one is a commercial Lumasense MWIR camera, model MC320MHT, which operates in the range of 3–5 μm and uses a 16-bit uncooled FPA of 320 × 240 pixels. The third camera is a quantum-dots infrared photodetector (QDIP) camera, which operates in the mid- and long-wave IR regions (3–5 μm and 8–12 μm), is equipped with an InAs QDs/InGaAs/GaAs cooled FPA of 356 × 240 pixels, and quantizes data in 14-bits. For this experiment, blackbody samples of each camera at two different temperatures were used. These samples were taken without the use of built-in FPN compensation options. Specifically, blackbody samples were recorded by the Amber camera at temperatures of 20 °C and 30 °C, by the Lumasense camera at 100 °C and 350 °C, and by the QDIP camera at 26 °C and 50 °C. By employing these data and a TPC version of the L f , the noise pattern of these technologies is artificially superimposed on the clean L f using a first-order approximation for the response of the camera13,14,15, Subsequently, the L f with FPN is generated, and the proposed method is applied (see Fig. 6).
From top to bottom, Fig. 6a shows the outcome of refocusing the RoI for the Amber, Lumasense, QDIP and Cedip Technologies. The first image of each row corresponds to an image of the L f with the noise of each transducer technology, and the images to the right show the RoI at different refocusing planes that are separated from each other by a ΔN = 10. The last image corresponds to the best-refocused plane, which is for N = 174. In all of the cases, we considered the use of a single camera when generating the corresponding L f ; namely, for each IR technology, a unique noise pattern was rendered on all the L f images. Note that the Cedip thermal transducer generates the strongest FPN. Indeed, for this case, the scene object cannot be distinguished from the noise. Figure 6b shows the SSIM index versus digital refocused planes for the four analyzed technologies. In all the cases, the object in the RoI is focused at the plane N = 174, and for all the photonic transducers, the SSIM is higher than 0.8 at that object plane. For the Cedip thermal transducer technology, the \(SSIM=0.49\) at N = 174. Note that an SSIM index value greater 0.4 means that objects can be distinguished in the image. Among all the technologies, the largest reduction of FPN is achieved for the photonic transducer QDIP camera, which begins with an index of structural similarity \(SSIM=0.04\) and later reaches a value of \(SSIM=0.90\) (close planes). The greatest FPN reduction is achieved for the photonic transducer Amber technology with a \(SSIM=0.98\).
Furthermore, to show the robustness of the proposed method, we conducted three complementary performance evaluations. The first one considers another way to record an IR L f . In Fig. 6c, we plot the variation of the spatial SSIM index for two different architectures when generating the L f for the QDIP technology (with photonic transducers). The first architecture is a single camera and a scanning system, and therefore, a single noise pattern is replicated in all the images of the L f (dotted line). In contrast, the second architecture considers a different camera of the same technology for each image of the L f (solid line). This design is simulated by using the procedure15 for synthesizing different FPN samples that preserve the technology-dependent spatial structure of the real FPN of a camera. As shown in Fig. 6c, the method performs equally well under both types of FPA architectures for generating the corresponding L f . This result provides clear evidence of the robustness of the proposed method under different L f architectures.
The second and third complementary performance evaluations were conducted using the Cedip camera and scanning. To study how reducing the L f size affects FPN cancellation, three cases are presented in the second evaluation. Specifically, a light field of 16 × 16 images was reduced to 8 × 8 and 4 × 4 images while maintaining a step between adjacent images of 15 mm.
Table 1 indicates that a reduction of the L f size produces a reduction of the SSIM index from \(SSIM=0.49\) to \(SSIM=0.47\). Regarding FPN reduction performance, we see that a smaller L f shows little diversity of spatial frequencies and generates refocused images with greater depth, which makes smaller L f relatively limited in regards to forming a discrete plane of refocused elements. This constraint does not allow a precise selection of information; therefore, the residual noise of the FPN structure remains even when one focuses only on the object of interest located in a near optical plane. However, in all cases, the SSIM index is similar to the ones computed for an L f of 16 × 16 images. Note that this phenomenon occurs even in the event of the smallest L f that contains only 4 × 4 images.
In the third case, the performance of the method is studied on 16 × 16-image light fields with a different step size between adjacent images. Three different L f with steps of 15 mm, 10 mm and 5 mm were experimentally acquired. The results of applying the proposed method to each one of these light fields are shown in Table 2.
As shown in Table 2, by reducing the separation between images, the number of planes between optical infinity and the best-refocused plane for the closest object decreases in number, from 174 planes to only 63. This decrease coarsens the depth of field at each discrete plane, which limits the refocusing precision on specific objects, although the SSIM index is not affected significantly, and the quality of the images is maintained.
Discussion
We have proposed a new technology-transducer and application-independent method for FPN reduction that does not rely on the use of physical information of the FPA or on using external reference scenes (such as calibrated blackbody sources). In particular, we have experimentally demonstrated the effectiveness of the proposed method in the case of a scanning plenoptic IR system with a single camera. The method was tested by varying the size of the captured L f and varying the displacement between consecutive images in the L f . These tests allow one to appreciate that noise reduction is robust under both variations of the experimental configuration. Notably, a small L f of 4 × 4 images produces a high increase of the SSIM index, which makes this method very promising for future IR imaging and a visible-IR-hyperspectral imaging system with L f ability.
Furthermore, we have demonstrated the robustness of the method for three other IR transducer technologies, using digital addition of real FPN characteristics from each technology in our L f free of FPN. In all cases, we obtained high noise cancellation, which strongly suggests that the method is technology independent. For studied technologies, the most notable FPN reduction occurs for the QDIP camera, for which the SSIM concerning the FPN free reference increases by 0.86. We also performed simulations of plenoptic L f obtained with an array of IR cameras, where each one of the L f images was captured by an independent camera. Here, each image has different FPN structures, and the remarkable property of the method is that even in such cases, the noise reduction is efficient and comparable to the case of using a single camera and scanning system for capturing the L f (same FPN for all the images).
In the next few years, the proposed method is expected to have technological implications in the development of new cameras, for low-energy photons, such as IR and terahertz cameras, and for visible and IR spectral selected photons (hyperspectral cameras) with the ability to perform denoising by digital refocusing.
Methods
One method to acquire a scene L f is through a two-dimensional transverse scanning employing a single camera, see Fig. 7a. This L f capture procedure was implemented with the following equipment: an optical bench, a high-precision linear translation stage with a displacement range of 600 mm with DC motor and rotary encoder (model: IMS600CC, Newport), a high-precision vertical translation stage (minimal translation of 1.25 μm) with a displacement range of 300 mm and with a DC motor (IMS-V Series, model: IMS300V, Newport) mounted on a right angle bracket (model: EQ120, Newport), and a 2-axis universal driver with an Ethernet connection (model: XPS-Q2, Newport) for controlling of vertical and horizontal translation stages. The digital camera was a Jade-UC, Cedip Inc., which integrates the incident radiance using an uncooled microbolometer FPA sensor of 320 × 240 pixels that operates in the range of 8–12 μm and has a dynamic range of 14-bits. This camera was operated without using its built-in FPN compensation options (raw-data imaging), and the formation of the image on the sensor used a focal length lens of f = 24 mm and an F-number of F = 1.1.
Through the overall equipment arrangement and operation, as shown in Fig. 7, the light field \({L}_{f}(x,y;m,n)\) was generated. The coordinates (x, y) correspond to the image pixels, and the coordinates (m, n) correspond to the position of each L f image. In our case, the images were of 320 × 240 pixels in an arrangement of 16 × 16 images. The offset between (m, n) coordinates of the L f was 15 mm for the vertical and horizontal directions.
The digital refocusing was accomplished by means of the Fourier slice photography theorem6:
Using the digital refocusing operator, \({{\mathscr{P}}}_{\alpha }\), it is possible to select a scene’s spatial frequency information from L f to reconstruct images focused on different object planes determined by the α parameter, where small α values correspond to far object planes. Thus, an effective refocused image on a fixed object plane is obtained through the two-dimensional inverse Fourier transform, \({ {\mathcal F} }^{(-\mathrm{2)}}\), from a two-dimensional slice information, \({{\mathfrak{B}}}_{\alpha }\), from the Fourier transform in four-dimensions, \({ {\mathcal F} }^{\mathrm{(4)}}\). For this particular work, the α values are related to different focal planes obtained in such a way that every \(N=\mathrm{1,\; 2,\; 3,}\cdots \) is associated with \({\alpha }_{1}=\mathrm{0.999,}{\alpha }_{2}=\mathrm{0.998,}{\alpha }_{3}=0.997\ldots ,{\alpha }_{N}=1-0.001\times N\), which take values equally spaced between them. Finally, 300 effective refocused images were reconstructed and arranged from the scene image farthest object plane to the closest object plane, i.e., from N = 1 to N = 300.
To quantify the quality assessment of the proposed method for denoising FPN, we used the SSIM index12, which allows one to quantify the SSIM between a reference image and a distorted image by considering their luminance, contrast, and structure by considering that these three parameters are relatively independent. This index has been developed for the visible region of the electromagnetic spectrum by considering that human perception is adapted to extract structural information of scenes. In our case, as reference images in a particular refocused plane, we used those refocused images obtained from the BBRTP corrected light field, and the distorted images are refocused images from the raw-data light field.
References
Scribner, D. A., Kruer, M. R. & Killiany, J. M. Infrared focal plane array technology. Proceedings of the IEEE. https://doi.org/10.1109/5.64383 (1991).
Holst, G. C. CCD Arrays, Cameras and Displays (Society of Photo Optical, 1998), second edn.
Rogalski, A. Infrared detectors: An overview. Infrared Physics and Technology 43, 187–210, https://doi.org/10.1016/S1350- 4495(02)00140-8 (2002).
Cao, Y. & Tisse, C.-L. Single-image-based solution for optics temperature-dependent nonuniformity correction in an uncooled long-wave infrared camera. Optics letters 39, 646–648, https://doi.org/10.1117/12.2027492 (2014).
Goodall, T., Bovik, A. C., Vikalo, H. & Paulter, N. G. Non-uniformity Correction of IR Images using Natural Scene Statistics. IEEE Global Conference on Signal and Information Processing (GlobalSIP). https://doi.org/10.1109/Global- SIP.2015.7418191 (2015).
Ng, R. Fourier slice photography. ACM Transactions on Graphics 24, 735–744, https://doi.org/10.1145/1073204.1073256 (2005).
Luan, Y., He, X., Xu, B., Yang, P. & Tang, G. Automatic calibration method for plenoptic camera. Optical Engineering 55, https://doi.org/10.1117/1.OE.55.4.043111 (2016).
Liu, P. & Zhang, R. A new freeform depth of field controlling method based on focused plenoptic camera. Journal of Computational Methods in Science and Engineering 16, 187–195, https://doi.org/10.3233/JCM-160609 (2016).
Heinze, C., Spyropoulos, S., Hussmann, S. & Perwass, C. Automated Robust Metric Calibration Algorithm for Multifocus Plenoptic Cameras. IEEE Transactions on Instrumentation and Measurement 65, 1197–1205, https://doi.org/10.1109/TIM.2015.2507412 (2016).
Georgiev, T. & Lumsdaine, A. Focused plenoptic camera and rendering. Journal of Electronic Imaging 19, 021106, https://doi.org/10.1117/1.3442712 (2010).
Georgiev, T. & Lumsdaine, A. Reducing plenoptic camera artifacts. Computer Graphics Forum, https://doi.org/10.1111/j.1467- 8659.2010.01662.x (2010).
Wang, Z., Bovik, A. C., Sheikh, H. R. & Simoncelli, E. P. Image quality assessment: From error visibility to structural similarity. IEEE Transactions on Image Processing 13, 600–612, https://doi.org/10.1109/TIP.2003.819861 (2004).
Perry, D. & Dereniak. Linear theory of nonuniformity correction in infrared staring sensors. Optical Engineering 32, 185–1859, https://doi.org/10.1117/12.145601 (1993).
Schulz, M. & Caldwell, L. Nonuniformity correction and correctability of infrared focal plane arrays. Infrared Physics & Technology. https://doi.org/10.1016/1350-4495(94)00002-3 (1995).
Perez, F., Pezoa, J., Figueroa, M. & Torres, S. N. Empirical frequency domain model for fixed-pattern noise in infrared focal plane arrays. Infrared Physics & Technology 67, 413–426, https://doi.org/10.1117/12.145601 (2014).
Brenner, J. F. et al. An automated microscope for cytologic research a preliminary evaluation. The Journal of Histochemistry and Cytochemistry 24, 100–111, https://doi.org/10.1177/24.1.1254907 (1976).
Santos, A. et al. Evaluation of autofocus functions in molecular cytogenetic analysis. J Microsc 188, 264–272, https://doi.org/10.1046/j.1365-2818.1997.2630819.x (1997).
Coelho, P., Tapia, J. E., Pérez, F., Torres, S. N. & Saavedra, C. Infrared light field datasets http://osf.io/f8tr7 (2017).
Torres, S. N. & Hayat, M. M. Kalman filtering for adaptive nonuniformity correction in infrared focal-plane arrays. Journal of the Optical Society of America. A, Optics, image science, and vision 20, 470–480, https://doi.org/10.1364/JOSAA.20.000470 (2003).
Hayat, M. M., Torres, S. N., Armstrong, E., Cain, S. C. & Yasuda, B. Statistical algorithm for nonuniformity correction in focal-plane arrays. Applied optics 38, 772–780, https://doi.org/10.1364/AO.38.000772 (1999).
Hardie, R. C., Hayat, M. M., Armstrong, E. & Yasuda, B. Scene-based nonuniformity correction with video sequences and registration. Applied Optics 39, 1241–1250, https://doi.org/10.1364/AO.39.001241 (2000).
Pezoa, J. E., Hayat, M. M., Torres, S. N. & Rahman, M. S. Multimodel Kalman filtering for adaptive nonuniformity correction in infrared sensors. Journal of the Optical Society of America. A, Optics, image science, and vision 23, 1282–1291, https://doi.org/10.1364/JOSAA.23.001282 (2006).
Godoy, S. E., Pezoa, J. E. & Torres, S. N. Noise-cancellation-based nonuniformity correction algorithm for infrared focal-plane arrays. Applied Optics 47, 5394–5399, https://doi.org/10.1364/AO.47.005394 (2008).
Acknowledgements
The authors acknowledge the financial support both from CONICYT grants, Basal Financing Program PFB0824 and FONDECYT 1160613. P.A. Coelho and J.E. Tapia acknowledge CONICYT Ph.D. scholarship.
Author information
Authors and Affiliations
Contributions
P.C., J.T., and F.P. have executed the experiment and processed the data. S.T. and C.S. designed and assisted in the experiment. All authors agree to the contents of the paper.
Corresponding authors
Ethics declarations
Competing Interests
The authors declare that they have no competing interests.
Additional information
Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Coelho, P.A., Tapia, J.E., Pérez, F. et al. Infrared light field imaging system free of fixed-pattern noise. Sci Rep 7, 13040 (2017). https://doi.org/10.1038/s41598-017-13595-7
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-017-13595-7
This article is cited by
-
Infrared light field imaging system free of fixed-pattern noise
Scientific Reports (2017)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.