The development of new technologies and tools in light microscopy has had an important role in making biomedical research more quantitative and interdisciplinary. This allows researchers to tackle more challenging scientific questions and obtain a more in-depth understanding of complex biological systems. However, with the rapid deployment of many new technologies, it is difficult for researchers to maintain in-depth knowledge of their capabilities and limitations. In addition, it is not always clear to users of advanced imaging techniques how specific limitations in hardware and software configurations for each modality might lead to errors that could affect scientific conclusions and reproducibility. Accordingly, the methods section of a publication requires complete and accurate information on the imaging conditions used to allow experimental replication. However, Marqués et al.1 recently showed that lack of detail in describing image acquisition is a widespread problem in biomedical publications, especially when considering that imaging is critical in biomedical research.

Many publications focus on guidelines to successfully perform a microscopy experiment, including sample preparation, method validation and appropriate controls to ensure reproducibility2,3,4,5. However, there are still few resources to guide and assist researchers in writing rigorous and reproducible microscopy methods. This Perspective focuses on filling this need by proposing minimal guidelines to ensure rigor and reproducibility in fluorescence light microscopy.

Scope of the guidelines

The scope of this Perspective is to emphasize the importance of appropriate microscopy methods reporting and help educate researchers about microscopy components and parameters that impact data and conclusions. We showcase examples to demonstrate and highlight errors that can arise from insufficient reporting and provide resources to assist researchers with this task: an educational poster (Extended Data Fig. 1), comprehensive checklists (Supplementary Exhibits 14) and a customizable checklist-generating tool, MicCheck (Supplementary Exhibit 5). These guidelines are in line with the tier-based system developed in the OME-4DN project for metadata collection to ensure consistency across the microscopy community6.

The checklists and MicCheck enumerate and describe in detail the essential and recommended metadata for the most common fluorescence light microscopy applications: widefield, laser scanning (confocal and multiphoton) and spinning disk confocal experiments, for live and fixed samples. The essential, or minimally required, metadata are critical for the correct interpretation of microscopy data and must be reported to ensure rigor and reproducibility. The recommended metadata represent best practices, particularly when using custom-built microscopes and nonstandard image acquisition strategies, but these may not substantially impact the conclusions that may be drawn. Finally, we include a list of resources and initiatives to improve image-based reproducibility and a list of definitions for many of the terms discussed in this Perspective (Supplementary Tables 1 and 2) that can serve as a reference for more novice microscopy users.

We intend these materials to be generally appropriate for many of the most common light microscopy configurations and applications7 but encourage researchers to use their own discretion and apply the recommendations as appropriate for their specific application and imaging system.

Guidelines on reporting instrument metadata

Illumination, light collection and wavelength selection

Illumination is critical to assess (1) how efficiently a fluorophore will be excited, (2) the relative illumination intensity for each fluorophore, (3) how reproducible intensity measurements are over time, (4) the probability of excitation cross-talk and (5) the compatibility of excitation filters8. How the excitation and collection of light from fluorophores is achieved can vary widely. Thus, the choice of hardware used to illuminate the sample and to collect the fluorescence signal emitted from the fluorophores will have a profound effect on data interpretation and the ability to make meaningful comparisons between datasets (Box 1). Specifically, the light source and the excitation and emission wavelength bandwidth and hardware (for example, the excitation filter and dichromatic mirror) used are essential metadata. The power density (irradiance) at the sample impacts the excitation efficiency of fluorophores, the photobleaching rate and the cumulative amount of light that can be collected from the sample during acquisition. While irradiance is critical for reproducibility and ideally should be carefully documented, it is challenging to measure, requiring dedicated tools and protocols (Box 1).

The signal that can be collected is determined by the hardware used. Well-matched filters produce brighter fluorescence at shorter exposures or lower excitation light intensities and improve overall detection, signal-to-noise ratio (SNR) and downstream data analysis. Under otherwise identical imaging conditions, the SNR of the fluorescent foci and membrane localization of the fusion protein in Fig. 1a are below the detection threshold using an emission filter that is not well matched to the fluorophore. This will lead to conflicting results if not well documented. Additionally, the sample requires higher illumination intensity to excite and detect the appropriate fluorescence signal to determine the localization pattern. This can impact photobleaching and affect the health of the sample, which in turn could affect the localization pattern, leading to irreproducibility.

Fig. 1: Illumination and wavelength selection impact signal detection, image quality and cross-talk between channels.
figure 1

a, Fluorescence images (using Fire LUT) of Escherichia coli cells expressing ZipA–sfGFP acquired under identical acquisition settings and excitation wavelengths with different emission filters. Left, the emission filter is well matched to sfGFP (‘green’; Semrock FF01-515/30). Center, the emission filter is less well matched to the fluorescence filter (‘yellow’; Semrock FF01-544/24). The yellow arrowheads indicate localization at the membrane and discrete foci. Right, violin plots of the fluorescence intensity of individual images. Encircled dots indicate the population mean (‘green’, n = 432 cells; ‘yellow’, n = 432 cells). Scale bars, 5 μm and 1 μm (magnification). b, Fluorescence images (using a grayscale LUT) of U2OS cells expressing Mito-RFP and labeled with SYTO Deep Red (nuclei) acquired with excitation at 555 nm (left) or 600 nm (center left) and emission collected from 605–680 nm. White arrowheads indicate SYTO Deep Red nuclear emission signal collected. Center right, same as in center left but with emission collection adjusted down to 605–645 nm, which almost completely removes the SYTO Deep Red cross-talk, indicated by white arrowheads. Right, SYTO Deep Red image. Scale bar, 10 μm. c, Fluorescence images (using a Fire LUT) of BPAE cells stained with DAPI and Alexa Fluor 488 phalloidin (right) imaged using a narrow-bandpass emission range of 415–475 nm (left and center left) or a broad-bandpass emission range (center right). Center left, same as in left but with higher laser intensity. Scale bar, 10 μm. d, Two-color fluorescence imaging of Convallaria using a single multi-bandpass dichroic mirror for both channels (left) or a different dichroic mirror for each channel (center). Overlay images are pseudocolored green (488-nm excitation) and magenta (561-nm excitation). White indicates overlap between the two signals. Right, magnified views show the extent of xy shift between the two images using the same (top) or different (bottom) dichroic mirrors. Scale bars, 25 μm and 5 μm (magnification).

In multicolor imaging, in which the excitation and emission spectra of the different fluorophores may be partially overlapping, the appropriate selection and documentation of excitation and emission wavelengths (Supplementary Table 1) is critical to spectrally separate and quantify signals in different channels. In Fig. 1b, the localization pattern and intensity of the Mito-RFP signal differ depending on the excitation wavelength used. Using 600-nm light to illuminate the sample results in an erroneous localization pattern and intensity levels, as they are partially due to cross-talk from a different fluorophore in the sample (nuclear SYTO Deep Red). This cross-talk can be minimized by illuminating the sample with 555-nm light. Figure 1c shows that using a narrow-bandpass emission range minimizes cross-talk between two channels (compare center right and left), even when the laser intensity is increased to improve SNR (center right). These examples highlight the need for appropriate controls to evaluate the extent of cross-talk between channels in multicolor experiments and how the lack of documentation of excitation and emission wavelengths may lead to irreproducibility.

Experiments in which multiple fluorophores need to be acquired simultaneously or with little temporal or spatial shift between channels (for example, ratiometric imaging and colocalization experiments) require the use and documentation of specific hardware to minimize chromatic shift (Box 2). This shift may be due to an improper angle of mounting for dichroic mirrors and imperfections on their thin-film interference coatings or to poor instrument calibration. Channel misregistration greatly impacts the conclusions from colocalization experiments and experiments aimed at quantifying distances between objects or intensity ratios between channels (Fig. 1d). Controls are critical in assessing the extent of chromatic shift and correcting it9,10,11 (Box 2 and Supplementary Table 1).

Objective lenses

Objective lenses are essential for image formation and impact all aspects of imaging; thus, they are one of the most critical components that need to be reported. Magnification is an important aspect of digital microscopy, as it contributes to the size of the field of view and, for camera-based microscopes, the pixel size (spatial sampling). Many researchers understand this, and in most cases this aspect is documented. However, objective lenses possess other characteristics that have an even greater impact on image formation and must be reported for others to reproduce an experiment.

Objective lenses contain spherical surfaces that focus the light into the sample and collect emitted fluorescence to form an image. These curved surfaces create different types of aberrations12,13,14. Manufacturers design objective lenses applying different aberration corrections depending on the intended application. These corrections are made assuming specific conditions such as coverslip thickness, refractive index of the medium used, temperature and illumination wavelengths. Some objective lenses are designed with correction collars, which allow for fine-tuning of the performance of the objective lens to compensate for sample-induced aberrations15,16. Of note, however, these aberrations are not completely eliminated. High-quality sample preparation is essential to ensure the best performance of any objective lens, and appropriate controls and corrections need to be applied for best practices3,5,9,17 (Box 2 and Supplementary Table 1).

The aberration corrections in an objective lens will have a profound effect on the interpretation of the microscopy data collected. Using a highly color-corrected lens will provide more precise measurements in experiments quantifying relative distance in multicolor three-dimensional (3D) imaging or colocalization (Fig. 2a). In the example provided, the extent of colocalization between channels depends on the objective correction, which, if not reported, could result in irreproducibility.

Fig. 2: Objective lenses and immersion medium determine brightness and optical resolution and impact data interpretation.
figure 2

a, A 3D rendering of the 3D sphere pattern on the ArgoLight-SIM calibration slide acquired with two channels (488 nm, green; 561 nm, magenta) using either an Olympus Plan Apo 40×/1.3 NA DIC (left) or an Olympus Plan Fluor 40×/1.3 NA DIC (center) objective. Right, insets show the extent of overlap between signals (white). Overlay images are pseudocolored in green (488-nm excitation) and magenta (561-nm excitation). Scale bars, 5 µm. b, Fluorescence images (using a Fire LUT) of BPAE cells stained with MitoTracker Red CMXRos acquired under identical imaging conditions using a Plan Fluor ELWD 40×/0.6 NA Ph2 (left), a Plan Apo 40×/0.95 NA DIC (center left) or a Plan Fluor 40x/1.3 NA oil-immersion DIC (center right) objective. Images are displayed at the same brightness/intensity scale. Right, violin plots of fluorescence intensity quantification of individual regions of interest. Long black bars indicate the mean fluorescence intensity for each population, and vertical bars indicate the s.d. Scale bar, 5 µm. c, The same sample as in b imaged with the same objectives as in b. Acquisition was optimized for each objective to obtain high-SNR images. Insets in each panel demonstrate the difference in resolution obtained when using objectives with different NA. Scale bars, 5 µm (main panel) and 2 μm (magnification). d, Representative xz projections of z stacks obtained by imaging 100-nm fluorescent beads passively adhered to #1.5 HT coverslips and mounted with glycerol using a Nikon Plan Apo 100×/1.45 NA Ph3 objective with either Olympus Type F (left) or Nikon NF (center) immersion oil. Insets show the midplane displayed at the same scale using the Fire LUT. Right, dot plot illustrating the axial FWHM of individual beads quantified using the PSFj plugin when either the Type F or Type NF oil-immersion objective was used for imaging. The bars represent the population mean. Scale bar, 1 µm.

Numerical aperture (NA) is a critical characteristic of an objective lens because it indicates the amount of light that can be collected. Higher-NA lenses, under identical imaging conditions, produce brighter images than low-NA lenses9,11,18 (Fig. 2b). A low-NA lens will require longer exposure times or higher excitation light intensities to detect the fluorescence signal, which will impact temporal resolution, photobleaching and/or phototoxicity, leading to irreproducibility. Most importantly, the NA, and not the magnification, determines the resolving power of the system. Consequently, the NA of the objective lens will determine whether two objects of interest can be resolved or even detected under certain conditions (Fig. 2b,c). In the example, the high-NA objective lens produces sharper and more resolved images, improving the detection and detail of individual mitochondria.

Objective lenses are designed to work with a specific immersion medium (for example, air, oil, silicone oil, water or glycerol). There are many types of immersion oil, which vary in viscosity, refractive index and dispersion under different conditions. The type of immersion medium affects the performance of an objective lens, by inducing or minimizing spherical aberration and sometimes introducing fluorescence background signal. Therefore, two different immersion oils may result in different axial resolution and overall brightness of the image, thus affecting the ability to resolve objects axially or to detect the object of interest (Fig. 2d). The wrong immersion medium may also induce chromatic aberrations due to differences in dispersion19. Even though the type of immersion medium is critical to performance of the objective lens, it is overlooked in most methods sections. This reporting is especially critical for multi-immersion objective lenses, which can be used with multiple immersion media.


The detector used is critical in all aspects of quantitative microscopy. It impacts the likelihood that the fluorescence signal will be detected (sensitivity and noise level), the temporal resolution (frame rate) and the spatial resolution (digital resolution) of the system20,21. As such, the choice of detector influences the capability to provide a conclusion faithful to the underlying biology. Thus, proper documentation ensures that similar observations can be made and similar conclusions can be drawn in different systems or experimental replicates.

The digital resolution of the system determines the ability to resolve the object of interest and will influence the downstream analysis workflow. In camera-based systems, it is determined by the physical size of each photodiode (pixel) and the total magnification (Supplementary Exhibits 14). Data collected with two different cameras attached to the same microscope and acquired using the same objective lens can lead to fundamentally different interpretations of the results (Fig. 3a). The camera with a large photodiode size (16 µm) cannot resolve the line pairs in the particular sample observed and instead creates a new pattern that does not correspond to the ground truth (aliasing). Pixel binning decreases the digital resolution and therefore must be reported20,21. Any additional magnification to the objective lens also needs to be reported to ensure measurement accuracy and precision, as this reduces light output (reduced SNR) and affects digital resolution (Supplementary Exhibits 14). In many cases, magnification changes are not accounted for in image calibrations and metadata, although not documenting them can lead to imprecision and irreproducibility in distance measurements (Box 2). This is true for any hardware component that is introduced manually into the lightpath, as it would not be recorded into the image metadata.

Fig. 3: The sensitivity of detectors varies with wavelength and determines the digital resolution and image quality.
figure 3

a, Fluorescence images of a pattern consisting of lines with incrementally increasing spacing on the ArgoLight-SIM calibration slide acquired with the same objective and a camera with either a 6.5-µm (left) or 16-µm (right) photodiode size. Lower row, line intensity profile plots. Scale bar, 2.5 µm. b, Fluorescence images (using a Fire LUT) of BPAE cells stained with Alexa Fluor 488 phalloidin acquired under identical imaging conditions with a single-point scanning confocal microscope using either a GaAsP (left) or multialkali PMT (MA; center) detector. Images are displayed with the same intensity scale. Right, fluorescence intensity quantification; bars represent the population mean. Scale bar, 10 µm. c, Fluorescence images (using a Fire LUT) of cultured cells prepared by Leica and imaged under identical conditions with a Leica Stellaris 8 Power HyD S (left column), Power HyD X (center-left column) or Power HyD R (center-right column) show a difference in intensity according to wavelength (right). Green, HyD S; purple, HyD X; yellow, HyD R. Scale bars, 10 µm. d, Fluorescence images of a pattern consisting of a repeating series of lines with progressively decreasing intensity (pattern C) on the ArgoLight-SIM calibration slide were acquired with an Andor Zyla sCMOS 4.2 plus camera under low-light conditions (left panels) or high-light conditions (right panels). Images were acquired under identical conditions in each case, with the exception of varying the amplifier gain amplification. Top left and top right, low gain; bottom left and bottom right, high gain. Bottom panels, plots of fluorescence intensities along a vertical line scan through the center of the pattern (purple, low gain amplification; green, high gain amplification). Scale bar, 5 µm.

Image quality and SNR are greatly impacted by the sensitivity (quantum efficiency, QE) and noise level of the detector. These specifications determine the probability of detection and minimal detectable signal. The peak QE of detectors varies widely22 (70–95% for cameras and 20–45% for photomultiplier tubes, PMTs). Under otherwise identical conditions, a more sensitive detector will detect more light, producing a brighter image (Fig. 3b). Thus, the sensitivity and noise level of a particular detector will determine whether the signal can be detected at all and the dynamic range of the measurements, which in turn impacts their precision and accuracy. A less sensitive detector will require increased illumination intensity to detect the signal, which can induce photobleaching, fluorophore saturation and/or photodamage, leading to irreproducibility (Box 1).

Furthermore, the sensitivity of a detector varies with wavelength22. Many systems are equipped with various types of detectors that offer different sensitivity at different wavelengths, such as in single-point scanning confocal and multiphoton microscopes. In the example provided in Fig. 3c, a researcher using a HyD X detector (or another detector with similar specifications) to collect the emission of a near-infrared fluorophore would conclude that a protein of interest is not expressed in a particular cell type. Thus, the choice of detector in an experiment must be reported to ensure reproducibility.

The properties described above are inherent to a given detector; therefore, providing information on the specific manufacturer and model (for cameras) or the specific type of non-camera-based detector (for example, Multialkali PMT) will satisfy the minimal metadata requirements to ensure reproducibility. However, many detectors offer different acquisition settings that may affect the sensitivity, noise level and dynamic range of the detector, thus impacting the ability to detect the signal of interest. These settings should be documented, as they will impact reproducibility.

The readout mode or digitizer of the detector and the bit depth are good examples of such settings that critically impact image acquisition. Many cameras and other detectors offer a choice of which digitizer (and associated gain) to use during acquisition, which can be optimized depending on the amount of light generated by the sample. Higher gain settings result in higher intensity levels per electron than lower gain settings. Under low-light conditions (that is, for dim samples), a high gain setting may improve signal detection in comparison to a low gain setting (Fig. 3d). However, high gain settings also increase noise and reduce the dynamic range of the detector, so they are not ideal for bright samples as these samples will reach the saturation level sooner than when using low gain settings (Fig. 3d). Some cameras enable charge amplification (electron-multiplied charge-coupled device, EM-CCD20), thereby improving the detection of very dim signals for applications including single-molecule fluorescence. Thus, researchers should specify the particular camera settings or readout modes used during acquisition.

Microscope stand and relevant motorized components

Microscopy experiments usually require the acquisition of multidimensional data (for example, multiple planes, positions, channels, time points, etc.). The microscope stand and peripheral motorized components required depend on the goals of the experiment and the sample used (Supplementary Exhibits 15). The precision of the motorized components and how the software manages the image acquisition workflow may significantly impact the conclusions from the experiment, as they can introduce inaccuracies and limitations in measurements. Thus, knowledge of the hardware configuration and acquisition settings is required to collect precise, accurate and reproducible data. Consequently, the specifications of each component need to be verified as ideal for the application at hand, validated through appropriate controls (Box 2 and Supplementary Table 1) and accurately reported.

In multicolor experiments, the hardware used and whether the images are acquired sequentially or simultaneously will have a profound impact on several aspects of quantitative microscopy. Simultaneous acquisition will greatly increase the temporal resolution of the experiment but may also result in an increase in signal contamination between channels (Fig. 4a). In this case, simultaneous acquisition would lead to erroneous results.

Fig. 4: Data acquisition mode and instrument precision impact data reproducibility and interpretation.
figure 4

a, Fluorescence images of BPAE cells stained with DAPI and Alexa Fluor 488 phalloidin acquired either simultaneously in the same track (center right and right) or sequentially in two different tracks (left and center left). Scale bar, 15 µm. b, Dot plots of the displacement between centroids of 1-µm beads imaged repetitively with two different translational precision settings of a Nikon linear-encoded motorized stage (green, open, lower precision setting; purple, precise, highest precision setting). Multiple xy positions were imaged repetitively over 15 loops. Bars represent the mean and s.e.m c, The xyz projections of TetraSpeck beads imaged by acquiring a z stack in each channel (left) or both channels in each focal plane of the z stack (center). Right, displacement plots of beads acquired using either the nosepiece Z drive (TiZ) or a PI piezo Z drive (piezo). Bars represent the mean and s.e.m. Scale bars, 2 µm. d, Fluorescence images of a Convallaria section showing the midplane (left) and xz and yz projections from z stacks acquired while varying the size of the z step (0.4 μm, 2 μm and 3 μm) (right). Scale bars, 10 µm. e, Same sample as in a imaged using a bidirectional scanning mode (left) or a unidirectional scanning mode (center). Inset images indicate a pronounced pixel shift in the image acquired using the bidirectional scanning mode when compared to the image acquired using the unidirectional scanning mode. Right, this shift can be corrected following calibration of the scanner. Scale bars, 5 µm (main panel) and 1 μm (magnification) . f, Same sample as in a acquired while varying the offset setting (top row, offset of 1.0 V (no offset); bottom row, offset of −118 V). Magnified views show processing steps to segment the mitochondria. Segmentation using images collected with improper offset settings results in identification of smaller objects (green arrowheads) and loss of objects (yellow arrowhead). Scale bars, 5 µm (main panel and magnification).

Experiments that rely on the repeatability of positions, such as those tracking moving objects in multiple positions over time, are greatly impacted by the speed and precision of the lateral movement of motorized stages. In some cases, these parameters can be changed in the software (Supplementary Appendix), and overlooking and not documenting these settings will impact reproducibility. For example, the displacement between the centroids of stationary fluorescent beads imaged repeatedly in a multiple-position experiment can vary over fourfold depending on the precision settings of a motorized stage (Fig. 4b). This also applies to the axial location of the beads when acquiring a z stack and will inevitably introduce error in the measurements (Fig. 4c). Furthermore, the order in which the acquisition is performed in a multidimensional experiment (for example, multicolor and z stack) can also impact the results (Supplementary Appendix). Acquiring a z stack of each wavelength instead of acquiring all the wavelengths at each z step can lead to higher imprecision. Therefore, consideration of the focusing device used and the order in which the data are collected is critical during experimental design and must be reported (Fig. 4c and Box 2).

Furthermore, the z-step interval and total volume (and how it is selected) in a z stack are critical to resolve objects in the axial dimension and need to be specified and reported. Larger step intervals than required (undersampling) can result in insufficient axial resolution. This will compromise downstream analysis and the conclusions of the experiment19 (Fig. 4d).

The success of a time-lapse imaging experiment depends on the ability to maintain focal and planar positioning and collect data at the appropriate temporal resolution and SNR to permit downstream image analysis without compromising the health of the sample. Irradiance is one of the key factors that affect sample health, as it can induce phototoxicity (Box 1). The effects of phototoxicity tend to be underestimated or overlooked, even though they are critical to properly interpret the observations4,23,24,25,26. Implementing and reporting hardware and acquisition settings that reduce irradiance is critical to allow meaningful comparisons between experiments (Supplementary Exhibits 14). Notably, the nominal time interval set in the software may not be consistent with the actual acquisition time interval, which depends on the shutter speed, delays due to the electronics that drive the acquisition and the acquisition software used. However, the timestamp found in the image metadata is likely accurate and should be used for quantification. Additionally, the actual average time interval and standard deviation should be included in the methods reported in a publication.

Guidelines on reporting acquisition software

Modern microscope systems are connected to a computer and controlled by acquisition software, which can be commercially available (for example, NIS Elements) or open source (for example, µManager). Software programming allows advanced automated image acquisition. Flexibility and integration depend on the specific software, the version (including relevant hotfixes) and available modules. Often, in the case of commercial packages, additional modules and/or upgrades have to be purchased individually, and the software can therefore vary substantially over time and between systems with almost identical hardware. It is critical to report software, modules and versions used. Any custom acquisition code should be reported and made available to the scientific community, including the URL or website that contains information on its source and capabilities.

Guidelines on reporting metadata specific to microscopy modality

Optical sectioning techniques, such as confocal and multiphoton, are designed to enhance the contrast of a wide variety of biological specimens, including tissue sections or whole organs and organisms. They use vendor-specific methods to configure the light path, which may vary by instrument. Consequently, each microscopy modality will have dedicated hardware and software acquisition configurations that will impact results. In this section, we describe the metadata specific to single-point scanning confocal, spinning disk confocal and multiphoton microscopy. The metadata applicable to widefield microscopy are described in “Guidelines on reporting instrument data” and Supplementary Exhibit 1. Other microscopy modalities, such as light-sheet and super-resolution microscopy, require separate and specific guidelines and considerations and are beyond the scope of this Perspective.

Single-point scanning confocal microscopy

The light path in a laser scanning confocal microscope is more complex than that in a conventional widefield microscope and entails additional hardware and settings that will impact acquisition and, ultimately, the interpretation of the results.

The hardware components required for optical sectioning (scanning mechanism, illumination and wavelength selection configuration, pinhole and detection) are usually integrated into a scan unit. The specific hardware components vary between different vendors and even in models available from the same vendor. Reporting the scan unit model will specify the capabilities and limitations of each system, enabling other researchers to reproduce or interpret a particular experiment. Some scan units provide different options (for example, type of scanning mechanism and detectors) that can impact maximum acquisition frame rate, noise level and overall SNR and thus should be reported.

Although reporting the scan unit provides fixed information on the hardware configuration, there are many user-defined settings that contribute to image quality, quantification and conclusions and also need to be reported to ensure reproducibility (Supplementary Exhibit 2). Unfortunately, many of these settings are not well understood and are incorrectly applied and/or incompletely reported, which undoubtedly affects reproducibility. Below are some examples of the most commonly overlooked or misunderstood settings that have a great impact on the image and require reporting.

In a laser scanning confocal microscope, the digital resolution is not limited by a physical photodiode size. Instead, the detector analog signal is digitized at regular intervals by the pixel clock. The final pixel size is determined by the total magnification, the pixel clock or frame size (how many intervals/pixels the scan area can be divided into) and the size of the scanned area (zoom factor). This offers flexibility but results in increased complexity, leading to irreproducibility if not properly reported. In many cases, researchers provide the frame size (for example, 1,024 × 1,024) instead of the actual pixel dimensions. This can be due to how the information is displayed in the acquisition software or to a poor understanding of the instrument. Regardless, the frame size alone does not provide crucial or reproducible information on the digital resolution of the image.

Most commercial laser scanning confocal microscopes can be used in uni- or bidirectional scanning modes. While the acquisition time can be reduced by half, bidirectional scanning may induce a pixel shift, especially at faster scan rates and higher zoom factors. This shift may not be apparent in the whole image, but careful inspection will show an offset between adjacent lines (Fig. 4e, left and inset, and Supplementary Appendix). Often, the scan mirrors can be carefully calibrated to reduce this pixel shift (Fig. 4e, right and inset), but reporting scanning directionality is still best practice.

A major setting that is often misused and under-reported is the offset. It may be tempting to adjust the offset to reduce fluorescence background and amplify the contrast for signal from the desired object. However, an improperly adjusted offset will lead to clipping of data intensities and removal of signal. If data clipping is substantial, entire structures within the sample can disappear from the image (Fig. 4f). Not reporting the offset will produce imprecise and conflicting quantitative measurements, such as the number of mitochondria per area, the width and area of each mitochondria and the overall intensity.

Another critical parameter to report is the pinhole diameter. It determines the thickness of the optical section, axial resolution and overall brightness and contrast. This information is typically given in Airy units (AU; Box 2 and Supplementary Table 2). It is important to note that the thickness of the optical section obtained by setting the pinhole to 1 AU will vary with wavelength. The pinhole diameter (and wavelength, if stated in AU) should be reported.

The complexity in reporting the acquisition settings and configurations in a laser scanning confocal microscope is compounded by the lack of standardization of the nomenclature utilized by different vendors. For example, the adjustable voltage of the detector, which modulates the amount of gain or amplification of signal, can be referred to as master gain, HV (high voltage) and gain or even displayed as the percentage of amplification (Supplementary Appendix). In some acquisition software, adjustment of the scan rate (pixel dwell time) is provided on an arbitrary unit scale instead of in the appropriate units of microseconds per pixel or Hz. Caution is essential when reporting the scan rate/dwell time in these systems as inaccurate reporting would contribute to a lack of reproducibility (Supplementary Appendix). Current important initiatives to improve metadata standardization are listed in Supplementary Table 1.

Multiphoton microscopy

Multiphoton microscopes share most of the hardware components of a laser-scanning electron microscope and therefore will be impacted by many of the same parameters. A notable difference is the absence of a pinhole, as optical sectioning is provided by the localized excitation produced by the nearly simultaneous absorption of multiple photons at the focal plane27.

In this modality, a pulsed laser is necessary to achieve the high peak intensity required to excite fluorophores by nearly simultaneous absorption of multiple photons. The pulse length (femtosecond scale) and repetition frequency are important parameters as they determine the peak intensity28. Oftentimes tunable lasers are used; the laser model, excitation wavelength, pulse length and repetition frequency should be reported to ensure reproducibility.

Many multiphoton systems have two light paths to collect and detect the emitted light. The first light path coincides with the traditional confocal light path (scan unit), whereas the second light path uses specialized detectors (non-descanned detectors, NDDs) that collect emitted light more efficiently than detectors within the scan unit. The image produced by the NDD will have higher SNR than that from a scan unit detector, other things being equal, and therefore the specific hardware and configurations need to be specified to ensure reproducibility (Supplementary Exhibit 3).

Spinning disk confocal microscopy

In spinning disk confocal microscopes, the sample is scanned simultaneously through multiple pinholes of fixed diameter on a disk. The spacing of these pinholes greatly impacts optical sectioning capabilities. Closely spaced pinholes reduce optical sectioning and contrast, whereas disks with pinholes spaced farther apart improve optical sectioning in thicker specimens but reduce sensitivity and frame rate and require longer integration times or higher illumination levels to generate images with high SNR.

The manufacturer and model of the scan unit determine the design of the disk, the diameter of the pinholes and the spacing between them. Some scan units provide multiple pinhole size and spacing options. Reporting the specific pinhole size and spacing used in the experiment is critical to ensure reproducibility (Supplementary Exhibit 4).

Guidelines on reporting sample preparation

Sample preparation is critical in fluorescence microscopy and will greatly influence the quality of the final image and its quantification29. There are many publications that focus on troubleshooting sample preparation and appropriate controls required for the interpretation of microscopy data in both fixed and live specimens3,5,9,17,29,30,31,32,33. Detailed sample preparation methods including fixation, permeabilization, labeling and mounting of the sample (Supplementary Exhibits 14) and validation steps should be documented when writing a manuscript. Here we highlight some examples of sample-related metadata that are critical for image acquisition in both fixed and live samples.

Of particular importance to report is the specific fluorophore used in any given experiment. Not only will the fluorophore’s properties heavily dictate the experimental design and the hardware required to accomplish a particular experiment34,35,36, but, notably, the fluorophore can also impact the outcome of the experiment itself. This is especially true for specific variants of a fluorescent protein. Even if spectral properties are similar, characteristics such as brightness, photostability, monomeric quality and maturation time are variant specific37. The choice of fluorescent protein and how the fluorescent protein is attached to the target protein (C terminus, N terminus, or specific linker sequences and lengths) may alter the localization, concentration, lifetime and/or functionality of the protein of interest, leading to potentially erroneous measurements and different experimental outcomes38,39. Consequently, reporting GFP as the fluorophore (instead of sfGFP or mGFPmut3) does not provide sufficient information to reproduce a particular set of experiments and can result in critically different results. The original papers describing the specific fluorescent protein’s development should also be cited if possible.

Sample optical properties contribute to the performance of the optics in the microscope and need to be documented. Objective lens corrections are engineered with a specific coverslip thickness in mind (usually 0.17 mm or grade no. 1.5). Use of a coverslip with a different thickness/grade can therefore negate those corrections13. Coverslip thickness tolerance (for example, 0.17 ± 0.005 mm) can be quite variable. Reporting the coverslip vendor and reference number therefore supports experimental reproducibility.

The mounting medium serves to homogenize the refractive index across the sample and match it to the requirements for the objective lens. The specifications of mounting media vary widely (refractive index, compatible fluorophores, curing, etc.). Mounting media can therefore impact the overall brightness, axial resolution and, in some cases, the final thickness of the sample, which will affect many aspects of quantification. Additionally, the refractive index of a mounting medium may change with time, and imaging in the same general window of time after mounting is recommended. Understanding how the mounting medium affects the sample and reporting its specifications will improve reproducibility.

Live-cell imaging enables the study of biological processes as they occur in real time though time-lapse imaging. In this application, maintaining the health of the sample should be the top priority to ensure rigor and reproducibility (Supplementary Exhibits 14 and Box 1). Most biological samples require an optimal temperature range to carry out biological processes. Precise control of the environmental conditions is critical to ensure normal cell physiology and function, including growth rate and molecular dynamics. The hardware used to maintain temperature, gas and humidity vary widely and will offer different stability and levels of control over these conditions. Thus, a description of the specific environmental conditions and how these are maintained will greatly improve accuracy, precision and reproducibility. The specific imaging medium and any additional components that may decrease phototoxicity or photobleaching (for example, antioxidants and reactive oxygen species scavengers) should be reported. Some components of the medium such as phenol red, fetal bovine serum, riboflavins and vitamins can produce substantial fluorescence background signal, limiting the ability to detect the signal of interest and impacting the accuracy and precision of quantitative measurements. Additionally, the concentration of a fluorescent dye and the solvent used or transfection reagents and expression of fluorescent protein fusions37 in a live sample may affect cell function40,41,42, induce synergistic effects with the conditions used (for example, drug treatments), and therefore impact reproducibility.

There is a wealth of publications offering insight on live-cell imaging that researchers can consult to better understand, optimize and troubleshoot aspects related specifically to this application4,18,33,43,44.

Notes on method validation

Understanding the capabilities and limitations of the particular application is essential in designing a successful microscopy experiment, as it enables researchers to identify the sources of measurement imprecision and mitigate its effects. In addition to carefully reporting microscopy methods, authors should aim to report critical steps for validation of the methodology. Box 2 describes some of the important validation steps that should be included in any experimental design and in the methods section, and Supplementary Table 1 lists selected resources for method validation. There are several outstanding publications that provide more information on method validation approaches and protocols3,5,17. Additionally, several initiatives in the microscopy field focus on the importance of quality control and instrument performance assessment to validate microscopy methods (Supplementary Table 1).

Notes on image format and processing and their impact on metadata

Image processing is used to enhance visualization and extract information from the digital image. While image processing is important for gleaning useful information (for example, during segmentation), it can irreversibly change the intensity levels and introduce nonlinear changes in shape and intensity across the image, compromising accurate quantification. Thus, all imaging processing steps and software settings must be reported45,46.

Most acquisition software options include image processing shortcuts to enhance image contrast (for example, haze reduction or smoothing) or other aspects of the image automatically or with a few convenient clicks of a button, often intended for presentation purposes while compromising the quantitative integrity of the image. Importantly, the resulting processed image is generally no longer faithful to the raw data collected, compromising reproducibility and accuracy in subsequent measurements, especially if the raw data are not saved alongside the processed image. Researchers should pay careful attention to software controls/modules used and ask experts (vendors or core scientists) for guidance in image processing. Most critically, the raw data image should always be saved.

Another aspect to consider is the file format used to acquire and export/save the images. Most commercial software produces a raw, uncompressed proprietary image format that contains all accompanying metadata recorded by the system. However, images can be exported in other file formats that may compress or otherwise scale or filter the data, reduce bit depth, alter the intensity levels acquired by the detector and lose the metadata, often irreversibly (for example, jpeg, mov, AVI). Ensuring a copy of the raw image in its original suggested proprietary file format is saved and backed up before converting to or exporting as a different file format is absolutely critical for quantitative microscopy. No matter how tempting it is to work with small file sizes, a compromise, often significant, has been made, even if unintentionally.

Resources for reporting microscopy methods

The examples provided in this Perspective serve as cautionary tales of how the lack of proper methods reporting can lead to conflicting data interpretation and thus irreproducibility. To facilitate proper microscopy methods reporting, we have assembled comprehensive checklists for the most common fluorescence light microscopy modalities (Supplementary Exhibits 14). Each checklist contains essential and recommended metadata to include in the methods section for any manuscript presenting data that were acquired using these modalities. Common examples are provided for each term, and a reference text example of a methods section write-up is also included after the checklist.

Additionally, to help researchers simplify and customize the essential metadata list according to their specific imaging and experimental design, we developed a microscopy metadata checklist generator tool (MicCheck). This tool guides researchers through simple questions related to their imaging choices and dynamically generates a checklist of essential and optional metadata that can then be copied and pasted into a text editor or downloaded in pdf format (Supplementary Exhibit 5). MicCheck is hosted online at, and relevant source files are available at the GitHub repository at In addition to online use, by downloading and modifying the example text file, core facilities or laboratories are also able to create their own versions of MicCheck with custom metadata examples specific to their microscope systems.

These checklists can also be used by researchers when designing imaging experiments to ensure that the relevant hardware and settings are properly identified and configured. They can also be used for reporting imaging methods, including during manuscript preparation or in the experimental design sections of funding proposals. Additionally, the checklists can be used by editors and reviewers to ensure sufficient detail is included to consider the experiments rigorous and reproducible.

Finally, for education and to help improve the adoption of these guidelines, we provide a visual infographic that can be printed and displayed in laboratories, core facilities or microscopy rooms to aid, remind and encourage researchers to compose detailed methods sections in their publications (Extended Data Fig. 1). The poster summarizes the essential metadata described in the checklists and the figures, depicting examples of how these hardware and settings can impact image quality and reproducibility.

Recommendations to improve methods reporting

While the description of minimal microscopy reporting guidelines is an essential first step toward improving methods reporting, it is not sufficient to solve the microscopy reproducibility crisis. The imaging scientific community needs to work together in a coordinated effort to improve methods reporting and the overall quality and reproducibility of image-based research, including researchers, imaging scientists, institutions, granting agencies, scientific journals and vendors. More rigorous and reproducible science results in better use of resources, improved data integrity and, therefore, fewer retracted publications and the ability to extend research by mining published results. Thus, there is a need to establish standards and provide resources to educate, inform and improve microscopy methods metadata reporting. Greater awareness and education will improve how microscopy data are collected, shared, validated, analyzed and reported.

Role for researchers

Researchers can contribute to improvement in microscopy methods reporting by (1) improving their knowledge of the microscopy techniques and appropriate validation steps used throughout their research; (2) upholding and requiring guidelines and recommendations set by the imaging community when reviewing manuscripts and grant applications; (3) involving imaging scientists during all steps of the research process, from experimental setup to data interpretation and manuscript preparation; (4) acknowledging core facilities and imaging scientists or considering collaborations that lead to authorship, as a way to enable the critical role core facilities fulfill in support of the scientific community; and (5) using data repositories to improve transparency and reproducibility and enhance data sharing47 (Supplementary Table 1).

Role for imaging scientists and core staff

Imaging scientists, especially those in core facilities, have a key role by sharing their technical expertise and providing intellectual contributions in all aspects of image-based science. In addition, they serve an educational mission by disseminating resources and tools to improve image-based research. Educating researchers on the importance of appropriate methods reporting will encourage best practices and participation by everyone. While core staff and imaging scientists have a critical educational role, this is not a substitute for formal training of graduate students and other researchers in proper metadata collection and reporting. Such training can take place at home institutions and in summer or other intensive programs focused on microscopy methodology and will greatly improve education and awareness broadly.

Imaging scientists are encouraged to stay informed about resources on education, rigor and reproducibility and to contribute to the development of guidelines and standards by many initiatives in the imaging community6,48,49,50. These initiatives (including those in Supplementary Table 1, as well as many others) can help in this task by presenting the resources in conferences, workshops and other educational initiatives they support.

Role for scientific journals and funding agencies

Scientific journals are urged to update their instructions to authors and have them reviewed by the imaging community, to ensure that adequate microscopy metadata information is included in the materials and methods. Removing word count limits from the methods section will allow authors to fully describe microscopy image collection and analysis, which should be written and reviewed with the same rigor as the entire paper. If space constraints cannot be adjusted, detailed methods descriptions can be included in supplementary information.

Encouraging or requiring proper recognition of the contribution of imaging scientists, as appropriate (acknowledgment or authorship), will facilitate researchers working closely with imaging scientists as standard practice. Incorporating imaging scientists in the peer review process, perhaps as technical experts, would help ensure that the experimental design, methodology and reporting of microscopy data support the conclusions of the manuscript.

Lastly, requiring appropriate analysis and quantification of imaging data (in addition to ‘representative images’) is long overdue. Quantification and validation of microscopy data will reduce bias and irreproducibility as well as the publication of artifactual results.


Rigorous and unbiased experimental design and analysis workflows are critical to provide accurate insight into the biological process under investigation3,5,9,17. Sample preparation, choice of instrument and related hardware, and image acquisition parameters (that is, metadata) have a profound effect on the image data validity and interpretation and therefore must be reported in the methods section of a published manuscript. However, microscopy methods reporting is often overlooked throughout the peer review process. In this work, we describe specific examples that highlight how lack of reporting of comprehensive information can affect the integrity and reproducibility of microscopy results. We present guidelines, checklists and resources to help researchers identify the critical metadata that should be included in their methods section, according to their specific experiment. The goal is not to put an unnecessary burden on researchers, but rather to help give readers of research papers enough information to assess the validity of biological findings and reproduce those findings independently. We hope that these reporting guidelines become second nature when carrying out microscopy experiments and reporting microscopy data and that implementing these guidelines will help improve the design of future experiments. We also hope that these guidelines provide a starting point for journal editors and peer reviewers when assessing microscopy data.

We recognize that addressing microscopy reproducibility is a complex, multifaceted issue that will require an ongoing and coordinated effort from everyone involved in scientific research. In particular, we recognize the contributions from core facilities and imaging scientists who are uniquely positioned to assist researchers in addressing these challenges by educating and supporting the scientific community3,5,6,9,17,44,45,48,49,50,51,52 (Supplementary Table 1). While this Perspective has focused on reporting of microscopy methodology, we emphasize that appropriate method validation and experimental design are critical to ensure the quality of quantitative microscopy data and the continued progress of feasible, rigorous and reproducible image-based science.

Reporting Summary

Further information on research design is available in the Nature Research Reporting Summary linked to this article.