Image reconstruction through a multimode fiber with a simple neural network architecture

Multimode fibers (MMFs) have the potential to carry complex images for endoscopy and related applications, but decoding the complex speckle patterns produced by mode-mixing and modal dispersion in MMFs is a serious challenge. Several groups have recently shown that convolutional neural networks (CNNs) can be trained to perform high-fidelity MMF image reconstruction. We find that a considerably simpler neural network architecture, the single hidden layer dense neural network, performs at least as well as previously-used CNNs in terms of image reconstruction fidelity, and is superior in terms of training time and computing resources required. The trained networks can accurately reconstruct MMF images collected over a week after the cessation of the training set, with the dense network performing as well as the CNN over the entire period.

Optical fibers have proven to be extremely useful for endoscopy and related applications 1,2 . Present commercial methods for transmitting images through fibers are based on single-mode fiber bundles 3,4 , consisting of thousands of fibers each transmitting a single pixel. It would be advantageous to instead transmit images in multimode fibers (MMFs), which are easy to fabricate and thinner than single-mode fiber bundles, and could potentially carry much more information. However, there is a serious drawback: due to mode-mixing and modal dispersion, any image coupled into a MMF is transformed into a complex speckle pattern at the output 5 . Researchers have devised various methods for reconstructing the input images from the speckle patterns, based on finding the complex transmission matrix of the MMF 6-10 or phase retrieval algorithms [11][12][13] . However, such methods generally require extra apparatus for measuring the optical phase, or have difficulty scaling to large image sizes.
Another promising approach is to use a training set of a priori known inputs to teach an artificial neural network (NN) how to map MMF output images to input images. This would not require additional interferometric equipment, and can potentially scale up to large image sizes. The idea was proposed and investigated decades ago [14][15][16] , but only in recent years has it been shown to perform well for reconstructing images of reasonable complexity [17][18][19][20][21] , aided by improvements in computational power and NN software.
These recent advances in NN-aided MMF image reconstruction have focused on deep convolutional neural networks (CNNs) [17][18][19][20][21][22] . Unlike traditional dense NNs 23 , CNNs use convolution operations instead of general matrix multiplication within the NN layers 24 , inspired by biological processes in visual perception. CNNs have enjoyed immense recent success in computer vision 25 , making it natural to investigate using them for MMF image reconstruction. They have also been applied to the related problem of image reconstruction in scattering media [26][27][28][29] . However, there are grounds to question how well-suited CNNs are for analyzing speckle patterns such as those produced by MMFs, which are very different from the natural images commonly dealt with in computer vision. In MMF images, information is encoded not just locally but in the global distribution of speckles 22,30,31 , whereas the localized receptive fields in convolutional layers are designed to extract relevant local features (such as edges) in natural images, rather than long-range spatial structures 32  www.nature.com/scientificreports/ NNs have previously been employed for controlling spatiotemporal nonlinearities in MMFs 33 , and for position sensing in a chaotic cavity 34 . This paper investigates the performance of dense NNs and CNNs for MMF image reconstruction. Whereas the earliest papers on NN-aided MMF image reconstruction used dense NNs [14][15][16] , most recent studies have concentrated on using CNNs [17][18][19][20][21][22] 35 ). To our knowledge, there has been no direct comparison between the two NN architectures in the context of MMF image reconstruction. The recent popularity of CNNs for this task is predicated on the local feature extraction capability of CNNs being useful for descrambling MMF images. Our comparison of dense NNs and CNNs should be useful to the community in testing this assumption.
Our principal comparison is between (i) the single hidden layer dense neural network (SHL-DNN), one of the simplest dense NN architectures, and (ii) U-Net, a CNN originally developed for biomedical imaging 36 , which has recently been used for MMF image reconstruction 18 . We do not compare very deep CNNs such as Resnet 18 or generative adversarial networks 37 , as these require much greater computational resources and longer training times, and thus seem ill-suited to the MMF image reconstruction problem. After optimizing both types of NNs (SHL-DNN and U-Net), we find that the SHL-DNN achieves a similarly high reconstructed image fidelity with shorter training time and less network complexity than the U-Net. For one of our reference datasets, SHL-DNN achieves a saturation Structural Similarity Index Measure (SSIM) 38 of 0.775 in 16 minutes and the U-Net achieves a saturation SSIM of 0.767 in 3.5 h on the same computer. The SHL-DNN has 20 million trainable parameters, and the U-net has 31 million trainable parameters. We also validated both NNs using images collected up to 235 h after the images in the training set; both NNs continue to perform well in image reconstruction. Moreover, we tested a "VGG-type" NN, which combines convolutional and dense layers, and found that it offers no additional performance advantage over the SHL-DNN.

Experimental setup
Multimode fiber image reconstruction. The optical setup is shown in Fig. 1a. A collimated beam from a diode laser with an operating wavelength of 808 nm (Thorlabs LP808-SF30) is expanded and directed onto a spatial light modulator (SLM) (Hamamatsu X13138-02). Along with two orthogonal polarizers, the SLM generates a programmable spatial modulation in the intensity of the light beam.
The modulated beam is coupled into a one meter long multimode fiber (MMF) (Thorlabs FT400EMT) via a matching collimator (NA 0.39). The distal end of the MMF is imaged with a CMOS camera (Thorlabs DC1545M). The camera images consist of complicated speckle patterns, as shown in the left panel of Fig. 1b, with no apparent relation to the ground truth images from the SLM. The camera images have 1280 × 1080 pixel resolution; to obtain a tractable dataset, we crop and downsample to 64 × 64 using the Lanczos algorithm 39 , as shown in the right panel of Fig. 1b.
By operating the SLM with a refresh rate of 0.9 Hz (which allows for the generation of stable and distortionfree images), we accumulate one dataset of 61524 MMF images collected over approximately 19 h for training and several datasets that spans across 235 h. The ground truth images are drawn equally from (i) the MNIST digit dataset containing handwritten digits in various styles 40 , and (ii) the MNIST-Fashion dataset containing images of clothing and apparel 41 . The MNIST digit dataset is used for most of the experiment; the MNIST-Fashion dataset is used in "Transfer learning and alternate image set" section.
The MNIST and MNIST-Fashion ground truth images are 28 × 28 , whereas the MMF-derived images in the dataset are 64 × 64 . Conceptually, there is no reason to restrict the MMF images (NN inputs) to the same size as the ground truth images (and NN outputs), as was the practice in earlier studies 17,18 . Intuitively, higher resolutions for the MMF images should be advantageous, as the image reconstruction algorithm is given more information to work with, subject to the constraints of trainability and computer memory capacity. The effects of varying the input size are studied in "Image reconstruction fidelity" section. Both the ground truth and MMF images have 8 bits of dynamic range.
Neural networks. We mainly investigate and compare two NN architectures for efficacy in MMF image reconstruction: a single hidden layer dense neural network (SHL-DNN) and the convolutional neural network U-Net. (A third architecture, a hybrid convolutional/dense network, is briefly discussed in "Hybrid neural network" section).
Dense NNs are the most elementary architecture for NN-based machine learning. The earliest papers on NN-aided MMF image reconstruction utilized dense NNs 14-16 , but were constrained by the lower levels of computational power then available. We implement the SHL-DNN shown in Fig. 1c, featuring a hidden layer of 4096 nodes sandwiched between input and output layers, with dense interlayer node connections. Each 64 × 64 input image is flattened and inserted into the input layer, which has 64 2 = 4096 nodes. The hidden layer and output layer have sigmoid activation functions. The result from the output layer (which has 28 2 = 784 nodes) is reshaped into a 28 × 28 image that can be compared to the ground truth image.
Convolutional neural networks (CNNs) have been applied to the MMF image reconstruction problem by several recent authors [17][18][19][20][21][22] . Here, we employ the U-Net architecture, which Rahmani et al. have previously used for MMF image reconstruction with the MNIST digit dataset 18 . As shown in Fig. 1d, the input is 64 × 64 and the output is 28 × 28 , the same as for the SHL-DNN. The network consists of a sequence of convolutional and pooling layers leading to a 4 × 4 × 1024 intermediary layer, followed by a sequence of convolutional upsampling layers. A 64 × 64 × 64 convolutional layer followed by a 2 × 2 max pooling layer downsamples the 64 × 64 input image to 32 × 32 before the 32 × 32 × 128 convolutional layer of the U-Net. Batch normalization is applied after www.nature.com/scientificreports/ each convolutional layer. Each convolutional layer has a ReLU activation function, and the output layer has a sigmoid activation function (similar to the SHL-DNN). We follow the typical U-Net architecture design rule 36 wherein a halving of the layer dimensions is accompanied by a doubling of the number of filters (image depth), and vice versa. There are also auxiliary skip connections that aid image localization 36 . The U-Net architecture contains numerous hyperparameters such as the number of layers, convolutional filter depths, batch size, etc. We tested the effects of varying these hyperparameters, and the "baseline" configuration shown in Fig. 1d gives the best results. Notably, in this configuration the filter depths are four times what was used in Rahmani et al. 18 .
Each NN is trained using Adam optimization with a batch size of 256 images, and an early stopping condition of 100 epochs after validation losses stop improving. We find that batch-normalization regularization is crucial for the U-Net to perform well, but dropout regularization is better for the SHL-DNN. Little performance improvement is observed when the batch size adjusted between 128 and 1024; a much larger batch size (27685) drastically lengthens training times. For the objective function, the NN output is compared against the ground truth (MNIST digit or MNIST-Fashion) image via the Structural Similarity Index Measure (SSIM) 38 , a wellestablished metric for quantifying the similarity between structured images (seeSupplementary Materials). All training was performed on the same computer (Intel Xeon Gold 5218 with NVIDIA Quadro RTX 5000 GPU).

Results
Image reconstruction fidelity. We train the SHL-DNN and U-Net using 30762 MMF images from the first 19 h of the data collection run. The ground truth images drawn from either the MNIST digit dataset 40 or the MNIST-Fashion dataset 41 ; separate instances of each network are trained for the two respective datasets. In each case, we assign 27685 images for training and the remaining 3077 for validation. The training and validation images are initially drawn randomly from across the collection period (the role of collection time will be investigated later, in "Performance over time" section). Figure 2 shows the results of MMF image reconstruction for six representative images from the validation set, three from the MNIST digit datasets and three from the MNIST-Fashion sets. The fully-trained SHL-DNN www.nature.com/scientificreports/ and U-Net both recover the ground truth images with remarkable fidelity (Fig. 2a,c-d), despite the lack of human-discernable patterns in the MMF images (Fig. 2b). The SHL-DNN and U-Net achieve similar fidelity, as corroborated by the similar SSIM scores, for both types of images. Figure 3a shows the SHL-DNN and U-Net training curves for the MNIST digit dataset (the results for MNIST-Fashion are similar; see Supplementary Materials). We plot the training curves against elapsed time to allow for fairer comparisons, since the two networks have very different training times per epoch. The performance of the SHL-DNN saturates at SSIM 0.775, comparable to SSIM 0.767 for the U-Net. Figure 3b,c, we compare the performance of both networks on two other common metrics: the mean squared error (MSE) and the resulting classification error for the validation set (however, the training still uses SSIM for the objective function). The classification error is meant to characterize the overall legibility of the reconstructed digits, and is obtained by passing the NN outputs to an auxiliary digit classifier (mnist_cnn.py from Keras 42 ). The results from these alternative measures are similar to what was obtained from the SSIM. The SHL-DNN achieves MSE 2.25 × 10 −2 and classification accuracy 0.90, while the U-Net achieves MSE 2.48 × 10 −2 and classification accuracy 0.90.
Although the two networks yield similar image reconstruction fidelity, the SHL-DNN can be trained more quickly. To reach its saturation SSIM (i.e., triggering of the stopping condition), the SHL-DNN takes 462 epochs and 16 minutes, whereas the U-Net takes 318 epochs and 3.5 h. The training time per epoch is 20 times faster for the SHL-DNN.
We systematically investigated the effects of various NN settings, and found that no further major performance improvements are achievable without increasing the training set size. (For these hyperparameter studies, a smaller training set of 8709 images was utilized.) For the SHL-DNN, the choice of input image size appears to play an important role. As shown in Fig. 3d, for a smaller input image size ( 28 × 28 ) the SSIM saturates at a lower value, which can be ascribed to the NN having less information available for image reconstruction. But having inputs that are too large, such as 84 × 84 , also leads to a lower SSIM compared to our baseline choice of 64 × 64 . The SHL-DNN performance decreases when the number of hidden layer nodes is reduced below the baseline value, as shown by the red curve in Fig. 3d for the 512 node case. On the other hand, further increasing the number of hidden layer nodes increases the training time without significant improvement in the saturated SSIM (see Supplementary Materials). Moreover, the number of hidden layer nodes seems to have negligible influence on the optimal input image size.
As for the U-Net, one setting that notably affects performance is the number of convolutional filters. We denote the number of U-Net filters used in Rahmani et al. 18 as " 1× ". The saturated SSIM score increases as the number of filters in increased up to 4× , which is the baseline value that we adopted. Further increases in the number of filters leads to a substantial increase in training time, without significant performance improvement. Another possible setting is the number of convolutional layers; we verified that deeper or shallower U-Net structures adversely affect the performance. Moreover, we find that removing the skip connections leads to a www.nature.com/scientificreports/ slight decrease in performance slightly; hence, the skip connections are included in our baseline configuration (although the reason these connections are advantageous for the present task is somewhat unclear). Some of these comparisons are shown in Fig. 3e. After these optimization studies, we arrive at the SHL-DNN and U-Net configurations shown in Fig. 1c,d. In these configurations, the SHL-DNN has 20 million trainable parameters and takes 39.9 million FLOPs per forward pass, and the U-net has 31 million trainable parameters and takes 62.8 million FLOPs per forward pass.  Fig. 1d] (brown), removed skipping layers (light green), " 3× " filters (dark green), and " 1× " filters (pink). In the last case, convergence is achieved about as quickly as the SHL-DNN, but at significantly lower SSIM. www.nature.com/scientificreports/ Performance over time. It is interesting to ask whether the image reconstruction ability of the NNs is persistent, or whether it degrades over time due to a drift in the MMF's transmission characteristics. Such temporal changes can be caused by thermal and mechanical perturbations of the environment, which induce minute deformations of the fiber.
To address this question, we validate the NNs (trained using images from the first 19 h of the dataset) against images collected during the subsequent 235 h. The results are shown in Fig. 4. The validation data are sorted by collection time and batched into 5 minute intervals.
In terms of both SSIM and digit classification accuracy, the image reconstruction performance for both NNs fluctuates over time, but is overall remarkably robust. It can be noticed in Fig. 4a,c that the performance fluctuations for the SHL-DNN and U-Net are correlated over time. In fact, their SSIM scores have a correlation coefficient of 0.950. This implies that the performance fluctuations are caused by the MMF undergoing physical fluctuations in its transmission characteristics (relative to the training set), which simultaneously impacts the performance of both NNs. Over the 235 h period, we observe only a slight long-term degradation in performance (both in terms of SSIM and digit classification accuracy), indicating that there is neligible sustained "drift" in the MMF's transmission characteristics. Over the entire experimental period, the SHL-DNN and U-Net consistently have similar performance, with SSIM variance of about 0.02.
Hybrid neural network. Rahmani et al. 18 studied the use of another type of NN for unscrambling MMF images: a hybrid convolutional and dense network of the type pioneered by Oxford's Visual Geometry Group (VGG). VGG-type networks are typically used for classification 43 , and they were used in Ref. 18 for digit classification with the MNIST digit dataset. In this paper, we are mainly interested in image reconstruction rather than classification. Nonetheless, it is helpful to study the performance of a VGG-type network for this purpose, as a further test of the usefulness of convolutional layers for extracting structural information from MMF images. www.nature.com/scientificreports/ We implement a simple VGG-type network as shown in Fig. 5a, consisting of two convolutional layers, a hidden dense layer with N h nodes, and a dense output layer. Figure 5b shows the training curves for VGG-type networks with several choices of N h , as well as for the baseline SHL-DNN. When N h is equal to the number of hidden layer nodes in the SHL-DNN, the saturated SSIM is 0.71-comparable to but certainly not better than the SHL-DNN (SSIM 0.775). For smaller values of N h , the performance is substantially worse. We also investigated reversing the configuration by placing the dense layers at the input and the convolutional layers at the output, but this not produce any improvement. These results seem to bolster the case that convolutional input layers do not provide additional benefits for MMF image reconstruction, a point that will be further discussed in "Discussion" section.
Transfer learning and alternate image set. Transferability is a common concern in machine learning.
In the present context, one may ask whether NNs trained using one kind of ground truth image-say, MNIST digits-can successfully reconstruct more general images. In other words, are the networks broadly capable of undoing the effects of mode mixing in the MMF, or are they merely recognizing patterns that are highly specific to the sort of images in the training set?
To investigate this, we train the SHL-DNN by withholding one digit from the MNIST digit dataset, and validating it against the omitted digit. Figure 6a shows representative results for the case of an omitted digit '9' . Although this SHL-DNN has not seen any examples based on the digit '9' , it reconstructs the images reasonably well, albeit with lower SSIM, as shown in Fig. 6b. Here, the training set (with '9' excluded) has 14565 images, and the other network settings are the same as in the baseline network described in "Image reconstruction fidelity" section. Over 1000 instances of the digit '9' , the mean SSIM is 0.72, compared to SSIM 0.86 for a validation set  www.nature.com/scientificreports/ of 2913 images that exclude the digit '9' . The performance of U-net is quite similar to SHL-DNN, as shown in Fig. 6c: the mean SSIM is 0.70 over 1000 instances of the digit '9' . When we attempt to reconstruct MNIST-Fashion images using a SHL-DNN trained on MNIST digits, or vice versa, the results are extremely poor (SSIM close to zero). Likewise, when we attempt to reconstruct images consisting of random uncorrelated pixel intensities, all three trained networks (SHL-DNN, U-Net, and VGGtype) give very poor results; over 1000 images, the MSE is in the range of 0.08-0.09 for all the three networks, comparable to the nascent training stage of Fig. 3b.

Discussion
We find that CNNs offer no performance advantage over the traditional dense NN architecture for MMF image reconstruction. In fact, the tested SHL-DNN is able to produce the same results as U-Net with much shorter training time and less network complexity, and this seems to be robust over various different NN settings. Moreover, a VGG-type hybrid convolutional/dense NN offers no obvious improvement over the SHL-DNN. These results suggest that the SHL-DNN, though a simple architecture, already gives a ceiling for the performance of neural networks on MMF image reconstruction. For practical real-time imaging applications, simpler NN architectures may be desirable as they can be trained more quickly and with fewer computational resources.
Our interpretation of the situation is that convolutional layers, though well-suited to extracting local features in natural images, do not provide any special advantage in processing speckle patterns of the sort produced by MMFs 30,31 . These speckle patterns are known to contain global correlations created by the superposition and dispersion of different transmission modes in the MMF 44,45 . By design, CNNs excel at processing local features and do not perform as well in processing global features 46 . It would be interesting to explore modifications to the CNN scheme, or preprocessing schemes for the speckle pattern, to improve performance 22 .
The trained NNs can reliably reconstruct images collected long after the training set. Specifically, after training over a period of 19 h, robust image reconstruction can be achieved over the subsequent 235 h, with no degradation corresponding to a long-term drift in the fiber transmission characteristics. The implication is that the 19 h training set is sufficiently representative of the perturbations (thermal fluctuations and weak dynamic disturbances) that can occur during the validation period 19,47 . Fan et al. have suggested that NNs can even be trained to perform image reconstruction over large varieties of deformations and configurations, such as twisting the MMF in different ways 19 .
The NNs perform poorly on images that are too different from those in the training set, which is a common problem with NN-based machine learning. Recently, Caramazza et al. have demonstrated using an optimization algorithm to learn the complex transmission matrix for MMF image reconstruction 35 , which bypasses the transfer learning limitations of the NN approach. However, this method requires much more computer memory, and the resulting image fidelity is lower; from our testing based on the same dataset from MNIST digits, the SSIM scores are ∼ 0.42 , compared to ∼ 0.75 for the SHL-DNN. In the future, it would be interesting to attempt to combine these two approaches in a way that overcomes their individual limitations.