Abstract
Free-space optical communications systems suffer from turbulent propagation of light through the atmosphere, attenuation, and receiver detector noise. These effects degrade the quality of the received state, increase cross-talk, and decrease symbol classification accuracy. We develop a state-of-the-art generative neural network (GNN) and convolutional neural network (CNN) system in combination, and demonstrate its efficacy in simulated and experimental communications settings. Experimentally, the GNN system corrects for distortion and reduces detector noise, resulting in nearly identical-to-desired mode profiles at the receiver, requiring no feedback or adaptive optics. Classification accuracy is significantly improved when these generated modes are demodulated using a CNN that is pre-trained with undistorted modes. Using the GNN and CNN system exclusively pre-trained with simulated optical profiles, we show a reduction in cross-talk between experimentally-detected noisy/distorted modes at the receiver. This scalable scheme may provide a concrete and effective demodulation technique for establishing long-range classical and quantum communication links.
Similar content being viewed by others
Introduction
The field of free-space optical (FSO) communications provides an exciting route forward in wireless information transfer by making use of various multiplexing schemes, including frequency and wavelength division multiplexing, and more recently spatial multiplexing1,2,3,4,5,6. A common method to implement the latter involves making use of orbital angular momentum (OAM), which is a degree of freedom that is in principle unbounded, thus permitting for the application of large alphabets in optical communication schemes. For example, by generating and transmitting various superpositions of OAM states, which result in different “petal pattern” images in the spatial domain, the alphabet size of the communications system may be significantly increased. An integral aspect of any communication protocol, however, is the ability to effectively demodulate the signal at the receiver, which in this case corresponds to determining which OAM superposition was sent and received. In real-world schemes, FSO communication systems comprise of propagation of the beams through random turbulence, attenuation, and non-negligible amount of dark noise at the receiver. As a consequence, the distorted, noisy received signals (images) can result in a lower accuracy, significantly deteriorating their practical implementation7,8.
State-of-the-art generative models have recently been applied to molecular design, radiotherapy, geophysics, speech recognition, and tomography9,10,11,12,13,14. Generative neural networks (GNNs) involve creating a new state at their output, given a noisy or unknown state as the input. One example of the generative learning approach is making use of an autoencoder in a variety of reconstructing scenarios15,16,17. In addition, supervised learning and artificial neural networks as classifiers have been successfully implemented in the context of FSO communications18,19,20,21,22,23,24,25,26. However, in order to train the network as a classifier, all preceding communication set-ups involving only supervised learning schemes require a huge amount of pre-labeled distorted optical profiles at the receiver, which is computationally less efficient and requires more time to generate and process the data. Additionally, in the presence of random distortions, attenuation, and detector noise, it may not be always feasible to tag the optical profiles for supervised learning. This restricts the classification efficiency of the set-up with respect to randomly varying turbulent effects and attenuation in realistic FSO systems.
In this article, we expand significantly upon these works and develop a communication scheme using a generative machine learning approach and demonstrate its robustness and ability to significantly mitigate the effects of turbulence, attenuation, and noise on the accuracy in both simulated and experimental communication settings. This receiver-end system is shown to be effective for a wide range of turbulence and detector noise strengths and requires no feedback to the transmitter of the communication link or any adaptive optics components. Our scheme with the developed generative network first creates new, significantly less distorted optical profiles at the receiver that are later demodulated using a convolutional neural network (CNN) classifier. Note that CNN is solely trained with only undistorted optical modes with added dark noise to simulate a realistic, non-ideal quantum efficiency detector, alone. Additionally, we calculate the cross-talk between the noisy, distorted modes at the receiver and show a significant enhancement when the GNN system is used. Furthermore, our network architecture is portable, cost-effective, and can be pre-trained before performing the communications, which circumvents the general technical issues that are present in adaptive optics.
Results
Model
The general experimental set-up is shown schematically in Fig. 1. A laser is incident on a spatial light modulator (SLM) with a given phase mask, such that the resultant optical spatial mode profile is in a desired superposition of OAM values ranging from 0 (Gaussian) to ±10 (a petal pattern with 20 lobes). The optical profile is then transmitted through turbulence via a second SLM, resulting in a distorted profile, after which it travels through a variable attenuator. Finally, detector dark noise is unavoidably added at the receiver. This noisy, attenuated, distorted image is then fed into the GNN. The GNN generates a new mode profile that may either be used to directly calculate its mean squared error (MSE) from the target or be fed into a CNN that classifies which mode was sent and received (that is, which letter of the communication alphabet). In this work, we produce data in three manners: test and training images both completely simulated (see “Numerical simulation of long distance turbulent propagation” in “Methods”), test and training images both experimentally generated (i.e., with a laser and two SLMs as shown in Fig. 1), and a combination of the two (i.e., simulated training images and experimentally generated test images).
As opposed to the traditional autoencoder27, convolutional denoising autoencoders are able to reconstruct a clean, corrected input from those that are partially distorted28. The idea behind using such a network design is to learn a hidden representation and extract the important features that are robust to noise or distortion present in the inputs. The generative network described here encodes and compresses the extracted crucial features from the input data into a smaller size layer. This smaller dimensional layer is a latent space. The encoded information in the latent space is then forwarded to the decoder, which eventually generates the corrected undistorted, clean modes as shown in Fig. 2a. We have built an encoder with convolutional layers (green blocks in Fig. 2a) with a kernel size of 5 × 5, with zero padding, rectified linear unit (ReLU) activation, stride length of 2, and 3 feature mappings followed by a max-pooling layer (red block) with a pool size of 2 × 2 and fully connected layer (blue circles). After this, the encoder finally stores the features into a latent space using the ReLU activation. Likewise, the decoder is built and starts with a fully connected layer (blue circles), which then forward the information to a convolutional layer with the same parameter settings as described above. Furthermore, in order to regain the original size of the input, a deconvolutional layer (magenta block) is applied, again with the same parameter values. In the end, a convolutional block with a single feature mapping generates a clean, corrected mode profile. Note that we apply a dropout with a rate of 5% after each layer, except at the fully connected layer at the end of the encoder and final convolutional layer of the decoder. We apply the small dropout rate to avoid overfitting, as well as the possible loss of features extracted from the convolution. The size of the fully connected layer is same as that of the latent space. As a classifier, we implement a CNN to demodulate the generated, reconstructed clean mode profiles. The same CNN is also implemented to demodulate the uncorrected detector received profiles at the receiver. This network comprises of a convolutional unit with a kernel of size 5 × 5, zero padding, ReLU activation, stride length of 2, and a single feature mapping followed by a max-pooling with a 2 × 2 filter attached to a fully connected layer (28 × 28 neurons) and an output layer as shown in Fig. 2a. No dropout is employed in this network.
Numerical results
Here we describe results from the numerical simulations. Examples of the simulated distorted, noisy images and GNN-generated corrections are shown in Fig. 2b–d, for varying degrees of signal-to-noise ratios (SNRs), communication link distances (Z), and turbulence strengths (\({C}_{n}^{2}\)), respectively. This process is repeated many times for all spatial modes (with random turbulences and noises added), and the accuracy of the system is calculated and compared to the accuracy when the GNN detection system is not used. Note that accuracy measures the ratio of number of correctly classified unknown OAM profiles to the total number of received modes at the receiver. In the first place, we optimize the latent space dimension and evaluate an accuracy for various SNRs of the detected OAM profiles. In order to generate training sets, we simulate 99 random turbulent phase screens with a strength of \({C}_{n}^{2}\) of 5 × 10−14 m−2/3 and transmission distance of 500 m. Note that the 99 simulated phase screens are all different from one another with respect to their phase distributions, such that two different turbulent phase screens produce different scintillation effects on OAM mode propagation even if they have the same turbulence strength. As a result, we have 99 different distorted optical profiles for each superposition OAM mode ranging from ℓ = 0 to ±10 for a total of 1089 images. The resolution of the images is fixed to 128 × 128 pixels for all of the simulations performed in this paper. Additionally, in order to simulate a transmitter with a fixed transmission intensity, we normalize the total intensity before we propagate the OAM modes into the turbulence. We append random additive Gaussian noise to the distorted, detected optical profiles at the receiver. Then the SNR of the final noisy, distorted optical profiles is measured as discussed in ref. 24. Note that we train the GNN separately for separate SNR image sets (and that the SNR is decreased by increasing the amount of added detector noise, as the transmitter intensity is held constant). Next, in order to train the networks, the set of images is divided into separate training and test sets. As a result, 50 images for each value of ℓ for a total of 550 images are used in training the network, whereas 49 images, again, for each value of ℓ for a total of 539 images are used for testing the networks. Consequently, we optimize the GNN with these training sets with a learning hyper-parameter (rate) set at 0.008. Then unknown test sets are fed to the optimized GNN that generates nearly ideal corrected optical profiles as the output, some examples of which for \({C}_{n}^{2}\ =\ 7\, \times 1{0}^{-14}\ {m}^{-2/3}\) are shown in Fig. 2b. The left, middle, and right columns in Fig. 2b represent the noisy received modes (R) and reconstructed (C) profiles with a GNN when the average SNR of images are 0.11 dB (σ = 20), −3.87 dB (σ = 50), and −5.91 dB (σ = 80), respectively. Next, in order to calculate the accuracy, the noisy test optical images (without corrections) and reconstructed/generated images from the GNN are forwarded to a pre-trained CNN and the corresponding accuracies are measured. Note that the pre-trained GNN–CNN networks reconstruct and then classify an optical profile on the order of milliseconds. The accuracy of the test set images with and without the GNN at various SNR levels of 3.11 to −5.91 dB with respect to different latent sizes of the GNN is shown in Fig. 3a. We find an improvement in accuracy from 0.87 to 0.93 for the image sets with SNR = 3.11 dB even at a small latent size of 4 × 4. As expected, we obtain better reconstructions and more improved accuracies as we increase the latent size up to 32 × 32, after which it begins to saturate. The reconstructed images shown in Fig. 2b are from the GNN with a 32 × 32 latent size. Finally, we achieve an improvement in accuracy from 0.87 to 0.97 at a latent size of 24 × 24, from 0.84 to 0.95 at a latent size of 24 × 24, from 0.77 to 0.92 at a latent size of 24 × 24, and from 0.68 to 0.86 at a latent size of 72 × 72 for the image sets with SNR = 3.11, 0.11, −3.87, and −5.91 dB, respectively.
Next we find the accuracy improvement with respect to different latent sizes over various communication link distances. The same strength of turbulence as described in previous paragraphs is used but again with different, random phase patterns. Here all optical mode profiles are assumed to be detected with an average SNR = −3.87 dB at the receiver. Noisy and reconstructed OAM profiles from the pre-trained GNN for distances of 400, 600, and 800 m are shown in Fig. 2c. The improvement in accuracy with respect to various latent sizes of the GNN at communication link distances from 200 to 800 m is shown in Fig. 3b. Here we find significant improvements in the accuracies from 0.976 to 0.998 at a latent size of 40 × 40, from 0.85 to 0.95 at a latent size of 88 × 88, from 0.72 to 0.88 at a latent size of 40 × 40, and from 0.67 to 0.85 at a latent size of 24 × 24 for the communication link distances of 200, 400, 600, and 800 m, respectively.
Next, we turn to reconstructing severely distorted OAM profiles due to various turbulence strengths in the channel. Turbulence strengths are varied from strong (\({C}_{n}^{2}\ =\ 1\times 1{0}^{-13}\ {m}^{-2/3}\)) to weak (\({C}_{n}^{2}\ =\ 9\times 1{0}^{-15}\ {m}^{-2/3}\)) levels with a fixed Z = 500 m, for a total of 7 different turbulence strength scenarios, with random phase screens for each class and different noise strengths at the receiver. Some examples of simulated received noisy OAM modes and reconstructed corrected modes using the GNN with a latent size of 32 × 32 for \({C}_{n}^{2}\ =\ 5\times 1{0}^{-14}\ {m}^{-2/3}\), \({C}_{n}^{2}\ =\ 7\times 1{0}^{-14}\ {m}^{-2/3}\), and \({C}_{n}^{2}\ =\ 1\times 1{0}^{-13}\ {m}^{-2/3}\) at a dark noise strength of σ = 50 are, respectively, shown in left, middle, and right column of Fig. 2d. Again with the pre-trained GNN with a latent size of 32 × 32, we show a significant improvement in accuracies for various turbulence strengths for noise strengths of σ = 10, σ = 20, σ = 50, and σ = 80 present at the receiver, as seen in Fig. 3c. Significant increase in accuracies are achieved with the GNN, for example, increasing the accuracy from 0.83 to 0.971 at a noise level of σ = 80 and turbulence strength of \({C}_{n}^{2}\ =\ 3\times 1{0}^{-14}\ {m}^{-2/3}\). Furthermore, for an intermediate turbulence strength of \({C}_{n}^{2}\ =\ 1\times 1{0}^{-14}\ {m}^{-2/3}\), we find an uncorrected accuracy of 0.995 and 0.945, respectively, at σ = 50 and σ = 80, which are dramatically enhanced to 1.0 with the GNN. Additionally, the accuracy of 0.989 at σ = 80 for a weak turbulence of \({C}_{n}^{2}\ =\ 9\times 1{0}^{-15}\ {m}^{-2/3}\) has also been corrected to an accuracy of 1.0 with the GNN as shown in the lower inset of Fig. 3c.
Similarly, in order to have a latent size benchmark with respect to different turbulent strengths, we vary the latent size of the GNN and fix the noise strength at σ = 50. Again with the GNN pre-trained for various latent sizes, the enhancements in accuracy with respect to latent sizes are shown in Fig. 3d. We find an improvement in accuracy from 0.995 to 1.0 for every latent size at or above 16 × 16 for \({C}_{n}^{2}\ =\ 1\times 1{0}^{-14}\ {m}^{-2/3}\) (red curve). For example, the accuracy at a latent size of 24 × 24 are zoomed in and shown in the inset. Moreover, we show a significant accuracy enhancement from 0.77 to 0.92 at a latent size of 24 × 24 (blue curve) and from 0.57 to 0.74 at a latent size of 40 × 40 (green curve), respectively, for an intermediate turbulence strength of \({C}_{n}^{2}\ =\ 5\times 1{0}^{-14}\ {m}^{-2/3}\) and strong strength of \({C}_{n}^{2}\ =\ 1\times 1{0}^{-13}\ {m}^{-2/3}\) in Fig. 3d. Note that, from the previous benchmarks of simulated OAM profiles, we set the size of 32 × 32 as the optimized latent space dimension of the GNN for the results discussed in the following paragraphs.
Experimental results
With the network optimized, we now focus on implementing the GNN in the experimental set-up as shown in Fig. 1. Initially, we generate light in OAM superposition modes from ±l = 0 to 10 on the first spatial light modulator (SLM 1), then mimic propagation through turbulence via a random phase mask on SLM 2 (see Fig. 1) that imparts a turbulence strength from \({C}_{n}^{2}\ =\ 8\times 1{0}^{-11}\) to 80 × 10−11 m−2/3. In order to generate training and test sets, we randomly vary the phase mask on SLM 2 and record the corresponding intensity pattern on a charged coupled device (CCD) camera. For a given \({C}_{n}^{2}\), we record 50 images for each superposition OAM value, for a total of 550 intensity images. We split these 550 images into 495 (45 from each ±ℓ) for the training set and 55 (5 from each ±ℓ) for the unknown testing set. Note that, to create a target pattern, we place an empty (zero turbulence) phase mask on SLM 2 for all ±ℓ. With the pre-trained GNN, we find an efficient reconstruction of the noisy, distorted OAM modes: typically an order-of-magnitude reduction in the MSE, as shown in Fig. 4. Examples of uncorrected noisy/distorted images at the receiver (R) and the corresponding corrected optical profiles (C) for the turbulence strengths of \({C}_{n}^{2}\, =\, 22.4\times 1{0}^{-11}\ {m}^{-2{/}3}\) (left), \({C}_{n}^{2}\, =\, 51.2\times 1{0}^{-11}\, {m}^{-2{/}3}\) (center), and \({C}_{n}^{2}\ =\ 80\times 1{0}^{-11}\ {m}^{-2/3}\) (right) are shown in Fig. 4a. The superposition OAM value in each row is ±ℓ = 2, 3, 6, and 9 from top to bottom, respectively. In order to measure the detection success with (blue) and without (tan) the GNN at the receiver, we calculate the MSE between the desired and the corrected or received optical profiles. While there is an expected trend of increasing MSE with increase in turbulence strength, \({C}_{n}^{2}\), the GNN continues to efficiently reconstruct the noisy and distorted OAM modes for all values of \({C}_{n}^{2}\) as shown in Fig. 4b.
Using the GNN, we demonstrate a significant reduction in mean MSE from 330.2 to 34.2 and from 742.3 to 120.7 for the weak turbulence strength of \({C}_{n}^{2}\ =\ 8\times 1{0}^{-11}\ {m}^{-2/3}\) and strong turbulence of \({C}_{n}^{2}\ =\ 80\times 1{0}^{-11}\ {m}^{-2/3},\) respectively. In order to gain insight into the dependence of our GNN’s performance on OAM value, we show the received (tan) and corresponding corrected (blue) MSE values for each ±ℓ from 0 to ±10 with a fixed \({C}_{n}^{2}\) of 51.2 × 10−11 m−2/3 in the inset of Fig. 4b. We reiterate here that the test intensity images (or unknown random turbulence phase masks) are not contained in the training set.
To further mimic the detrimental effect of loss in FSO communications, we fix the turbulence strength at \({C}_{n}^{2}\ =\ 51.2\times 1{0}^{-11}\ {m}^{-2/3}\) and vary the attenuation ratio from −2.52 to 0 dB. Note that we randomly vary the turbulent phase masks on SLM 2 while keeping the strength parameter \({C}_{n}^{2}\) constant. As described in the previous paragraph, we record 50 images for each ±ℓ from 0 to ±10 for a total of 550, which is then split between the training and testing sets. With the GNN pre-trained, we successfully reconstruct the unknown noisy, distorted, and attenuated intensity OAM profiles at the receiver. Some examples of noisy-distorted as well as attenuated OAM modes (R) and the corresponding corrected modes (C) for the attenuation ratios of −0.18 dB (left), −1.93 dB (center), and −2.43 dB (right) are shown in Fig. 4c. The superposition OAM value in each row is ±ℓ = 1, 5, 7, and 10 from top to bottom, respectively. We find that the GNN successfully reconstructs the lost spatial features of the OAM profiles, even in the presence of strong attenuation. The MSEs between the noisy-attenuated and reconstructed OAM profiles at the receiver with respect to various attenuation ratios are shown in Fig. 4d. Here we show a significant reduction in average MSE, from 1.22 × 104 to 2.5 × 102 and from 9.64 × 102 to 52.8 for attenuation ratios −2.52 and −1.18 dB, respectively. Additionally, we calculate MSE for various OAM values at the attenuation of −1.93 dB, which are shown in left-top inset of Fig. 4d. As expected, the GNN gives an increasing trend of MSE with increasing OAM value. Again, the GNN significantly reduces the MSE in a consistent manner.
Lastly, we train the GNN with only simulated OAM images (see “Methods”) and reconstruct the experimentally distorted and attenuated OAM profiles. Then the reconstructed modes are classified with our CNN, which is again solely pre-trained with the simulated target patterns with Gaussian noise of strength σ = 1. In order to illustrate the success of the GNN in correcting individual OAM profiles and limiting cross-talk with nearby OAM values, we select an experimental test set with moderately strong turbulence \({C}_{n}^{2}\ =\ 51.2\times 1{0}^{-11}\ {m}^{-2/3}\) and an attenuation ratio of −1.93 dB. Here the test set contains 50 distorted experimental optical images for each OAM value of ±ℓ ranging from 0 to ±10, for a total of 550. With the optimized GNN trained only with numerically simulated images, we reconstruct 50 unknown experimental images for each OAM mode and predict the corresponding mode value using the CNN. The mode values predicted by the CNN without (a) and with (b) the GNN versus the target OAM value are shown in the cross-talk plot of Fig. 5. We find the majority of the uncorrected optical profiles with OAM mode values ℓ = ±1, ±3, ±4, ±10 are incorrectly demodulated at the receiver as shown in Fig. 5a. However, with GNN reconstruction, the incorrect predictions have been significantly reduced from 18 to 3, from 33 to 9, from 48 to 8, and from 26 to 8 out of 50 test images for each OAM mode value of ℓ = ±1, ±3, ±4, and ±10, respectively, as shown in Fig. 5b.
Discussion
In this work, we have developed a GNN and CNN system that efficiently improves received signals in a FSO communication scheme that have been severely deteriorated due to the effects of turbulence, attenuation, and detector noise. We demonstrate the scheme’s ability to significantly improve the accuracy at a receiver and correct for varying levels of distortion, both experimentally and via simulations. We also show such accuracy improvements as the FSO link distance and dark noise at the receiver are varied. By training the GNN solely with simulated images, the system is able to decrease the cross-talk between experimental OAM mode profiles at the receiver. Moreover, by using our GNN, we have significantly improved the classification accuracy of the CNN, which is solely trained with undistorted modes with dark noise at the receiver, allowing for a nontrivial reduction in the training set size as compared to previous CNNs used in FSO communications systems, which require training on an unbounded number of distorted mode profiles. The current system also bypasses the need for active and adaptive optics at the receiver of a FSO communications platform. Furthermore, the addition of this unsupervised learning scheme may be implemented to build an autonomous technique and extended to demodulate more complex optical profiles, which are difficult to label and classify with current supervised techniques. The flexibility and generality of the developed neural networks will allow for the straightforward integration into current FSO communication systems, with the ability to be directly applied to quantum communication systems as well29,30,31,32,33,34,35,36.
Methods
Turbulence simulation
We use a Kolmogorov phase with the Von Karman spectrum effects model37 to simulate the turbulence in the atmosphere, which is given by Eq. (1),
where r0 = \({(0.423{k}^{2}{C}_{n}^{2}Z)}^{-3/5}\) is the Fried parameter for a propagation distance Z, and k = 2π/λ is the wave-vector for a given wavelength λ of light. Here κ is the spatial frequency, κm = 5.92/l0, and κ0 = 2π/L0, with inner (l0) and outer (L0) scales of turbulence. Finally, we generate random phase screens using the inverse Fourier transformation as given in Eq. (2),
where ℜ means taking only the real part, \({{\mathcal{F}}}^{-1}\) is the inverse Fourier transformation, \({{\mathbb{C}}}_{NN}\) is a complex random normal number with zero mean and unit variance, and \(\sqrt{{\Phi }_{NN}(\kappa )}\) is the square root of the phase distributions given by Eq. (1) over the sampling grid of size N × N.
Numerical simulation of long-distance turbulent propagation
The turbulent environment is simulated by placing a turbulent phase screen (SLM 2) generated by Eq. (2) a meter away from the SLM 1 plane. Then we propagate the Gaussian beam, G(x, y, w0), through all the phase screens (SLM 1 and SLM 2) and have the intensity profile, \({I}_{r}^{t}\), with additive dark noise, N(0, σ), at the receiver, Z ≥ 200 m away from the SLM 2 plane, using Eq. (3),
where \({\Theta }^{({\ell }_{1},{\ell }_{2})}\) represents the OAM phase mask on SLM 1 and H1 and H2 are transfer functions from SLM 1 to SLM 2 plane and from SLM 2 plane to receiver and CCD, respectively. In order to generate the turbulence discussed here, we use w0 = 4 cm, N = 128, λ = 1550 nm, l0 = 1 mm, L0 = 200 m, and \({C}_{n}^{2}\) varies from 9 × 10−15 m−2/3 to 1 × 10−13 m−2/3.
Numerical simulation of laboratory turbulent propagation
In order to simulate various intensity patterns for OAM modes of ±ℓ = 0–10 to pre-train the GNN as mentioned in the description of Fig. 5 in the text, we first randomly generate 100 turbulent phase mask, again, using Eq. (2) with N = 800, λ = 795 nm, l0 = 1 mm, L0 = 25 m, and \({C}_{n}^{2}\ =\ 36.8\times 1{0}^{-11}\ {m}^{-2/3}\). Then we find the corresponding intensity pattern at the receiver using Eq. (3) with w0 = 0.45 mm, 0.2 m as the distance between SLM 1 to SLM 2, and 0.6 m as the distance between SLM 2 to CCD. After that, we add Gaussian noise with mean 0 and standard deviation of 0.1 and normalize them to 8-bit images. The 100 noisy-distorted images for each OAM mode, as a total of 1100, is split into a training set with 90 images for each OAM, as a total of 990, and a testing set with 10 images for each OAM, as a total of 110 images. Once the GNN is optimized and stored, we feed unknown noisy, distorted, and attenuated experimental OAM images into pre-trained GNN, which finally reconstructs them as its output. Note that we perform the experiment at \({C}_{n}^{2}=\ 51.2\times 1{0}^{-11}\ {m}^{-2/3}\) with an attenuation ratio of −1.93 dB scenario, whereas for numerical simulation we take \({C}_{n}^{2}=\ 36.8\times 1{0}^{-11}\ {m}^{-2/3}\) with dark noise as mentioned above.
Generative convolutional denoising autoencoder
The distorted received optical profiles, \({I}_{r}^{t}\), as given by Eq. (3) are fed to the encoder of the GNN, which compresses them into a latent space S as expressed in Eq. (4),
where θ represents a parameter space of \({w}_{1}^{k}\), \({w}_{2}^{k}\) (weights) and \({b}_{1}^{k}\), \({b}_{2}^{k}\) (biases) of kth feature mappings of the first and second convolutional layers, respectively, where “Max” corresponds to a max-pooling operation. Also, W and B represent the weight and bias of a fully connected layer, and for convenience “⋆” represents the convolutional/transpose-convolutional operation. Note that we apply the ReLU activation after each convolutional operation. The resulting latent space S is then forwarded to the decoder that maps it to reconstructed pixels I of the input space as given by Eq. (5):
where primes represent the parameter space of the decoder corresponding to nth feature mappings. Here each training optical profile \({{I}_{r}^{t}}^{(i)}\) is successively mapped into a corresponding latent space S(i) and a generation I(i). After that, a square reconstruction loss L\(({I}^{(i)},{I}_{r}^{(i)})\) is evaluated, where \({I}_{r}^{(i)}\) is an undistorted optical profile at the receiver. Finally, in order to optimize the parameters, we minimize the average reconstruction loss given by Eq. (6) using adamoptimizer of tensorflow38:
CNN as a demodulator
In order to train a CNN to classify the generated modes, we simulate optical modes at the receiver without any turbulence using Eq. (3) for each OAM superposition value ranging from ℓ = 0 to ±10. We then manually add random Gaussian noise with σ = 2 (σ = 1 for the result shown in Fig. 5). Here we keep low noise in the training and testing sets of the CNN to estimate how closely the generated modes by the GNN fit with the target mode. Finally, we simulate 150 noisy images for each value of OAM for a total of 16,500 images. The image set is split into a training set with 130 images and a test set with 20 images, again, for each OAM profile. Then the parameter space of the CNN is optimized by minimizing a softmax cross-entropy loss using adamoptimizer. Note that pre-trained CNN network has an unity accuracy with respect to the test images.
Attenuation ratio
We use the following relation in order to calculate the attenuation ratio:
Data availability
The data that support the findings of this study are available from the corresponding authors on reasonable request.
Code availability
The code used in this study are available from the corresponding authors on reasonable request.
References
Huang, H. et al. 100 tbit/s free-space data link enabled by three-dimensional multiplexing of orbital angular momentum, polarization, and wavelength. Opt. Lett. 39, 197–200 (2014).
Ren, Y. et al. Free-space optical communications using orbital-angular-momentum multiplexing combined with mimo-based spatial multiplexing. Opt. Lett. 40, 4210–4213 (2015).
Qu, Z. & Djordjevic, I. B. Beyond 1 Tb/s free-space optical transmission in the presence of atmospheric turbulence. In 2017 Photonics North (PN) 1 (IEEE, 2017).
Nejad, R. M. et al. Orbital angular momentum mode division multiplexing over 1.4 km rcf fiber. In 2016 Conference on Lasers and Electro-Optics (CLEO) 1–2 (IEEE, 2016).
Milione, G. et al. 4×20 gbit/s mode division multiplexing over free space using vector modes and a q-plate mode (de) multiplexer. Opt. Lett. 40, 1980–1983 (2015).
Wang, J. et al. Terabit free-space data transmission employing orbital angular momentum multiplexing. Nat. Photonics 6, 488 (2012).
Kaushal, H. & Kaddoum, G. Optical communication in space: challenges and mitigation techniques. IEEE Commun. Surv. Tutor. 19, 57–96 (2017).
Malik, M. et al. Influence of atmospheric turbulence on optical communications using orbital angular momentum for encoding. Opt. Express 20, 13195–13200 (2012).
Sanchez-Lengeling, B. & Aspuru-Guzik, A. Inverse molecular design using machine learning: Generative models for matter engineering. Science 361, 360–365 (2018).
Jang, B., Chang, J., Park, A. & Wu, H. Ep-2092: Generative model of functional rt-plan chest ct for lung cancer patients using machine learning. Radiother. Oncol. 127, S1149 (2018).
Li, Z., Meier, M.-A., Hauksson, E., Zhan, Z. & Andrews, J. Machine learning seismic wave discrimination: application to earthquake early warning. Geophys. Res. Lett. 45, 4773–4779 (2018).
Donahue, C., Li, B. & Prabhavalkar, R. Exploring speech enhancement with generative adversarial networks for robust speech recognition. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 5024–5028 (IEEE, 2018).
Torlai, G. et al. Neural-network quantum state tomography. Nat. Phys. 14, 447 (2018).
Lohani, S., Kirby, B. T., Brodsky, M., Danaci, O. & Glasser, R. T. Machine learning assisted quantum state estimation. Mach. Learn. Sci. Technol. 1, 035007 (2020).
Gondara, L. Medical image denoising using convolutional denoising autoencoders. In 2016 IEEE 16th International Conference on Data Mining Workshops (ICDMW) 241–246 (IEEE, 2016).
Fichou, D. & Morlock, G. E. Powerful artificial neural network for planar chromatographic image evaluation, shown for denoising and feature extraction. Anal. Chem. 90, 6984–6991 (2018).
Cheng, X., Zhang, L. & Zheng, Y. Deep similarity learning for multimodal medical images. Computer Methods Biomech. Biomed. Eng. Imaging Vis. 6, 248–252 (2018).
Doster, T. & Watnik, A. T. Machine learning approach to OAM beam demultiplexing via convolutional neural networks. Appl. Opt. 56, 3386–3396 (2017).
Lohani, S., Knutson, E. M., Zhang, W. & Glasser, R. T. Dispersion characterization and pulse prediction with machine learning. OSA Contin. 2, 3438–3445 (2019).
Park, S. R. et al. De-multiplexing vortex modes in optical communications using transport-based pattern recognition. Opt. Express 26, 4004–4022 (2018).
Lohani, S. & Glasser, R. T. Turbulence correction with artificial neural networks. Opt. Lett. 43, 2611–2614 (2018).
Tian, Q. et al. Turbo-coded 16-ary OAM shift keying FSO communication system combining the CNN-based adaptive demodulator. Opt. Express 26, 27849–27864 (2018).
Li, J., Zhang, M., Wang, D., Wu, S. & Zhan, Y. Joint atmospheric turbulence detection and adaptive demodulation technique using the CNN for the OAM-FSO communication. Opt. Express 26, 10494–10508 (2018).
Lohani, S., Knutson, E. M., O’Donnell, M., Huver, S. D. & Glasser, R. T. On the use of deep neural networks in optical communications. Appl. Opt. 57, 4180–4190 (2018).
Zhao, Q. et al. Mode detection of misaligned orbital angular momentum beams based on convolutional neural network. Appl. Opt. 57, 10152–10158 (2018).
Jiang, S. et al. Coherently demodulated orbital angular momentum shift keying system using a CNN-based image identifier as demodulator. Opt. Commun. 435, 367–373 (2019).
Bengio, Y. Learning deep architectures for ai. Found. Trends Mach. Learn. 2, 1–127 (2009).
Vincent, P., Larochelle, H., Bengio, Y. & Manzagol, P.-A. Extracting and composing robust features with denoising autoencoders. In Proc. 25th International Conference on Machine Learning 1096–1103 (ACM, 2008).
Hughes, R. J., Nordholt, J. E., Derkacs, D. & Peterson, C. G. Practical free-space quantum key distribution over 10 km in daylight and at night. N. J. Phys. 4, 43 (2002).
Paterson, C. Atmospheric turbulence and orbital angular momentum of single photons for optical communication. Phys. Rev. Lett. 94, 153901 (2005).
Swaim, J. D. & Glasser, R. T. Squeezed-twin-beam generation in strongly absorbing media. Phys. Rev. A 96, 033818 (2017).
Tyler, G. A. & Boyd, R. W. Influence of atmospheric turbulence on the propagation of quantum states of light carrying orbital angular momentum. Opt. Lett. 34, 142–144 (2009).
Gupta, P., Horrom, T., Anderson, B. E., Glasser, R. & Lett, P. D. Multi-channel entanglement distribution using spatial multiplexing from four-wave mixing in atomic vapor. J. Mod. Opt. 63, 185–189 (2016).
Krenn, M. et al. Communication with spatially modulated light through turbulent air across vienna. New J. Phys. 16, 113028 (2014).
Vallone, G. et al. Free-space quantum key distribution by rotation-invariant twisted photons. Phys. Rev. Lett. 113, 060503 (2014).
Iandola, F. N. et al. Squeezenet: Alexnet-level accuracy with 50x fewer parameters and <0.5 mb model size. Preprint at https://arxiv.org/abs/1602.07360 (2016).
Bos, J. P., Roggemann, M. C. & Gudimetla, V. R. Anisotropic non-Kolmogorov turbulence phase screens with variable orientation. Appl. Opt. 54, 2039–2045 (2015).
Abadi, M. et al. TensorFlow: large-scale machine learning on heterogeneous systems. http://tensorflow.org (2015).
Acknowledgements
This material is based on research supported by, or in part by, the U. S. Office of Naval Research under award number N000141912374. We also acknowledge funding from the National Science Foundation Graduate Research Fellowship under grant number DGE-1154145 and Northrop Grumman - NG NEXT.
Author information
Authors and Affiliations
Contributions
S.L. developed the set-up and implemented the neural networks and ran all the simulations. E.M.K performed the experiments, and R.T.G. conceived of and led the project. All authors contributed to analyzing the data and writing the manuscript.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Lohani, S., Knutson, E.M. & Glasser, R.T. Generative machine learning for robust free-space communication. Commun Phys 3, 177 (2020). https://doi.org/10.1038/s42005-020-00444-9
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s42005-020-00444-9
This article is cited by
-
Dimension-adaptive machine learning-based quantum state reconstruction
Quantum Machine Intelligence (2023)
-
Optimized decision strategy for quadrature phase-shift-keying unambiguous states discrimination
Quantum Information Processing (2022)
-
Transition technologies towards 6G networks
EURASIP Journal on Wireless Communications and Networking (2021)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.