## Introduction

### Context and objectives

Imaging with low-light-levels is central to a large range of applications from biology and security to high precision metrology. In particular, imaging delicate objects such as biological cells or photosensitive chemicals that requires low light illumination to preserve their integrity. The ultimate goal for such applications is to recover as much image information as possible per photon. For that purpose, extracting both phase and intensity information in low-light imaging is a crucial aim. Indeed in some circumstances phase contains important information about an object that is not present in the intensity information alone. For example, most cells are effectively transparent, but their shape can still be inferred from the phase information1.

Traditional imaging schemes suffer from noise either due to the sensor noise or due to parasitic light from the surroundings. In very low light regimes these source of noise become predominant leading to images presenting poor signal to noise ratio (SNR). The use of quantum imaging in this context allows the improvement of the images quality. For a review on quantum imaging techniques and the kind of improvement they allow see ref.2. In particular, a way to remedy the aforementioned noise issue is to implement a particular quantum imaging scheme called ‘heralded imaging scheme’3,4. In heralded imaging (HI) a heralding detector records one photon from a photon pair and triggers the imaging detector to record exactly when the second photon is incident on the sensor. The imaging detector, usually an ICCD camera, is gated for a few picoseconds, and records only the photons originating from a pair. This therefore eliminates most of the undesired events such as background light and dark counts from camera since they are not correlated with heralding events. The use of such techniques to reach very low intensities were so far limited to intensity only imaging or qualitative phase contrast only imaging5,6, limiting therefore the amount of information extracted about an object. Here we demonstrate that quantitative phase-amplitude imaging with a small number of photons is possible by using a Fourier ptychographic acquisition technique enabled by quantum correlations through a heralded imaging scheme.

Fourier ptychography (FP) is a synthetic aperture technique which uses low-resolution objectives and tilt-shift illumination7 to synthesize high-resolution, amplitude and phase images8. It replaces the traditional interferometric methods9 with a computational iterative algorithm8. This simplifies the experimental setup, enabling for example the development of 3D printed low-cost microscopes10,11. FP has notedly been used to demonstrate quantitative phase imaging via Fourier ptychographic microscopy12, 3D refocusing13, and highly-resolved in vitro microscopy through the acquisition of gigapixel-scale phase and amplitude images1. For a review on Ptychographic image acquisition technique see ref.14.

In FP, a plane wave is used to coherently illuminate an object. By changing the angle of this plane wave illumination (with respect to the object), different images can be recorded. A wide range of illumination angles are generally used to capture several images, which are then combined using FP reconstruction algorithms. As we will show bellow, Fourier ptychography can be combined with quantum correlations to obtain low photon number phase and intensity images. Our work is enabled by using parametric down-conversion as the illumination source which exhibits Einstein-Podolsky-Rosen (EPR) correlations between signal and idler photons. A pump beam of sufficiently large diameter is used such that the down-coverted twin photons are created over a wide range of different spatial modes, resulting in an extended illumination source, thereby allowing the acquisition of various images of the same object corresponding to different illumination directions without having to change the actual illumination conditions of the object. The mechanism allowing such an acquisition uses the transverse spatial correlations exhibited by the two particles. When the momentum of one of the particles is measured with a precise value, the other particle is projected onto a state with a precisely defined opposed momentum. Therefore, by post-selecting in an image the photons whose twin photon were detected with a given momentum, one ensures that all the post selected photons had the same state corresponding to a plane wave when incident on the imaged object. This ensures that such post-selected images will virtually be that of the object acquired under a particular illumination direction. Post-selecting images on different momentum measurement outcome for the twin photon allows the acquisition of images virtually illuminated from many distinct directions. These images are combined together using an iterative FP algorithm to synthesise a wide-field of view, high-resolution amplitude and phase image. In the experimental demonstration reported here, the acquired images exhibit Poissonian noise due to the low number of photons in each image pixel. To minimise the impact of this Poisonian noise on the reconstruction we use a sequential Gauss-Newton optimisation method to perform the reconstruction. Such an optimisation method has previously been proven to be robust in the presence of Poissonian noise in the context of Fourier ptychography15.

The illumination source used in the present realisation are photons generated through the spontaneous parametric down-conversion (SPDC) process. This type of source has been used as illumination in several low-light level applications and techniques16,17,18. In particular SPDC has been used as a way to generate the optical manifestation of the EPR state in its original domain i.e. in the spatial variables of position and momentum19,20,21,22. These EPR states have been used in single-photon imaging, for example to implement quantum ghost imaging3,4,23, which can be extended to trans-wavelength imaging schemes24. SPDC can also be used to record video of single-photon double-slit interference25 and we used it recently to compare the resolution limits of quantum ghost imaging to classical imaging26. EPR states have also been used to perform quantum imaging with undetected photons27 and importantly for the presently reported realisation to implement heralded imaging4.

The nature of SPDC is that the signal and idler photons exhibits correlations in their transverse position or momentum, this results in peculiar features concerning its optical spatial coherence. Illuminating an object with one of the SPDC beams and detecting the intensity of that beam gives an incoherent imaging scheme. It was shown that by detecting coincidences between the signal and idler SPDC beams, by using on one beam a single mode detector, leads to retrieval of the underlying coherence of the state28. In the present work this coherence is needed as the Fourier Ptychography method requires the acquisition of a spatially coherent image. We show in the following section how correlations generated with an SPDC source can be used to perform Fourier ptychography.

## Results

### Principle of the acquisition

We show here that it is possible to use both the spatial coherence, manifest through coincident detection, and the SPDC spatial extent to realize a Fourier ptychographic reconstruction which requires the spatially coherent illumination of object from a range of different directions. For a classical source, its extended nature would lead to a loss of this spatial coherence.

Let us consider an EPR state that can be written in the following way:

$$|EPR\rangle \propto {\int }_{-\infty }^{\infty }{|x\rangle }_{s}{|x\rangle }_{i}dx\propto {\int }_{-\infty }^{\infty }{|k\rangle }_{s}{|-k\rangle }_{i}dk$$
(1)

where we use the proportionality to avoid tracking normalisation coefficients and i (s) stand for signal (idler). The associated density matrix has the following form:

$${{\boldsymbol{\rho }}}_{EPR}\propto \int {\int }_{-\infty }^{\infty }{|x\rangle }_{s}{|x\rangle }_{i}{\langle x^{\prime} |}_{s}{\langle x^{\prime} |}_{i}dxdx^{\prime}$$
(2)
$$\propto \int {\int }_{-\infty }^{\infty }{|k\rangle }_{s}{|-k\rangle }_{i}{\langle k^{\prime} |}_{s}{\langle -k^{\prime} |}_{i}dkdk^{\prime}$$
(3)

The state of the signal alone can be retrieved by tracing this operator on the idler:

$${{\boldsymbol{\rho }}}_{s}={{\rm{tr}}}_{i}({{\boldsymbol{\rho }}}_{EPR})\propto {\int }_{-\infty }^{\infty }{|k\rangle }_{s}{\langle k|}_{s}dk$$
(4)

which is a statistical mixture state of plane waves i.e. a spatially incoherent illumination. Now if one considers the following measurement operator corresponding to the detection of the idler photon with a given wave vector noted kf,

$${\bf{P}}={I}_{s}\otimes {|{k}_{f}\rangle }_{i}{\langle {k}_{f}|}_{i}$$
(5)

After projection and tracing out the state on the idler, one can show that the state of the signal becomes,

$${{\boldsymbol{\rho }}}_{s}={{\rm{tr}}}_{i}({\bf{P}}{{\boldsymbol{\rho }}}_{EPR}{\bf{P}})\propto {\int }_{-\infty }^{\infty }{e}^{-i{k}_{f}x}{|x\rangle }_{s}{\langle x^{\prime} |}_{s}{e}^{i{k}_{f}x^{\prime} }dxdx^{\prime}$$
(6)

corresponding to a pure state of the form,

$$|{{\rm{\Psi }}}_{s}\rangle =|{k}_{f}\rangle \propto {\int }_{-\infty }^{\infty }{e}^{-i{k}_{f}x}|x\rangle dx$$
(7)

This is a plane wave whose tilt compared to the optical axis is parametrised by the position of detection of the idler photon in the wave vector plane kf. Equivalently if the idler position is detected at the position xf then the state of the signal photon becomes:

$$|{{\rm{\Psi }}}_{s}\rangle =|{x}_{f}\rangle \propto {\int }_{-\infty }^{\infty }{e}^{i{x}_{f}k}|k\rangle dk$$
(8)

Such a state corresponds to an inclined plane wave incident on an object placed in the far-field of the crystal. Such wave planes are exactly the illumination that is needed to perform a Fourier ptychographic (FP) reconstruction. By illuminating an object with an SPDC signal beam and detecting correlations between the momentum of the idler photons (defined by the transverse position of the single-mode detector) and the positions of the correlated signal photons (after the object) one can, in principle, record a complete Fourier ptychographic reconstruction without changing the illumination of the object. A proposal to harness EPR-like correlations in a similar way was recently proposed in the context of plenoptic imaging29.

Our aim in the following section is to experimentally demonstrate such a protocol, and to show that one can use this scheme to perform a phase and intensity reconstruction of an object with images acquired with a low number of photons per image pixel.

### Experimental implementation

The setup used for the experimental demonstration is presented in Fig. 1. The EPR source is formed through spontaneous parametric down-conversion of a UV Laser. We use an ICCD camera triggered by a single-photon avalanche photo-diode (SPAD) to detect images of heralded single photons. The photons detected by the SPAD are first collected with a single-mode fiber, which ensures that the detected idler photons are collected at a given position (xf, yf) inside the beam. We therefore post-select the results for a certain position (xf, yf) for the idler photons i, which is simply done by triggering the ICCD camera conditionally on the detection of a photon by the SPAD. This triggering ensures that the acquired images correspond to images obtained by illuminating the object with the state described in Eq. (8). We then change the position of the fiber within the spatial extent of the idler beam (xf, yf) by changing the period of the blazed phase grating displayed on an SLM, thereby changing the deflection angle of the idler beam. A set of images is acquired, each image corresponding to different values of (xf, yf) and therefore to different illumination angles (θx, θy) of the object. By recording many such heralded images, One can then perform a complete Fourier Ptychographic acquisition.

The plane wave synthesis in our system can be explained using Klyshko advanced wave picture30,31. In this picture the heralding detector is replaced by a classical light source contra-propagating in the heralding arm up to the crystal that then acts simply as a mirror reflecting the light through the rest of the system to the camera. The light originating from a single point on the heralding detector side forms a collimated beam in the object plane in the imaging arm. The single mode fiber, defines a single point in the heralding arm, resulting in a single plane wave in the object plane to form a spatially coherent image on the camera. The position of this single mode fiber determines the angle of illumination and the spatial frequencies sampled. Figure 2a shows few of images of an object acquired for various periods of blazed gratings.

With this experimental setup we can acquire the images in two different ways. The first method is the heralded imaging (HI) described above, through which we perform the Ptychographic acquisition. In this case, the camera is triggered by the external trigger pulses corresponding to detection events of idler photons. But for comparison, we can also acquire images through direct imaging (DI) by changing only the way the camera is triggered. In this later case the camera is triggered using its internal trigger mechanism resulting in images generated by a random subset of photons that pass through the object and arrive at the camera sensor during the camera trigger window. Since either beam of the SPDC light is spatially incoherent when considered independently from the other, the DI image of the object that is obtained is an image resulting from spatially incoherent illumination. Figure 2 compares the image acquired with the HI method (a, b and c) and the one acquired with DI (d).

The images obtained through the DI method are acquired in a continuous trigger mode and we use the camera detector as a conventional linear detector. We acquire the DI images by accumulating frames that are not photon sparse, and due to the long integration time have a high background that we must subsequently, numerically, subtract. In contrast, the images acquired through heralded imaging are composed of photo detection events and do not need to be background subtracted. This is visible in Fig. 2, the colorbar in panel d goes bellow 0 due to the supplemental noise present in the background and the subtraction of the mean background. This is not the case in the other panels n Fig. 2 for which no background substraction have to be performed. A full comparison between heralded, ghost and direct imaging is out of the scope of the present article, for a discussion and comparison of the performances of heralded and direct imaging see4.

One can observe in Fig. 2 how switching from HI to DI result in the loss of the spatial coherence. Indeed one can see more details in the images acquired through DI and background subtracted Fig. 2d compared to the one acquired through the HI Fig. 2b,c. Moreover, one can see in Fig. 2 that by summing the detected intensity of the the different HI acquisitions for different equivalent illumination angles (see Fig. 2e), one retrieves an image that looks very similar to the DI image with the difference that no background subtraction has had to be performed in the HI case. The comparison is consistent with the notion that the DI image is a equivalent to of an incoherent sum of the HI images i.e. a spatially incoherent imaging scheme.

However, because in our scheme the individual HI images are accessible, it allows us to perform Fourier Ptychographic reconstructions with very low photon numbers as exemplified by the fact that no background subtraction has to be performed on any of the images. As dicussed above, this is done by triggering the camera with a very short gate-time of only a few nanoseconds synchronised to when a single photon event is likely to be incident on the sensor. By doing this one removes the detector dark noise that would occur when using the long exposure time normally required for low light imaging. This protocol allows us to acquire images with a very high signal to noise ratio relative to the low number of detected events. It should be noted that this present experimental demonstration of principle presents a low efficiency in terms of photon detected compared to the number of photons illuminating the object (about 1%), however, our proposed scheme could potentially lead to single snapshot Fourier ptychography with a very weak illumination of the sample by replacing the single mode detector with a second camera, the coincident HI image associated with each pixel of which would correspond to a different illumination angle of the object.

### Reconstructed images

The heralded images acquired as described above can be processed using a Fourier ptychography reconstruction algorithm. It is based on the alternate projection type algorithm developed by Gerchberg-Saxton32 and later improved by Fienup33. In these algorithms the amplitude of the complex field is known in one plane and a support structure for the amplitude and phase is known in the other plane. In an imaging system the support structure is the lens pupil shape, i.e., all the pixel values outside the lens pupil area should be zero. During the reconstruction, a random value is assigned for the unknown phase in the first plane and the complex field is propagated to the second plane. In this plane all the values outside the lens pupil area are set to zero and this distribution is then back propagated to the first plane. The amplitudes in the new complex field will be replaced by the known amplitudes from the experiment. These back and forth propagations are performed several times until the iterative algorithm finds the right phase.

In FP, the complex field is propagated between the Fourier plane and the image plane. Here, the alternate projection is performed for every illumination angle in sequence to complete one iteration, where the pupil position (support constraint) depends on the illumination angle. Each illumination angle updates a unique set of spatial frequencies in the object’s Fourier plane, hence by combining all the illuminations a large synthetic aperture is formed. In these datasets, an overlap between the spatial frequencies sampled by adjacent illumination angles is maintained to ensure the convergence34. This overalp redundancy deals system errors such as misalignment, aberrations and noise35,36. In this work the images are subject to photon shot-noise arising since the photon counts are extremely low, around 1 photon per pixel (see Fig. 2(c)), resulting in low signal to noise ratio. We implement a second order quasi Gauss-Newton based method for updating spatial frequencies in the Fourier plane15. This combined with an adaptive step-size approach37 provides a robust convergence of our data.

In Fig. 3a USAF resolution test target is imaged to demonstrate the resolution improvement quantitatively and in 2 a part of a wasp wing is imaged to demonstrate the resolution improvement and the phase recovery. In both cases we show that we have been able to obtain a reconstruction of the object in both intensity (Figs 3c and 4c) and phase (Figs 3d and 4d). We also show for comparison in Figs 3b and 4b) the direct sum of all the low photon number images acquired with the 225 different SLM gratings which is similar to an incoherent image. Also, we show the low-resolution, coherent image obtained for a normal illumination in Figs 3a and 4a. It can be seen that the images reconstructed through FP have higher-resolution compared to the low-resolution images and they also show an improved resolution compared to the equivalent incoherent imaging scheme. The phase image of the wasp wing in Fig. 4c shows features that are not clear in the intensity image – the dark regions in the intensity image still contain phase information showing that the wing is not discontinuous.

Figure 3c is used to determine the resolution achieved. Using the slanted edge modulation transfer function method we find a resolution of 23 μm. The resolution achieved by the incoherent imaging is found to be of of 30 μm. In both cases the obtained field of view is of ~2 mm.

Finally, it is worth noting that that even-though the resolution and field of view of our optical system are here relatively limited, as observable on Fig. 3, due to the technical choices we made here to implement the proposed technique, it should be noted that the technique imposes no inherent limit on the resolution of the system and that imaging/microscopy systems could be build that would harness our technique while achieving a much better resolution.

## Discussion

We have demonstrated the possibility of harnessing EPR-type quantum correlations to produce Fourier ptychographic (FP) images. We use FP combined with a heralded single-photon imaging scheme to allow the acquisition of images at extremely low light level and with a good signal to noise ratio. This is made possible by the implementation of a ‘heralded imaging scheme’ through the use of time-gated ICCD camera that temporally selects the heralded photons thereby removing parasitic contributions, e.g. dark noise, that would otherwise lead to a degradation of the image quality. By enabling the acquisition of low light level images with high signal to noise ratio, this promising technique allows the implementation of extremely low light level imaging in the context of microscopy. The proposed technique could therefore find applications in the context of delicate biological or chemical samples imaging. By enabling shot noise limited phase-amplitude imaging, our technique allows to image with a reduced exposure compared to technical noise limited classical technique. Moreover, further developments of our technique could exploit the sub-shot noise nature of the SPDC source to perform sub-shot noise imaging extracting both phase and amplitude about an object38,39. It enables the extraction of both intensity and phase with only a limited number of detected photons with the help of Fourier ptychographic reconstruction. Finally, although the exact implementation reported in the present work relied on scanning the position of the heralding detector, it would be possible to implement a non scanning FP scheme based on a modification of the present scheme where the heralding photons would also be detected by a camera or a SPAD array placed in a Fourier plane of the object. The detection of correlations between different pixels of the two spatially resolved detectors would then enable a similar FP reconstruction as the one presented here to be performed but without the need for scanning and hence improving the overall efficiency.

## Methods

### Experimental setup

The principle of the image acquisition in this work is similar to the heralded imaging configuration reported in4. The source is a 3 mm long β-Barium Borate (BBO) crystal, cut for type I phase matching, pumped by a quasi continuous laser at 355 nm. The parametric down-converted light is spectrally filtered to select photons at the degenerate frequency by an interference filter with a 10 nm bandwidth centred on 710 nm that is placed just after the BBO crystal (filter not shown on Fig. 1). The photon-pairs are then stochastically separated into a camera arm and a heralding arm by using a pellicle beam-splitter (BS). In the present configuration the object is placed in the far-field of the crystal output in the camera arm. It is illuminated by the signal (by convention) multi-mode SPDC beam with a full-width half-maximum of 156 μm. The camera is positioned in an image plane of the object, with a magnification $$M=\frac{4}{1.5}$$ between the object plane and the camera image plane. In the heralding arm, a spatial light modulator (SLM) is placed in the far field of the crystal, i.e. in a plane equivalent to the object plane with identical magnification as the object plane. The idler photons reflected off the SLM are then collected into a single-mode fiber (SMF) using a fiber coupler set up such that retro-propagated light outputting the SMF input is collimated after the coupler. The output of the optical fiber is then coupled to the heralding SPAD, the output from which triggers the ICCD in order to acquire images heralded by the detection of an idler photon. Because the SMF fiber is collecting photons at a given position in a plane that corresponds to a Fourier plane of the object plane, the images acquired by the triggered ICCD correspond to images obtained by illuminating the object with the state described in Eq. (8). This is due the fact that when the camera is triggered conditionally on the detection of a photon by the SPAD, the state of the SPDC light is projected onto a post selected state for which the idler has a determined position. Note that the object is here positioned in the far field of the crystal and the SMF selects the transverse position of the idler photon (xf, yf) that parametrise the angle of incidence of the illumination on the object (θx, θy).

Changing the angular position of the fiber within the spatial extent of the idler can be implemented by changing the period of the blazed grating displayed on the SLM thereby changing the deflection angle of the idler beam. One can then perform a complete Fourier Ptychographic acquisition, by recording many such heralded images, each being acquired for a particular period and orientation of grating. Each time the SPAD detects a photon, the camera intensifier is triggered with a gate width of a few nanoseconds. During this window, the camera single photons events hitting the intensifier are amplified and such events are then recorded by the CCD array. The intensifier can be triggered many times during the CCD chip exposure time such that the frames that are obtained are accumulations of all the single-photon events acquired during the exposure time. The HI image is then formed from summing these individual frames, each one of which is photon-sparse. There exists an electronic delay in the ICCD triggering that needs to be compensated to ensure that the photons detected in the images are from the same pair as the herald photons. This delay is achieved by introducing a 22 m image-preserving delay line in the camera arm as shown in Fig. 1. The heralded images (as shown in Fig. 2a) correspond to 1000 frames each of 1 second of exposure, during which time the camera intensifier is triggered for every heralding detector pulse. The CCD sensor is air cooled to −30 °C. The images are thresholded to generate binary images that corresponds to detection of single photons. We can calculate the threshold over which a pixel is considered to correspond to a photo-detection and the noise probability per pixel by acquiring frames with the camera optical input blocked. The dark-count probability per pixel arising from the camera readout noise, is calculated to be around 5 · 10−5 per frame. The images obtained through the DI method are acquired in a continuous trigger mode and we use the ICCD camera as a linear detector. Indeed, to retrieve a comparable number of photons acquired in comparison to the HI method, the total triggering time needs to be considerably longer for DI compared to HI as the photons are not heralded and are randomly detected by the camera. This results in most of the detected photons in DI being background noise events, resulting in a very poor signal to noise ratio compared to the HI (see4 for a comparison between HI and DI when both are used in single photon counting mode). For this reason, to obtain good quality DI image based on thresholded frames one would need to acquire a very high number of frames since each frame needs to be photon sparse and that most of the photo-detection are then noise. This would mean a very long acquisition to obtain the DI images. To avoid this extended measurement time we do not acquire the DI images in a single-photon counting mode and instead simply acquire the DI images by accumulating frames that are not photon sparse and use the ICCD as a linear detector by simply subtracting the strong background of the images evaluated with the camera optical input blocked.

To perform the FP reconstruction a set of 15 × 15 images is acquired with an increment of Δk ≈ 4.2 rad · mm−1 in term of transverse wave vector for the plane wave illuminating the object, which corresponds to an increment Δθ ≈ 5 · 10−4rad of the angle of incidence of the light. The magnification of the system from the SLM to the Camera is $$M=\frac{4}{1.5}$$, this means that for each increment of the angle Δθ corresponds to a tilt of $${\rm{\Delta }}\theta ^{\prime} =\frac{{\rm{\Delta }}\theta }{M}$$, the mean direction of propagation of the post selected photons impinging on the camera. Such a tilt corresponds to a shift of 3 pixels in the Fourier domain obtained by proceeding to a Fast Fourier Transform on a 874 camera pixels region of interest.

### Reconstruction procedure

The reconstruction procedure starts by creating an initial estimate of the high-resolution image by interpolating the central low-resolution image (the high-resolution image contains more pixels due to the improved resolution). This interpolated image is Fourier transformed to obtain the high-resolution frequency spectrum of the object O(θ). Starting with this first evaluation, one chooses to work with one of the recorded illumination angles θi, in our case this corresponds to a particular angular position of the single-mode fiber in the heralding beam. The complex spectrum O(θ) is then shifted according to the illumination angle θi. In the subsequent step the complex spatial spectrum is filtered with the pupil function P of the imaging system. The resulting spectrum Otemp(θ) = O(θ − θi) × P(θ) is propagated to the image plane through an inverse Fourier transform, where the image amplitude is then updated with the square root of the image Ii recorded for the fiber position equivalent to the illumination angle θi. Another Fourier transform is performed to propagate the resulting complex image back to the spatial frequency domain where the resulting new object spatial frequencies evaluation O(θ) is used to update the most current shifted reconstruction of the object O(θ − θi). The result can then be shifted back by θi to obtained the new current evaluation of the object complex spectrum O(θ). The same steps are then performed again with the remainder of the images taken at different angles of illumination Ii, which completes one iteration of the reconstruction. The choice of the reconstruction method that is used determines the cost function used to update the object spatial frequencies using O(θ). As mentioned before, in order to minimise the impact of the strong shot-noise present in the images, we use a cost function corresponding to a sequential quasi Gauss-Newton optimisation method15. Similar cost function can also be used to update the pupil function to correct for any aberrations present in the system35. The reconstruction process is iterated until the difference between the current reconstruction and the reconstruction from previous iteration falls below a predefined convergence criterion. The final high-resolution spatial frequency spectrum O(θ) is inverse Fourier transformed to obtain the high-resolution amplitude and phase reconstruction of the object.

### Instrumentation

To perform the experiment we used an Andor ICCD camera (Model iStar DH334T-18U-A3). The SPAD is an SPCM from Perkin-Elmer (SPCM-ARQ-13-FC). The SLM used to perform the scanning of the relative position of the beam and the fibre is an Hamamatsu LCOS-SLM (X10468). The Laser is a JDSU xCyte (Model CY-355-150).