## Abstract

Quantum circuits with hierarchical structure have been used to perform binary classification of classical data encoded in a quantum state. We demonstrate that more expressive circuits in the same family achieve better accuracy and can be used to classify highly entangled quantum states, for which there is no known efficient classical method. We compare performance for several different parameterizations on two classical machine learning datasets, Iris and MNIST, and on a synthetic dataset of quantum states. Finally, we demonstrate that performance is robust to noise and deploy an Iris dataset classifier on the ibmqx4 quantum computer.

## Introduction

Neural networks offer state-of-the-art performance in a wide number of machine learning tasks including computer vision, natural language processing, generative modeling, and reinforcement learning.^{1} The hierarchical structure of deep neural networks can allow them to match the expressiveness of shallower models with exponentially fewer parameters.^{2,3,4} In recent years, there has been much interest in translating the success of neural networks to the quantum computing context.^{5} Despite this, there are many open questions regarding the advantages that quantum computation can bring to machine learning.^{6,7} Do quantum algorithms offer a clear speed-up over classical approaches for inference and training? Are quantum machine learning algorithms robust to noise? What are the best quantum circuit layouts to carry out machine learning tasks? An exciting way to explore these questions is through experimentation on available quantum hardware, and simulation on classical hardware.

Tensor networks are a method for representing an intractable high rank tensor as a decomposition of tractable lower rank tensors connected by contraction. They are widely used in many-body physics for the simulation of strongly correlated quantum systems, and can be used to represent both quantum states and quantum circuits.^{8,9,10,11} Tensor networks with hierarchical structure exhibit many similarities with neural networks and in some cases have been shown to be equivalent.^{12,13} Given that tensor networks can be used to represent both neural networks and quantum circuits, they are a natural choice for exploring the intersection of both fields. In this work, we consider the supervised machine learning tasks of classifying classical and quantum data on a quantum computer using hierarchical quantum circuits.

To perform classification on a quantum computer the input data must be encoded in a quantum state. Two ways in which this can be achieved are by encoding the data in the amplitudes of individual qubits in a fully separable state (qubit encoding), or in the amplitudes of an entangled state (amplitude encoding). We test classifiers using both encoding methods. For classical data we perform qubit encoding using single qubit rotations. For quantum data we assume that the data arrives from another quantum device and is already an entangled amplitude encoded state.

Once the data has been encoded, the classifier consists of a series of unitary operations applied to the initial quantum state. Then, a measurement is carried out on a target qubit. In practice, multiple runs are required to approximate the expectation of the measurement outcome, and the most frequent outcome is taken as the predicted class. More runs increase the classifier confidence.

In addition to the pipeline just described, we need to specify the layout of the hierarchical circuit, and the algorithm for learning its parameters. The circuits we use here are tree-like and can be parameterized with a simple gate-set that is compatible with currently available quantum computers. The first of these circuits is known as a tree tensor network (TTN).^{9} We then consider a more complex circuit layout known as the multi-scale entanglement renormalization ansatz (MERA).^{10} MERAs are similar to TTNs, but make use of additional unitary transformations to effectively capture a broader range of quantum correlations. Both one-dimensional (1D) and two-dimensional (2D) versions of TTN and MERA circuits have been proposed in the literature.^{14,15}

In the 1D case, TTN and MERA circuits can be evaluated efficiently using classical techniques when the input data is qubit encoded. Evaluating such circuits on amplitude encoded data is likely to be classically intractable. In 2D, the TTN circuit is efficiently simulatable when using qubit encoding, whereas the 2D MERA circuit is not. Because we cannot simulate large 2D MERA circuits, we restrict our experiments to the 1D case. In all experiments we find that 1D MERA outperforms 1D TTN, suggesting that 2D MERA could in principle outperform 2D TTN. Such a hypothesis should be tested with future experiments as suitably large near-term quantum computers become available. Classifiers that possess a 1D structure could be used for sequential data, such as time-series, while classifiers that possess a 2D structure would be the natural choice for 2D data, such as a natural images.

Optimizing the circuits can be accomplished by stochastic gradient descent. In the case of efficiently simulatable networks, it makes sense to use the analytic gradient. For circuits that cannot be efficiently simulated it is possible to use a quantum computer and estimate gradients numerically. Moreover, a hybrid approach that involves a classical pre-training step to initialize some of the gates has been previously proposed.^{16} We empirically validate this approach by initializing a 1D MERA with a pre-trained 1D TTN. Such pre-training reduces the average number of training steps needed until convergence on a model with comparable accuracy, a benefit for implementations on near-term quantum computers.

We demonstrate our techniques using TTNs and MERAs and compare performance for a number of parameterizations. The first of these uses only single-qubit rotations and fixed CNOT gates. The second uses more general two-qubit gates. The third uses three-qubit gates, where the additional ancilla qubits allow for non-linear operations. Both real and complex parameterizations are compared.

We test the ability of each classifier to predict binary labels on two canonical machine learning datasets, Iris,^{17} and MNIST handwritten digits,^{18} and on synthetic quantum datasets. We also use the IBM Quantum Experience^{19} to test robustness to depolarizing noise, and to deploy the model on the ibmqx4 quantum computer.

The structure of the article is the following. Section 2 contains a description of the hierarchical quantum classifiers and the results of experiments on classical data and quantum states. Section 3 contains a discussion of the results, a comparison to existing methods and directions for future work.

## Results

### Data encoding

Classification consists of assigning a category to an observation. In machine learning, an inference model is trained to minimize the classification error on a finite set of data, also known as the training set. The actual performance of the classifier, the generalization error, is then estimated on a set of data points not used for training, also known as the test set. The functional form of the inference model is often critical to the success of the classifier. State-of-the-art models for high-dimensional datasets with complex structure are typically hierarchical or compositional.^{1} These ideas can be translated to the paradigm of quantum computation using the framework of tensor networks. Before describing the tensor network architectures used in this work, namely TTN and MERA, it is important to first clarify what datasets are considered in this paper to gauge the performance of these networks, and how they are prepared.

Let us first consider the case of classical data. A classical dataset for binary classification is a set \({\cal D} = \left\{ {\left( {{\boldsymbol{x}}^d,y^d} \right)} \right\}_{d = 1}^D\), where \({\boldsymbol{x}}^d \in {\Bbb R}^N\) are *N*-dimensional input vectors, and *y*^{d} ∈ {0, 1} are the corresponding class labels. Classifying classical data on a quantum computer requires that the input vectors be encoded in a quantum state. There are a variety of ways to accomplish this and different algorithms require different encoding methods. The most efficient approach in terms of space is to encode classical data in the amplitudes of a superposition, that is, using *N* qubits to encode a 2^{N}-dimensional data vector. However, in the general case and depending on the quantum classifier used, the computational cost of preparing data as a superposition can negate the speedup obtained during classification.^{7} A simpler method is to encode each element of a classical data vector in the amplitude of a single-qubit. This type of encoding requires *N* qubits to encode an *N* dimensional data-vector and, therefore, is less efficient in terms of space. However, the state preparation is clearly efficient in terms of time as it only requires single-qubit rotations. We opt for this type of encoding for classical data. In particular, we first re-scale the data vectors element-wise to lie in \(\left[ {0,{\textstyle{\pi \over 2}}} \right]\). Then, we encode each vector element in a qubit using the following scheme:^{20}

The final data vector is written as \(\psi ^d = \otimes _{n = 1}^N\psi _n^d\), and is ready to be used in a quantum algorithm.

Let us now consider the case of quantum data. A quantum dataset for binary classification is a set \({\cal D} = \left\{ {\psi ^d,\left. {y^d} \right)} \right\}_{d = 1}^D\), where \(\psi ^d \in {\Bbb C}^{2^N}\) are 2^{N}-dimensional input vectors of unit length, and *y*^{d}∈{0,1} are the corresponding classes. In contrast to classical data, quantum data, such as the output of a quantum circuit or a quantum sensor, may already be in superposition. That is, the quantum states are used as-is, and there is no relevant cost for the preparation.

### Circuit architecture

We now discuss the quantum circuit architectures for classification. The first circuit architecture is inspired by TTNs, specifically binary trees. The TTN circuit begins by applying a set of two-qubit nearest-neighbor unitaries to the input. We then discard one of the qubits output from each unitary, halving the number of qubits in the next layer of the circuit. In the following layer we again apply two-qubit unitaries to the remaining qubits before discarding half of them. This process is repeated until only one qubit remains. The network in full consists of measuring a single-qubit expectation value on this remaining qubit

where \(\hat U_{{\mathrm {QC}}}\left( {\left\{ {U_i} \right\}} \right)\) is the quantum circuit made up of unitaries *U*_{i}(*θ*_{i}), ** θ** = {

*θ*

_{i}} is the set of parameters which define the unitaries, and \(\hat M\) is the single-qubit operator whose expectation we are calculating. A circuit diagram of an eight-qubit TTN is shown in Fig. 1a. The solid lines encompass the circuit, while the dashed lines represent its conjugate transpose.

The MERA network is closely related to the TTN. All of the unitaries that make up a tree network are maintained with an additional layer of two qubit unitaries added before each layer of the TTN. These additional unitaries, {*D*_{i}}, each operate on one qubit of neighboring unitaries in the upcoming TTN layer. In a conventional MERA network, the addition of these unitaries allows quantum correlations on a particular length scale to be captured at the same layer of the network.^{10} A circuit diagram of an eight-qubit MERA is shown in Fig. 1b.

### Unitary parameterization

We have explored a number of different ways to parameterize the unitaries used in these circuits. Some of the input data used is purely real, we therefore tested the effect of restricting the unitaries to be real too. That is, we chosen unitaries such that *U*_{i} ∈ *SO*(·) ⊂ *SU*(·). We also consider general, complex valued unitaries *U*_{i} ∈ *SU*(·). As has been observed in the context of the time-dependant variational principle applied to tensor networks, the use of complex weights often prevents optimization from getting stuck in local minima.^{21,22}

We also explored a number of other methods for parameterizing the unitaries; Fig. 2 illustrates three such paramaterizations. In Fig. 2a, the unitary block is composed of two arbitrary single-qubit rotations and a CNOT_{ij} gate, where *i* and *j* are control and target qubit, respectively. Note that in some cases the direction of the CNOT_{ij} may be reversed in order to respect the causal structure. For example, in our eight-qubit implementations we reverse the control and target qubits for blocks *U*_{2}, *U*_{4}, and *U*_{6} lying in the lower part of the circuit. In the case of the restriction to *SO*(4) the single-qubit rotations are simply *Y*-rotations.

In Fig. 2b, the unitary block consists of an arbitrary two-qubit gate. It is interesting to explore this much more general setting in simulations, although a practical implementation of such unitary may be costly. That is, the two-qubit unitary needs to be compiled into low-level hardware-dependent gates.

Finally, Fig. 2c shows a three-qubit gate involving an ancilla qubit. By tracing out the ancilla qubit we can effectively implement a rich class of non-linear functions, e.g. step functions,^{23} closely resembling the operations of classical neural networks. Again, in practice a significant overhead is expected due to compilation.

The measurement \(\hat M\) is performed on a specific qubit and consists of a simple Pauli measurement in a chosen direction. This can be implemented in practice by an additional single-qubit rotation followed by the projective measurement onto \(\left| 0 \right\rangle \left\langle 0 \right|\). This is sufficient for a binary classification task; by computing and thresholding the expectation value of *M*, TTN and MERA classify the input *ψ*^{d} into one of the two classes. In our example in Fig. 1, the measurement is performed on qubit number six.

### Learning process and complexity

We now discuss the learning process. In principle, the circuit parameters would be adjusted to directly maximize the classification accuracy on the training set or, in other words, minimize the classification error. Optimizing such an objective function is highly non-trivial and it is common to optimize a bound instead. Here we choose to minimize the mean square error between predictions and true class labels

where *ψ*^{d} are inputs, *y*^{d} are class labels, *D* is the number of training data points, and ** θ** groups all the adjustable parameters of the circuit as described above. Although there exist several approaches to carry out this optimization, artificial neural networks are commonly optimized by stochastic gradient descent algorithms. At each iteration

*t*, we estimate the gradient ∇

*J*

^{(t)}and choose a learning rate

*η*

^{(t)}. Parameters are then updated via a rule of the kind

*θ*^{(t+1)}←

*θ*^{(t)}+

*η*

^{(t)}∇

*J*

^{(t)}. This algorithm is stochastic because at each iteration the gradient is estimated on a small batch rather than on the full training set. Beside speeding up the calculation, this noisy gradient may help in escaping from local minima. Much literature and experimentation has been dedicated to improving stochastic gradient descent algorithms. In this work, we employ a variant called Adaptive Moment Estimation (Adam).

^{24}

The cost function is a function of the measurement outcome of the circuit being trained. In order to obtain these measurement outcomes, the circuit itself must be evaluated. In Table 1 we summarize the complexity of obtaining the measurement outcomes at the end of the different types of circuits in this paper. The complexity stated is in terms of the number of multiplications of scalar numbers required to perform the task. The complexities in the two-dimensional cases are stated for a grid of *N* × *N* qudits. The complexities stated for the 2D networks use the network architecture introduced in refs ^{10,25}

In the case of efficiently contractable networks we can compute the exact gradient using off-the-shelf automatic differentiation software (e.g., TensorFlow^{26}). This applies to many 1D networks including TTNs and MERA. For networks that cannot be efficiently contracted a finite-difference method or an approximation to the true gradient must be used.^{27} These strategies introduce additional noise due to finite-sampling error, and intrinsic noise of near-term quantum devices. We begin exploring the impact of these with simulations in Section 2.7. Note that all of the circuits we train in this paper can be evaluated efficiently on quantum hardware.

### Experimental results: Iris dataset

In this experiment, we tested the ability of a TTN to classify varieties of Iris. The Iris dataset^{17} consists of 150 examples in total of three varieties of Iris flowers. Each example of Iris is described by four real-valued attributes *x*_{1–4}. We encoded the four attributes into four qubits using Eq. (1). We then parameterized unitaries using the simple gate shown in Fig. 2a, and restricted the single-qubit rotations to be real (i.e., *Y*-rotations). To allow for binary classification, three binary datasets were extracted from the original set. In each subset, each class comprised 1/2 of the examples. For each class, 1/3 of examples were used as a test set and used to compute the accuracy. Mean accuracy and one standard deviation computed on five random initializations are given by Table 2. As shown, TTN performed extremely well in all cases.

#### Experimental results: Handwritten digits (MNIST)

In this experiment we tested the ability of TTN and MERA classifiers on a number of handwritten digit recognition tasks and compared the performance of different parameterizations. MNIST^{18} is a canonical data-set consisting of 70,000 labeled gray-scale images of handwritten digit from 0 to 9. From this dataset we generated four binary classification tasks. In the first we kept only images containing 0 or 1, and for the second task, only 2 or 7. For the third tasks we re-labeled all images as even or odd. For the final task we divided the images into those that were >4 or not. MNIST images are 28 × 28 pixels. To allow for simulation using eight qubits, we performed principal component analysis on the images for each task and kept only the eight components with highest variance. Finally, we used Eq. (1) to encode the data.

Of the 70,000 examples 55,000 were used for training, 5000 for validation and 10,000 for testing. Training was performed using the Adam optimizer^{24} with batches of 20 examples. Validation and test accuracy were recorded every 10 training batches, and training was stopped when validation set accuracy did not increase for 30 consecutive tests. Figure 3 shows typical learning curves for train and test datasets.

Mean accuracy and one standard deviation computed on five random initializations are given by Table 3. The ‘Classifier’ column describes if the circuit was a TTN, MERA, or hybrid, that is, a MERA pre-trained with TTN. The ‘Unitaries’ column describes if the circuit was parameterized using a simple, general or ancilla gate set as described by Fig. 2. The ‘Rotations’ column specifies the type of rotation used, either real, *SO*(4), or complex, *SU*(4).

Some remarks are in order. First, we note that the restriction to simple unitaries led to significantly lower accuracy than when using general unitaries. Complex rotations improved the accuracy of the classifiers in all tasks except for task ‘0 or 1’ where accuracy was already >99.5% with real rotations. It is notable that this is the case despite the input data being real-valued. Second, the MERA classifiers achieved higher accuracy than TTN classifiers in all cases, demonstrating the power of the additional unitaries. Third, the hybrid classifier achieved accuracy comparable to that of the standard MERA. On average, hybrid classifiers required 2.45 times more training steps until convergence than standard MERA. However, the number of post-training steps required was only 0.825 times the number of training steps of standard MERA. This indicates that classical pre-training may lead to a reduction in the number of training steps carried out on the quantum computer, a potential advantage in the near-term. Finally, all networks outperformed the logistic regression benchmark except those using the simple gate-set.

One may wonder whether the accuracy on some of the tasks can be made more competitive with the state-of-the-art results on MNIST. In order to efficiently simulate all the circuits, each 28 × 28 image was reduced to an eight-dimensional vector using PCA, thereby discarding a lot of information that could be useful for classification. To verify this we ran logistic regression without PCA on the most difficult of the four tasks, “Is > 4”. This model achieved a test accuracy of 87.09%, a significant improvement over the logistic regression on the PCA reduced data which achieved 70.7% instead (Table 3). We concluded that reducing the dimensionality of the data can have a detrimental effect on the model accuracy and therefore we expect TTN and MERA classifiers to perform better when using more principal components, or even raw data.

### Experimental results: Quantum data

We now consider the problem of classifying quantum data, that is, quantum states generated by different physical processes. A physical process can be simulated by a quantum circuit. By setting up two different quantum circuit layouts, we can generate synthetic classification tasks. Let us first define the *building block* for our quantum circuit layouts.

Our building block consists of single-qubit rotations *U*_{i} for all qubits *i* ∈ {0, …, *N*}, followed by all the possible CNOT_{ij} gates where *i* and *j* are control and target qubits, respectively, and *i* < *j*. The angles of the single-qubit rotations are the only parameters of our building block. By stacking several of these building blocks, we can generate deeper and more complex circuits. In particular, we chose to identify the class with the number of building blocks in the stack (e.g., class 5 consists of 5 building blocks).

Now, for each class, we can generate a quantum state by randomizing all the single-qubit gates, and then executing the circuit on initial state \(\left| 0 \right\rangle\). This is repeated many times in order to generate a dataset. As discussed in Section 2, we assume that each quantum state in the dataset can be directly fed into the quantum computer where the classifier is executed, hence not requiring any pre-processing. The tasks of the classifier is to determine which of two circuit layouts a state was generated from.

Here, we work with circuits of *N* = 8 qubits. We generated datasets of *D* = 5000 quantum states for each of the classes *y* ∈ {1, 2, 3, 5, 10}. To make sure that the synthetic classification task was well defined, we first looked for a strategy that could correctly classify the states most of the time. For each state, we computed the maximum bipartite entanglement entropy, \(\max _AS\left( {\rho _A} \right) = \max _BS\left( {\rho _B} \right)\), over all possible partitions *A*, *B* of the eight qubits. Figure 4 shows histograms of this quantity for three classification tasks. By inspecting the overlap of distributions we can find an optimal threshold that would classify states correctly most of the time. This shows that the classification task is meaningful. We would like to stress that this is an intractable strategy. The only purpose is to demonstrate that, in principle, there is a feature of the state that correlates with the class. The hope is that a hierarchical quantum classifier can find equally successful strategies in a tractable way.

The classifier used for this task was a TTN like the one shown in Fig. 1. We considered two parameterizations; the first uses general gates such as the one shown in Fig. 2b. The second uses arbitrary three-qubit gates where one of the qubits is an ancilla initialized in the state \(\left| 0 \right\rangle\), as illustrated in Fig. 2c. The data described above was divided into training, validation, and test sets. Each of these sets were balanced, that is, they had an equal number of states from each class. A set of 1000 examples from each class was held out as a test set. Training was performed for 4000 iterations with batches of 40 states and test accuracy was recorded every 50 iterations. The best test accuracy was recorded for each task.

Table 4 reports mean classification accuracy and one standard deviation computed on five random initializations. Results for the TTN with general two-qubit gates are no better than random class assignment in all tasks, indicating the need for a more expressive model. Indeed, when using gates augmented by an ancilla qubit, TTN was able to classify quantum states with some accuracy, suggesting that those may play a key role. The classification accuracy is higher for the ‘1 or 10’ task; this is somewhat expected as the overlap of classes 1 and 10 shown in Fig. 4a is less than that of the other tasks shown in Fig. 4b, c.

Finally, as a proof of principle, we verified the performance of a classical logistic regression model. We fed the vector of amplitudes to the model and trained with off-the-shelf software. The test accuracy was close to 50%, that is, no better than random. We shall stress that this approach is not feasible in practice, since only providing the input in classical form would require full tomography of the quantum dataset.

### Experimental results: Characterizing the effect of noise on classification performance

Many machine learning models including neural networks are highly robust against the negative effects of noise. Some kinds of noise can even help with convergence and generalization.^{28,29} In this experiment, we tested the effect of depolarizing noise on the quantum classifier by simulating a depolarizing channel. It consists of a completely positive map Δ_{λ} parametrized by *λ* from a 2^{N}-dimensional state *ρ* to a linear combination of *ρ* and a maximally mixed state

We used one of the TTN classifiers for classes 1 and 2 of the Iris dataset (see Section 2.5) and simulated the noisy circuit using the IBM Quantum Experience. The depolarizing channel was applied to the system after the application of each unitary gate in the circuit, that is, after each single-qubit rotation and CNOT gate. The entire test set was used to evaluate accuracy.

In order to make a realistic case, we used a finite number of measurements to estimate the class predictions. For each data point, we took 401 measurements in the computational basis and obtained the most likely class by majority vote. The 401 measurements may not be sufficient to estimate the output of the circuit with high confidence when the probability assigned to both classes is close to 0.5. In other words, repeating the 401 measurements and taking the majority vote could lead to a different class assignment for the very same data point. Therefore, we repeated the computation of the accuracy 200 times and obtained error bars. Finally, we increased the amount of noise *λ* from 0 to 0.2 in increments of 0.01.

Figure 5 shows mean and one standard deviation of the classification accuracy on the test set. We first noticed that finite sampling led to some error even when no depolarizing noise was used. Indeed, we obtained a mean accuracy of 96.5% with *λ* = 0; the very same model achieved 100% accuracy under exact computation (see results for “1 or 2” in Table 2). Second, the mean accuracy reduced as we injected depolarizing noise, but it remained above 95% for depolarizing noise up to *λ* = 0.07 showing some level of resilience. Finally, as we increased the noise further, the standard deviation of the accuracy increased as well. This is expected: as the output state gets closer to the maximally mixed state according to Eq. (4), the probability assigned to both classes gets closer to 0.5. Hence, a larger number of measurements would be needed to estimate the class.

### Experimental results: Deployment on a quantum computer

In this experiment, we deployed the Iris classifier for classes 1 and 2 (see Section 2) on the ibmqx4 quantum computer available in the IBM Quantum Experience. As shown in Fig. 6, this TTN classifier has three CNOT gates and seven rotations in the *Y* direction. A test set of 34 unseen examples was used to determine accuracy. For each example, the circuit was run 401 times, and the samples were used to compute the most likely class. The circuit correctly classified 100% of the test set, and achieved a test cost function value of 0.0811 (Eq. (3)).

## Discussion

Combining the success of deep neural networks and other machine learning methods with the power of quantum computation is a tantalizing prospect. Much work to date has focused on modifying classical machine learning algorithms to incorporate quantum linear algebra subroutines, thus inheriting their speed-ups. One such subroutine is the quantum algorithm for solving linear systems, also known as HHL.^{30} The algorithm is exponentially faster than the best known classical alternative, although this comes with some caveats.^{7} Quantum classifiers that use HHL include the quantum support vector machine^{31} and the kernel least squares.^{32} Whilst promising, these algorithms also inherit the limitations of HHL, in particular, the requirement that classical data be efficiently prepared in amplitude encoding. Another quantum subroutine that can be readily embedded in a quantum classification model is Grover’s algorithm which, for example, has been used to improve both computational and statistical complexity of the perceptron model.^{33}

While the above proposals assume availability of universal quantum computers, much of the recent literature has been focusing on algorithms for noisy intermediate-scale quantum technologies.^{34} These consist of hybrid quantum-classical algorithms where the quantum computer is used to execute ansatz circuits and to measure observables of interest. A classical optimization routine is used to adjust the ansatz circuit in order to minimize a cost function. Originally proposed for quantum chemistry and combinatorial optimization, these approaches have been recently investigated for supervised^{27} and unsupervised^{35,36} machine learning. The underlying ansatz circuits are often inspired by the structure of classical neural networks, but without explicit reference to tensor networks.

Refs ^{16,37} propose training tree-like tensor networks to be classifiers. In particular, ref. ^{37} demonstrates that TTNs can be used to classify images of handwritten digits and to encode classes of images in quantum many-body states. The framework proposed in ref. ^{16} examines the role of training TTNs to be classifiers in a quantum computing context and provides numerical evidence that TTNs can be used to perform supervised and unsupervised machine learning with the support of a quantum computer. Our work extends these ideas in a number of respects. Firstly, more complex networks, such as MERA are studied and their superiority relative to simpler networks is demonstrated. Secondly, it is shown that these networks can be used to classify quantum mechanical data in addition to classical data. Thirdly, networks are demonstrated that are constructed from simple two-qubit gates that can be natively implemented on available hardware. Finally, a trained tensor network is successfully deployed on a real quantum device (ibmqx4).

In this report, we have demonstrated that hierarchical quantum circuits can be used to classify classical and quantum data. Circuits based on the MERA outperform simpler tree-like circuits known as TTNs. These circuits can be parameterized with a simple gate set that can be easily implemented on existing quantum computers. A trained model is shown to be resistant to depolarizing noise and is successfully deployed on the ibmqx4 quantum computer.

Both MERA and TTN are naturally extendable to larger inputs. In 1D each additional layer doubles the dimensionality of the input. It is less clear how to increase or decrease the modeling power of a circuit without changing the dimension of the input. In classical neural networks this is achieved by increasing the depth and breadth of the network. One possibility for accomplishing this with quantum hierarchical classifiers is to use *χ*-level quantum systems (qudits) for some suitable *χ* > 2 as the unit of computation, rather than qubits (*χ* = 2). Ref. ^{37} demonstrates that model expressiveness in tensor network classifiers can be increased by increasing the input and internal bond dimensions. This is equivalent to performing computation using qudits. Data can be encoded in qudits using a generalization of qudit encoding described in ref. ^{20}. Whilst it is possible to simulate qudits with qubits, there are practical considerations that can make this challenging.^{38}

Currently it is unclear what network architecture is ideal for a classification task, a thorough examination of the role entanglement plays in classification circuits may help illuminate this. Consider the case of a TTN circuit applied to a product state input. In this circuit the measurement qubit interacts with each other qubit in the circuit at most once and therefore its entanglement with the rest of the circuit will increase as unitaries are applied. If the measurement qubit is highly entangled with the rest of the network it will struggle to minimize the cost function Eq. (3) but clearly it is necessary to introduce some entanglement in the network for correlations between input qubits to be shared. Such a trade-off may limit the effectiveness of TTN circuits, especially as they are scaled to larger inputs.

Constraining machine learning models using regularization can help them to generalize better to unseen data. Indeed, parameters with large magnitude are a characteristic of overfitting. The unitary constraint of quantum circuits naturally prevents parameters from becoming large, and it is likely acting as a strong regularizer. Additional regularization methods from the machine learning literature will become important in future quantum machine learning work. For example, the addition of noise during training of classical neural networks can also have a regularizing effect^{28} and help the model to learn invariant representations.^{29} In our study, we did not simulate circuit noise during the training phase, but we did show high resistance to depolarizing noise during the prediction step.

Much of the success of convolutional neural networks comes from their ability to learn layers of translation invariant representations using a shared set of weights. Translation invariance can be enforced in TTN and MERA by restricting the unitaries within each layer to be the same. Similarly, scale invariance can be enforced by restricting the unitaries between different layers of the circuit to be the same. The role of weight sharing in hierarchical quantum classifiers is a question for future research.

In this report, we have identified two cases where the cost of classical simulation is thought to be exponentially harder than that on a quantum computer. The first of these, which we do not test, is when the hierarchical quantum classifier cannot be classically simulated even when the input is a product state, 2D MERA circuits being one such example. The second case is when the input data consists of entangled quantum states. Here, an entirely classical approach may require expensive tomography and become intractable as the system size grows. While there are many existing methods for classifying 2D classical data, developing methods for classifying quantum data is a promising research direction.

## Data availability

All data needed to evaluate the conclusions are available from the corresponding author upon reasonable request.

## Additional information

**Publisher’s note:** Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

## References

- 1.
LeCun, Y., Bengio, Y. & Hinton, G. Deep learning.

*Nature***521**, 436 (2015). - 2.
Lin, H. W., Tegmark, M. & Rolnick, D. Why does deep and cheap learning work so well?

*J. Stat. Phys.***168**, 1223–1247 (2017). - 3.
Mhaskar, H., Liao, Q. & Poggio, T. Learning functions: when is deep better than shallow. Preprint at arXiv:1603.00988 (2016).

- 4.
Telgarsky, M. Benefits of depth in neural networks. Preprint at arXiv:1602.04485 (2016).

- 5.
Schuld, M., Sinayskiy, I. & Petruccione, F. The quest for a quantum neural network.

*Quantum Inf. Process.***13**, 2567–2586 (2014). - 6.
Ciliberto, C. et al. Quantum machine learning: a classical perspective.

*Proc. R. Soc. Lond. A***474**, 20170551 (2018). - 7.
Aaronson, S. Read the fine print.

*Nat. Phys.***11**, 291 (2015). - 8.
Orús, R. A practical introduction to tensor networks: matrix product states and projected entangled pair states.

*Ann. Phys.***349**, 117–158 (2014). - 9.
Shi, Y.-Y., Duan, L.-M. & Vidal, G. Classical simulation of quantum many-body systems with a tree tensor network.

*Phys. Rev. A***74**, 022320 (2006). - 10.
Vidal, G. Class of quantum many-body states that can be efficiently simulated.

*Phys. Rev. Lett.***101**, 110501 (2008). - 11.
Verstraete, F., Murg, V. & Cirac, J. I. Matrix product states, projected entangled pair states, and variational renormalization group methods for quantum spin systems.

*Adv. Phys.***57**, 143–224 (2008). - 12.
Cohen, N. & Shashua, A. Convolutional rectifier networks as generalized tensor decompositions. In

*Proc. International Conference on Machine Learning*, 955–963 (2016). - 13.
Levine, Y., Sharir, O., Cohen, N. & Shashua, A. Bridging many-body quantum physics and deep learning via tensor networks. Preprint at arXiv:1803.09780 (2018).

- 14.
Cincio, L., Dziarmaga, J. & Rams, M. M. Multiscale entanglement renormalization ansatz in two dimensions: quantum ising model.

*Phys. Rev. Lett.***100**, 240603 (2008). - 15.
Evenbly, G. & Vidal, G. Entanglement renormalization in noninteracting fermionic systems.

*Phys. Rev. B***81**, 235102 (2010). - 16.
Huggins, W., Patel, P., Whaley, K. B. & Miles Stoudenmire, E. Towards quantum machine learning with tensor networks. Preprint at arXiv:1803.11537 (2018).

- 17.
Fisher, R. A. The use of multiple measurements in taxonomic problems.

*Ann. Hum. Genet.***7**, 179–188 (1936). - 18.
LeCun, Y., Cortes, C. & Burges, C. J. C.

*The MNIST Database of Handwritten Digits*(1998). http://yann.lecun.com/exdb/mnist/. - 19.
IBM Quantum Experience. https://www.research.ibm.com/ibm-q/. Accessed 29 Mar 2018.

- 20.
Stoudenmire, E. & Schwab, D. J. Supervised learning with tensor networks. In

*Advances in Neural Information Processing Systems*, 4799–4807 (2016). - 21.
Haegeman, J. et al. Time-dependent variational principle for quantum lattices.

*Phys. Rev. Lett.***107**, 070601 (2011). - 22.
Trabelsi, C. et al. Deep complex networks. Preprint at arXiv:1705.09792 (2017).

- 23.
Ho Wan, K., Dahlsten, O., Kristjánsson, H., Gardner, R. & Kim, M. S. Quantum generalisation of feedforward neural networks.

*npj Quantum Inf.***3**, 36 (2017). - 24.
Kingma, D. & Ba, J. Adam: a method for stochastic optimization. Preprint at arXiv:1412.6980 (2014).

- 25.
Vidal, G. Entanglement renormalization.

*Phys. Rev. Lett.***99**, 220405 (2007). - 26.
Abadi, M. et al. Tensorflow: a system for large-scale machine learning.

- 27.
Farhi, E. & Neven, H. Classification with quantum neural networks on near term processors. Preprint at arXiv:1802.06002 (2018).

- 28.
Graves, A. Practical variational inference for neural networks. In

*Advances in Neural Information Processing Systems*, 2348–2356 (2011). - 29.
Achille, A. & Soatto, S. On the emergence of invariance and disentangling in deep representations. Preprint at arXiv:1706.01350 (2017).

- 30.
Lloyd, S. Quantum algorithm for solving linear systems of equations. In

*APS March Meeting Abstracts*(2010). - 31.
Rebentrost, P., Mohseni, M. & Lloyd, S. Quantum support vector machine for big data classification.

*Phys. Rev. Lett.***113**, 130503 (2014). - 32.
Schuld, M., Sinayskiy, I. & Petruccione, F. Prediction by linear regression on a quantum computer.

*Phys. Rev. A.***94**, 022342 (2016). - 33.
Kapoor, A., Wiebe, N. & Svore, K. Quantum perceptron models. In

*Advances in Neural Information Processing Systems*, 3999–4007 (2016). - 34.
Preskill, J. Quantum computing in the NISQ era and beyond. Preprint at arXiv:1801.00862 (2018).

- 35.
Otterbach, J. S., et al. Unsupervised machine learning on a hybrid quantum computer. Preprint at arXiv:1712.05771 (2017).

- 36.
Benedetti, M., Garcia-Pintos, D., Nam, Y. & Perdomo-Ortiz, A. A generative modeling approach for benchmarking and training shallow quantum circuits. Preprint at arXiv:1801.07686 (2018).

- 37.
Liu, D. et al. Machine learning by two-dimensional hierarchical tensor networks: a quantum information theoretic perspective on deep architectures. Preprint at arXiv:1710.04833 (2017).

- 38.
Wang, D.-S. & Sanders, B. C. Quantum circuit design for accurate simulation of qudit channels.

*New J. Phys.***17**, 043004 (2015).

## Acknowledgements

E.G. is supported by the UK Engineering and Physical Sciences Research Council (EPSRC) [EP/P510270/1] and by Rahko Limited. M.B. is supported by EPSRC and by Cambridge Quantum Computing Limited (CQCL). S.S. is supported by the Royal Society, EPSRC and the grant ARO-MURI W911NF-17-1-0304 (US DOD, UK MOD, and UK EPSRC under the Multidisciplinary University Research Initiative). J.L., and A.H. are supported by EPSRC [EP/L015242/1]. We gratefully acknowledge the support of NVIDIA Corporation with the donation of the Titan Xp GPU used for this research.

## Author information

### Affiliations

#### Department of Computer Science, University College London, London, WC1E 6BT, UK

- Edward Grant
- , Marcello Benedetti
- , Joshua Lockhart
- & Simone Severini

#### Rahko Limited, London, N1C 4PF, UK

- Edward Grant

#### Cambridge Quantum Computing Limited, Cambridge, CB2 1UB, UK

- Marcello Benedetti

#### College of Computer Science and Technology, Zhejiang University, 310027, Hangzhou, China

- Shuxiang Cao

#### Clarendon Laboratory, Department of Physics, University of Oxford, Oxford, OX1 3PU, UK

- Shuxiang Cao

#### London Centre for Nanotechnology, 17-19 Gordon St, London, WC1H 0AH, UK

- Andrew Hallam
- & Andrew G. Green

#### Department of Physics and Astronomy, University College London, London, WC1E 6BT, UK

- Andrew Hallam

#### GTN Limited, London, EC1V 2PY, UK

- Vid Stojevic

### Authors

### Search for Edward Grant in:

### Search for Marcello Benedetti in:

### Search for Shuxiang Cao in:

### Search for Andrew Hallam in:

### Search for Joshua Lockhart in:

### Search for Vid Stojevic in:

### Search for Andrew G. Green in:

### Search for Simone Severini in:

### Contributions

E.G., A.H., J.L., and A.G. designed the classifiers. M.B., E.G., and J.L. designed the experiments. A.H. and J.L. performed the complexity analysis. S.C., E.G., M.B., and J.L. wrote the code. E.G. and S.C. performed the experiments on ibmqx4. M.B., E.G., A.H., and S.C. analyzed the experimental results. Supervision and guidance was provided by S.S., A.G., and V.S. All authors contributed to the final version of the manuscript.

### Competing interests

The authors declare no competing interests.

### Corresponding author

Correspondence to Edward Grant.

## Rights and permissions

**Open Access** This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.