## Abstract

Transcription factors regulate gene expression, but how these proteins recognize and specifically bind to their DNA targets is still debated. Machine learning models are effective means to reveal interaction mechanisms. Here we studied the ability of a quantum machine learning approach to classify and rank binding affinities. Using simplified data sets of a small number of DNA sequences derived from actual binding affinity experiments, we trained a commercially available quantum annealer to classify and rank transcription factor binding. The results were compared to state-of-the-art classical approaches for the same simplified data sets, including simulated annealing, simulated quantum annealing, multiple linear regression, LASSO, and extreme gradient boosting. Despite technological limitations, we find a slight advantage in classification performance and nearly equal ranking performance using the quantum annealer for these fairly small training data sets. Thus, we propose that quantum annealing might be an effective method to implement machine learning for certain computational biology problems.

## Introduction

Quantum computation has been the subject of intense scientific scrutiny for its potential to solve certain fundamental problems, such as factoring of integers^{1} or simulation of quantum systems,^{2,3} more efficiently than classical algorithms, by using unique quantum phenomena including entanglement and tunneling. More recently, there has been much interest in the potential of quantum machine learning to outperform its classical counterparts.^{4,5,6,7,8,9,10,11,12,13,14,15,16,17,18} Although different implementations and models of quantum computing are still in development, promising theoretical and experimental research indicates that quantum annealing (QA),^{19} or adiabatic quantum optimization,^{20} may be capable of providing advantages in solving classically-hard problems that are of practical interest (for a review see ref. ^{21}). QA is the only paradigm of quantum computation that currently offers physical implementations of a non-trivial size, namely the D-Wave (DW) processors.^{22,23,24}

The adiabatic theorem of quantum mechanics, which underlies QA, implies that a physical system will remain in the ground state if a given perturbation acts slowly enough and if there is a gap between the ground state and the rest of the system’s energy spectrum^{25} (Fig. 1a). To use the adiabatic theorem to solve optimization problems, we can specify a beginning Hamiltonian, *H*_{B}, whose ground state is easy to find (typically a transverse field), and a problem Hamiltonian, *H*_{P}, that does not commute with *H*_{B} and whose ground state encodes the solution to the problem we are seeking to optimize.^{26} We then interpolate from *H*_{B} to *H*_{P} by defining the combined Hamiltonian *H*(*s*) = *A*(*s*)*H*_{B} + *B*(*s*)*H*_{P}, with 0 ≤ *s* = *t*/*t*_{
f
} ≤ 1, where *A*(*s*) and *B*(*s*) are, respectively, decreasing and increasing smoothly and monotonically, *t* is time, and *t*_{
f
} is the total evolution, or annealing time. The adiabatic theorem ensures that the ground state of the system at *t* = *t*_{
f
} will give the desired solution to the problem, provided the interpolation is sufficiently slow, i.e., *t*_{
f
} is large compared to the timescale set by the inverse of the smallest ground state gap of *H*(*s*) and by d*H*(*s*)/d*s*^{27} (Fig. 1a). When QA is implemented in a physical device, temperature and other noise effects play an important role; thermal excitation and relaxation cannot be neglected and affect performance.^{28,29,30}

QA algorithms were implemented on the D-Wave Two X (DW2X) processor installed at the Information Sciences Institute of the University of Southern California. The problem Hamiltonians that are used for DW can be described as Ising spin models with tunable parameters.^{31} The Ising model assumes a graph *G* = (*V*, *E*) composed of a set of vertices, *V*, and edges, *E*. Each of the *N* spins is a binary variable located at a unique vertex. For the DW2X, *N* = 1098, the spins are represented by superconducting flux qubits, and *G* is the so-called Chimera graph (see Supplementary Material, Fig. S1). The problem, or Ising, Hamiltonian for this system can be written as

where the local fields {*h*_{
i
}} and couplings {*J*_{
ij
}} define a problem instance, and are programmable on the DW2X to within a few percent Gaussian distributed error. The ${\sigma}_{i}^{z}$ represent both binary variables taking on values ±1, and the Pauli *z*-matrices. Given a spin configuration $\left\{{\sigma}_{i}^{z}\right\}$, *H*_{P} is the total energy of the system. The beginning Hamiltonian is a transverse magnetic field: ${H}_{\mathrm{B}}={\sum}_{i}{\sigma}_{i}^{x}$, where ${\sigma}_{i}^{x}$ is the Pauli *x*-matrix acting on qubit *i*. The Ising Hamiltonian can be easily transformed into a quadratic unconstrained binary optimization (QUBO) problem by applying the following transformation: ${w}_{i}=\frac{{\sigma}_{i}^{z}+1}{2}\in \left\{0,1\right\}$. Problems submitted to DW are automatically scaled so that all *h*_{
i
} and *J*_{
ij
} values lie between −1 and 1, and DW returns a set of spin values $\left\{{\sigma}_{i}^{z}=\pm 1\right\}$ that attempts to minimize the energy given by Eq. (1) (a lower energy indicates better optimization). Much attention has been paid to whether the DW devices are capable of delivering quantum speedups.^{32,33,34,35,36} Here we sidestep this question and instead use the DW2X as a physical device that implements QA for the purpose of solving a problem in machine learning, while focusing on performance measures other than speedup.

In order to probe the potential of a machine learning approach that is based on QA, we have used the DW2X processor to solve the simplified formulation of a biologically relevant problem: transcription factor (TF)-DNA binding (see Fig. 1b). TFs are a key component in the regulation of gene expression, yet the mechanisms by which they recognize their functional binding sites in a cell and thereby activate or repress transcription of target genes are incompletely understood. Nucleotide sequence, flexibility of both TFs and binding sites, the presence of cofactors, cooperativity, and chromatin accessibility are all hallmarks that affect the binding specificity of TFs in vivo.^{37,38} As a first step to gaining insight into TF binding, it is valuable to understand the intrinsic binding specificity of the TFs for DNA, which is optimally gained from in vitro data. Widely used methods to gain such an understanding and represent the DNA sequence preferences of TFs are based on position weight matrices (PWM) or PWM-like models.^{39} In the simplest of these models, the binding preference of a TF for each of the four nucleotides of the DNA alphabet {A, C, G, T} of a sequence of length *L* is represented as a 4 × *L* matrix. Such models implicitly treat each position in the DNA sequence as being independent, so that each element of the matrix can be thought of as the contribution of a nucleotide at the corresponding position to the overall binding affinity. Since the independence of the nucleotide positions is in many cases a valid approximation and also because of current restrictions on the size of the DW processors, in this work we have used a model consisting of single-nucleotide sequence features to show a proof of principle of the use of machine learning via QA in biology. Despite technological limitations of emerging quantum technology, we concurrently demonstrate cases in which this form of machine learning using QA outperforms classical machine learning when training with small data sets. This is among the very first successful applications of quantum hardware to a realistic, though simplified problem in computational biology.

## Results

Experimental data sets on TF-DNA binding for a specific TF consist of *N* sequences of fixed length *L* and *N* values that express a measure of the binding affinity of the chosen TF to each sequence: ${\left\{\left({\overrightarrow{x}}_{n},{y}_{n}\right)\right\}}_{n=1}^{N}$. In other words, the *n*th sequence is represented by the vector ${\overrightarrow{x}}_{n}=\left({x}_{n,1},{x}_{n,2},\dots ,{x}_{n,L}\right)$ with *x*_{n,j} ∈ {A, C, G, T}, for *j* = 1, …, *L*, and *y*_{
n
} is the corresponding measure of binding affinity. For instance, ${\overrightarrow{x}}_{n}$ may be ACAACTAA, with *y*_{
n
} = 4.95. In this work we used binding from three genomic-context protein binding microarray (gcPBM) experiments, which use fluorescence intensity as a measure of binding affinity,^{40} and two high-throughput systematic evolution of ligands by exponential enrichment (HT-SELEX)^{41,42,43} experiments, which report relative binding affinity. After preprocessing, the three gcPBM data sets consisted of *N* ≈ 1600 sequences of *L* = 10 base-pairs. The two HT-SELEX data sets consisted of *N* ≈ 3200 and 1800 sequences of length *L* = 12 after preprocessing (see Methods for a brief descrption of the preprocessing procedure). We used the following one-hot encoding to represent the sequence as a vector of binary variables: A = 1000, C = 0100, G = 0010, T = 0001, and thus transformed ${\overrightarrow{x}}_{n}$ into a feature vector ${\overrightarrow{\varphi}}_{n}\equiv {\left({\varphi}_{n,1},\dots ,{\varphi}_{n,4L}\right)}^{T}$. This encoding scheme^{44} was used so that all combinations of inclusion and exclusion of the four nucleotides may be identified. Similar to previous studies^{44,45,46,47} the goal of the present work is to identify patterns within the data to qualitatively assess whether the strength of a TF binding to a particular unseen sequence is above a certain threshold (classification) or to rank sequences in terms of binding affinity (ranking).

To identify conditions in which machine learning with existing QA devices may be of use for studying a simplified biological problem, we report results obtained by solving a learning protocol with six different strategies: (i) an adiabatic quantum machine learning approach formulated in refs. ^{4,5} (DW), (ii) simulated annealing^{48} (SA) (using the implementation given in ref. ^{49}), (iii) simulated QA^{50} (SQA), a classical algorithm that can represent the potential of a noiseless thermal quantum annealer, (iv) L_{2} regularized multiple linear regression (MLR), (v) Lasso^{51} and (vi) a scalable machine learning tool known as XGBoost (XGB).^{52} DW, SA and SQA are probabilistic approaches. SQA is a (classical) path integral Monte Carlo method that has performed very similarly to QA and captures some of its main advantages.^{53} MLR is a deterministic method with a closed-form solution that returns the weights that best minimize the objective function (defined below). Lasso is a method for linear regression that uses an L_{1} norm (see description of objective function below for more details). XGB uses boosted trees and has been applied to a variety of machine learning tasks in physics, natural language processing and ad-click prediction (e.g., ref. ^{54}).

Given a transformed feature vector ${\overrightarrow{\varphi}}_{n}$ that represents a DNA sequence, the goal of each method is to compute a predicted binding score $f\left({\overrightarrow{\varphi}}_{n}\right)$ that best matches the actual binding score. To carry out the task, an *objective function* must be optimized. The objective function consists of two parts: a training loss function and a regularization term that helps avoid overfitting. We may write the objective function as

where *R* is the training loss, Ω is the regularization term, and $\overrightarrow{w}$ is the set of feature weights to be determined by the six learning algorithms: DW, SA, SQA, MLR, Lasso and XGB. The mean squared error was used as the loss function for all six methods; namely, $R\left(\overrightarrow{w}\right)={\sum}_{n}{\left({y}_{n}-{f}_{\overrightarrow{w}}\left({\overrightarrow{\varphi}}_{n}\right)\right)}^{2}$, where *y*_{
n
} is the actual binding score of the *n*th sequence, and ${f}_{\overrightarrow{w}}\left({\overrightarrow{\varphi}}_{n}\right)$ is the predicted binding score. The regularization term was $\Omega \left(\overrightarrow{w}\right)=\lambda {\u2225\overrightarrow{w}\u2225}_{1}$ for DW, SA, SQA and Lasso, $\mathrm{\Omega}\left(\overrightarrow{w}\right)=\lambda {\u2225\overrightarrow{w}\u2225}_{2}^{2}$ for MLR, and $\mathrm{\Omega}\left(\overrightarrow{w}\right)=\gamma S+\frac{1}{2}\lambda {\sum}_{j=1}^{S}{w}_{j}^{2}$ for XGB, where the ${\u2225\cdot \u2225}_{1}$ norm is the number of 1’s (Hamming weight), the ${\u2225\cdot \u2225}_{2}^{2}$ norm is the square of the Euclidean norm, and *S* is the number of leaves.^{52} The calibration of the hyper-parameters *λ*,* γ* and *S* is discussed below. The loss function should be minimized and the regularization term generally controls model complexity by penalizing complicated models; the strength of the regularization was determined using a 100-fold Monte Carlo cross-validation. All six methods assume a linear model for the predicted binding affinity, i.e., ${f}_{\overrightarrow{w}}\left({\overrightarrow{\varphi}}_{n}\right)={\overrightarrow{w}}^{T}{\overrightarrow{\varphi}}_{n}={\sum}_{j}{w}_{j}{\varphi}_{n,j}$. DW, SA and SQA return binary weights and are probabilistic methods, that is, they return a distribution of weights with different energies [values of the Hamiltonian in Eq. (1)]. In order to utilize the distribution of weights returned, while not sacrificing the discrete nature of the QUBO approach, up to twenty of the best weights were averaged (see Supplementary Material, Sec. SID for a description of how excited-state solutions were included and Fig. S2 for an example of the decrease in the objective function).

Our computational procedure consisted of three main phases: (1) calibration of hyper-parameters, (2) training, and (3) testing (Fig. 2). About 10% of the data were held out for testing during the testing phase ('test data' or ${\mathcal{D}}^{\mathrm{TEST}}$); these test data were not seen during calibration and training stages. Calibration and training were carried out using the remaining 90% of the data ('training data' or ${\mathcal{D}}^{\mathrm{TRAIN}}$). Due to the discrete nature of the weights returned in the QUBO approach, as well as technological limitations of the DW2X device, calibration of hyper-parameter *λ* was carried out by repeatedly sampling a small number of sequences, about 2% and 10% of ${\mathcal{D}}^{\mathrm{TRAIN}}$, corresponding to about 30 and 150 sequences, respectively. In particular, in the calibration phase we determined the hyper-parameters by using 100-fold Monte Carlo (or split and shuffle) cross-validation with training splits of 2% and 10% of the training data, varying *λ* from 2^{−3} to 2^{6}. Monte Carlo cross-validation was used so that hyper-parameters would be tuned on a similar number of sequences as used in the training phase (in contrast, *n*-fold cross-validation trains on $\frac{n-1}{n}\times 100\%$ of the data). The same calibration procedure was applied to tune *λ* for SA, SQA, MLR and Lasso: the resulting values of *λ* are listed in the Supplementary Material, Tables S1 and S2. In order to demonstrate good performance for XGB, *γ*, *S*, and several additional parameters needed to be tuned (see Methods). In the training phase we used a bagging (bootstrap aggregating) procedure,^{55} randomly sampling with replacement 2% and 10% of the training data, namely about 30 and 150 sequences for the gcPBM data sets. Each subset of about 30 or 150 sequences formed a training 'instance', and the mapping of a subset of data to the *h*_{
i
} and *J*_{
ij
} seen by DW and SA is given in the Methods. Each learning approach (DW, SA, SQA, MLR, Lasso and XGB) was trained on the same set of instances. To collect statistics, 50 instances were randomly selected with replacement, for each training size. In the testing phase, the predictive power was assessed in terms of classification performance (the mean area under the precision-recall curve or AUPRC) and ranking performance (the median Kendall’s *τ*) on the test data unseen during calibration and training phases. AUPRC is a measure of classification performance that may help discern between similar algorithms when there is a high degree of class imbalance; i.e., when the data set contains many more false labels than true labels.^{56} Kendall’s *τ* is a rank correlation coefficient that counts all mismatches equally.^{57} Additional methodological details are given in Methods and in the Supplementary Material, Sec. SI.

### Performance on gcPBM Data

To quantify the relative performance of the algorithms in capturing DNA–protein binding preferences, we first present results for high-quality gcPBM^{40} data of three TFs from the basic helix-loop-helix (bHLH) family: the Mad1/Max heterodimer (‘Mad’), the Max homodimer (‘Max’), and the c-Myc/Max heterodimer (‘Myc’).^{44} bHLH proteins typically recognize and bind as dimers to the enhancer box (E-box), which is of the form CANNTG, where N denotes any of the four nucleotides (A, C, G, or T). Mad, Max, and Myc are part of a gene network that controls transcription in cells; a mutation of Myc has been associated with many forms of cancer.^{58} For the work here, these three data sets were modified to consist of about 1600 sequences of ten base pairs (bp) in length with the E-box located at the central 6 bp.

In Fig. 3 we present the AUPRC and Kendall’s *τ* obtained with the different algorithms when training with about 30 (2%) and 150 (10%) sequences. To compute the AUPRC, a threshold of the data was introduced: for a threshold at the *p*th percentile of the data, *p*% of the total number of sequences have binding affinities below the threshold and were set as negatives ('false'), and the (1 − *p*)% of the sequences that have binding affinities above the threshold were set as positive ('true'); see Supplementary Material, Sec. SID for a more detailed explanation of the procedure to threshold the data and to generate and calculate the AUPRC. During the calibration phase, we tuned hyper-parameters with a single threshold at the 80th percentile of the data, and during the testing phase we evaluated performance between the 70th and the 99th percentiles of the data. Kendall’s *τ* was evaluated between the predicted and measured binding affinity. A higher AUPRC indicates a better ability to correctly classify sequences that would be strongly bound by a TF, and a higher *τ* indicates a better ability to accurately rank the binding affinities for different sequences.

For the AUPRC, when training on instances with 2% of the data (left column in Fig. 3a), DW, SA and SQA perform very similarly, with DW slightly outperforming SA on the Myc data, and are somewhat better than MLR at the 70th and 80th percentiles. MLR tends to do better at the higher thresholds: this behavior could be affected by the fact that, during the calibration phase, we selected the *λ* that gave the best performance at the 80th percentile. Lasso, which uses the same L_{1} norm as DW, SA, and SQA, performs better than XGB but worse than the other methods. XGB, which has been successfully applied to a growing number of learning tasks, does poorly with small training sizes. When training with 10% of the data (right column in Fig. 3a), the trends of relative classification performance are quite different. XGB and MLR perform very similarly, though XGB does slightly better for the Max data set. DW tends to perform better than SA and SQA, especially at higher thresholds. DW’s mean AUPRC is normally worse than MLR and XGB’s, though there is overlap between the error bars. SA and SQA generally perform worse than the other methods, but not conspicuously so. A more thorough analysis of DW’s classification performance in comparison to SA and SQA with the same problem parameters is reported in the Supplementary Material, Figs. S3–S6 and related text in Sec. SIIA. Lasso’s performance is in general comparable to DW, SA, and SQA and generally seems to perform the worst with 10% of training data.

For Kendall’s *τ* (Fig. 3b), Lasso and XGB’s performance are the least favorable when training with 2% of the data. SQA generally performs the best over the three TFs, though MLR’s median *τ* is marginally greater than SQA’s for Mad. SA’s performance is very close to SQA’s and DW’s performance is slightly worse than the other two annealing schemes, though generally better than the typical machine learning algorithms. With 10% of the data, DW performs the worst; SA and SQA perform very similarly, with SQA being slightly better on two of the three data sets; MLR and Lasso perform very similarly, though MLR looks slightly better; and XGB performs the best.

The fact that for Mad and Max with 2% of the training data there is very little variation in Kendall’s *τ* for SA and SQA (and to a lesser extent, DW), is a consequence of the choice of hyper-parameters. The specific values of the hyper-parameters that gave optimal value of Kendall’s *τ* during the calibration phase are shown in Supplementary Material, Table S2, but we note here that the value of *λ* is quite high. *λ* controls the model complexity and is closely related to the bias-variance tradeoff, which states that it is impossible to simultaneously minimize errors from both bias and variance. A large value of *λ* introduces a large bias;^{59} consequently, for the cases where there is no or little variance, SA and SQA are essentially extracting the same pattern from all the training data. For the ranking tasks shown here with training on about 30 sequences, this gives the best performance for SA and SQA. It may be unsurprising, however, that a large value of *λ* be appropriate for small data sets; over-fitting may be a greater concern with smaller amounts of data.

The results presented in Fig. 3 suggest a precise case where current quantum technology may offer slight performance advantages relative to classical computational approaches; that is, when there is only a small amount of experimental training data available (about 30 sequences in our specific cases). In both classification and ranking tasks, DW performs comparably to SA and SQA and better than Lasso and XGB. MLR performs comparably with the annealing methods, but its error bars are much larger, indicating that its performance is less stable and more dependent on the training data. Moreover, the similarity between DW and SQA suggests that for small training sizes DW is functioning very nearly like a noiseless quantum annealer as captured by quantum Monte Carlo simulations. On a larger size of the training data DW’s performance decreases relative to the classical approaches for all three TFs, though results are still competitive. The decrease in the performance of all annealing methods (DW, SA, and SQA) seems to indicate a limitation on using methods with discrete weights, which enforce simpler models. Such models may be more advantageous with a small number of training samples because they prevent overfitting. However, with larger amounts of training data, a simpler model may not adequately learn the variation within the data and hence suffer worse performance. Nevertheless, the fact that Lasso uses the same L_{1} norm as the annealing methods (i.e., DW, SA and SQA), yet does not perform as well, indicates an advantage of such annealing methods when training with a small number of sequences. This is consistent with the finding reported in ref. ^{18}.

### Weight logos from feature weights

Since the one-hot encoding was used with a linear model to represent DNA sequence, the weights returned by DW, SA, and MLR reflect the relative importance of the corresponding nucleotide at a position in the sequence for the binding score. The magnitude of these feature weights for DW, SA, and MLR can be visualized as a 'weight logo' and are presented in Fig. 4 for the Mad, Max, and Myc gcPBM data sets. XGB, which finds an ensemble of trees, does not assign weights to individual nucleotides and hence does not easily lend itself to visualization. Similar plots for SQA and Lasso are shown in the Supplementary Material Sec. SIIB and Fig. S7. The weight logos show the contribution of nucleotides at particular positions to the strength of binding. The contribution of a nucleotide at a particular position in the sequence is represented by its height; nucleotides with the smallest weights are at the bottom and those with the largest weights are at the top. These weight logos in Fig. 4 were obtained by averaging the weights from the 50 training instances of the same number of sequences with the AUPRC as the objective. In other words, the logo represents the average of the weights that give the AUPRCs shown in Fig. 3a. DW, SA, and MLR all perform very similarly and give weight logos that are in good agreement with the expected consensus sequence, CANNTG. This demonstrates that all methods are able to capture biologically relevant information.

### Performance on HT-SELEX Data

HT-SELEX^{41,43} is a method for investigating the relative binding affinity of a TF for a particular sequence of DNA, an in vitro technique complementary to PBM. We present results for the Max homodimer and TCF4, another member of the bHLH family with consensus sequence CANNTG, using data from HT-SELEX experiments.^{47} The Max data set consisted of 3200 sequences of 12 bp in length, and the TCF4 data set was modified to contain 1800 sequences of 12 bp in length.

The procedure for splitting each data set into test and training data was similar to that described earlier for the gcPBM data sets (see Fig. 2). There was no overlap between training and testing data. The quantitative results for classification and ranking performance of the six different machine learning approaches are summarized in Fig. 5a,b, and the weight logos for DW, SA, and MLR in Fig. 5c. As with the gcPBM data, when training with about 30 sequences (1% of training data for Max and 2% of the training data for TCF4), DW, SA and SQA exhibit the best performance on the test data set. MLR matches the annealing protocols with a threshold at the 70th and 80th percentile of the data, but does worse at the higher percentiles of the data (left column in Fig. 5a,b). Lasso and XGB have the poorest performance. When training with about 150 (5% of training data for Max and 10% of the training data for TCF4) sequences, XGB performs very well, as on the gcPBM data sets with more training data, and MLR does well on the Max data set but rather poorly on the TCF4 data set; Lasso is comparable to MLR. DW’s performance is worse than the best performing method (XGB), but comparable to the other methods (right column in Fig. 5,b). XGB’s performance on the TCF4 data set is much better than the other methods, except when thresholding at the 99th percentile.

In terms of Kendall’s *τ* (Fig. 5a,b, bottom), all methods have similar performance when training with about 30 sequences, with the exception of XGB which does not do as well on the Max data set. When training with about 150 sequences, XGB gives the best ranking performance, as it did with the gcPBM data, and the other methods all perform similarly. Finally, the weight logos in Fig. 5c indicate that DW, SA, and MLR capture patterns in the data that give good agreement with the expected consensus sequence. The weight logos for the Max and TCF4 HT-SELEX data sets from SQA and Lasso are reported in Supplementary Material Fig. S8.

## Discussion

In this work we have explored the possibility of using a machine learning algorithm based on QA to solve a simplified but actual biological problem, the classification and ranking of TF-DNA binding events. This is the first application of QA to real biological data.

We have shown that DW performs comparably or slightly better than classical counterparts for classification when the training size is small, and competitively for ranking tasks. This trend is consistent with results on older sets of gcPBM and HT-SELEX data for various TFs, which are reported in Supplementary Material, Sec. SIII, Figs. S9–S14. Moreover, these results are consistent with a similar approach for the Higgs particle classification problem,^{18} where DW and SA both outperformed XGB with small training sizes, with a slight occasional advantage for DW over SA. This robustness across completely different application domains suggests that these findings represent real present-day advantages of annealing approaches over traditional machine learning in the setting of small-size training data. In areas of research where data sets with a small number of relevant samples may be more common, a QUBO approach such as QA realized via DW may be the algorithm of choice. On the other hand, when data is plentiful, some of the other state-of-the-art classical algorithms may be a better choice.

We have also demonstrated that the feature weights obtained by DW reflect biological knowledge; the weight logos for the TF-DNA binding data from gcPBM and HT-SELEX are consistent with the consensus binding site. This gives some confidence that QA is learning relevant biological patterns from the data. Yet, the approach is not without limitations. One limitation comes from the use of a single-nucleotide model to encode the DNA binding sites. In fact, we implicitly used a simple model that assumes independence between positions in the sequence. This is not always a valid approximation; higher-order 'k-mer' features or other 'shape' features that account for interdependencies between nucleotide positions may enhance model precision.^{44,45,47,60,61} We are limited to this simple model because of major technological constraints on the number of available qubits, which limits the number of features that can be used and thus the length of sequences that can be examined. The DW2X processor used for this study has 1098 functional qubits, but because of a sparse connectivity between qubits, only 40 or so features can actually be implemented on the device and in our study (see Supplementary Material, Sec. SIA for more details). Another serious limitation is the use of discrete weights. Discrete weights seem to be advantageous with a small number of training samples, as they enforce simpler models and are less prone to overfitting. However, as the amount of training data increases, these simpler models do not fare as well as some of the classical methods, which allow for greater numerical precision in the weights.

Despite these limitations, it is encouraging to see competitive performance for the simplified problem we have studied here. Although the performance advantage from annealing-type optimizers makes it difficult to solely attribute the performance to quantumness, this work may inspire future investigations into the power of QA devices. As quantum technology continues to develop and advance, it is possible that some of the practical limitations will be addressed and the range of problems that can be explored will be expanded.

## Methods

### QUBO mapping of TF–DNA binding problem

After processing the experimental data sets of *N* sequences of fixed length *L* and a measure of the binding affinity, we obtained the restricted data sets to which we applied six different machine learning strategies. Data sets were formulated as ${\left\{\left({\overrightarrow{\varphi}}_{n},{y}_{n}\right)\right\}}_{n=1}^{N}$, where ${\overrightarrow{\varphi}}_{n}\equiv {\left({\varphi}_{n,1},\dots ,{\varphi}_{n,4L}\right)}^{T}$ is the transformed feature vector, and *y*_{
n
} is the binding affinity. Solving for the simplest model is equivalent to finding a vector of binary weights $\overrightarrow{w}=\left({w}_{1},\dots ,{w}_{4L}\right)$, where *w*_{
i
} ∈ {0, 1}, such that the quantity

is minimized. The problem can then be specified as finding a ${\overrightarrow{w}}_{\mathrm{opt}}$ such that

where *λ* is a regularization (penalty) term included to prevent overfitting and ${\u2225\overrightarrow{w}\u2225}_{1}={\sum}_{m}{w}_{m}$ is the number of non-zero weights. To represent the above as an Ising problem, note that we can rewrite Eq. (4) as follows:

where

Constants that do not affect the optimization are dropped in the latter step. This procedure demonstrates that the problem of TF-DNA binding can be formulated as a QUBO problem, which in turn can easily be transformed into an Ising Hamiltonian of the form in Eq. (1) and passed to DW. The data normalization procedure is described in Supplementary Material, Sec. SIC.

### Technical details of algorithms

In order to solve practical problems of interest on DW, an embedding procedure must be used (see Supplementary Material, Sec. SIA). Some additional preprocessing was also performed for DW and SA to ensure that all response values were feasible (see Supplementary Material, Sec. SIC). DW, SA, SQA, MLR, Lasso and XGB were run on the same set of instances for assessment of the quantum annealer on the chosen problem. The experimental quantum processor, DW2X, was designed and built by D-Wave Systems, Inc. For each instance a total of 10,000 anneals ('runs') were collected from the processor, run with an annealing time of 20 μs. SA and SQA are classical analogs of QA that perform annealing on a classical and path integral Monte Carlo simulation of the Ising spin glass, respectively. SA and SQA were run with 10,000 sweeps (each sweep is an update of all spins) per repetition (or 'anneals') with an initial inverse temperature of 0.1 and a final inverse temperature of 3, for a total of 10,000 repetitions. The SA code was adapted from ref. ^{49}, and an in-house version of SQA was used. MLR is a widely used technique to minimize the loss function shown in Eq. (4), with the convex penalty term $\lambda {\u2225\overrightarrow{w}\u2225}_{2}^{2}$ instead of the linear penalty term. Lasso has the linear penalty term;^{51} XGB uses boosted trees.^{52} The weights $\overrightarrow{w}$ returned by MLR, Lasso and XGB are real-valued, whereas the weights returned by DW, SA and SQA (which solve a QUBO/Ising problem) are binary. In addition, DW, SA and SQA are probabilistic, meaning that a distribution of weights with different energies [the value of *H*_{P} in Eq. (1)] are returned. Up to 20 of the lowest energy weights were included for both DW, SA and SQA (see Supplementary Material, Sec. SID for more details). The lower the energy, the better the particular solution is at minimizing Eq. (4). In contrast, MLR, Lasso and XGB are deterministic and return a single solution.

In the calibration phase, only one hyper-parameter, *λ* was tuned for DW, SA, SQA, MLR and Lasso. All five methods were tuned separately for both classification and ranking tasks, resulting in different optimal *λ* for each method (see Supplementary Material Tables S2 and S3 for final values of *λ*). With an older data set we varied both the number of sweeps for SA and the value of *λ* but results were not significantly different; hence, here we only vary *λ* for SA. SA also has various other parameters that are related to the algorithm itself, including number of runs, initial and final temperature, and the cooling schedule, all of which affect the optimization performance. These parameters were not tuned. Similar additional parameters for DW, including annealing time and number of runs, were not tuned either. XGB’s performance depends on several hyper-parameters, and more careful tuning was necessary in order to give competitive performance. XGB parameters^{52} that were considered include *γ*, the max_depth, and min_child_weight (all of which control model complexity), subsample, colsample_bytree, (which add randomness to make training robust to noise), as well as learning rate, η. Rather than doing a full grid search over all these parameters, parameters were tuned sequentially; i.e., one value of *η* was fixed, then the best value of max_depth and min_child_weight were found. The optimal *γ* for those values was then found; and finally subsample and colsample_bytree tuned. *η* was then varied and the process repeated. *η* was varied from 0.05 to 0.3, max_depth from 3 to 20, min_child_weight from 1 to 20, *γ* from 0 to 1, and subsample and colsample_bytree both from 0.6 to 1.

In the testing phase, we evaluated performance based on two metrics: the AUPRC for classification performance and Kendall’s *τ* for ranking performance. For the AUPRC, we reported mean values with standard deviations as error bars, whereas for Kendall’s *τ* the median value was presented.

### Data processing and availability

Original probes for the gcPBM^{40} data contained 16,000–18,000 sequences of 36 bp in length with the fluorescence intensity as a measure of binding affinity. The same data is used in^{44} and may be downloaded from GEO (https://www.ncbi.nlm.nih.gov/geo/) under accession number GSE59845. Because of current limitations of the architecture of the DW device that limit the number of features that may be used, the data was truncated to the central 10 bp. For each sequence of 10 bp, we calculated its average gcPBM signal. In other words, all sequences in the data sets were unique. The final Mad, Max, and Myc data sets consisted of 1655, 1642, and 1584 sequences, respectively, of length 10 bp, and the logarithm base 2 with fluorescence intensities was used. The HT-SELEX data came from mammalian TFs^{42} that was re-sequenced with on average 10-fold increase in sequencing depth.^{47} The sequencing data is available at the European Nucleotide Archive (ENA—https://www.ebi.ac.uk/ena; study identifier PRJEB14744) and was pre-processed following the protocol in ref. ^{47} After this first step of pre-processing, the Max and TCF4 data sets consisted of 3209 and 15,556 sequences of length 12 and 14 bp, respectively. The Max data set did not require further truncation, but one bp on the left and right flanks were trimmed for the TCF4 data set, giving a modified data set of 1826 sequences of length 12 bp. As with the gcPBM data, the relative affinity was averaged for each truncated sequence.

## Additional information

**Publisher's note:** Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

## References

- 1.
Shor, P. W. Polynomial-time algorithms for prime factorization and discrete logarithms on a quantum computer.

*SIAM J. Comput.***26**, 1484 (1997). - 2.
Feynman, R. P. Quantum mechanical computers.

*Opt. News***11**, 11–20 (1985). - 3.
Lloyd, S. Universal quantum simulators.

*Science***273**, 1073–1078 (1996). - 4.
Neven, H., Denchev, V. S., Rose, G. & Macready, W. G. Training a binary classifier with the quantum adiabatic algorithm. Preprint at https://arxiv.org/abs/0811.0416 (2008).

- 5.
Pudenz, K. L. & Lidar, D. A. Quantum adiabatic machine learning.

*Quantum Inf. Process***12**, 2027–2070 (2013). - 6.
Lloyd, S., Mohseni, M. & Rebentrost, P. Quantum algorithms for supervised and unsupervised machine learning. Preprint at https://arxiv.org/abs/1307.0411 (2013).

- 7.
Wittek, P.

*Quantum Machine Learning: What Quantum Computing Means to Data Mining*. (Elsevier, AP, 2014). - 8.
Lloyd, S., Mohseni, M. & Rebentrost, P. Quantum principal component analysis.

*Nat. Phys.***10**, 631–633 (2014). - 9.
Rebentrost, P., Mohseni, M. & Lloyd, S. Quantum support vector machine for big data classification.

*Phys. Rev. Lett.***113**, 130503 (2014). - 10.
Wiebe, N., Kapoor, A., and Svore, K. M. Quantum deep learning. Preprint at https://arxiv.org/abs/1412.3489 (2014).

- 11.
Schuld, M., Sinayskiy, I. & Petruccione, F. An introduction to quantum machine learning.

*Contemp. Phys.***56**, 172–185 (2015). - 12.
Wiebe, N., Kapoor, A. & Svore, K. Quantum algorithms for nearest-neighbor methods for supervised and unsupervised learning.

*Quantum Inf. Comput.***15**, 0318–0358 (2015). - 13.
Aaronson, S. Read the fine print.

*Nat. Phys.***11**, 291–293 (2015). - 14.
Adachi, S. H. & Henderson, M. P. Application of quantum annealing to training of deep neural networks. Preprint at https://arxiv.org/abs/1510.06356 (2015).

- 15.
Amin, M. H., Andriyash, E., Rolfe, J., Kulchytskyy, B. & Melko, R. Quantum boltzmann machine. Preprint at https://arxiv.org/abs/1601.02036 (2016).

- 16.
Biamonte, J. et al. Quantum machine learning. Preprint at https://arxiv.org/abs/1611.09347 (2016).

- 17.
Benedetti, M., Realpe-Gómez, J., Biswas, R. & Perdomo-Ortiz, A. Quantum-assisted learning of graphical models with arbitrary pairwise connectivity. Preprint at https://arxiv.org/abs/1609.02542 (2016).

- 18.
Mott, A., Job, J., Vlimant, J.-R., Lidar, D. & Spiropulu, M. Solving a higgs optimization problem with quantum annealing for machine learning.

*Nature***550**, 375 EP– (2017). - 19.
Kadowaki, T. & Nishimori, H. Quantum annealing in the transverse Ising model.

*Phys. Rev. E***58**, 5355 (1998). - 20.
Farhi, E. et al. A quantum adiabatic evolution algorithm applied to random instances of an NP-complete problem.

*Science***292**, 472–475 (2001). - 21.
Albash, T. & Lidar, D. A. Adiabatic quantum computing. Preprint at https://arxiv.org/abs/1611.04471 (2016).

- 22.
Harris, R. et al. Experimental investigation of an eight-qubit unit cell in a superconducting optimization processor.

*Phys. Rev. B***82**, 024511 (2010). - 23.
Johnson, M. W. et al. Quantum annealing with manufactured spins.

*Nature***473**, 194–198 (2011). - 24.
Bunyk, P. I. et al. Architectural considerations in the design of a superconducting quantum annealing processor.

*IEEE Trans. Appl. Supercond.***24**, 1–10 (2014). - 25.
Kato, T. On the adiabatic theorem of quantum mechanics.

*J. Phys. Soc. Jpn.***5**, 435 (1950). - 26.
Farhi, E., Goldstone, J., Gutmann, S. & Sipser, M. Quantum computation by adiabatic evolution. Preprint at https://arxiv.org/abs/quant-ph/0001106 (2000).

- 27.
Jansen, S., Ruskai, M.-B. & Seiler, R. Bounds for the adiabatic approximation with applications to quantum computation.

*J. Math. Phys.***48**, 102111 (2007). - 28.
Childs, A. M., Farhi, E. & Preskill, J. Robustness of adiabatic quantum computation.

*Phys. Rev. A***65**, 012322 (2001). - 29.
Amin, M. H. S., Averin, D. V. & Nesteroff, J. A. Decoherence in adiabatic quantum computation.

*Phys. Rev. A***79**, 022107 (2009). - 30.
Albash, T. & Lidar, D. A. Decoherence in adiabatic quantum computation.

*Phys. Rev. A***91**, 062320 (2015). - 31.
Barahona, F. On the computational complexity of Ising spin glass models.

*J. Phys. A Math. Gen.***15**, 3241 (1982). - 32.
Rønnow, T. F. et al. Defining and detecting quantum speedup.

*Science***345**, 420–424 (2014). - 33.
Katzgraber, H. G., Hamze, F. & Andrist, R. S. Glassy chimeras could be blind to quantum speedup: Designing better benchmarks for quantum annealing machines.

*Phys. Rev. X***4**, 021008– (2014). - 34.
Venturelli, D. et al. Quantum optimization of fully connected spin glasses.

*Phys. Rev. X***5**, 031040– (2015). - 35.
Hen, I. et al. Probing for quantum speedup in spin-glass problems with planted solutions.

*Phys. Rev. A***92**, 042325– (2015). - 36.
Amin, M. Searching for quantum speedup in quasistatic quantum annealers. Preprint at https://arxiv.org/abs/1503.04216 (2015).

- 37.
Slattery, M. et al. Absence of a simple code: how transcription factors read the genome.

*Trends Biochem. Sci.***39**, 381–399 (2014). - 38.
Shlyueva, D., Stampfel, G. & Stark, A. Transcriptional enhancers: from properties to genome-wide predictions.

*Nat. Rev. Genet.***15**, 272–286 (2014). - 39.
Stormo, G. & Zhao, Y. Determining the specificity of dna-protein interactions.

*Nat. Rev. Genet.***11**, 751–760 (2010). - 40.
Gordân, R. et al. Genomic regions flanking e-box binding sites influence DNA binding specificity of bHLH transcription factors through DNA shape.

*Cell Rep.***3**, 1093–1104 (2013). - 41.
Jolma, A. et al. Multiplexed massively parallel selex for characterization of human transcription factor binding specificities.

*Genome Res.***20**, 861–873 (2010). - 42.
Jolma, A. et al. Dna-binding specificities of human transcription factors.

*Cell***152**, 327–339 (2013). - 43.
Liu, J. & Stormo, G. D. Combining selex with quantitative assays to rapidly obtain accurate models of protein-dna interactions.

*Nucleic Acids Res.***33**, e141 (2005). - 44.
Zhou, T. et al. Quantitative modeling of transcription factor binding specificities using DNA shape.

*Proc. Natl. Acad. Sci.***112**, 4654–4659 (2015). - 45.
Yang, L. et al. TFBSshape: a motif database for DNA shape features of transcription factor binding sites.

*Nucleic Acids Res.***42**, D148–D155 (2014). - 46.
Abe, N. et al. Deconvolving the recognition of DNA shape from sequence.

*Cell***161**, 307–318 (2015). - 47.
Yang, L. et al. Transcription factor family-specific DNA shape readout revealed by quantitative specificity models.

*Mol. Syst. Biol.***13**, 910 (2017). - 48.
Kirkpatrick, S., Gelatt, C. D. & Vecchi, M. P. Optimization by simulated annealing.

*Science***220**, 671–680 (1983). - 49.
Isakov, S., Zintchenko, I., Rønnow, T., and Troyer, M. Optimized simulated annealing for ising spin glasses. Preprint at https://arxiv.org/abs/1401.1084 (2014).

- 50.
Santoro, G. E., Martoňák, R., Tosatti, E. & Car, R. Theory of quantum annealing of an Ising spin glass.

*Science***295**, 2427–2430 (2002). - 51.
Tibshirani, R. Regression shrinkage and selection via the lasso.

*J. R. Stat. Soc. Series B Stat. Methodol.***58,**267–288 (1996). - 52.
Chen, T. & Guestrin, C. in XGBoost: a scalable tree boosting system.

*Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining*. 785–794 (ACM, New York, NY, 2016). - 53.
Crosson, E. & Harrow, A. Simulated quantum annealing can be exponentially faster than classical simulated annealing. Preprint at https://arxiv.org/abs/1601.03030 (2016).

- 54.
Chen, T. & He, T. in Higgs boson discovery with boosted trees.

*Proceedings of the 2014 International Conference on High-EnergyPhysics and Machine Learning***42**, 69–80 (2014). - 55.
Breiman, L. Bagging predictors.

*Mach. Learn.***24**, 123–140 (1996). - 56.
Davis, J. & Goadrich, M. in The relationship between precision-recall and roc curves

*Proceedings of the 23rd International Conference on Machine Learning*. 233–240 ACM: New York, NY, 2006. - 57.
Kendall, M. A new measure of rank correlation.

*Biometrika***30**, 81–93 (1938). - 58.
Grandori, C., Cowley, S., James, L. & Eisenman, R. The Myc/Max/Mad network and the transcriptional control of cell behavior.

*Annu. Rev. Cell Dev. Biol.***16**, 653–699 (2000). - 59.
Cucker, F. & Smale, S. Best choices for regularization parameters in learning theory: on the bias-variance problem.

*Found. Comput. Math.***2**, 413–428 (2002). - 60.
Mordelet, F., Horton, J., Hartemink, A. J., Engelhardt, B. E. & Gordân, R. Stability selection for regression-based models of transcription factor-dna binding specificity.

*Bioinformatics***29**, i117–i125 (2013). - 61.
Zhou, T. et al. DNAshape: a method for the high-throughput prediction of DNA structural features on a genomic scale.

*Nucleic Acids Res.***41**, W56–W62 (2013).

## Acknowledgements

The authors thank Lin Yang for assistance with pre-processing the gcPBM and HT-SELEX data. The authors would also like to thank Tameem Albash for providing the SQA code. This work was supported by the USC Women in Science and Engineering Program (to R.D.F), the National Institutes of Health grants R01GM106056 and U01GM103804 (to R.R.), ARO grant number W911NF-12-1-0523 and NSF grant number INSPIRE-1551064 (to D.L.). R.R. is an Alfred P. Sloan Research Fellow. D.L. is a Guggenheim Foundation Fellow and a Caltech Moore Scholar in Physics. The research is based upon work (partially) supported by the Office of the Director of National Intelligence (ODNI), Intelligence Advanced Research Projects Activity (IARPA), via the U.S. Army Research Office contract W911NF-17-C-0050. The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies or endorsements, either expressed or implied, of the ODNI, IARPA, or the U.S. Government. The U.S. Government is authorized to reproduce and distribute reprints for Governmental purposes notwithstanding any copyright annotation thereon.

## Author information

### Affiliations

#### Department of Chemistry, University of Southern California, Los Angeles, CA, 90089, USA

- Richard Y. Li
- , Remo Rohs
- & Daniel A. Lidar

#### Computational Biology and Bioinformatics Program, Department of Biological Sciences, University of Southern California, Los Angeles, CA, 90089, USA

- Richard Y. Li
- , Rosa Di Felice
- & Remo Rohs

#### Center for Quantum Information Science & Technology, University of Southern California, Los Angeles, CA, 90089, USA

- Richard Y. Li
- & Daniel A. Lidar

#### Department of Physics and Astronomy, University of Southern California, Los Angeles, CA, 90089, USA

- Rosa Di Felice
- , Remo Rohs
- & Daniel A. Lidar

#### Center S3, CNR Institute of Nanoscience, Via Campi 213/A, 41125, Modena, Italy

- Rosa Di Felice

#### Department of Computer Science, University of Southern California, Los Angeles, CA, 90089, USA

- Remo Rohs

#### Department of Electrical Engineering, University of Southern California, Los Angeles, CA, 90089, USA

- Daniel A. Lidar

### Authors

### Search for Richard Y. Li in:

### Search for Rosa Di Felice in:

### Search for Remo Rohs in:

### Search for Daniel A. Lidar in:

### Contributions

R.L, R.D.F, R.R. and D.L designed and conceived the study. R.L. implemented and executed DW, SA, SQA, MLR, Lasso and XGB, and analyzed the results. R.R. and D.L directed the study. R.L, R.D.F, R.R., and D.L wrote the manuscript.

### Competing interests

The authors declare no competing financial interests.

### Corresponding authors

Correspondence to Remo Rohs or Daniel A. Lidar.

## Electronic supplementary material

**Open Access** This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.