Article | Open

Inferring time derivatives including cell growth rates using Gaussian processes

  • Nature Communications 7, Article number: 13766 (2016)
  • doi:10.1038/ncomms13766
  • Download Citation
Published online:


Often the time derivative of a measured variable is of as much interest as the variable itself. For a growing population of biological cells, for example, the population’s growth rate is typically more important than its size. Here we introduce a non-parametric method to infer first and second time derivatives as a function of time from time-series data. Our approach is based on Gaussian processes and applies to a wide range of data. In tests, the method is at least as accurate as others, but has several advantages: it estimates errors both in the inference and in any summary statistics, such as lag times, and allows interpolation with the corresponding error estimation. As illustrations, we infer growth rates of microbial cells, the rate of assembly of an amyloid fibril and both the speed and acceleration of two separating spindle pole bodies. Our algorithm should thus be broadly applicable.


Estimating the time derivatives of a signal is a common task in science. A well-known example is the growth rate of a population of cells, which is defined as the time derivative of the logarithm of the population size1 and is used extensively in both the life sciences and biotechnology.

A common approach to estimate such derivatives is to fit a mathematical equation that, say, describes cellular growth and so determine the maximum growth rate from the best-fit value of a parameter in the equation2. Such parametric approaches rely, however, on the mathematical model being a suitable description of the underlying biological or physical process and, at least for cellular growth, it is common to find examples where the standard models are not appropriate3.

The alternative is to use a non-parameteric method and thus estimate time derivatives directly from the data. Examples include taking numerical derivatives4 or using local polynomial or spline estimators5. Although these approaches do not require knowledge of the underlying process, it can be difficult to determine the error in their estimation5 and to incorporate experimental replicates, which with wide access to high-throughput technologies, are now the norm.

Here we develop a methodology that uses Gaussian processes to infer both the first and second time derivatives from time-series data. One advantage of using Gaussian processes over parametric approaches is that we can fit a wider variety of data. Rather than assuming that a particular function characterizes the data (a particular mathematical equation), we instead make assumptions about the family of functions that can describe the data. An infinite number of functions exist in this family and the family can capture many more temporal trends in the data than any one equation. The advantages over existing non-parametric methods are that we can straightforwardly and systematically combine data from replicate experiments (by simply pooling all data sets) and predict errors both in the estimations of derivatives and in any summary statistics. A potential disadvantage because we use Gaussian processes is that we must assume that the measurement noise has a normal or log-normal distribution (as do many other methods), but we can relax this assumption if there are multiple experimental replicates.

To illustrate how our approach predicts errors and can combine information from experimental replicates, we first focus on inferring growth rate from measurements of the optical density of a growing population of biological cells. Plate readers, which are now widespread, make such data easy to obtain, typically with hundreds of measurements and often at least three to ten replicates. We will also, though, show other examples: estimating the rate of in vitro assembly of an amyloid fibril and inferring the speed and acceleration of two separating spindle pole bodies in a single yeast cell.


An overview of Gaussian processes

A Gaussian process is a collection of random variables for which any subset has a joint Gaussian distribution6. This joint distribution is characterized by its mean and its covariance.

To use a Gaussian process for inference on time series, we assume that the data can be described by an underlying, or latent, function and we wish to infer this latent function given the observed data. For each time point of interest, we add a random variable to the Gaussian process. With n time points, there are therefore n corresponding random variables in the Gaussian process. The latent function is given by the values taken by these random variables (Fig. 1a). Without losing any generality, we set the mean of each random variable to be zero6.

Figure 1: An overview of inference with Gaussian processes.
Figure 1

(a) A graphical model of a Gaussian process6. Squares denote known variables (times, ti, and data points, yi); circles denote unknown variables (the underlying, latent function, fi). We associate a variable in the Gaussian process to each time point and the value of this variable gives the value of the latent function. Each observed data point, yi, depends only on the corresponding latent variable, fi. Each f variable, however, depends on all the other f variables (they covary). (b) Four examples of latent functions with a squared exponential covariance function. The functions are strictly only defined at the time points of the observations (shown with black semi-circles on the x axis) but are drawn with a continuous line for clarity. (c) Four examples of latent functions after conditioning on the data (data are shown as blue dots). Although each individual function is smooth, there is more variation between functions where the data is more spread. Averaging many latent functions gives the best fit. The hyperparameters of the covariance matrix are the same as those in b.

How each random variable in the Gaussian process depends on the other random variables both at earlier and later times, that is, how the variables covary, determines the shape of the latent function. For example, if each random variable does not covary with any other (the covariance matrix of the Gaussian process is the identity matrix), then the latent function will randomly jump back and forth around zero. If each random variable covaries equally with every other random variable (all the entries of the covariance matrix are one), then the functions sampled will be straight horizontal lines starting at the value of the random variable associated with the first time point. More pertinently, if the covariance for each random variable is positive for those random variables close in time and tending to zero for random variables far away in time, then the functions generated vary, but do so smoothly.

To proceed, we therefore must choose the type of covariance function for the Gaussian process and, in doing so, we necessarily make some assumptions about the latent functions that underlie the data. We will often use a squared exponential covariance, which imposes little restriction on the shape of the latent function other than to assume that it is smooth (infinitely differentiable; Fig. 1b). The Matern covariance function (paramaterized with ν=5/2) relaxes this smoothness assumption and imposes that the latent function is only twice differentiable6. Another alternative is the neural network covariance, which tends to generate sigmoid-shaped latent functions6. We emphasize that choosing a covariance function to describe the latent function puts much less restriction on its shape than the more traditional choice of selecting a particular mathematical equation to model the latent function.

Each covariance function is parameterized in its own way and we must find the appropriate values of these parameters given the data. More correctly, the parameters are called hyperparameters (Methods) and determining the hyperparameters is the computationally intensive part of the inference.

Once the hyperparameters are optimized, we can sample from the Gaussian process to generate a latent function that is consistent with our data (Fig. 1c). The mean latent function, which we could find from averaging many samples but can also calculate directly, gives the ‘best-fit’ and the variance of the latent function provides an estimation of the error in the inference.

Following on from earlier work7, we adapt this standard inference procedure to also allow the inference of the first two time derivatives of the latent function, because these time derivatives are, in many applications, of more interest than the latent function itself. Errors in inference of the latent function are automatically carried through to the errors in inferring time derivatives.

Verification of the algorithm

To verify our algorithm’s inference of first and second time derivatives, we followed the tests of De Brabanter et al.8. Gaussian measurement noise was added to the same analytic functions chosen by De Brabanter et al.8 for which time derivatives can be found exactly and the mean absolute difference between the inferred derivative and the exact derivative was used to score the inference (see ref. 8 for details—the end points are not included). We show the distribution of scores for 100 different data sets each with a different sample of the measurement noise (Fig. 2).

Figure 2: The inference method can perform better than alternatives.
Figure 2

(a) Inference of the first derivative. A box plot of error scores (related to the mean absolute difference between the inferred and exact derivative) for inference of the first derivative. We use either a squared exponential covariance function or a neural network covariance function for the Gaussian process (GP) and compare with local polynomial regression (with p=3) and a quintic penalized smoothing spline (data for both from ref. 8). Top left shows one sample data set (in red with 500 data points), the true underlying function (in green) and the inferred latent function using a neural network covariance function—the best fit (in blue); top right shows the corresponding first derivative (with here an error score of 0.64): exact (in green) and inferred (in blue). Equivalent plots for the alternative inference methods are given by De Brabanter et al.8. Errors (in light blue) are s.d. (b) Inference of the second derivative. A box plot of scores for inference of the second derivative. The two alternatives are local polynomial regression (with p=5) and a septic penalized smoothing spline (data for both from ref. 8). Top right shows one sample data set (in red with 1,500 data points), the underlying function (in green) and the inferred latent function using a neural network covariance function (in blue); top left shows the corresponding second derivative (with here an error score of 26.2): exact (in green) and inferred (in blue).

For these tests, our method outperforms established alternatives. For illustration, we show results for both the squared exponential covariance function and the neural network covariance function. Independent of the choice, the method performs at least as well as alternatives (Fig. 2).

Estimating cellular growth rates

We now consider the inference of microbial growth rates (strictly, we infer the specific growth rate: the time derivative of the logarithm of the population size). The population size as a function of time is commonly fit to a parametric equation2, although these equations are restrictive and describe only a particular type of growth3. Therefore, to provide a further test of our algorithm, we considered a linear sum of two growth equations—the Gompertz9 and Richards10 models—to generate a growth curve that cannot in principle be fit by either, but where an exact expression for the first derivative can still be found. We compare our results with smoothing splines, an established non-parametric alternative3.

For these data sets and this magnitude of measurement noise, both methods perform equally well, but the inference using Gaussian processes becomes more robust as the number of data points increase (Fig. 3). We note that we have artificially favoured the smoothing spline, because the smoothing parameter for the spline is set with the variance used to generate the synthetic measurement noise. The Gaussian process methodology, in contrast, infers this variance. Despite the advantage of the spline-based inference, its median error is 45% higher when n=1,000.

Figure 3: Inference of growth rates with Gaussian processes becomes more robust with increasing numbers of data points.
Figure 3

(a) We show the inferred growth rate (the best fit with a squared exponential Gaussian process in blue with errors in light blue and from a cubic smoothing spline in purple) and the exact growth rate (in green) for synthetic data sets with either 10, 100 or 1,000 time points (insets, with the underlying growth curve in green). Even though we favour smoothing splines by setting the smoothing parameter to be proportional to the exact variance of the measurement noise (whereas the Gaussian process infers this parameter), both methods perform similarly. (b) A box plot of errors scores for inference of the growth rate from 100 data sets with randomly generated log-normal measurement noise. For the Gaussian process, the distribution of error scores tightens with increasing number of data points.

Turning to experimental measurements, we fit optical densities, which are proportional to the number of cells if properly calibrated11,12, and show that we can infer growth rates for two cases that cannot be easily described by parametric approaches3. The first exhibits a diauxic shift with two distinct phases of growth and the second shows an exceptionally long lag (Fig. 4). We infer the growth rate and the estimated errors in our inference as a function of time using all experimental replicates. Data from replicate measurements are pooled together and the algorithm applied as for a single replicate.

Figure 4: Microbial growth rates can be inferred as a function of time.
Figure 4

(a) A growth curve of S. cerevisiae in a mixture of 0.4% glucose and 1% galactose showing a diauxic shift (7 replicates, each with n=115). The best-fit (mean) latent function is shown in dark blue and the inferred growth rate is shown below. All error bars (light blue) are s.d. The inset shows, as an example, four sample estimates of the growth rate as a function of time (samples of the first derivative of the latent function—the corresponding samples of the latent function itself are not shown). (b) Growth of E. coli in hyperosmotic conditions with an unusually long lag and short growth period (two replicates, each with n=646) and the inferred growth rate. The magnitude of the measurement noise is here allowed to vary with time and empirically estimated across the replicates (Methods). Error bars (light blue) are s.d.

Having the inferred growth rate over time can make identifying different stages of the growth curve substantially easier than making this identification from the optical density data alone. For example, the local minimum in the growth rate of Fig. 4a is expected to indicate a shift from cells using glucose to using galactose. Inferring a time-dependent growth rate should increase the robustness of high-throughput automated studies, which usually focus on identifying exponential growth13,14.

Often summary statistics are used to describe a growth curve, such as the maximum growth rate and the lag time2, and we can estimate such statistics and their associated errors. From our inference, we can sample latent functions that are consistent with the data. Each sample provides an example of a latent function that ‘fits’ the data. To estimate errors in statistics, we generate say 100 samples of the latent function and its time derivatives (Fig. 4a inset). For each sample, we calculate the statistic for that sample, such as the maximum growth rate. We therefore obtain a probability distribution for the statistic and report the mean and s.d. of this distribution as the best-fit value and the estimated error (0.16±0.002 h−1 for the maximum growth rate for the data in Fig. 4a). A similar approach applies for any statistic that can be calculated from a single growth curve (Methods).

The data for Fig. 4b are considerably noisier than the data for Fig. 4a and the spread of data is larger at short times than at long times. The magnitude of the measurement noise changes with time. We typically assume that the measurement noise can be described by a Gaussian distribution with zero mean and a constant s.d. The magnitude of the measurement noise is determined by this s.d. and the s.d. here, for the data of Fig. 4b, appears to be time dependent (it is largest at early times). To empirically estimate the relative scale of this change, we calculate the variance across replicates at each time point. We assume that the magnitude of the measurement noise is a time-independent constant multiplied by this time-dependent relative scale and we fit that constant (Methods).

Further applications

As additional examples, we first infer the rate of assembly of an amyloid fibril as a function of time from in vitro data (Fig. 5a)15. Despite each replicate having high measurement noise compared with the microbial data, the rate of fibril assembly can be inferred accurately because of the multiple replicates. The second example is one where both the first and the second derivative are useful: estimating the speed of separation of the spindle poles during anaphase in a single cell of budding yeast (Fig. 5b). We demonstrate that we can infer both time derivatives and their errors from a single replicate. As expected, the size of the estimated error increases for the first derivative relative to the error in the regression and increases again for the second derivative. Changes in the speed of separation (extrema in the second derivative) are used to characterize anaphase16 into the fast, pause and slow elongation phases17. We chose a Gaussian process with a neural network covariance function for this data rather than the squared exponential covariance function used for the others: a difference that is important here because we only have a single replicate with few data points. The latent functions generated then tend to be flatter either side of the increase in separation, which leads to smoother inferences of the acceleration.

Figure 5: The algorithm has wide application.
Figure 5

(a) Inferring the in vitro rate of assembly of an amyloid fibril. Fluorescence data reporting the formation of fibrils in bovine insulin (at a concentration of 0.1 mg ml−1) by the binding of the dye Thioflavin T are shown (red dots) with 15 replicates (each with n=91)15. The best fit (top) and the inferred rate of fibril assembly (bottom) are shown in dark blue. We empirically estimate the magnitude of the measurement noise across the replicates. Errors (in light blue) are s.d. (b) Inferring the speed and acceleration of separation of the spindle poles in S. cerevisiae. The distance, s, between the two spindles in a single cell is plotted in microns as a function of time (red dots with n=57). The best fit and the inferred speed (middle) and acceleration (bottom) are shown in dark blue. The triangles denote turning points in the acceleration and separate anaphase into stages with fast and slow elongation separated by a pause16. Errors are s.d.


To conclude, we have introduced a non-parametric method that uses Gaussian processes to infer first and second derivatives from time-series data. In tests, our approach is at least as accurate as others (Figs 2 and 3), but has several advantages: it systematically estimates errors, both for the regression and the inferred derivatives; it allows interpolation with the corresponding error estimation (Gaussian processes were developed for interpolation6); and it allows sampling of the latent function underlying the data and so can be used to estimate errors in any statistic of that function by calculating the statistic for the samples.

For fitting growth curves, several alternatives exist3,18,19,20, which, although mostly focusing on parametric approaches, do allow spline fitting3 and polynomial regression18,20. Both approaches have been criticized, being sensitive to outliers and potentially having systematic biases5, and at least in the case of splines appear less robust (Fig. 3). Further, our software performs inference using all replicates, can infer second derivatives and rigorously estimates errors. Where error estimation in summary statistics has been addressed3, bootstrapping of the data is used. This approach is perhaps less suited for time-series data than our approach of sampling latent functions, because it leads to some randomly chosen data points being weighted more than others when generating sample fits.

Of the three we considered, we find that the squared exponential function is generally the best choice of covariance function when estimating time derivatives, because it typically results in the inference of first and second derivatives with a smoothness that is consistent with a priori expectations of the nature of the underlying dynamics. Although the Matern covariance is not as restrictive, because it constrains the smoothness of the latent functions less, it can lead to the inference of rough, fluctuating derivatives, in particular for the second derivative and if the magnitude of the measurement noise is high. For example, using the Matern covariance gives poor results for the data in Fig. 2a (with median error scores that are 60% higher than those for the squared exponential covariance), but performs slightly better (medians within 10%) for the less noisy data in Fig. 3. Finally, the neural network covariance, although perhaps the least prone to the inference of rough time derivatives, can be more sensitive to prior information: the hyperparameter controlling the flexibility of the latent function is optimized to its upper bound more often than for the other covariance functions. All three covariance functions are implemented in our code and can be tested for a new type of data.

Similar to any Bayesian method, prior information on bounds for the hyperparameters of the covariance function can affect the inference, although these bounds can typically be set so that the best-fit values are far from the bounds. In particular, how closely the latent function follows the data depends both on its flexibility and on the size of the measurement noise. An outlier can be followed if the flexibility is high or if the measurement noise is low. When there is not sufficient data, the algorithm, rightly in our opinion, requires prior information to make this choice. Alternative methods also require prior specification, such as the degree of smoothness in fitting with either splines or a local polynomial method, such as LOESS (locally weighted scatter plot smoothing). For a particular type of data, the bounds typically need to be set once allowing high-throughput analyses.

Measuring cellular growth rates is a daily task in many laboratories, but, if using a non-parameteric approach, researchers often follow the method developed in the early days of molecular biology: finding the gradient of a line fit to the portion of the growth curve that appears most straight on a semi-log plot. Our methodology takes advantages of advances in machine learning, to allow inference not only of the maximum growth rate but of the growth rate as a function of time. Time-dependent growth rates must capture more of the underlying biology, such as the time of the diauxic shift in Fig. 4a, but they have been little exploited. We believe that using more advanced inference techniques, such as the one based on Gaussian processes that we present here, in combination with developments in high-throughput technologies will transform our understanding of cellular growth and the factors that control it.


Using a Gaussian process to fit time-series data

In the following, we will denote a Gaussian distribution with mean μ and covariance matrix Σ as (μ, Σ) and use the notation of Rasmussen and Williams6 as much as possible.

Prior probability

For n data points yi at inputs xi (each xi is a time for a growth curve), we denote the underlying latent function as f(x). We define a covariance matrix k(x, x′), which has an explicit dependence on hyperparameters θ, and obeys

where the expectations are taken over the distribution of latent functions (samples of f(x)).

We interpret equation (1) as giving the prior probability distribution of the latent functions f(X), where were we use X to denote the inputs xi, such that

where K(X, X) is the n × n matrix with components k(xi, xj). With f denoting [f(x1), ..., f(xn)], this prior probability can be written as

noting the dependence of k(x, x′; θ) on the hyperparameters θ.

Marginal likelihood

After choosing a covariance function, to use Gaussian processes in regression, we must optimize the covariance function’s hyperparameters given the observed data. We will do so by maximizing the marginal likelihood, where the marginalization is made by integrating over all possible latent functions6. Once the parameters of the covariance function have been determined, we can sample latent functions given the data. We consider the squared exponential, Matern (with ν=5/2) and neural network covariance functions.

To optimize the hyperparameters given the data, we therefore consider the likelihood P(y|θ, X), which, more correctly, is a marginal likelihood

where the marginalization is over all choices of the latent function f evaluated at X.

If we assume that for all yi, yi=f(xi)+ where each is an independent Gaussian variable with zero mean and a s.d. of σi=σ for simplicity, then

where I is the n × n identity matrix. Equations (3 and 5) imply that the marginal likelihood is also Gaussian:

We use a maximum-likelihood method to find the hyperparameters and maximize the marginal likelihood equation (6). We have two hyperparameters for the squared exponential covariance function and the parameter, σ, which characterizes the measurement noise. We assume a bounded, uniform prior probability for each of these hyperparameters and use the Broyden–Fletcher–Goldfarb–Shanno algorithm4 to find their optimum values. Although one optimization run from random initial choices of the hyperparameters is usually sufficient, choosing the best from multiple runs can prevent the algorithm finding local maxima.

Making predictions

Given the optimum choice of the hyperparameters, we would like to generate sample latent functions at points X*, which to include the possibility of interpolation need not be the same as X, by sampling from P(f*|X, y, θ, X*). Using equation (6) and that the distribution of the latent function evaluated at X* is also Gaussian, we can write the joint probability of y and f* as, following ref. 6,

where K(X, X*) is the n × n* matrix with components k(xi, ).

Conditioning equation (7) on the data y, standard results for Gaussian distributions6 give that the probability distribution P(f*|X, y, θ, X*) is also Gaussian with mean

and covariance matrix

We use equations (8) and (9) to sample f*.

Inferring the first and second time derivatives

To determine the time derivative of the data, we use that the derivative of a Gaussian process is another Gaussian process6. We can therefore adapt standard techniques for Gaussian process to allow time derivatives to be sampled too.

Building on the work of Boyle7, we let g(x) and h(x) be the first and second derivatives with respect to x of the latent function f(x). If f(x) is a Gaussian process then so are both g(x) and h(x). Writing 1 and 2 for the partial derivatives with respect to the first and second arguments of a bivariate function, we have

and that

as well as

following ref. 21.

Consequently, the joint probability distribution for y and f*, g* and h* evaluated at points X* is again Gaussian (cf. equation (7)):

where we write K=K(X, X) and K*=K(X*, X*) for clarity.

The covariance function is by definition symmetric: k(xi, xj)=k(xj, xi) from equation (1). Therefore, and so

for all positive integers k and . Consequently, the covariance matrix in equation (13) is also symmetric.

Conditioning on y now gives that the distribution P(f*, g*, h*|X, y, θ, X*) is Gaussian with mean

and covariance matrix

Equation (16) includes equation (9) and shows that

which gives the error in the estimate of the first derivative7. Similarly,

is the error in estimating the second derivative.

Using an empirically estimated measurement noise

Although our derivation is given for a Gaussian process where the measurement errors in the data are independent and identically distributed with a Gaussian distribution of mean zero, the derivations are unchanged if the measurement noise has a different s.d. for each time point6.

When the magnitude of the measurement noise appears to change with time, we first empirically estimate the relative magnitude of the measurement noise by the variance across all replicates at each time point. We then smooth this estimate over time (with a Gaussian filter with a width of 10% of the total time of the experiment, but the exact choice is not important) and replace the identity matrix, I, in equations (6), (15) and (16) by a diagonal matrix with the relative measurement noise on the diagonal in order to make predictions.

Estimating the growth characteristics

From the growth curve, we estimate the maximum growth rate as the maximum time derivative of the logarithm of the growth curve2:

where we denote the growth curve as y(t). The doubling time is ln(2) times the inverse of the growth rate. We define the lag time as the intercept of the line parallel to the time axis that passes through the initial OD, y(0), and the tangent to the logarithm of the growth curve from the point on the growth curve with maximum growth rate (a standard choice2). If this point of maximum growth rate is at t=t*, then

For each characteristic, we can estimate measurement error through calculating the characteristic for 100 s of sampled latent growth curves.

Implementation and GUI

The code for our algorithm is freely available and written in Python 3 using NumPy22, SciPy, Matplotlib23, and the Pandas data analysis library (all available via the free Anaconda package) and is compatible with Microsoft’s Excel. We give an example script and data set and have written a GUI that runs on Windows, OS X, and Linux.

Bounds used for the hyperparameters are given in Table 1.

Table 1: Ranges of hyperparameters used for the examples.

Code availability

The software and instructions for its use are at

Generating synthetic data

To generate the synthetic data shown in Fig. 3, we use a weighted sum of a Gompertz model (parameters: A=1.1, μm=0.6 and λ=2.3; weight: 0.3) and a Richards model (parameters: A=1.5, μm=0.3, λ=4.3 and ν=0.8; weight: 0.7) using the notation of ref. 2. We added log-normal measurement noise with zero mean and a s.d. of σm=0.03 and used the SciPy implementation of a cubic spline and its time-derivative, setting the smoothing parameter to be m2 where n is the number of data points.

The error score in Fig. 3 is the mean absolute deviation of the inferred growth rate from the exact growth rate ignored 5% of the data points both at the beginning and end of the time series, to avoid endpoint effects potentially dominating the error8.

Experimental methods

Data for Fig. 4a was gathered using a Tecan Infinity M200 plate reader and a BY4741 strain of Saccharomyces cerevisiae growing in synthetic complete media supplemented with 0.4% glucose and 1% galactose at 30 °C, following an established protocol24. Optical density was measured at an absorbance wavelength of 595 nm every 11.4 min.

Data for Fig. 4b was gathered using a Spectrostar Omega microplate reader and a BW25113 strain of Escherichia coli growing in MM9 (sodium–sodium instead of sodium–potassium) media with 0.1% glucose and 1,106 mOsm sucrose at 37 °C. Optical density was measured at an absorbance wavelength of 600 nm every 7.5 min.

Data for Fig. 5a is from ref. 15.

Data for Fig. 5b was gathered using a custom spinning disk confocal microscope for 20 min in 20 s time steps with 50 ms exposure time per focal plane. Spindle pole bodies were labelled with Spc42-Cerulean. An image stack of 30 z-planes with 300 nm step size was gathered for each time point to allow the position of the spindle poles to be fitted to three-dimensional Gaussian distributions and tracked in time. Imaging, fitting and tracking followed an established protocol16.

Data availability

Data generated in this work is available at

Additional information

How to cite this article: Swain, P. S. et al. Inferring time derivatives including cell growth rates using Gaussian processes. Nat. Commun. 7, 13766 doi: 10.1038/ncomms13766 (2016).

Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.


  1. 1.

    The growth of bacterial cultures. Ann. Rev. Microbiol. 3, 371–394 (1949).

  2. 2.

    , , & Modeling of the bacterial growth curve. App. Environ. Microbiol. 56, 1875–1881 (1990).

  3. 3.

    , , , & grofit: Fitting biological growth curves with R. J. Stat. Softw. 33, 1–21 (2010).

  4. 4.

    , , & Numerical Recipes: The Art of Scientific Computing Cambridge Univ. Press (2007).

  5. 5.

    & in Proceedings of the 22nd International Workshop on Statistical Modelling (2007).

  6. 6.

    & Gaussian Processes for Machine Learning MIT Press (2006).

  7. 7.

    Gaussian Processes for Regression and Optimization (PhD thesis, Victoria Univ. Wellington (2007).

  8. 8.

    et al. Derivative estimation with local polynomial fitting. J. Mach. Learn. Res. 14, 281–301 (2013).

  9. 9.

    On the nature of the function expressive of the law of human mortality, and on a new mode of determining the value of life contingencies. Philos. Trans. R. Soc. Lond. 115, 513–585 (1825).

  10. 10.

    A flexible growth function for empirical use. J. Exp. Bot. 10, 290–300 (1959).

  11. 11.

    & Automated screening in environmental arrays allows analysis of quantitative phenotypic profiles in Saccharomyces cerevisiae. Yeast 20, 53–67 (2003).

  12. 12.

    , , , & General calibration of microbial growth in microplate readers. Sci. Rep. (in the press).

  13. 13.

    et al. Compensation for differences in gene copy number among yeast ribosomal proteins is encoded within their promoters. Genome Res. 21, 2114–2128 (2011).

  14. 14.

    et al. Quantifying the determinants of evolutionary dynamics leading to drug resistance. PLoS Biol. 13, e1002299 (2015).

  15. 15.

    et al. Mechanistic and environmental control of the prevalence and lifetime of amyloid oligomers. Nat. Commun. 4, 1891 (2013).

  16. 16.

    et al. Distinct roles for antiparallel microtubule pairing and overlap during early spindle assembly. Mol. Biol. Cell 24, 3238–3250 (2013).

  17. 17.

    , & Kinetics of spindle pole body separation in budding yeast. Proc. Natl Acad. Sci. USA 92, 9707–9711 (1995).

  18. 18.

    , , & BGFit: management and automated fitting of biological growth curves. BMC Bioinformatics 14, 283 (2013).

  19. 19.

    IPMP 2013—a comprehensive data analysis tool for predictive microbiology. Int. J. Food Microbiol. 171, 100–107 (2014).

  20. 20.

    et al. Modeling microbial growth curves with GCAT. Bioenerg. Res. 8, 1022–1030 (2015).

  21. 21.

    , , , & Derivative observations in Gaussian process models of dynamic systems. Adv. Neural Inf. Process Syst. 15, 1033–1040 (2003).

  22. 22.

    Python for scientific computing. Comput. Sci. Eng. 9, 10 (2007).

  23. 23.

    Matplotlib: a 2D graphics environment. Comput. Sci. Eng. 9, 90 (2007).

  24. 24.

    , , & Unmixing of fluorescence spectra to resolve quantitative time-series measurements of gene expression in plate readers. BMC Biotechnol. 14, 11 (2014).

Download references


We thank Guido Sanguinetti and Nacho Molina for advice on Gaussian processes, Ryan Morris and Cait MacPhee for providing the data in Fig. 5a, Alejandro Granados for technical assistance, and our funders: a BBSRC iCASE award (K.S.), the Wellcome Trust and Conacyt (L.F.M.-G.) and the CIHR (A.L. and J.V.). P.S.S., I.B.N.C. and T.P. gratefully acknowledge the support of the UK Research Councils Synthetic Biology for Growth programme and are members of a BBSRC/EPSRC/MRC funded Synthetic Biology Research Centre.

Author information


  1. SynthSys—Synthetic and Systems Biology, School of Biological Sciences, University of Edinburgh, Mayfield Road, Edinburgh EH9 3BF, UK

    • Peter S. Swain
    • , Keiran Stevenson
    • , Luis F. Montano-Gutierrez
    • , Ivan B.N. Clark
    •  & Teuta Pilizota
  2. Department of Biology, McGill University, Montreal, Quebec, Canada H3G 0B1

    • Allen Leary
    •  & Jackie Vogel
  3. Integrated Quantitative Biology Initiative, McGill University, Montreal, Quebec, Canada H3G 0B1

    • Jackie Vogel


  1. Search for Peter S. Swain in:

  2. Search for Keiran Stevenson in:

  3. Search for Allen Leary in:

  4. Search for Luis F. Montano-Gutierrez in:

  5. Search for Ivan B.N. Clark in:

  6. Search for Jackie Vogel in:

  7. Search for Teuta Pilizota in:


P.S.S. developed the algorithm, and with K.S. and A.L. developed the computer code. K.S., A.L., L.F.M.-G., I.B.N.C. and J.V. generated the data. T.P. and P.S.S. wrote the manuscript.

Competing interests

The authors declare no competing financial interests.

Corresponding author

Correspondence to Peter S. Swain.


By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.

Creative Commons BYThis work is licensed under a Creative Commons Attribution 4.0 International License. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in the credit line; if the material is not included under the Creative Commons license, users will need to obtain permission from the license holder to reproduce the material. To view a copy of this license, visit