Abstract
Accurate, highresolution tracking of influenza epidemics at the regional level helps public health agencies make informed and proactive decisions, especially in the face of outbreaks. Internet users’ online searches offer great potential for the regional tracking of influenza. However, due to the complex data structure and reduced quality of Internet data at the regional level, few established methods provide satisfactory performance. In this article, we propose a novel method named ARGO2 (2step Augmented Regression with GOogle data) that efficiently combines publicly available Google search data at different resolutions (national and regional) with traditional influenza surveillance data from the Centers for Disease Control and Prevention (CDC) for accurate, realtime regional tracking of influenza. ARGO2 gives very competitive performance across all US regions compared with available Internetdatabased regional influenza tracking methods, and it has achieved 30% error reduction over the best alternative method that we numerically tested for the period of March 2009 to March 2018. ARGO2 is reliable and robust, with the flexibility to incorporate additional information from other sources and resolutions, making it a powerful tool for regional influenza tracking, and potentially for tracking other social, economic, or public health events at the regional or local level.
Introduction
Internet users’ online records contain the footprints of the activities of millions of individuals in nearly every aspect of life, and offer the potential for realtime tracking of public health and social events^{1,2,3}, including influenza epidemics^{4,5}, at the regional level^{6,7,8}. Accurate, uptodate regional tracking of influenza epidemics, which cause up to 500,000 deaths a year worldwide^{9}, helps personnel including clinicians, epidemiologists, and public health officials, as well as relevant agencies to make informed and proactive decisions^{10,11,12}, especially in face of outbreaks^{13}. However, owing to the complexity of the data structure and low Internet data quality at the regional level, few existing methods provide satisfactory performance in estimating the regional flu activities^{6,7,10,14,15,16}, which is of particular concern as most public health decisions, interventions, and resource allocations are conducted at the regional or local level. Here we present a novel method ARGO2 (standing for 2step Augmented Regression with GOogle data) that gives accurate, realtime influenza tracking at the US Health and Human Services (HHS) regional level (see Fig. 1 for an illustration of the ten US HHS regions; e.g. Region 1 contains six northeastern states of US: CT, MA, ME, NH, RI, VT).
Traditionally, the situational awareness of influenza (flu) activities relies on the Influenzalike Illness (ILI) reports from nationwide surveillance networks of healthcare centers. CDC’s Influenzalike Illness Surveillance Network (ILINet), in particular, aggregates the percentage of outpatients with ILI, and reports this ILI percentage (%ILI) at both the US national and the Health and Human Services (HHS) regional levels. Owing to the time needed for administrative aggregation and processing, CDC’s ILI report often lags behind realtime for 1 to 2 weeks (with less accuracy for more recent weeks), which is far from optimal for decisionmaking. To alleviate the lag of CDC’s ILI report, a new approach, digital disease detection, has been developed to provide decision makers with more systematic, objective, and timely information. Digital flu detection utilizes statistical or mechanistic models to estimate and forecast current and future %ILI at national and/or regional levels based on information from Internetderived data, such as Google search data, as well as traditional surveillance data, such as CDC’s ILI reports. Examples of digital flu detection methods include Google Flu Trends (GFT)^{5}, a susceptibleinfectiousrecoveredsusceptible model with ensemble adjustment Kalman filter (SIRSEAKF)^{6,17,18,19,20}, an empirical Bayes framework with GFT^{14}, an epidemiology model with GFT^{8}, a wisdomofcrowds humanbased forecast^{10}, an ensemble penalized regression^{21}, an artificialtreeenhanced neural network^{22}, and a Gaussian process nonlinear estimator^{23,24}, among others. Yang et al.^{25,26} proposed a method named ARGO (AutoRegression with GOogle search data) that works well at tracking %ILI at the US national level.
Despite the progress in digital flu detection over the last few years^{27,28}, the regional %ILI estimation results, compared to the national ones, are considerably less accurate and still far from satisfactory, as documented in Biggerstaff et al.^{11}. Furthermore, most public health decisions, interventions, and resource allocations (e.g., vaccine campaigns and outbreak responses) are carried out at the regional or local levels rather than at the national level. Many digital flu estimation methods^{14,15,18,22} apply the same method for both the national level estimation and the regional level estimation, even though the latter exhibits a strong spatial structure. Attempts have been made to consider the dependent geographical structure^{12,28,29}; Zou et al.^{29}, in particular, proposed a multitask nonlinear regression method for regional estimation, which showed some good results. Nevertheless, most methods are still insufficient for accurate estimation of the %ILI at the US HHS regional level. In fact, many of the methods underperform a “naive” persistence forecast method. The naive method simply uses CDC’s reported regional %ILI of the previous week as the %ILI estimate for the current week without any modeling effort. See Supplementary Table S13 for further details on performance of methods in the 2015–2016 CDC’s Epidemic Prediction Initiative (FluSight challenge, predict.phiresearchlab.org, predict.cdc.gov).
To effectively address these difficulties, we introduce a novel method ARGO2 that gives accurate and robust realtime %ILI estimates at the regional level. ARGO2 introduces a statistically principled framework that combines multiresolution, multisource information with the regional dependence structure. It operates in two steps. In step one, ARGO2 extracts online search query information (using publicly available data from Google Trends) at two resolutions: at each individual region and at the national level. This information is then used to generate preliminary estimates for each region and for the nation as a whole. In step two, ARGO2 utilizes three components to boost performance: (1) the dependence structure between the different regions, (2) the temporal trend, and (3) the national estimate as the baseline. The twostep procedure of ARGO2 has the following features: (i) It automatically selects the most relevant search terms and filters out highsparsity terms, which overcomes the lowerquality issue found in Google’s regional search data. It incorporates (ii) the lowerresolution, national %ILI estimate as the baseline, (iii) the shortterm momentum of flu activity, and (iv) crossregional influence (correlations) to boost estimation accuracy on highresolution, regional estimation. (v) It adopts a twoyear sliding window for model training, which intends to capture the evolution in people’s search patterns, Google’s search engine, epidemic activity, and other patterns that change over time^{30}.
Results
We apply ARGO2 to produce retrospective estimates of flu activity in the ten US HHS regions for the time period of March 29, 2009 to March 17, 2018. Our estimation target is CDC’s weekly weighted %ILI. In a given week, the most recent CDC’s ILI report typically reflects the %ILI of the previous week. At every week, to estimate the %ILI of the current week, we use the information that would have been available at that time, including the historical CDC’s ILI reports (which give the %ILI up to the previous week) and the flurelated search query volumes from Google Trends (which are available in realtime). We compare our estimates with the actual %ILI subsequently revealed by CDC weeks later and evaluate the estimation accuracy using multiple metrics, including mean squared error (MSE), mean absolute error (MAE), mean absolute percentage error (MAPE), and correlation with the true %ILI. We also compare the performance of our method with several benchmark methods (detailed in the Methods section), including: (i) GFT (last estimate available: the week ending on August 15, 2015), (ii) estimates by the lag1 vector autoregressive model (VAR model), (iii) estimates by the lag1 vector autoregressive model with regional GFT estimates as an exogenous variable, denoted as GFT + VAR, and (iv) naive estimates, which for each US HHS region simply use CDC’s reported %ILI of the previous week as the estimate for the current week. Since ARGO2 uses a twoyear sliding window in its model training, for fair comparison, models (ii) and (iii) are also trained with the twoyear sliding window. It is worth emphasizing that the methodology of ARGO2 was frozen on December 26, 2016, so 2017–2018 is a strict validation period for ARGO2.
We summarize the overall results in Fig. 1 and Table 1, averaging over ten US HHS regions. For MSE and MAE, ARGO2 uniformly outperforms all other methods for all flu seasons (2009–2018) considered in this study. ARGO2 also gives competitive performance in MAPEs and correlations. Specifically regarding MSE, ARGO2 statistically significantly outperforms all the alternatives at the 0.05 significance level (detailed in Supplementary Information and Supplementary Table S14): it gives over 35% error reduction from the naive method in MSE (ARGO2’s MSE = 0.149, 64.4% relative to naive, see also Supplementary Table S16), and 30% reduction from the best alternative method in comparison, i.e., VAR, (VAR’s MSE = 0.212) in the whole period of 2009–2018. Regarding MAE, ARGO2 reaches a 16% error reduction compared to the naive method (ARGO2’s MAE = 0.224, 83.6% relative to the naive method) in the whole period, and achieves a 12% improvement from the best alternative method in comparison (VAR’s MAE = 0.256). ARGO2 has the smallest MAPE in all the evaluation periods except the 2010–2011 flu season, and it holds the lead in the correlation metric in all evaluation periods except the 2009 H1N1 outbreak. Close examination of the H1N1 period shows that GFT significantly overshoots the true peak, resulting in spurious high correlation but poor error metrics, whereas ARGO2 performs consistently both in correlation and in the error metrics (Supplementary Figs S4–S13).
Compared to the benchmark methods, ARGO2’s advantages are apparent. Autoregressionbased methods, VAR and GFT + VAR, hardly outperform the naive method. GFT performs rather poorly. It is worth noting that systematically outperforming the naive method for regional %ILI estimation is not easy, as shown in Supplementary Table S13, where most methods did worse than the naive method in 2015–2016 CDC’s Epidemic Prediction Initiative. When we apply ARGO2 to the data for the 2015–2016 CDC’s Epidemic Prediction Initiative, we see that ARGO2 is the only method that uniformly outperforms the naive method across all ten US HHS regions (see Supplementary Information for more details).
ARGO2 also shows the most consistent error reduction across all periods, notably during the H1N1 outbreak when regular flu seasonality did not hold (see Supplementary Table S16). Such consistency indicates the reliability of ARGO2 for accurate flu tracking in response to potential variations of flu epidemics. Particularly, for the flu outbreak during the 2017–2018 season, which was after the methodology of ARGO2 was frozen in December 2016, ARGO2 still maintains superior performance (MSE = 0.278, 41.5% relative to the naive method). This validates the robustness and flexibility of ARGO2, showcasing its capability to generate accurate estimates for irregular flu epidemics.
In addition to the weekly estimates, ARGO2 also gives confidence intervals. Table 2 and Supplementary Figs S14–S23 show the coverage of the confidence intervals for the ten US HHS regions. The nominal 95% confidence interval has an actual 94.96% coverage on average, suggesting that our confidence intervals closely measure the accuracy of our weekly estimates.
Detailed results of each of the ten US HHS regions (Supplementary Tables S1–S10 and Supplementary Figs S4–S13) show that ARGO2 gives rather robust estimation across various geographical regions (i.e., consistent error reduction across all time periods and regions). Across all regions, ARGO2 reaches at least 20% error reduction in MSE, and close to 10% in MAE compared with the naive method for the whole period, and it maintains the lead over all other benchmark methods. Breaking down different periods, ARGO2 holds the lead in most of the cases (about 75%). We also observe that Region 5 (containing US states IL, IN, MI, MN, OH, WI) gives the highest error reduction in the whole period (relative MSE = 0.487, relative MAE = 0.722 to naive), which may be attributed to the high correlation of Region 5 with neighboring regions (as shown in Fig. 1). This finding is consistent with^{28} and^{29}, where the authors concluded that a network model is best in structurally central regions that have strong connections with neighboring regions.
Discussion
ARGO2 efficiently combines multisource, multiresolution information and provides accurate, reliable, realtime flu tracking at the regional level. ARGO2 aggregates publicly available search data from Google at different resolutions (both national and regional) and traditional influenza surveillance data from the CDC. ARGO2 also incorporates crossregional and temporal correlation of influenza activities for accurate estimation. Across all the regions, ARGO2 outperforms most previously available Internetsearchbased regional influenza tracking methods. Such highprecision, regionally differential surveillance information, as provided by ARGO2, enables public health officials to make timely decisions regarding the changes in flu epidemics and to optimize the allocations of medical and personnel resources across the nation. Furthermore, ARGO2 also constructs confidence intervals as a measure of reliability for its weekly %ILI estimates. The reliable estimation by ARGO2 gives public health agencies and health care providers a head start in response planning for potential flu emergencies or outbreaks in the future.
The challenges of digital flu detection at the regional level include three important components: (1) Regional Internetderived data are often of inferior quality to the national counterparts, typically containing weaker signals. For example, Google Trends search frequency data are much sparser (i.e., have more zero counts) at the regional level than at the national level (as illustrated in Supplementary Fig. S1). (2) The correlation structure of regional %ILI is complex and timeheterogeneous, which requires careful modeling efforts; see Fig. 1, which illustrates the strong spatial correlation across US HHS regions. (3) CDC’s regional ILI report is based on surveys with much smaller sample sizes than the national level, making CDC’s regional %ILI noisier than its national counterpart. Here we discuss in detail how ARGO2 address these challenges to achieve improved estimation accuracy.
In contrast to the lowresolution national level data, we note that the quality of the publicly available highresolution Google Trends regional data is not as satisfactory. The high sparsity of Google Trends data (i.e., much more zero entries) for the majority of the search terms observed at the regional level (see Supplementary Fig. S1) severely impacts the contribution from Google search data to the %ILI estimation. We address this issue in our method by adopting an L_{1} penalty in step one to filter out lowquality sparse terms that correlate poorly with CDC’s historical %ILI.
In addition to the incorporation of Google search data, another key contributor to the improved accuracy of our estimation is the aggregation of multisource, multiresolution information. We include (i) the national baseline %ILI estimates (obtained from the ARGO method^{25)} based on the Google search data at the national level, (ii) Google search information at the regional level, and (iii) the regionspecific %ILI timeseries pattern and variation by considering the (timeseries) increments of %ILI at the regional level. The national level %ILI estimates by ARGO capture not only the general average trends of flu activities, but also the seasonality.
We model the crossregional and crossresolution dependence through structuring the covariance matrix of the (timeseries) %ILI increments, which effectively enhances the accuracy and efficiency in estimating the large covariance matrix with limited data in the twoyear sliding window. The crosscorrelation on the timeseries increments captures the geographic spread of the flu and aggregates regional connectivity factors such as public transportation, geographic proximity, climatic patterns, and vaccination coverage. In fact, the crossregional covariance matrices identified by ARGO2 (Supplementary Fig. S2) indicate stronger connectivity among central regions and northeastern regions, which agrees with the “regional clusters” found in^{29}.
One of the limitations for our analysis is using CDC’s %ILI as our prediction target. We note that the reported %ILI is only a proxy for the actual flu incidence in the population. In fact, due to the reduced sample sizes, CDC’s reported %ILI at the regional level is even more susceptible to sampling bias, for instance, from enrolled health care providers in CDC’s ILINet or patients’ willingness to visit the providers. Nevertheless, despite the limitation of CDC’s %ILI itself, accurate estimation of CDC’s %ILI at the regional level can still help public health officials allocate resources in preparation for potential surges of patient visits to healthcare facilities^{13,31,32}.
Another limitation of ARGO2 is for tracking rare events at the regional level. At the national level, the Google search data have low sparsity (i.e., few zero entries) and have been successfully utilized to track rare diseases such as Dengue fever in tropical countries^{33,34}. However, the high sparsity (i.e., much more zero entries) in Google search data at the regional level (see Supplementary Fig. S1) could limit the usage of ARGO2 for regional rare disease tracking. A possible solution is to further extend ARGO2 by embracing alternative data sources such as electronic health records or Twitter text data.
The ARGO2 procedure has general applicability. The separation of the two steps offers the flexibility to substitute the first step (currently obtained from using Google Trends data) with the output from other models or data sources, while preserving the second step’s capability of multiresolution spatialtemporal boosting. A wide variety of digital disease detection methods, including a Wikipediabased mechanistic epidemic model^{35}, a Google Trendsbased Gaussian Process estimator^{23}, a Twitterbased linear estimator^{36}, a Twitterbased wordembedded nonlinear estimator^{24}, or the ensemble of these estimators^{37}, can be fitted into the crossregional boosting step of ARGO2. The first step of ARGO2 can also incorporate other potentially relevant factors/predictors (e.g., weather) into the prediction model. In addition, such incorporation of multiple sources of information can potentially further extend ARGO2 to track other rare events (e.g. epidemics of rare diseases). The ARGO2 framework could be readily adapted to various spatial and temporal scales for tracking or forecasting other diseases and public health/social events that leave footprints on Internet users’ online records.
Methods
Google data
Google Trends is publicly available at trends.google.com. For a search query term specified by a user, Google Trends gives an integer valued (weekly) time series ranging from 0 to 100, which corresponds to the search intensity of the query term, where 100 represents the highest historical search volume. The publiclyavailable integervalued time series is based on sampling Google’s raw search records. For benchmarking, we also downloaded the discontinued Google Flu Trends (GFT) data (www.google.org/flutrends/about/data/flu/us/data.txt). GFT has regional level prediction for %ILI from September 28, 2003 to August 15, 2015. GFT was discontinued in the week ending August 15, 2015.
The search query terms that we use are identified from Google Correlate (www.google.com/trends/correlate), which gives the most highly correlated search terms with a time series supplied by a user (detailed in Supplementary Information). Supplementary Tables S11 and S12 in Supplementary Information list these search terms.
Aggregation of statelevel Google search data
Google Trends only provides statelevel, rather than regional, search frequency data. We use a simple approach to estimate the regional search frequency based on aggregation of state level data. Specifically, to estimate each search query term’s regional frequency, we use a weighted average of the integervalued time series returned by Google Trends for each of the states in that region, where the weights are proportional to the state population according to the 2010 US census.
CDC’s ILINet data
CDC’s ILINet invites more than 2,800 enrolled outpatient health care providers around the US to report the number of patients with influenzalike illness (defined as fever and cough and/or a sore throat without a known cause other than influenza) together with the total number of all patients seen (www.cdc.gov/flu/weekly/overview.htm). The reports from health care providers are later aggregated at the US HHS regional level as well as at the national level. Each Friday, CDC releases a weekly ILI report for the previous week, where the %ILI are reported at both the US national level and at the HHS region level. The initial publication is subject to later revision as CDC receives updated data or late submissions from the health care providers. Consequently, CDC’s ILI report always lags behind realtime for 1 to 2 weeks and is less accurate for more recent weeks. CDC’s weekly %ILI are publicly available at gis.cdc.gov/grasp/fluview/fluportaldashboard.html.
The model of ARGO2
Step One: extracting Internet search information
At this step, we obtain a preliminary estimate for the regional level %ILI based completely on regionwise Google search data. This preliminary estimate will be the input for the second step. For a given region m, m = 1, …, 10, let x_{i,t,m} be the logarithm of the aggregated regionlevel Google search frequency of search term i at week t; let y_{t,m} be the logittransformation of CDC’s (weighted) %ILI at time t for region m. At week T, owing to the time delay of CDC’s ILI reports, y_{t,m} would be available only up to T−1, whereas the variables x_{i,t,m} are available up to time T.
For each region m, to estimate y_{T,m}, we use a L_{1} regularized linear estimator in the first step:
$${\hat{y}}_{T,m}={\hat{\beta }}_{0,m}+{{\bf{X}}}_{T,m}^{T}{\hat{{\boldsymbol{\beta }}}}_{m},$$where the vector X_{T,m} = (x_{i,T,m}), and \(({\hat{\beta }}_{0,m},{\hat{{\boldsymbol{\beta }}}}_{m})\) is obtained via
$$argmi{n}_{{\beta }_{0,m},{{\boldsymbol{\beta }}}_{m}}\sum _{t=TN}^{T1}{({y}_{t,m}{\beta }_{0,m}{{\bf{X}}}_{t,m}^{T}{{\boldsymbol{\beta }}}_{m})}^{2}+\lambda {\Vert {{\boldsymbol{\beta }}}_{m}\Vert }_{1}.$$N in the above equation is the training window. We set N = 104, i.e., a twoyear window. The use of this moving window, as recommended in previous studies^{25,30}, is to account for changes in people’s search patterns as well as changes in the search engine itself. The L_{1} penalty^{38} adds sparsity in estimated coefficients \({\hat{{\boldsymbol{\beta }}}}_{m}\) for all input search query terms. Due to the penalty, noisy terms with zero coefficients are eliminated from the model, and the most relevant search terms with nonzero coefficients are automatically selected. We set λ through crossvalidation. In addition, we also obtain an accurate estimate \({\hat{y}}_{T}^{nat}\) for the national %ILI by using the ARGO method^{25}, which uses national level Google search data.
Step Two: crossregional boosting
The second step aggregates multiresolution and time series information to boost the regionallevel estimation. It derives the best linear predictor based on the time series increments of %ILI, the preliminary regional %ILI estimates, and the accurate national %ILI estimates, while also accounting for crosssource, crossregion and temporal correlations.
Let p_{t} = (p_{t,1}, …, p_{t,10})^{T} denote CDC’s %ILI of the ten US HHS regions; they are related to y_{t,m} through the inverse logit transformation p_{t,m} = exp(y_{t,m})/(1 + exp(y_{t,m})). Our preliminary estimate for p_{t} from the first step is \({\hat{{\boldsymbol{p}}}}_{t}^{GT}={({\hat{p}}_{t,1},\ldots ,{\hat{p}}_{t,10})}^{T}\), where \({\hat{p}}_{t,m}=\exp ({\hat{y}}_{t,m})/(1+\exp ({\hat{y}}_{t,m}))\). Our estimate of the national %ILI from the first step is \({\hat{p}}_{t}^{nat}=\exp ({\hat{y}}_{t}^{nat})/(1+\exp ({\hat{y}}_{t}^{nat}))\). Let the boldfaced \({\hat{{\boldsymbol{p}}}}_{t}^{nat}\) denote the length10 vector \({\hat{{\boldsymbol{p}}}}_{t}^{nat}={({\hat{p}}_{t}^{nat},\ldots ,{\hat{p}}_{t}^{nat})}^{T}\).
Estimating p_{t} is equivalent to estimating the time series increment Δp_{t} = p_{t} − p_{t−1}. We denote Z_{t} = Δp_{t} for notational simplicity. For the estimation of Z_{t}, we want to incorporate the crossregional, crosssource correlations. Three predictors for Z_{t} after the first step are thereby included: (i) Z_{t−1} = Δp_{t−1}, (ii) \({\hat{{\boldsymbol{p}}}}_{t}^{GT}{{\boldsymbol{p}}}_{t1}\), and (iii) \({\hat{{\boldsymbol{p}}}}_{t}^{nat}{{\boldsymbol{p}}}_{t1}\); they represent time series information, information from the regional level Google search, and the information from the national level Google search, respectively. Let W_{t} denote the collection of these three predictors: \({{\bf{W}}}_{t}={({{\bf{Z}}}_{t1}^{T},{({\hat{{\boldsymbol{p}}}}_{t}^{GT}{{\boldsymbol{p}}}_{t1})}^{T},{({\hat{{\boldsymbol{p}}}}_{t}^{nat}{{\boldsymbol{p}}}_{t1})}^{T})}^{T}\). To combine these predictors, we use the best linear predictor formed by them:
where μ_{Z} and μ_{W} are the mean vectors of Z and W respectively, and Σ_{ZZ}, Σ_{ZW} and Σ_{WW} are the covariance matrices of and between Z and W. The best linear predictor gives the optimal way to linearly combine the three predictors to form a new one. It also provides the variance estimate as
Consistent with the first step, we adopt a sliding twoyear training window to estimate μ_{Z}, μ_{W}, Σ_{ZZ}, Σ_{ZW} and Σ_{WW} in Equation (1). For μ_{Z} and μ_{W}, we use the empirical means of the corresponding variables as the estimators. However, for the covariance matrices, due to their large sizes – 30 × 30 for Σ_{WW} and 10 × 30 for Σ_{ZW} – and the small number of observations – 104 weekly training data points in the twoyear window, we need to structure the covariance matrices for reliable estimation. We assume the following structure:

1.
The covariance between the time series increments satisfies cov(Z_{t}) = cov(Z_{t−1}) = Σ_{ZZ} and cov(Z_{t}, Z_{t−1}) = ρΣ_{ZZ}, where 0 < ρ < 1: This essentially assumes that the time series increments are stationary and have a stable autocorrelation across time and region. We could more generally assume cov(Z_{t}, Z_{t−1}) is Σ_{ZZ} multiplied by a diagonal matrix. However, our numerical investigation suggested that the additional flexibility would not help in the estimation performance, mainly due to the substantial extra variance incurred by these additional parameters. Therefore, we keep the simple assumption of cov(Z_{t}, Z_{t−1}) = ρΣ_{ZZ}.

2.
(i) Independence between the time series increment and the estimation error of the firststep regional estimates, i.e., \({{\bf{Z}}}_{t}\perp ({\hat{{\boldsymbol{p}}}}_{t}^{GT}{{\boldsymbol{p}}}_{t})\), and (ii) crossregional independence between the estimation error of the firststep regional estimates: (i) and (ii) are based on the fact that the firststep preliminary regional estimates are carried out separately for each region based solely on regional Google search information of that region alone. Mathematically, (i) and (ii) imply that \(D=cov({\hat{{\boldsymbol{p}}}}_{t}^{GT}{{\boldsymbol{p}}}_{t})\) is diagonal, \(cov({{\bf{Z}}}_{t},{\hat{{\boldsymbol{p}}}}_{t}^{GT}{{\boldsymbol{p}}}_{t1})={\Sigma }_{ZZ}\), and \(cov({\hat{{\boldsymbol{p}}}}_{t}^{GT}{{\boldsymbol{p}}}_{t1})={\Sigma }_{ZZ}+D\).

3.
Independence between the time series increment and the regional deviates from national baseline, i.e., \({{\bf{Z}}}_{t}\perp ({\hat{{\boldsymbol{p}}}}_{t}^{nat}{{\boldsymbol{p}}}_{t})\): Thereby, we have \(cov({{\bf{Z}}}_{t},{\hat{{\boldsymbol{p}}}}_{t}^{nat}{{\boldsymbol{p}}}_{t1})={\Sigma }_{ZZ}\) and \(cov({\hat{{\boldsymbol{p}}}}_{t}^{nat}{{\boldsymbol{p}}}_{t1})={\Sigma }_{ZZ}+{\Sigma }^{nat}\), where \({{\rm{\Sigma }}}^{nat}=cov({\hat{{\boldsymbol{p}}}}_{t}^{nat}{{\boldsymbol{p}}}_{t})\).

4.
Independence between the errors from two separate sources of estimation, i.e., (\({\hat{{\boldsymbol{p}}}}_{t}^{GT}{{\boldsymbol{p}}}_{t})\perp ({\hat{{\boldsymbol{p}}}}_{t}^{nat}{{\boldsymbol{p}}}_{t})\): Mathematically, this implies \(cov({\hat{{\boldsymbol{p}}}}_{t}^{GT}{{\boldsymbol{p}}}_{t1},{\hat{{\boldsymbol{p}}}}_{t}^{nat}{{\boldsymbol{p}}}_{t1})={\Sigma }_{ZZ}\).
The covariance matrices are thereby simplified as:
$$\begin{array}{ccc}{{\rm{\Sigma }}}_{ZW} & = & (\begin{array}{ccc}\rho {{\rm{\Sigma }}}_{ZZ} & {{\rm{\Sigma }}}_{ZZ} & {{\rm{\Sigma }}}_{ZZ}\end{array}),\\ {{\rm{\Sigma }}}_{WW} & = & (\begin{array}{ccc}{{\rm{\Sigma }}}_{ZZ} & \rho {{\rm{\Sigma }}}_{ZZ} & \rho {{\rm{\Sigma }}}_{ZZ}\\ \rho {{\rm{\Sigma }}}_{ZZ} & {{\rm{\Sigma }}}_{ZZ}+D & {{\rm{\Sigma }}}_{ZZ}\\ \rho {{\rm{\Sigma }}}_{ZZ} & {{\rm{\Sigma }}}_{ZZ} & {{\rm{\Sigma }}}_{ZZ}+{{\rm{\Sigma }}}^{nat}\end{array}).\end{array}$$The assumptions made are supported by empirical evidence: (1) close agreement is observed between our estimated structural matrix and the empirical one (Supplementary Fig. S2), and (2) stationary bootstrap indicates statistical acceptance of the null hypothesis of our assumed covariance structure (Supplementary Fig. S3).
To further control the estimation stability, we incorporate a ridgeregression^{39} inspired shrinkage to the linear predictor in Equation (1), replacing the 40 × 40 joint covariance matrix of \({({{\bf{Z}}}_{t}^{T},{{\bf{W}}}_{t}^{T})}^{T}\) by the average of the structured covariance matrix and its empirical diagonal. Effectively, in Equation (1), Σ_{ZW} is replaced by \(\frac{1}{2}{\Sigma }_{ZW}\), and Σ_{WW} is replaced by \((\frac{1}{2}{\Sigma }_{WW}+\frac{1}{2}{D}_{WW})\), where D_{WW} is the diagonal of the empirical covariance of W_{t}:
$${\hat{{\bf{Z}}}}_{t}={{\boldsymbol{\mu }}}_{Z}+\frac{1}{2}{\Sigma }_{ZW}{(\frac{1}{2}{\Sigma }_{WW}+\frac{1}{2}{D}_{WW})}^{1}({{\bf{W}}}_{t}{{\boldsymbol{\mu }}}_{W}).$$Similarly, we update Equation (2) with above substitution as well:
$${\rm{V}}{\rm{a}}{\rm{r}}({{\bf{Z}}}_{t}{{\bf{W}}}_{t})={\Sigma }_{ZZ}\frac{1}{2}{\Sigma }_{ZW}{(\frac{1}{2}{\Sigma }_{WW}+\frac{1}{2}{D}_{WW})}^{1}\frac{1}{2}{\Sigma }_{WZ}.$$Σ_{ZZ}, Σ^{nat}, D, and D_{WW} are estimated by the corresponding sample covariance from the data in the most recent 2year training window; ρ is estimated by minimizing the Frobenius norm (L_{2} distance) between the empirical correlation and structured correlation.
Our final regional %ILI estimate for week T after step two is:
$${\hat{{\boldsymbol{p}}}}_{T}={{\boldsymbol{p}}}_{T1}+{\hat{{\boldsymbol{\mu }}}}_{Z}+{\hat{\Sigma }}_{ZW}{({\hat{\Sigma }}_{WW}+{\hat{D}}_{WW})}^{1}({{\bf{W}}}_{T}{\hat{{\boldsymbol{\mu }}}}_{W}),$$with a corresponding 95% confidence interval as:
$$[{\hat{{\boldsymbol{p}}}}_{T}\pm 1.96\times \sqrt{{\hat{\Sigma }}_{ZZ}\frac{1}{2}{\hat{\Sigma }}_{ZW}{({\hat{\Sigma }}_{WW}+{\hat{D}}_{WW})}^{1}{\hat{\Sigma }}_{WZ}}].$$Accuracy metrics
The MSE, MAE, MAPE and correlation of estimator \(\tilde{p}\) to the target p for US HHS region i are: \({\rm{MSE}}({\tilde{p}}_{i},{p}_{i})=\frac{1}{n}\sum _{t=1}^{n}\,{({\tilde{p}}_{i,t}{p}_{i,t})}^{2}\); \({\rm{MAE}}({\tilde{p}}_{i},{p}_{i})=\frac{1}{n}\sum _{t=1}^{n}\,{\tilde{p}}_{i,t}{p}_{i,t}\); \({\rm{MAPE}}({\tilde{p}}_{i},{p}_{i})=\frac{1}{n}\sum _{t=1}^{n}\,{\tilde{p}}_{i,t}{p}_{i,t}/{p}_{i,t}\). The correlation is the Pearson correlation between p_{i,t} and \({\mathop{p}\limits^{ \sim }}_{i,t}\) across index t. Note that all the metrics are computed on the original %ILI scale.
Benchmark methods
We compare ARGO2 with benchmark methods including VAR, GFT, GFT + VAR, and naive estimation. VAR is the lag1 vector autoregressive model (VAR1 model) on the 10region multivariate logittransformed %ILI, i.e. \({{\boldsymbol{y}}}_{t}={\boldsymbol{c}}+{A}_{1}{{\boldsymbol{y}}}_{t1}+{{\boldsymbol{\epsilon }}}_{t}\). GFT is the GFT estimates produced by Google (discontinued in the week ending on August 15, 2015). GFT + VAR is the lag1 vector autoregressive model with GFT estimate as exogenous variable, i.e., \({{\boldsymbol{y}}}_{t}={\boldsymbol{c}}+{A}_{1}{{\boldsymbol{y}}}_{t1}+B{{\boldsymbol{y}}}_{t}^{GFT}+{{\boldsymbol{\epsilon }}}_{t}\), where \({{\boldsymbol{y}}}_{t}^{GFT}\) is the logittransformed GFT estimates. The naive method uses the previous week’s CDC observation as the estimate for the current week.
Additional information
Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
 1.
Scott, S. L. & Varian, H. R. Predicting the present with Bayesian structural time series. Int. J. Math. Model. Numer. Optim. 5, 1–2 (2014).
 2.
Scott, S. & Varian, H. In Economic Analysis of the Digital Economy (eds Goldfarb, A., Greenstein, S. M. & Tucker, C. E.) 119–135 (University of Chicago Press, 2015).
 3.
Porter, A. T., Holan, S. H., Wikle, C. K. & Cressie, N. Spatial FayHerriot models for small area estimation with functional covariates. Spat. Stat. 10, 27–42 (2014).
 4.
Polgreen, P. M., Chen, Y., Pennock, D. M. & Nelson, F. D. Using Internet Searches for Influenza Surveillance. Clin. Infect. Dis. 47, 1443–1448 (2008).
 5.
Ginsberg, J. et al. Detecting influenza epidemics using search engine query data. Nature 457, 1012–1014 (2009).
 6.
Shaman, J. & Karspeck, A. Forecasting seasonal outbreaks of influenza. Proc. Natl. Acad. Sci. USA 109, 20425–20430 (2012).
 7.
Wu, L. & Brynjolfsson, E. In Economic analysis of the digital economy (eds Avi, G., Greenstein, S. M. & Tucker, C. E.) 89–118 (University of Chicago Press, 2015).
 8.
Dukic, V., Lopes, H. F. & Polson, N. G. Tracking epidemics with Google Flu trends data and a statespace SEIR model. J. Am. Stat. Assoc. 107, 1410–1426 (2012).
 9.
World Health Organization. Influenza fact sheet. Available at, http://www.who.int/mediacentre/factsheets/fs211/en/ (2016).
 10.
Farrow, D. C. et al. A human judgment approach to epidemiological forecasting. PLoS Comput. Biol. 13, e1005248 (2017).
 11.
Biggerstaff, M. et al. Results from the centers for disease control and prevention’s predict the 2013–2014 Influenza Season Challenge. BMC Infect. Dis. 16, 1–10 (2016).
 12.
Yang, W., Olson, D. R. & Shaman, J. Forecasting influenza outbreaks in boroughs and neighborhoods of New York City. PLoS Comput. Biol. 12, e1005201 (2016).
 13.
Lipsitch, M. et al. Improving the evidence base for decision making during a pandemic: the example of 2009 influenza A/H1N1. Biosecur. Bioterror. 9, 89–115 (2011).
 14.
Brooks, L. C., Farrow, D. C., Hyun, S., Tibshirani, R. J. & Rosenfeld, R. Flexible modeling of epidemics with an empirical Bayes framework. PLoS Comput. Biol. 11, e1004382 (2015).
 15.
Chakraborty, P. et al. Forecasting a moving target: Ensemble models for ILI case count predictions. In Proceedings of the 2014 SIAM International Conference on Data Mining 262–270, https://doi.org/10.1137/1.9781611973440.30 (2014)
 16.
Butler, D. When Google got flu wrong. Nature 494, 155–156 (2013).
 17.
Yang, W., Lipsitch, M. & Shaman, J. Inference of seasonal and pandemic influenza transmission dynamics. Proc. Natl. Acad. Sci. 112, 2723–2728 (2015).
 18.
Shaman, J., Karspeck, A., Yang, W., Tamerius, J. & Lipsitch, M. Realtime influenza forecasts during the 2012–2013 season. Nat. Commun. 4, 2837 (2013).
 19.
Yang, W., Karspeck, A. & Shaman, J. Comparison of filtering methods for the modeling and retrospective forecasting of influenza epidemics. PLoS Comput. Biol. 10, e1003583 (2014).
 20.
Shaman, J. & Kandula, S. Improved discrimination of influenza forecast accuracy using consecutive predictions. PLoS Curr. Outbreaks, https://doi.org/10.1371/currents.outbreaks.8a6a3df285af7ca973f (2015).
 21.
Guo, P. et al. Monitoring seasonal influenza epidemics by using internet search data with an ensemble penalized regression model. Sci. Rep. 7, 1–11 (2017).
 22.
Hu, H. et al. Prediction of influenzalike illness based on the improved artificial tree algorithm and artificial neural network. Sci. Rep. 8, 1–8 (2018).
 23.
Lampos, V., Miller, A. C., Crossan, S. & Stefansen, C. Advances in nowcasting influenzalike illness rates using search query logs. Sci. Rep. 5, 12760 (2015).
 24.
Lampos, V., Zou, B. & Cox, I. J. Enhancing feature selection using word embeddings: The case of flu surveillance. In Proceedings of the 26th International Conference on World Wide Web 695–704, https://doi.org/10.1145/3038912.3052622 (2017).
 25.
Yang, S., Santillana, M. & Kou, S. C. Accurate estimation of influenza epidemics using Google search data via ARGO. Proc. Natl. Acad. Sci. 112, 14473–14478 (2015).
 26.
Yang, S. et al. Using electronic health records and Internet search information for accurate influenza forecasting. BMC Infect. Dis. 17, 332 (2017).
 27.
Santillana, M., Zhang, W. D., Althouse, B. & Ayers, J. W. What can digital disease detection learn from (an external revision to) Google Flu Trends? Am. J. Prev. Med. 47, 341–347 (2014).
 28.
Davidson, M. W., Haim, D. A. & Radin, J. M. Using networks to combine “big data” and traditional surveillance to improve influenza predictions. Sci. Rep. 5, 8154 (2015).
 29.
Zou, B., Lampos, V. & Cox, I. MultiTask Learning Improves Disease Models from Web Search. Proc. 2018 World Wide Web Conf. World Wide Web  WWW ’18 87–96 https://doi.org/10.1145/3178876.3186050 (2018).
 30.
Burkom, H. S., Murphy, S. P. & Shmueli, G. Automated time series forecasting for biosurveillance. Stat. Med. 26, 4202–4218 (2007).
 31.
Nsoesie, E. O., Brownstein, J. S., Ramakrishnan, N. & Marathe, M. V. A systematic review of studies on forecasting the dynamics of influenza outbreaks. Influenza Other Respi. Viruses 8, 309–316 (2014).
 32.
Chretien, J. P., George, D., Shaman, J., Chitale, R. A. & McKenzie, F. E. Influenza forecasting in human populations: A scoping review. PLoS One 9, e94130 (2014).
 33.
Chan, E. H., Sahai, V., Conrad, C. & Brownstein, J. S. Using web search query data to monitor dengue epidemics: A new model for neglected tropical disease surveillance. PLoS Negl. Trop. Dis. 5, e1206 (2011).
 34.
Yang, S. et al. Advances in using Internet searches to track dengue. PLoS Comput. Biol. 13, e1005607 (2017).
 35.
Hickmann, K. S. et al. Forecasting the 2013–2014 influenza season using Wikipedia. PLoS Comput. Biol. 11, e1004239 (2015).
 36.
Paul, M. J., Dredze, M. & Broniatowski, D. Twitter improves influenza forecasting. PLoS Curr. Outbreaks, https://doi.org/10.1371/currents.outbreaks.90b9ed0f59bae4ccaa683a3 (2014).
 37.
Santillana, M. et al. Combining search, social media, and traditional data sources to improve influenza surveillance. PLoS Comput. Biol. 11, e1004513 (2015).
 38.
Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B Stat. Methodol. 58, 267–288 (1996).
 39.
Hoerl, A. E. & Kennard, R. W. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 12, 55–67 (1970).
Acknowledgements
This work was supported in part by National Science Foundation [grants number DMS1510446 and DMS1810914]. The R package “argo” that implements the ARGO2 method, together with all datasets analyzed in the current study, is available in the Harvard Dataverse repository, https://doi.org/10.7910/DVN/TZEPN0. The R package would also be made available on the CRAN repository at https://CRAN.Rproject.org/package=argo.
Author information
Author notes
Shaoyang Ning and Shihao Yang contributed equally.
Affiliations
Department of Statistics, Harvard University, 1 Oxford Street, Cambridge, 02138, MA, USA
 Shaoyang Ning
 , Shihao Yang
 & S. C. Kou
Authors
Search for Shaoyang Ning in:
Search for Shihao Yang in:
Search for S. C. Kou in:
Contributions
S.N. and S.Y. contributed equally to this manuscript. S.N., S.Y. and S.C.K. designed the research; S.N., S.Y. and S.C.K. performed the research; S.N. and S.Y. analyzed data; and S.N., S.Y. and S.C.K. wrote the paper. S.N. and S.Y. are PhD candidates at Department of Statistics, Harvard University. S.C.K. is Professor of Statistics at Harvard University.
Competing Interests
The authors declare no competing interests.
Corresponding author
Correspondence to S. C. Kou.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.