This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

On the accuracy of short-term COVID-19 fatality forecasts

Nino Antulov-Fantulin anino@ethz.ch Computational Social Science, ETH Zurich, 8092, Zurich, Switzerland    Lucas Böttcher lucasb@g.ucla.edu Dept. of Computational Medicine, University of California, Los Angeles, 90095-1766, Los Angeles, United States Computational Social Science, Frankfurt School of Finance and Management, Frankfurt am Main, 60322, Germany
Abstract
Background

Forecasting new cases, hospitalizations, and disease-induced deaths is an important part of infectious disease surveillance and helps guide health officials in implementing effective countermeasures. For disease surveillance in the U.S., the Centers for Disease Control and Prevention (CDC) combine more than 65 individual forecasts of these numbers in an ensemble forecast at national and state levels.

Methods

We collected data on CDC ensemble forecasts of COVID-19 fatalities in the United States, and compare them with easily interpretable “Euler” forecasts serving as a model-free benchmark that is only based on the local rate of change of the incidence curve. The term “Euler method” is motivated by the eponymous numerical integration scheme that calculates the value of a function at a future time step based on the current rate of change.

Results

Our results show that CDC ensemble forecasts are not more accurate than “Euler” forecasts on short-term forecasting horizons of one week. However, CDC ensemble forecasts show a better performance on longer forecasting horizons.

Conclusions

Using the current rate of change in incidences as estimates of future incidence changes is useful for epidemic forecasting on short time horizons. An advantage of the proposed method over other forecasting approaches is that it can be implemented with a very limited amount of work and without relying on additional data (e.g., human mobility and contact patterns) and high-performance computing systems.

COVID-19; Forecasting; Numerical Analysis, Computer-Assisted; Epidemiological Monitoring

Over the course of the COVID-19 pandemic more than 65 international research groups contributed to an ensemble forecast of reported COVID-19 cases, hospitalizations, and fatalities in the U.S. [1]. These forecasts are a central source of information on the further development of the pandemic and used by various governmental and non-governmental entities including the Centers for Disease Control and Prevention (CDC) [2].

Different forecasting methods rely on different underlying models and assumptions. One may roughly divide forecasting models into three different classes: (i) mechanistic models [3, 4], (ii) purely data-driven models [5], and (iii) hybrid models. Most classical epidemic models are mechanistic and aim at describing disease dynamics in terms of interacting individuals in a population. Such models are usually applied to describe the influence of certain factors (e.g., population density, demographics, contact patterns, mobility, etc.) on the dynamics of an epidemic. Data-driven or machine learning models make fewer assumptions about the underlying dynamics and are applicable to a broader range of forecasting problems, but they also come at the cost of less interpretability for policymakers and epidemiologists.

Here, we show that a very basic, model-free forecasting approach provides effective short-term forecasts of COVID-19 fatalities. We refer to this method as “Euler forecast”, owing to its mathematical connection to the Euler method [6, 7] that is used in computational mathematics to calculate the value of a function at a future time step based on the current rate of change.

Refer to caption
Figure 1: Comparison of predicted and reported weekly COVID-19 deaths in the U.S. (a,b) Forecasts of reported weekly COVID-19 deaths in the US for (a) 1-week and (b) 4-week forecasting horizons. Blue and red lines represent CDC ensemble forecasts [1] and regularized Euler forecasts [Eq. (3)] with λ=10\lambda=10, respectively. Reported COVID-19 fatalities (dashed black lines) are based on [8]. (c,d) 4-week moving averages of weekly forecasting errors of Euler–Lagrange and CDC ensemble forecasts. Solid lines indicate 4-week moving averages that are calculated based on the shown data points.

Methods

We collected data on CDC ensemble forecasts between June 2020 and June 2021 [1]. Ensemble forecasts are available for cumulative and weekly incidence numbers and a forecasting horizon between one to four weeks. All forecasts use data from the Johns Hopkins Coronavirus Resource Center [8] as reference. Forecasts are made for epidemiological weeks which run Sunday through Saturday. As an example, if forecasts with one and four-week forecasting horizons are being made on June 7, 2020 the corresponding target dates are June 13, 2020 and July 4, 2020 [9].

We compare CDC ensemble forecasts of COVID-19 fatalities with a simple and easily interpretable forecasting method. To do so, let y(t)y(t) be the incidence of COVID-19 fatalities at time tt. We use y˙(t)\dot{y}(t) to denote the rate of change of y(t)y(t) at time tt. Forecasting the incidence y(t+Δt)y(t+\Delta t) at a target time t+Δtt+\Delta t requires us to find an estimate of this quantity at an earlier time tt. A straightforward way to construct short-term forecasts is to use the current rate of change y˙(t)\dot{y}(t) and determine a forecast at time tk=t0+kΔtt_{k}=t_{0}+k\Delta t according to the Euler method [6, 7]

y(t0+kΔt)=y(t0)last incidence+kΔty˙(t0)incidence correction,y(t_{0}+k\Delta t)=\underbrace{y(t_{0})}_{\text{last incidence}}+\underbrace{k\Delta t\,\dot{y}(t_{0})}_{\text{incidence correction}}, (1)

where Δt\Delta t and k=1,2,k=1,2,\dots represent a time step (e.g., one week) and the number of time steps in the forecasting horizon, respectively. However, observed incidences are subject to observation noise that results from confounding factors including sampling bias, measurement errors, and reporting delays [10].

A possible way to “de-noise” observed data is to use previous weekly incidences instead of daily incidence levels. If observational noise can be reduced by averaging over a period of several days, daily errors are less pronounced on a weekly level. However, the local daily derivative is quite sensitive to noise and our incidence correction term is not helping in making accurate short-term forecasts. Therefore, we can impose some degree of regularity to reduce the level of noise with the following minimization

argmin{wk}k(ykwk)2+λk(wkwk1)2,\operatorname*{arg\,min}_{\{w_{k}\}}\sum_{k}(y_{k}-w_{k})^{2}+\lambda\sum_{k}(w_{k}-w_{k-1})^{2}\,, (2)

where yk=y(t0+kΔt)y_{k}=y(t_{0}+k\Delta t), wk=w(t0+kΔt)w_{k}=w(t_{0}+k\Delta t) is a regularized approximation of yky_{k}, and λ\lambda is a regularization parameter. In the limit λ0\lambda\rightarrow 0, the argument of Eq. (2) is minimized if w(t)w(t) approaches y(t)y(t). In the limit λ\lambda\rightarrow\infty, the argument of Eq. (2) is minimized if w(t)w(t) is constant (i.e., if wkwk1=0w_{k}-w_{k-1}=0). This optimization process has its equivalent Euler–Lagrange formulation for differentiation [11, 12]. Values of λ(0,)\lambda\in(0,\infty) yield functions w(t)w(t) that are smoothened versions of y(t)y(t) with respect to the discrete rate of change wkwk1w_{k}-w_{k-1}. Finally, the regularized Euler short-term forecast111All optimization procedures in this work are applied in a causal manner. That is, at the prediction time TT only historical data y(tT)y(t\leq T) is being used in the minimization (2). is given by

y(t+kΔt)=y(t)+k[w(t)w(tΔt)].y(t+k\Delta t)=y(t)+k\,[w(t)-w(t-\Delta t)]\,. (3)

In the following section, we utilize the regularized Euler method to generate forecasts of reported COVID-19 fatalities.

Our source codes are publicly available at [13].

Results

Figure 1 shows CDC ensemble forecasts (solid blue lines) of the weekly incidences of reported COVID-19 fatalities. The dashed black lines indicate reported COVID-19 fatalities. Between June and early November 2020, the majority of reported cases were close to the ensemble forecast. As COVID-19 deaths surged in November 2020, the forecasts of the ensemble method became less accurate than in previous months.

For a comparison between the CDC ensemble point estimates and those obtained with the regularized Euler method [Eq. (3)], Fig. 1 also shows Euler-method forecasts (solid red lines) of weekly incidences of COVID-19 fatalities in the U.S. We observe that one-week CDC ensemble forecast for the majority of data points are not more accurate than one-week Euler forecasts [Fig. 1(a)], which we use as a local-derivative-based forecasting benchmark. Although Euler and CDC forecasts still exhibit a similar structure for a four-week forecasting horizon [Fig. 1(b)], the Euler method is associated with larger deviations from the reported fatalities than the CDC ensemble method. To quantify differences in forecasting errors between the two methods, we use

δx,y(t)=|x(t)y(t)|\delta_{x,y}(t)=|x(t)-y(t)| (4)

to denote the absolute error between the Euler forecast x(t)x(t) and CDC forecast y(t)y(t) for target time tt.

Figure 1(c,d) show the 4-week moving averages of weekly forecasting errors δ(t)\delta(t) (solid lines) of the Euler (red) and CDC ensemble (blue) methods. As suggested by our above discussion of Fig. 1(a,b), we observe that the error of the Euler method is substantially smaller than that of the ensemble forecast for a one-week forecasting horizon. In about 61% of the forecasting instances shown in Fig. 1(a), the regularized Euler method has a smaller error than the CDC ensemble forecast. The cumulative forecasting errors are 49,925 (Euler) and 52,885 (CDC). Without correction term [i.e., for k=0k=0 in Eq. (3)], the cumulative forecasting error of the Euler method is 52,660, again smaller than that of CDC ensemble forecast. Note that no regularization corresponds to a simple shift of the incidence curve. For a 4-week forecasting horizon [Fig. 1(d)], the cumulative error of the CDC ensemble forecast is 87,717, about 35% smaller than that of the Euler method.

Discussion

Our results suggest that easily interpretable methods like the Euler method, a model-free local-derivative-based forecasting benchmark, provide an effective alternative to more complex epidemic forecasting frameworks on short-term forecasting horizons. Similar conclusions were drawn in a recent study [14] that compared Euler-like forecasts with those generated by Google Flu Trends. Regularized Euler forecasts have smaller errors with respect to CDC ensemble forecasts on one-week forecasting horizons in about 61% of all cases. Simple curve shifts without regularization provide better one-week forecasts in 63% of all cases, yet with a mean absolute error that is about 5% larger than that found for regularized Euler forecasts. For longer forecasting horizons, it is not surprising that CDC forecasts that rely on additional input data, and epidemiological and statistical models become more accurate than Euler-like forecasting benchmarks.

One clear advantage of Euler forecasting methods is that they are less labor and resource intensive than more complex forecasting models, which often rely on the knowledge of expert groups and require specialized computing infrastructure. In their simplest implementation, Euler forecasts use the currently observed incidence rate as an estimate of the incidence rate in the following week. The regularization methods (3) can help further improve such data-driven forecasts.

In agreement with [14], our results emphasize the importance of benchmarking complex forecasting models against simple forecasting baselines to further improve forecasting accuracy. Our study also points towards recent findings on algorithm rejection and aversion [15] that found that “people have diminishing sensitivity to forecasting error” and that “people are less likely to use the best possible algorithm in decision domains that are more unpredictable”. Finally, in highly uncertain and noisy forecasting regimes, simple methods tend to outperform more complex methods because of a more favorable bias-variance tradeoff [16].

References

  • [1] The COVID-19 Forecast Hub. https://covid19forecasthub.org/, 2021. Accessed: 2021-02-17.
  • [2] Evan L Ray, Nutcha Wattanachit, Jarad Niemi, Abdul Hannan Kanji, Katie House, Estee Y Cramer, Johannes Bracher, Andrew Zheng, Teresa K Yamana, Xinyue Xiong, Spencer Woody, Yuanjia Wang, Lily Wang, Robert L Walraven, Vishal Tomar, Katharine Sherratt, Daniel Sheldon, Robert C Reiner, B. Aditya Prakash, Dave Osthus, Michael Lingzhi Li, Elizabeth C Lee, Ugur Koyluoglu, Pinar Keskinocak, Youyang Gu, Quanquan Gu, Glover E. George, Guido España, Sabrina Corsetti, Jagpreet Chhatwal, Sean Cavany, Hannah Biegel, Michal Ben-Nun, Jo Walker, Rachel Slayton, Velma Lopez, Matthew Biggerstaff, Michael A Johansson, and Nicholas G Reich. Ensemble Forecasts of Coronavirus Disease 2019 (COVID-19) in the U.S. medRxiv, 2020.
  • [3] Matt J Keeling and Pejman Rohani. Modeling infectious diseases in humans and animals. Princeton University Press, 2011.
  • [4] Lucas Böttcher and Nino Antulov-Fantulin. Unifying continuous, discrete, and hybrid susceptible-infected-recovered processes on networks. Physical Review Research, 2(3):033121, 2020.
  • [5] Terence C Mills. Applied time series analysis: A practical guide to modeling and forecasting. Academic Press, 2019.
  • [6] Leonhard Euler. Institutiones calculi integralis, volume 4. Academia Imperialis Scientiarum, 1794.
  • [7] Alfio Quarteroni, Riccardo Sacco, and Fausto Saleri. Numerical Mathematics, volume 37. Springer Science & Business Media, 2010.
  • [8] Ensheng Dong, Hongru Du, and Lauren Gardner. An interactive web-based dashboard to track COVID-19 in real time. The Lancet Infectious Diseases, 2020.
  • [9] Data submission instructions. https://github.com/reichlab/covid19-forecast-hub/blob/master/data-processed/README.md, 2021. Accessed: 2021-06-23.
  • [10] Lucas Böttcher, Maria R D’Orsogna, and Tom Chou. Using excess deaths and testing statistics to determine COVID-19 mortalities. European Journal of Epidemiology, 36(5):545–558, 2021.
  • [11] Jane Cullum. Numerical differentiation and regularization. SIAM Journal on Numerical Analysis, 8(2):254–265, 1971.
  • [12] Rick Chartrand. Numerical differentiation of noisy, nonsmooth data. International Scholarly Research Notices, 2011, 2011.
  • [13] Euler CDC forecasting GitHub repository. https://github.com/ninoaf/epidemic_cdc_forecasts, 2021. Accessed: 2021-06-24.
  • [14] Konstantinos V Katsikopoulos, Özgür Şimşek, Marcus Buckmann, and Gerd Gigerenzer. Transparent modeling of influenza incidence: Big data or a single data point from psychological theory? International Journal of Forecasting, 2021.
  • [15] Berkeley J Dietvorst and Soaham Bharti. People reject algorithms in uncertain decision domains because they have diminishing sensitivity to forecasting error. Psychological Science, 31(10):1302–1314, 2020.
  • [16] Jerome Friedman, Trevor Hastie, Robert Tibshirani, et al. The elements of statistical learning, volume 1. Springer Series in Statistics, 2001.