ArXiv TLDR

EpiCastBench: Datasets and Benchmarks for Multivariate Epidemic Forecasting

🐦 Tweet
2605.11598

Madhurima Panja, Danny D'Agostino, Huitao Li, Tanujit Chakraborty, Nan Liu

cs.LGcs.AIcs.DBq-bio.QM

TLDR

EpiCastBench introduces 40 diverse multivariate epidemic datasets and a standardized benchmark for evaluating forecasting models.

Key contributions

  • Introduces EpiCastBench with 40 diverse multivariate epidemic datasets.
  • Establishes standardized evaluation settings for fair model comparison.
  • Benchmarks 15 forecasting models, including deep learning and foundation models.
  • Provides all datasets and code publicly for reproducibility.

Why it matters

This paper fills a critical gap by providing diverse multivariate epidemic datasets and a standardized benchmark. It enables robust development and fair comparison of advanced forecasting models, which is crucial for improving data-driven public health decision-making.

Original Abstract

The increasing adoption of data-driven decision-making in public health has established epidemic forecasting as a critical area of research. Recent advances in multivariate forecasting models better capture complex temporal dependencies than conventional univariate approaches, which model individual series independently. Despite this potential, the development of robust epidemic forecasting methods is constrained by the lack of high-quality benchmarks comprising diverse multivariate datasets across infectious diseases and geographical regions. To address this gap, we present EpiCastBench, a large-scale benchmarking framework featuring 40 curated (correlated) multivariate epidemic datasets. These publicly available datasets span a wide range of infectious diseases and exhibit diverse characteristics in terms of temporal granularity, series length, and sparsity. We analyze these datasets to identify their global features and structural patterns. To ensure reproducibility and fair comparison, we establish standardized evaluation settings, including a unified forecasting horizon, consistent preprocessing pipelines, diverse performance metrics, and statistical significance testing. By leveraging this framework, we conduct a comprehensive evaluation of 15 multivariate forecasting models spanning statistical baselines to state-of-the-art deep learning and foundation models. All datasets and code are publicly available on Kaggle (https://www.kaggle.com/datasets/aimltsf/epicastbench) and GitHub (https://github.com/aimltsf/EpiCastBench).

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.