What’s Wrong With The Current Time-Series Deep Learning?

Source

Core Claim

This position argues that time-series foundation models should optimize for maintaining and evolving a useful internal representation of system state, not merely for forecasting future observations. Forecasting remains useful, but the deeper target is understanding what regime the system is in, what constraints govern it, what changed, and which futures are plausible.

Key Contributions

  • Names forecasting bias as a field-level problem: time-series work is too often read as finance or observation forecasting rather than system understanding and reasoning.
  • Separates observation forecasting from latent-state prediction. Forecasting predicts future values in the observed target space; state prediction aims to maintain a representation of the system that can support derived observations, decisions, and eventually action-conditioned reasoning.
  • Treats context as part of the modeling interface, not decoration. The model needs the physical meaning of the data and the environment state in which the observations occur.
  • Reframes the enterprise and industrial setting as high-dimensional, real-time, and data-rich but useful-signal-poor. The hard problem is often finding rare meaningful events inside large volumes of repetitive normal-state data.
  • Identifies always-on operation as a state-maintenance problem: a useful model should keep its representation of the system’s internal state current as new observations arrive.

Relation To The Wiki

Use this source as the wiki’s time-series North Star. It does not replace forecasting sources, but it changes how they should be read: a forecasting result is strongest when it also improves state representation, context use, regime understanding, calibration, rare-event sensitivity, or the path toward action-conditioned world modeling.

The follow-up post adds a field-map observation from Alex’s ICLR 2026 analysis: time-series representation learning was visibly concentrated in EEG/ECG/neuro/physiology work, while the broader time-series field still leaned toward forecasting. Use ICLR 2026 Time-Series Classification Meta-Analysis for the local audit trail and caveats.

Limitations

  • This is a position source, not a benchmark or model paper.
  • The ICLR follow-up says the meta-analysis used Codex-assisted batch analysis with non-exhaustive hand verification, so the 32 out of 57 number should be treated as a directional field-map observation rather than a final bibliometric statistic.
  • The source is written from the perspective of foundation models; small task-specific time-series models can have different constraints.

Open Questions

  • Which training objectives most directly teach a model to maintain useful latent state rather than only optimize forecast error?
  • What benchmarks can separate observation forecasting from state understanding, context use, and rare-event sensitivity?
  • How should always-on state maintenance be evaluated in high-dimensional enterprise and industrial time series?