Temporal Taskification in Streaming Continual Learning: A Source of Evaluation Instability

2026-04-23Machine Learning

Machine Learning
AI summary

The authors study how dividing a continuous data stream into separate time-based tasks (called taskification) affects how well continual learning models perform. They show that changing where these task boundaries are placed can lead to very different evaluation results, even when using the same data and models. Their experiments with network traffic forecasting reveal that shorter tasks create more variability and instability in learning outcomes. This suggests that the way data streams are split into tasks is an important factor to consider when testing continual learning methods.

Streaming Continual LearningTaskificationPlasticityStabilityExperience ReplayElastic Weight ConsolidationLearning without ForgettingNetwork Traffic ForecastingBenchmark EvaluationBackward Transfer
Authors
Nicolae Filat, Ahmed Hussain, Konstantinos Kalogiannis, Elena Burceanu
Abstract
Streaming Continual Learning (CL) typically converts a continuous stream into a sequence of discrete tasks through temporal partitioning. We argue that this temporal taskification step is not a neutral preprocessing choice, but a structural component of evaluation: different valid splits of the same stream can induce different CL regimes and therefore different benchmark conclusions. To study this effect, we introduce a taskification-level framework based on plasticity and stability profiles, a profile distance between taskifications, and Boundary-Profile Sensitivity (BPS), which diagnoses how strongly small boundary perturbations alter the induced regime before any CL model is trained. We evaluate continual finetuning, Experience Replay, Elastic Weight Consolidation, and Learning without Forgetting on network traffic forecasting with CESNET-Timeseries24, keeping the stream, model, and training budget fixed while varying only the temporal taskification. Across 9-, 30-, and 44-day splits, we observe substantial changes in forecasting error, forgetting, and backward transfer, showing that taskification alone can materially affect CL evaluation. We further find that shorter taskifications induce noisier distribution-level patterns, larger structural distances, and higher BPS, indicating greater sensitivity to boundary perturbations. These results show that benchmark conclusions in streaming CL depend not only on the learner and the data stream, but also on how that stream is taskified, motivating temporal taskification as a first-class evaluation variable.