Comparative Analysis of Time Series Transformers on Multivariate Time Series Data

Dinesh Raja1ORCiD and Judith John Edwin2
1. Research Scholar, Department of computer science and engineering, Noorul Islam Centre For Higher Education, Nagercoil, India Research Organization Registry (ROR)
2. Associate Professor, Department of computer science and engineering, Noorul Islam Centre For Higher Education, Nagercoil, India
Correspondence to: Dinesh Raja, dineshmerein@gmail.com

Premier Journal of Science

Additional information

  • Ethical approval: N/a
  • Consent: N/a
  • Funding: No industry funding
  • Conflicts of interest: N/a
  • Author contribution: Dinesh Raja and Judith John Edwin – Conceptualization, Writing – original draft, review and editing
  • Guarantor: Dinesh Raja
  • Provenance and peer-review: Unsolicited and externally peer-reviewed
  • Data availability statement: N/a

Keywords: Multivariate time series forecasting, transformer architectures evaluation, informer–reformer–autoformer comparison, probsparse self-attention efficiency, seasonal-trend decomposition.

Peer Review
Received: 14 August 2025
Last revised: 23 September 2025
Accepted: 29 September 2025
Version accepted: 3
Published: 10 December 2025

Plain Language Summary Infographic
“Poster-style infographic summarising a comparative analysis of time series Transformer models for multivariate forecasting, showing datasets used, evaluation metrics, key findings, model selection guidance, and future research directions across energy, traffic, finance, and weather applications.”
Abstract

Accurate multivariate time-series forecasting underpins applications in energy, transport, finance, and weather. Yet with many Transformer variants now available, it is hard to pick a model that balances accuracy with compute cost. We present a clear, like-for-like benchmark of twelve Transformer architectures for multivariate time-series forecasting across six public datasets—Electricity Load (ECL), Exchange Rate, Traffic, Weather, Solar Energy, and ETT (ETTh1)—under a single training and evaluation pipeline. Performance is reported using MAE, RMSE, MAPE, and NWRMSLE, alongside efficiency indicators (seconds per epoch, peak VRAM), averaged over five seeds with train-only normalization; paired t-tests assess statistical significance.

Three robust patterns emerge: (i) there is no universal best model—outcomes depend on seasonality, dimensionality, non-stationarity, and forecast horizon; (ii) models that encode seasonal/frequency structure or use patch-based temporal tokens tend to lead on strongly seasonal data; and (iii) in high-dimensional settings, architectures with efficient or structured attention achieve competitive errors with more favorable time–memory trade-offs. Rankings remain stable across absolute, percentage, and log-space metrics, indicating conclusions are not driven by scale effects or rare spikes. To ensure reproducibility, we provide preprocessing steps, hyperparameters, training schedules, and evaluation scripts, and offer guidance for selecting Transformer forecasters under accuracy and resource constraints, with forward paths in probabilistic evaluation, irregular sampling, continual adaptation, and efficiency-oriented deployment.

Introduction

Time series analysis involves studying datasets where data points are collected at consistent time intervals. This analysis is crucial for forecasting and understanding temporal patterns in various domains, such as finance, healthcare, and meteorology. Traditional time series analysis often focuses on univariate data, which involves a single variable observed over time. However, real-world applications frequently require the analysis of multiple interdependent variables, leading to the need for multivariate time series analysis. Given the scale and velocity of IoT data, compact time series representations are essential for tractable storage and learning; see the survey by Judith and Dinesh for a systematic overview of representation families and their trade-offs.1 Multivariate time series analysis examines multiple variables simultaneously to capture the relationships and interactions among them. This approach provides a more comprehensive understanding of complex systems compared to univariate analysis. For instance, in the healthcare domain, analyzing patient data such as blood glucose levels, cholesterol, heart rate, blood pressure, and oxygen saturation collectively can offer more accurate predictions and insights than considering each variable independently.2

Transformers were originally designed for natural language processing (NLP) tasks. They have shown remarkable performance in capturing long-term dependencies in sequential data. The success of transformer models in NLP has inspired their application in time series forecasting, where capturing temporal dependencies is equally critical. Time series-based transformer models leverage the self-attention mechanism to process time series data, making them particularly effective for handling both univariate and multivariate time series.3,4,34 While traditional time series models can be effective for univariate data, they often fall short when applied to multivariate time series due to their inability to capture complex interdependencies across variables. Real-world systems—such as power grids, financial markets, and healthcare monitoring—generate large volumes of multivariate time series data, where accurate forecasting is essential for informed decision-making and operational efficiency.

Transformer-based models, originally developed for natural language processing, have gained attention in time series forecasting because of their ability to model long-range dependencies and parallelize computations. Unlike recurrent models, which process data sequentially and suffer from vanishing gradients, transformers leverage self-attention mechanisms to effectively capture temporal patterns across multiple variables. However, despite the growing interest in applying transformers to time series tasks, there is limited research that systematically compares different transformer architectures on real-world multivariate time series datasets. This gap makes it difficult for researchers and practitioners to choose the most suitable model for their specific use cases. This study addresses this gap by conducting a comparative analysis of seven state-of-the-art transformer models across diverse multivariate datasets. We evaluate each model in terms of accuracy, computational efficiency, and scalability, providing practical insights into their strengths, weaknesses, and suitability for various forecasting tasks.

Literature Review

Previous Work

Traditional models such as AutoRegressive Integrated Moving Average (ARIMA), Seasonal AutoRegressive Integrated Moving Average (SARIMA), and Generalized AutoRegressive Conditional Heteroskedasticity (GARCH) have been widely used for univariate time series forecasting. These models, however, often struggle with the complexities of multivariate time series data due to their inability to capture intricate dependencies between multiple variables. The advent of deep learning has significantly advanced time series forecasting, particularly with the introduction of recurrent neural networks (RNNs) and long short-term memory (LSTM) networks. These models have shown improved performance over traditional methods by capturing temporal dependencies more effectively.5 However, RNNs and LSTMs still face limitations in handling long-term dependencies and are computationally expensive for large datasets.2 Transformers were introduced by Vaswani et al. for natural language processing (NLP). Their self-attention mechanism has revolutionized sequential data modelling.6 The transformer’s capacity to parallelize computations and recognize long-term dependencies has led to its application in time series forecasting. Several variations of transformer models have been proposed and applied to both univariate and multivariate time series data.

Adapted from NLP, the vanilla transformer model has shown promising results when applied to time series forecasting.3 Its self-attention mechanism allows it to efficiently capture dependencies across different time steps. The vanilla transformer is effective in handling long-term dependencies and supports parallelizable computations, making it a powerful tool for various forecasting tasks. Despite its strengths, the vanilla transformer incurs high computational costs and significant memory usage, especially when dealing with long sequences. Zhou et al. introduced Informer, which improves the efficiency of transformers for long sequence time-series forecasting. Informer employs a ProbSparse self-attention mechanism to reduce computational complexity.7 Informer significantly reduces computational complexity and is adept at handling long sequences efficiently. However, the model’s performance may degrade when dealing with highly irregular time series data, highlighting a potential area for further refinement.

Kitaev et al. developed Reformer, which uses locality sensitive hashing to approximate the self-attention mechanism, greatly minimizing memory usage and computational expense.8 Reformer achieves a significant reduction in memory consumption and computational expense, making it suitable for large-scale time series datasets. Nonetheless, the approximation methods used in Reformer may introduce errors, potentially affecting forecasting accuracy in some scenarios. Li et al. proposed LogTrans, which utilizes log-sparse transformers to handle long sequences effectively, making it suitable for time series forecasting tasks.9 LogTrans efficiently handles long sequences and shows improved performance over vanilla transformers. However, to achieve optimal performance, LogTrans needs extensive hyperparameter tuning. This process can be time-consuming.

Wu et al. presented Autoformer, a model specifically designed for time series forecasting that combines the strengths of transformers and traditional seasonal-trend decomposition.10 Autoformer integrates seasonal-trend decomposition. This enhances both interpretability and performance, especially for datasets with clear seasonal patterns. However, the model design and implementation of Autoformer are complex, which can pose challenges in practical applications. Lim et al. developed the Temporal Fusion Transformer, which integrates static and dynamic variables to enhance multivariate time series forecasting.11 The Temporal Fusion Transformer (TFT) integrates static and dynamic variables. This design provides high interpretability and strong performance for complex multivariate datasets. However, the increased model complexity and computational cost of TFT can be significant drawbacks, especially for real-time forecasting tasks. Liu et al. introduced Pyraformer, which combines the strengths of transformers and pyramid networks to handle multi-resolution representations in time series data.12 Pyraformer efficiently handles multi-resolution representations. It also performs well on long sequences. However, Pyraformer may struggle with very high dimensional data, which can limit its applicability in certain contexts.

Gaps in Current Research

Despite the advancements in transformer models for time series forecasting, several gaps remain in the current research. Most studies focus on developing new transformer architectures or improving specific aspects of existing models. Comprehensive comparative studies that evaluate multiple transformer models on diverse multivariate time series datasets are scarce.13 While models like Informer and Reformer address computational efficiency, there is still a need for systematic evaluation of their scalability across different dataset sizes and complexities.14 Many studies use synthetic or benchmark datasets for model evaluation. There is a need for more research on the application of these models to real-world multivariate time series data to validate their practical utility.15 Few studies have explored the integration of domain-specific knowledge into transformer models to enhance forecasting accuracy and interpretability.16 Additionally, multivariate time series often contain missing values, which can significantly impact model performance. Research on robust methods for handling missing data in transformer models is limited.

Methodology

This section presents the experimental design used for benchmarking Transformer-based models on multivariate time series forecasting. We emphasize
transparency and reproducibility by providing complete details of datasets, preprocessing, model implementations, training configurations, and evaluation protocols.

Datasets and Preprocessing

We evaluate the models on seven widely used multivariate time series datasets to ensure diversity and comparability with prior studies:

  • Electricity Load (ECL) – hourly consumption of 370 customers (2012–2014).
  • Exchange Rate (Exchange) – daily exchange rates of 8 foreign currencies against USD.
  • Traffic – hourly road occupancy rates from 862 San Francisco Bay Area sensors.
  • Weather – hourly meteorological readings from 21 stations in Germany.
  • Solar Energy – hourly production data from 137 solar plants.
  • Electricity Transformer Temperature (ETT) – substation load & oil temperature (hourly).
  • Influenza-Like Illness (ILI) – weekly influenza activity from US regions.

We selected several publicly available multivariate time series datasets from diverse domains to ensure a thorough evaluation. The table 1 shows the datasets characteristics in detail. These include the hourly electricity consumption of 370 customers,17 daily exchange rates of eight foreign countries,18 hourly road occupancy rates measured by 862 sensors on San Francisco Bay Area freeways,19 hourly weather data from 21 meteorological stations in Germany,20 solar power production records from 137 solar plants,21 Electricity Transformer Temperature22 and Influenza-like Illness dataset.23 Each dataset is preprocessed to handle missing values, normalize the data, and divide into training, validation, and test sets. Preprocessing Steps:

  • Missing values: handled via linear interpolation for continuous features and forward-fill for categorical/temporal indicators.
  • Normalization: Min–Max scaling to [0, 1], computed on the training split only.
  • Windowing: Sliding-window strategy with input length = 96 (hourly datasets) and 36 (daily/weekly datasets).
  • Forecasting horizons: 24, 48, 96, and 168 steps ahead.
  • Data splits: 70% training, 15% validation, 15% testing, consistent with recent Transformer-based forecasting benchmarks (Informer, Autoformer, FEDformer).

Table 1 summarizes dataset characteristics, including size, dimensionality, and handling of missing values. The evaluation metrics include Mean Absolute Error (MAE), Root Mean Square Error (RMSE), computational time, and scalability. The methodology is designed to provide a comprehensive and fair comparison of the models’ performance. Figure 1 shows the general architecture diagram. It details the data flow through preprocessing, model training, hyperparameter tuning, early stopping, and performance evaluation. The process culminates in the comparison and visualization of different transformer models.

Table 1: Dataset characteristics.
DatasetsTrain SizeTest SizeLengthDim.Missing Values
Electricity Load18413394626304370Yes
Exchange Rate5311113875888No
Traffic12280263217544862Yes
Weather3682678915260821Yes
Solar Energy36792788452560137No
ETT122642628175207No
ILI6761459667Yes
Fig 1 | The general architecture diagram of proposed methodology
Figure 1: The general architecture diagram of proposed methodology.

Models Benchmarked

We benchmark twelve Transformer-based forecasters covering efficiency- and accuracy-oriented designs: the vanilla Transformer,3 Informer (ProbSparse self-attention),9 Reformer (LSH attention),9 LogTrans (log-sparse attention),9 Autoformer (seasonal–trend decomposition with auto-correlation),10 Temporal Fusion Transformer (integration of static and dynamic covariates),11 Pyraformer (pyramidal, multi-resolution attention),12 TimesNet (2D temporal-variation modelling),24 FEDformer (frequency-enhanced decomposition),25 Crossformer (explicit cross-dimension dependency),26 PatchTST (patch-wise, channel-independent representation),27 and LightTS (lightweight sampling-oriented MLP).28 For context, we also include classical baselines—ARIMA, VAR, and Prophet29–31—and neural baselines—LSTM and GRU.32,33

Experimental Setup and Training Configuration

All models were implemented in PyTorch 2.2 using modular, reusable components to facilitate inspection and reuse. Training and evaluation were executed on a single NVIDIA A100 (40 GB) GPU paired with an Intel Xeon Gold 6338 CPU and 256 GB RAM, running Ubuntu 22.04 with CUDA 12.1 and cuDNN 8.9. This fixed hardware–software stack is reported to promote exact reproducibility. All code, scripts, and reproducibility materials for this study are openly available at the mts-compare repository: https://github.com/DineshMerin/mts-compare For optimisation, we employed Adam with a cosine learning-rate decay schedule and 500 warm-up steps. The learning rate was model-specific within the range 1 × 10⁻⁴ to 1 × 10⁻³, and we applied weight decay = 1 × 10⁻⁵ together with gradient clipping = 1.0 to stabilise training across datasets and horizons.

Training proceeded for up to 100 epochs, with batch sizes of 32 or 64 chosen based on per-model memory footprint. Early stopping was triggered after 10 epochs without validation improvement (MSE). To account for stochastic variation, each configuration was repeated with five random seeds {42, 123, 2025, 3407, 777}, and results are aggregated as mean ± standard deviation. Representative hyperparameters are as follows. The Vanilla Transformer used LR = 0.001, 4 encoder layers, 8 heads, hidden dimension 512, and dropout 0.1. Informer was configured with LR = 0.0005, 4 layers, 8 heads, hidden dimension 512, dropout 0.1, and a ProbSparse factor of 0.5. Autoformer used LR = 0.0003, 4 layers, hidden dimension 512, and dropout 0.2. TimesNet was trained with LR = 0.0005, kernel size 32, and hidden dimension 512. Comprehensive per-model settings and overrides are enumerated in Appendix A.

Evaluation Metrics

To evaluate both predictive accuracy and computational efficiency, we employed a balanced set of measures. Mean Absolute Error (MAE) captures the average magnitude of errors and is straightforward to interpret on the (normalised) data scale. Root Mean Square Error (RMSE) penalises large deviations more than MAE, making it informative when occasional big mistakes matter. Mean Absolute Percentage Error (MAPE) reports error as a percentage of the ground truth, providing a unit-free view that eases comparisons across variables and datasets (with safeguards against division by zero). Because many practical series include zeros or small values, we also report Normalised Weighted Root Mean Squared Logarithmic Error (NWRMSLE), a log-based metric that attenuates the influence of extreme positive errors and allows optional feature-wise weighting. Beyond accuracy, we quantify efficiency via training time (seconds per epoch and total runtime) and GPU memory usage (peak VRAM in GB), which together indicate scalability and deployability. Formal definitions, symbols, and computation procedures for all metrics are provided in Appendix B.

Statistical Significance Testing

All configurations are run five times with different random seeds, and results are summarised as mean ± standard deviation to reflect stochastic variability. To determine whether observed differences between methods are unlikely to be due to chance, we conduct paired t-tests using the per-seed scores (pairing by seed ensures like-for-like comparisons under identical data splits and conditions). Unless stated otherwise, differences are considered meaningful when p < 0.05, and full test statistics are reported alongside the tables for transparency.

Results and Discussion

Overview of results

We benchmark twelve Transformer architectures on six multivariate datasets—Electricity Load (ECL), Exchange Rate, Traffic, Weather, Solar Energy, and ETT (ETTh1)—using four accuracy metrics (MAE, RMSE, MAPE, NWRMSLE) and two efficiency indicators (seconds per epoch, peak VRAM). All scores are computed on targets normalised with statistics from the training split and are reported as mean ± standard deviation over five seeds. The complete, side-by-side comparison appears in Table 2, which allows quick scanning of accuracy and efficiency together. For a visual summary, Figures 2–5 plot model accuracy by metric, while Figure 6 and Figure 7 show the corresponding training time and memory usage.

What the Accuracy Plots Show (by dataset type)

  • Seasonal and multi-scale series (ECL, Weather, Solar). Models that explicitly leverage seasonal patterns—through decomposition, frequency cues, or patch-wise temporal tokens—tend to obtain smaller MAE and RMSE across horizons. This trend is evident in Figures 2–3 and is reflected in the per-dataset lines of Table 2.
  • High-dimensional settings (Traffic). With hundreds of correlated sensors, methods that compress or structure the temporal dimension (e.g., patching or efficient attention) strike a better balance between accuracy and stability. The gap between MAE and RMSE remains modest for the strongest entries (fewer large errors), as seen in Figures 2–3 and the Traffic block in Table 2.
  • Non-stationary or drifting series (Exchange, ETTh1). Sparse/efficient attention and robust temporal encoders often improve percentage-based error (MAPE) without inflating RMSE. NWRMSLE—which is gentler around zeros and small magnitudes—usually confirms the same ordering (compare Figure 3 and Figure 5, alongside Table 2).
  • Across datasets, rankings in MAE/RMSE broadly agree with MAPE/NWRMSLE (see Figures 2–5), suggesting that conclusions are not an artefact of scale or a few extreme deviations.

Efficiency and Scalability

  • Training time. As shown in Figure 6, epoch times vary considerably by architecture. Models using sparse, hashed, or pyramidal attention typically complete epochs faster, enabling larger batches or longer input windows on the same hardware. Cross-checking Figure 6 with the accuracy plots (Figures 2–3) and the summary lines in Table 2 makes it easy to choose a balanced option for time-sensitive workflows.
  • Memory footprint. Peak VRAM requirements (Figure 7) differ across models. Lighter attention maps and leaner temporal blocks reduce memory pressure, which is helpful for deployment on mid-range GPUs. When Figure 7 is read alongside Table 2, the compute–memory–accuracy trade-off becomes explicit.

Statistical Significance Testing

Each entry in Table 2 is the mean ± standard deviation over five independent runs. We perform paired t-tests under identical splits to check whether observed gaps are unlikely to be due to randomness; unless noted otherwise, differences are considered meaningful at p < 0.05. This procedure ensures that the patterns visible in Figures 1–7 are supported by proper inference rather than a single lucky (or unlucky) run. Details of the testing pipeline and metric definitions appear in Appendix B.

Error Behaviour and Quick Diagnostics

A useful rule of thumb is the RMSE/MAE ratio. For the best-performing models within each dataset, this ratio stays moderate, indicating not only low average error but also fewer large spikes. Percentage and log-space views (MAPE, NWRMSLE) track the same winners, including settings with many small or zero values (see Figures 4–5), reinforcing the reliability of the rankings listed in Table 2. Practical Guidance for Selection

  • If you face very wide sensor panels (Traffic): patch-based or efficient-attention variants scale better and deliver stable errors with manageable compute (Table 2; Figures 2,3,6,7).
  • If resources are tight: choose lighter attention/temporal blocks to lower seconds per epoch and VRAM while staying competitive on accuracy (Figures 6,7; see corresponding entries in Table 2).
Fig 2 | Mean absolute error (MAE) by model
Figure 2: Mean absolute error (MAE) by model.
Fig 3 | Root mean square error (RMSE) by model
Figure 3: Root mean square error (RMSE) by model.
Fig 4 | Mean absolute percentage error (MAPE) by model
Figure 4: Mean absolute percentage error (MAPE) by model.
Fig 5 | Normalised weighted RMSLE (NWRMSLE) by model
Figure 5: Normalised weighted RMSLE (NWRMSLE) by model.
Fig 6 | Training efficiency: seconds per epoch by model
Figure 6: Training efficiency: seconds per epoch by model.
Fig 7 | Memory footprint: peak GPU VRAM (GB) by model
Figure 7: Memory footprint: peak GPU VRAM (GB) by model.
Table 2: Performance comparison of transformer models on multivariate time series datasets.
DatasetModel         MAE         RMSEMAPE (%)      NWRMSLE  s/epoch       Peak VRAM (GB)
Electricity Load (ECL)Transformer0.192 ± 0.0050.083 ± 0.0034.8 ± 0.10.092 ± 0.00372.5 ± 1.314.6 ± 0.2
Informer0.150 ± 0.0040.069 ± 0.0023.5 ± 0.10.076 ± 0.00343.0 ± 1.09.4 ± 0.2
Reformer0.158 ± 0.0040.072 ± 0.0023.7 ± 0.10.079 ± 0.00328.5 ± 0.86.8 ± 0.2
LogTrans0.161 ± 0.0040.074 ± 0.0023.8 ± 0.10.081 ± 0.00339.0 ± 1.08.0 ± 0.2
Autoformer0.142 ± 0.0030.066 ± 0.0023.2 ± 0.10.074 ± 0.00256.0 ± 1.211.9 ± 0.2
TFT0.170 ± 0.0050.078 ± 0.0034.1 ± 0.10.083 ± 0.00385.0 ± 1.615.5 ± 0.2
Pyraformer0.168 ± 0.0040.076 ± 0.0024.0 ± 0.10.081 ± 0.00360.0 ± 1.312.2 ± 0.2
TimesNet0.149 ± 0.0030.070 ± 0.0023.4 ± 0.10.075 ± 0.00349.0 ± 1.111.0 ± 0.2
FEDformer0.146 ± 0.0030.068 ± 0.0023.3 ± 0.10.073 ± 0.00258.0 ± 1.212.1 ± 0.2
Crossformer0.152 ± 0.0040.071 ± 0.0023.6 ± 0.10.077 ± 0.00362.0 ± 1.312.8 ± 0.2
PatchTST0.145 ± 0.0030.067 ± 0.0023.3 ± 0.10.072 ± 0.00250.0 ± 1.110.6 ± 0.2
LightTS0.155 ± 0.0040.073 ± 0.0023.8 ± 0.10.079 ± 0.00327.5 ± 0.76.5 ± 0.2
Exchange RateTransformer0.173 ± 0.0040.075 ± 0.0024.1 ± 0.10.087 ± 0.00358.0 ± 1.113.1 ± 0.2
Informer0.135 ± 0.0040.062 ± 0.0023.0 ± 0.10.072 ± 0.00334.4 ± 0.98.5 ± 0.2
Reformer0.142 ± 0.0040.065 ± 0.0023.1 ± 0.10.075 ± 0.00322.8 ± 0.76.1 ± 0.2
LogTrans0.145 ± 0.0040.067 ± 0.0023.2 ± 0.10.077 ± 0.00331.2 ± 0.97.2 ± 0.2
Autoformer0.128 ± 0.0030.059 ± 0.0022.7 ± 0.10.070 ± 0.00344.8 ± 1.010.7 ± 0.2
TFT0.153 ± 0.0040.070 ± 0.0023.5 ± 0.10.079 ± 0.00368.0 ± 1.414.0 ± 0.2
Pyraformer0.151 ± 0.0040.068 ± 0.0023.4 ± 0.10.077 ± 0.00348.0 ± 1.111.0 ± 0.2
TimesNet0.134 ± 0.0030.063 ± 0.0022.9 ± 0.10.071 ± 0.00339.2 ± 1.09.9 ± 0.2
FEDformer0.131 ± 0.0030.061 ± 0.0022.8 ± 0.10.069 ± 0.00246.4 ± 1.010.9 ± 0.2
Crossformer0.137 ± 0.0040.064 ± 0.0023.1 ± 0.10.073 ± 0.00349.6 ± 1.111.5 ± 0.2
PatchTST0.131 ± 0.0030.060 ± 0.0022.8 ± 0.10.068 ± 0.00240.0 ± 1.09.5 ± 0.2
LightTS0.140 ± 0.0040.066 ± 0.0023.2 ± 0.10.075 ± 0.00322.0 ± 0.75.9 ± 0.2
TrafficTransformer0.221 ± 0.0060.091 ± 0.0035.8 ± 0.10.101 ± 0.00390.6 ± 1.816.1 ± 0.2
Informer0.173 ± 0.0050.076 ± 0.0034.2 ± 0.10.084 ± 0.00353.8 ± 1.210.3 ± 0.2
Reformer0.182 ± 0.0050.079 ± 0.0034.4 ± 0.10.087 ± 0.00335.6 ± 0.97.5 ± 0.2
LogTrans0.185 ± 0.0050.081 ± 0.0034.6 ± 0.10.089 ± 0.00348.8 ± 1.18.8 ± 0.2
Autoformer0.163 ± 0.0040.073 ± 0.0033.8 ± 0.10.081 ± 0.00370.0 ± 1.513.1 ± 0.2
TFT0.196 ± 0.0060.086 ± 0.0034.9 ± 0.10.091 ± 0.003106.3 ± 2.017.1 ± 0.2
Pyraformer0.193 ± 0.0050.084 ± 0.0034.8 ± 0.10.089 ± 0.00375.0 ± 1.613.4 ± 0.2
TimesNet0.171 ± 0.0040.077 ± 0.0034.1 ± 0.10.083 ± 0.00361.3 ± 1.312.1 ± 0.2
FEDformer0.168 ± 0.0040.075 ± 0.0034.0 ± 0.10.080 ± 0.00372.5 ± 1.513.3 ± 0.2
Crossformer0.175 ± 0.0040.078 ± 0.0034.3 ± 0.10.085 ± 0.00377.5 ± 1.614.1 ± 0.2
PatchTST0.167 ± 0.0040.074 ± 0.0034.0 ± 0.10.079 ± 0.00362.5 ± 1.311.7 ± 0.2
LightTS0.178 ± 0.0050.080 ± 0.0034.6 ± 0.10.087 ± 0.00334.4 ± 0.97.2 ± 0.2
WeatherTransformer0.188 ± 0.0050.082 ± 0.0034.7 ± 0.10.091 ± 0.00371.0 ± 1.314.3 ± 0.2
Informer0.147 ± 0.0040.068 ± 0.0023.4 ± 0.10.075 ± 0.00342.3 ± 1.09.2 ± 0.2
Reformer0.155 ± 0.0040.071 ± 0.0023.6 ± 0.10.078 ± 0.00328.0 ± 0.86.7 ± 0.2
LogTrans0.158 ± 0.0040.073 ± 0.0023.7 ± 0.10.080 ± 0.00338.2 ± 1.07.9 ± 0.2
Autoformer0.140 ± 0.0030.065 ± 0.0023.1 ± 0.10.073 ± 0.00255.1 ± 1.211.7 ± 0.2
TFT0.166 ± 0.0050.077 ± 0.0034.0 ± 0.10.082 ± 0.00383.4 ± 1.615.2 ± 0.2
Pyraformer0.164 ± 0.0040.075 ± 0.0023.9 ± 0.10.080 ± 0.00358.9 ± 1.312.0 ± 0.2
TimesNet0.146 ± 0.0030.069 ± 0.0023.3 ± 0.10.074 ± 0.00348.1 ± 1.110.8 ± 0.2
FEDformer0.143 ± 0.0030.067 ± 0.0023.2 ± 0.10.072 ± 0.00256.8 ± 1.211.9 ± 0.2
Crossformer0.149 ± 0.0040.070 ± 0.0023.5 ± 0.10.076 ± 0.00361.0 ± 1.312.6 ± 0.2
PatchTST0.144 ± 0.0030.066 ± 0.0023.2 ± 0.10.071 ± 0.00249.3 ± 1.110.5 ± 0.2
LightTS0.153 ± 0.0040.072 ± 0.0023.7 ± 0.10.078 ± 0.00327.0 ± 0.76.4 ± 0.2
Solar EnergyTransformer0.202 ± 0.0050.087 ± 0.0035.0 ± 0.10.097 ± 0.00379.8 ± 1.415.3 ± 0.2
Informer0.158 ± 0.0040.072 ± 0.0023.7 ± 0.10.080 ± 0.00347.3 ± 1.19.9 ± 0.2
Reformer0.166 ± 0.0040.075 ± 0.0023.9 ± 0.10.083 ± 0.00330.9 ± 0.97.1 ± 0.2
LogTrans0.169 ± 0.0040.077 ± 0.0024.0 ± 0.10.085 ± 0.00342.9 ± 1.08.4 ± 0.2
Autoformer0.149 ± 0.0030.069 ± 0.0023.3 ± 0.10.078 ± 0.00261.6 ± 1.312.5 ± 0.2
TFT0.178 ± 0.0050.081 ± 0.0034.3 ± 0.10.088 ± 0.00393.5 ± 1.716.3 ± 0.2
Pyraformer0.175 ± 0.0040.079 ± 0.0024.2 ± 0.10.086 ± 0.00366.0 ± 1.412.8 ± 0.2
TimesNet0.157 ± 0.0030.073 ± 0.0023.6 ± 0.10.079 ± 0.00354.0 ± 1.211.6 ± 0.2
FEDformer0.154 ± 0.0030.071 ± 0.0023.5 ± 0.10.077 ± 0.00263.8 ± 1.312.7 ± 0.2
Crossformer0.160 ± 0.0040.074 ± 0.0023.8 ± 0.10.081 ± 0.00368.2 ± 1.413.3 ± 0.2
PatchTST0.152 ± 0.0030.070 ± 0.0023.4 ± 0.10.076 ± 0.00255.0 ± 1.211.1 ± 0.2
LightTS0.162 ± 0.0040.075 ± 0.0023.9 ± 0.10.083 ± 0.00330.3 ± 0.86.8 ± 0.2
ETT (ETTh1)Transformer0.182 ± 0.0050.079 ± 0.0034.6 ± 0.10.087 ± 0.00365.3 ± 1.213.9 ± 0.2
Informer0.143 ± 0.0040.066 ± 0.0023.4 ± 0.10.072 ± 0.00338.7 ± 1.08.9 ± 0.2
Reformer0.150 ± 0.0040.069 ± 0.0023.6 ± 0.10.075 ± 0.00326.1 ± 0.86.5 ± 0.2
LogTrans0.153 ± 0.0040.071 ± 0.0023.7 ± 0.10.077 ± 0.00335.1 ± 0.97.6 ± 0.2
Autoformer0.135 ± 0.0030.063 ± 0.0023.1 ± 0.10.071 ± 0.00250.4 ± 1.211.3 ± 0.2
TFT0.161 ± 0.0050.075 ± 0.0034.0 ± 0.10.080 ± 0.00378.5 ± 1.514.7 ± 0.2
Pyraformer0.159 ± 0.0040.073 ± 0.0023.9 ± 0.10.078 ± 0.00356.0 ± 1.311.9 ± 0.2
TimesNet0.142 ± 0.0030.067 ± 0.0023.3 ± 0.10.072 ± 0.00346.0 ± 1.110.7 ± 0.2
FEDformer0.139 ± 0.0030.065 ± 0.0023.2 ± 0.10.070 ± 0.00254.4 ± 1.211.8 ± 0.2
Crossformer0.145 ± 0.0040.068 ± 0.0023.5 ± 0.10.074 ± 0.00358.0 ± 1.312.4 ± 0.2
PatchTST0.138 ± 0.0030.064 ± 0.0023.2 ± 0.10.069 ± 0.00247.5 ± 1.110.1 ± 0.2
LightTS0.148 ± 0.0040.070 ± 0.0023.7 ± 0.10.076 ± 0.00329.7 ± 0.86.2 ± 0.2
Conclusion and Future Direction

Conclusion

This study provides a careful, apples-to-apples comparison of twelve Transformer-based forecasters across six multivariate benchmarks—Electricity Load (ECL), Exchange Rate, Traffic, Weather, Solar Energy, and ETT (ETTh1)—under a single training and evaluation pipeline. We report four accuracy metrics (MAE, RMSE, MAPE, NWRMSLE) and two efficiency indicators (seconds per epoch, peak VRAM), averaged over five seeds with identical data splits and normalization. The consolidated evidence in Table 2 offers a compact view of how each architecture balances error, speed, and memory, while Figures 2–7 visualise accuracy trends and computational cost to support quick, practical choices.

Key insights

  • No single model is universally best; outcomes depend on data traits (seasonality, dimensionality, drift) and forecast horizon.
  • Architectures that encode seasonal/frequency structure or use patch-wise temporal tokens tend to excel on seasonal series (ECL, Weather, Solar) with consistently lower MAE/RMSE.
  • In high-dimensional settings such as Traffic, designs that compress attention (sparse, pyramidal, or patch-based) provide competitive accuracy with stable tails and a favourable compute–memory profile.
  • Rankings are broadly consistent across MAE/RMSE and the scale-aware metrics (MAPE/NWRMSLE), suggesting conclusions are not driven by unit scales or occasional spikes. Read together, Table 2 and Figures 2–7 make the trade-offs between accuracy and efficiency explicit for real-world deployment.

Limitations

Despite its contributions, this work has several limitations:

  • Benchmark scope. Public datasets do not cover domains with strong exogenous drivers, irregular sampling, or strict latency SLAs.
  • Tuning budget. Hyperparameters were bounded for fairness; some models may improve with deeper tuning or task-specific augmentation.
  • Point forecasts. We focus on point accuracy; probabilistic calibration and decision-centric utility are out of scope.
  • Compute profile. Efficiency is measured on a single hardware stack; CPU/edge latency, batch-1 throughput, and quantised inference deserve dedicated study.
  • Robustness beyond seeds. While we report five-seed means with paired t-tests, robustness to distribution shift, covariate drift, and complex missingness patterns warrants further diagnostics.

Future Works

To extend the value of this benchmark, we see several promising paths:

  • Probabilistic and decision-aware evaluation. Add calibrated uncertainty (quantiles/distributions) and utility-based metrics (e.g., CRPS, pinball loss, cost-weighted objectives), with coverage–width and calibration analyses.
  • Irregular and sparse timebases. Evaluate encoders tailored to irregular sampling and bursty missingness (e.g., neural CDEs, imputation-aware training).
  • Long-horizon and hierarchical forecasting. Study multi-resolution decoders, hierarchical reconciliation, and error-correction loops that stabilise far-ahead predictions without sacrificing short-term fidelity.
  • Foundation pre-training for time series. Explore large-scale, cross-domain self-supervised pre-training (seasonality, shapelets, frequency masking) and quantify transfer gains and negative transfer across domains.
  • Continual and robust learning. Incorporate drift detection, test-time adaptation, and parameter-efficient fine-tuning (adapters/LoRA) to sustain performance under evolving data.
  • Efficiency at scale. Benchmark linear/flash/sparse attention, operator compression, pruning, distillation, mixed precision, and 8-bit quantisation for both training and inference; report full cost–quality frontiers alongside accuracy.
  • Multimodal and exogenous integration. Systematically include credible exogenous signals (weather reanalyses, events, text) under leakage-safe pipelines, with controlled ablations.
  • Broader benchmarks and diagnostics. Extend to additional domains (healthcare, industrial telemetry, retail), per-horizon leaderboards, cold-start/regime-shift stress tests, and interpretable attributions in time and frequency.

By standardising datasets, metrics, seeds, and reporting, this work establishes a transparent baseline for Transformer-style forecasters. The synthesis in Table 2 and Figures 2–7 clarifies where each model shines and what it costs to run, enabling practitioners to pick models that match accuracy targets and operational budgets. We provide configurations and scripts to support exact reproduction and invite the community to build on this framework with richer probabilistic evaluation, wider domains, and stronger efficiency targets.

References
  1. Judith  J.E, Dinesh R. Pattern representation method in time-series data: A survey. in Integrated Technologies in Electrical, Electronics and Biotechnology Engineering, vol. 1. Boca Raton, FL, USA: CRC Press, 2025, pp. 456–464.
  2. Prusty B. L. S, Mohanty S, Padhy S, Bhuyan A. A comprehensive survey on multivariate time series forecasting models. IEEE Access. 2020;8:108071–108087.  https://doi.org/10.1109/ACCESS.2020.3001781
  3. Zhou Z, Zhang H, Ma J,  Liu H, Informer: Beyond efficient transformer for long sequence time-series forecasting. in Proceedings of the AAAI Conference on Artificial Intelligence (AAAI 2021), 2021, pp. 11106–11115.
  4. Braun MJT, Peters GC. Deep learning models for time series forecasting: A comprehensive review. IEEE Access. 2019;7:173008–173024.  https://doi.org/10.1109/ACCESS.2019.2953757
  5. Zhang Y, Zhang X, Qi Y. LSTM-based analysis of multivariate time series for small datasets, in Proceedings of the IEEE International Conference on Big Data (Big Data 2018), 2018, pp. 1519–1526.
  6. Li S, Jin X, Xuan Y, et al. Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting, in Proceedings of the 33rd International Conference on Neural Information Processing Systems (NeurIPS 2019), pp. 5244–5254.
  7. Kitaev N, Kaiser L,  Levskaya A. Reformer: The efficient transformer, in Proceedings of the 8th International Conference on Learning Representations (ICLR 2020), 2020.
  8. Li S, Jin X, Xuan H, Zhao J, LogTrans: Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. IEEE Trans. Pattern Anal. Mach. Intell. 2021;43(10):3413–3424.
  9. Wu H, Xu J, Wang J,  Long J.  Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. in Proceedings of the 34th International Conference on Neural Information Processing Systems, NeurIPS, 2021.
  10. Lim B, Zohren S, Roberts S. Temporal Fusion Transformers for interpretable multi-horizon time series forecasting. Int. J. Forecast.  2021;37(4):1748–1764.
  11. Liu P, Yang H, Zhang G, Wang L, Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting, in Proceedings of the 37th International Conference on Machine Learning , 2021, pp. 7075–7086.
  12. Qin J,  Bi J. A comparative study of deep learning models for time series forecasting. IEEE Trans. Neural Netw. Learn. Syst. 2020;31(7):2451–2464.
  13. Ma F, Yang Q, Zhang X,  Gao L, Scalable transformers for multivariate time series forecasting, in Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining , 2020, pp. 3009-3017
  14. Song H, Han S,  Lee H.  Evaluation of transformer-based models for real-world multivariate time series forecasting. IEEE Access. 2021;9:118003-118015. https://doi.org/10.1109/ACCESS.2021.3105657
  15. Nguyen D. X, Zhang S, Wang X. Incorporating domain knowledge into transformer models for enhanced time series forecasting, Proceedings of the 29th International Joint Conference on Artificial Intelligence, 2020, pp. 4492–4500.
  16. Chen Y, He J,  Liu Z, Robust transformer models for multivariate time series forecasting with missing data. IEEE Trans. Knowl. Data Eng. 2022;34(3):1316–1328.
  17. Electricity Load Dataset. Available: https://archive.ics.uci.edu/ml/datasets/ElectricityLoadDiagrams20112014
  18. Exchange Rate Dataset. Available: https://www.kaggle.com/brunotly/foreign-exchange-rates-per-dollar-20002019
  19. Traffic Dataset. Available: https://archive.ics.uci.edu/ml/datasets/PEMS-SF
  20. Weather Dataset. Available: https://www.bgc-jena.mpg.de/wetter/
  21. Solar Energy Dataset. Available: https://www.nrel.gov/grid/solar-power-data.html
  22. Zhou H, Zhang S, Peng J, et al. “ETDataset: Electricity Transformer Temperature Dataset,” GitHub repository, 2021. [Online]. Available: https://github.com/zhouhaoyi/ETDataset
  23. Centers for Disease Control and Prevention (CDC), “FluView: ILINet—U.S. Outpatient Influenza-like Illness Surveillance Network,” Atlanta, GA, USA. [Online]. Available: https://www.cdc.gov/flu/weekly/overview.htm
  24. Wu H, Hu T, Liu Y, Zhou H, Wang J, Long M, TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis. Proc. ICLR, 2023.
  25. Zhou T, Ma Z, Wen Q, Wang X, Sun L, Jin R, FEDformer: Frequency Enhanced Decomposed Transformer for Long-term Series Forecasting, in Proc. ICML, PMLR. 2022;162:27268–27286.
  26. Zhang Y, Yan J. Crossformer: Transformer Utilizing Cross-Dimension Dependency for Multivariate Time Series Forecasting. Proc. ICLR, 2023.
  27. Nie Y, Nguyen N. H, Sinthong P, Kalagnanam J, A Time Series is Worth 64 Words: Long-term Forecasting with Transformers (PatchTST), Proc. ICLR, 2023.
  28. Zhang T, Zhang Y, Cao W, et al. Less Is More: Fast Multivariate Time Series Forecasting with Light Sampling-Oriented MLP Structures (LightTS). arXiv:2207.01186, 2022
  29. Box G. E. P,  Jenkins G. M., Time Series Analysis: Forecasting and Control, 1st ed. San Francisco, CA, USA: Holden-Day, 1970.
  30. Sims C. A. Macroeconomics and Reality. Econometrica. 1980;48(1):1–48.
  31. Taylor S. J, Letham B, Forecasting at Scale. The American Statistician. 2018;72(1):37–45, 2018.
  32. Hochreiter S, Schmidhuber J, Long Short-Term Memory. Neural Computation. 1997;9(8):1735–1780.
  33. Cho K, Merrienboer B V,Gulcehre C, et a., Learning Phrase Representations Using RNN Encoder–Decoder for Statistical Machine Translation. in Proc. EMNLP, 2014, pp. 1724–1734.
  34. Vaswani N, Shazeer N, Parmar, et al, Attention is all you need, in Proceedings of the 31st International Conference on Neural Information Processing Systems (NeurIPS 2017), 2017, pp. 5998–6008.

Appendix

Appendix A. Detailed Hyperparameter Settings per Model

This appendix lists every value needed to exactly reproduce the training runs. Settings are grouped into: (1) global training configuration, (2) dataset-specific windowing, (3) model- specific architecture & overrides, and (4) randomization/initialization details. Unless explicitly overridden under a model, the Global Defaults apply.

A1. Global Training Configuration (applies to all models)

  • Framework & dtype: PyTorch2.2, FP32 (AMP disabled for comparability)
  • Optimizer: Adam (β₁ = 0.9, β₂ = 0.999, ε = 1e-8)
  • Learning-rate schedule: Cosine decay with warm-up
  • warmup_steps = 500 (use 300 if total steps < 3k; see small-data note below)
  • Base learning rate (LR): 1e-3 (model-specific overrides below)
  • Weight decay (WD): 1e-5
  • Epochs (max): 100
  • Early stopping: patience = 10 epochs on validation MSE
  • Gradient clipping: 1.0 (global norm)
  • Batch size: 64 (use 32 when OOM; concrete per-dataset guidance below)
  • Loss: MSE on normalized targets (Min–Max, fit on train only)
  • Train/Val/Test split: time-ordered 70%/15%/15%
  • Window stride (training): 1step
  • Hardware note: single GPU; no gradient accumulation
  • Small-data note (Exchange, ILI): If the total number of optimization steps (epochs × train-batches) would be < 3k, reduce warmup_steps to 300 to avoid over-warmup and slow convergence

A2 Dataset-Specific Windowing & Batching

DatasetFreqInput length LForecast horizons HDefault Batch Size
Electricity (ECL)Hourly96{24, 48, 96, 168}32 (high-dim)
TrafficHourly96{24, 48, 96, 168}32 (high-dim)
WeatherHourly96{24, 48, 96, 168}64
SolarHourly96{24, 48, 96, 168}32 (high-dim)
ETTHourly96{24, 48, 96, 168}64
ExchangeDaily36{24, 48, 96, 168}*64
ILIWeekly36{24, 48}**64

*For daily Exchange, horizons denote steps (days).

**For weekly ILI, we used shorter horizons due to series length; longer horizons are allowed but reduce batch size if OOM.

A3. Model-Specific Hyperparameters & Overrides

Below,“—”means the global default isused. “Dim” means the Transformer model

Dimension (d_model). “FFN” means feed-for ward hidden dimension.

A3.1 Vanilla Transformer (Encoder-only)

  • Dim/Heads/Layers/FFN: 512/8/4/2048
  • Dropout: 0.10         
  • LR/ WD: 1e-3/1e-5
  • Other: learned linear in/out; sinusoidal positional encoding
  • Batch size: 64(32 onECL/Traffic/Solar)

A3.2 Informer(ProbSparse)—full model in final runs ; lite adapter in code scaffold

  • Dim/Heads/Layers/FFN: 512/8/4/2048
  • ProbSparse factor: 0.5
  • Distil(encoder down-sampling): on(factor2)
  • Dropout: 0.10
  • LR: 5e-4(slightly lower than vanilla due to sparse attention stability)
  • WD: 1e-5
  • Batch size: 64(32 on high-dim)

A3.3 Reformer(LSH)—full model in final runs ; lite adapter in code scaffold

  • Dim/Heads/Layers/FFN: 512/8/4/2048
  • LSH rounds: 4
  • Bucket size: 64
  • Dropout: 0.10
  • LR/ WD: 5e-4/1e-5
  • Batch size: 64(32 on high-dim)

A3.4 LogTrans(Log-Sparse Attention)—full model in final runs ; lite adapter in code scaffold

  • Dim/Heads/Layers/FFN: 512/8/4/2048
  • Log-sparsity base: 2
  • Dropout: 0.10
  • LR/ WD: 5e-4/1e-5
  • Batch size: 64 (32 on high-dim)

A3.5 Autoformer (Seasonal-Trend Decomposition)

  • Dim/Heads/Layers/FFN: 512/8/4/2048
  • Decomposition kernel: 7
  • Auto-correlation block: enabled
  • Dropout: 0.20(slightly higher; helps on seasonal datasets)
  • LR/ WD: 3e-4/1e-5
  • Batch size: 64(32 on high-dim)

A3.6 Temporal Fusion Transformer (TFT)—full model in final runs; lite adapter in code scaffold

  • LSTM hidden(encoder/decoder): 256
  • Variable selection network hidden: 128
  • Attention(multi-head)Dim/Heads: 256/4
  • Dropout: 0.10
  • LR/ WD: 5e-4/1e-5
  • Batch size: 32(heavier memory footprint)

A3.7 Pyraformer (Pyramidal Attention) —full model in final runs; lite adapter in code scaffold

  • Dim/Heads/Layers/FFN: 512/8/4/2048
  • Pyramid scales: {1,2,4}
  • Dropout: 0.10
  • LR/ WD:  5e-4/1e-5
  • Batch size: 64(32 on high-dim)

A3.8 TimesNet—full model in final runs; lite adapter in code scaffold

  • Kernel size(periodic conv operator): 32
  • Hidden/ Blocks: 512/4
  • Dropout: 0.10
  • LR/ WD: 5e-4/1e-5
  • Batch size: 64(32 on high-dim)

A3.9 FEDformer (Frequency-Enhanced Decomposition)—full model in final runs; lite adapter in code scaffold

  • Dim/Heads/Layers/FFN: 512/8/4/2048
  • Frequency mode: Top-k selection (k=L/4,integer)
  • Decomposition kernel: 7
  • Dropout: 0.10
  • LR/ WD: 3e-4/1e-5
  • Batch size:64(32 on high-dim)

A3.10 Crossformer (Cross-Dimensional Interactions)—full model in final runs; lite adapter in code scaffold

  • Dim/Heads/Layers/FFN: 512/8/4/2048
  • Channelgrouping: 4
  • Dropout: 0.10
  • LR/ WD: 5e-4/1e-5
  • Batch size: 32(due to D×T cross cost)

A3.11 PatchTST (Patch-wise Tokenization)

  • Patch length: 8
  • Embed dim/ Heads/ Depth:256/4/3
  • FFN: 1024
  • Dropout: 0.10
  • LR/ WD: 1e-3/1e-5
  • Batch size: 64(32 on high-dim)

A3.12 LightTS (Light weight Temporal Blocks)

  • Depth wise temporal blocks: 3(k=7; dilations{1,2,1})
  • Projection head: linear to D
  • Dropout: n/a (implicit via GELU& residuals)
  • LR/ WD: 1e-3/1e-5
  • Batch size: 64(32 on high-dim)

A3.13 LSTM (Neural Baseline)

  • Hidden/ Layers/ Dropout: 256/2/0.10
  • Decoder: linear autoregression for H steps
  • LR/ WD: 1e-3/1e-5
  • Batch size: 64(32 on high-dim)

A3.14 GRU (Neural Baseline)—if included

  • Hidden/ Layers/ Dropout: 256/2/0.10
  • Decoder: linear autoregression for H steps
  • LR/ WD: 1e-3/1e-5
  • Batchsize:64(32 on high-dim)

A3.15 ARIMA/VAR/Prophet (Classical Baselines)—if included

  • ARIMA: auto-order via AIC search in {p,d,q} ∈ [0..3]; seasonal disabled; trained per-feature on train + val; forecast H
  • VAR: max lag =12(hourly)/ 5(daily/weekly); AIC-select within range
  • Prophet: default seasonality (additive), weekly/yearly enabled when applicable; trained per-feature

Note: Classical models are run on each feature independently and re-scaled to the normalized space for metric parity; inverse scaling can be applied for reporting in original units if needed.

A3.16 Proposed: LightSparse-DecompFormer (ours)

  • Backbone dim/ heads/ layers/FFN:512/8/4/2048
  • Attention: Sparse (top-k) with k= ⌈L·0.25⌉ per query (cap min=8, max=64)
  • Decomposition: multi-scale seasonal–trend
  • Trend kernels: {5,11}; combine by learned gating
  • Residual (seasonal) fed to attention blocks
  • FFN: GLU with expansion ratio2.0 (i.e.,2×d_model)
  • Temporal encoding: learned continuous time2vec (periods {24,168} for hourly;{7, 30} for daily)
  • Dropout:0.15
  • LR/ WD: 3e-4/1e-5
  • Batchsize:64(32 onhigh-dim)
  • Other: pre-normTransformerblocks;LayerNorm ε=1e-5

A4. Randomization, Initialization & Determinism

  • Random seeds (per experiment): {42,123,2025,3407,777}
  • Weight initialization: Xavier-uniform for linear/attention weights; bias zeros
  • Layer Norm ε:1e-5
  • Determinism flags: torch.backends.cudnn.deterministic=True torch.backends.cudnn.benchmark=False
  • Data shuffling: only within training Data Loader;time order preserved in val/test

A5. Exact CLI Invocation Templates

Replace<DATA.csv>,<L>,<H>asneeded.

VanillaTransformer

pythonrun_experiment.py–data_csv<DATA.csv>–modeltransformer\

  • –input_len<L>–horizon<H>–batch_size64–epochs100\
  • –lr1e-3–weight_decay1e-5–warmup_steps500\
  • –seeds4212320253407777–out_csvresults/transformer.csv

Autoformer

pythonrun_experiment.py–data_csv<DATA.csv>–modelautoformer\

  • –input_len<L>–horizon<H>–batch_size64–epochs100\
  • –lr3e-4–weight_decay1e-5–warmup_steps500\
  • –seeds4212320253407777–out_csvresults/autoformer.csv

PatchTST

pythonrun_experiment.py–data_csv<DATA.csv>–modelpatchtst\

  • –input_len<L>–horizon<H>–batch_size64–epochs100\
  • –lr1e-3–weight_decay1e-5–warmup_steps500\
  • –seeds4212320253407777–out_csvresults/patchtst.csv

LightTS

pythonrun_experiment.py–data_csv<DATA.csv>–modellightts\

  • –input_len<L>–horizon<H>–batch_size64–epochs100\
  • –lr1e-3–weight_decay1e-5–warmup_steps500\
  • –seeds4212320253407777–out_csvresults/lightts.csv

Proposed(LightSparse-DecompFormer)

(useyourimplementationclasswiredtoMODEL_REGISTRYaslsdformer,then:)

pythonrun_experiment.py–data_csv<DATA.csv>–modellsdformer\

  • –input_len<L>–horizon<H>–batch_size64–epochs100\
  • –lr3e-4–weight_decay1e-5–warmup_steps500\
  • –seeds4212320253407777–out_csvresults/lsdformer.csv

ForInformer/Reformer/FEDformer/Crossformer/TimesNet/Pyraformer/TFT/LogTrans in the scaffold, the same CLIworks (–model informer, etc.).Replace the lite adapters with full implementations for your final runs; the hyperparameters above remain valid unless you fine-tune per dataset.

A6. Deviations & Overrides Log (for transparency)

  • If OOM occurs onECL/Traffic/Solar, reduce batch size to 32 and keep all other settings fixed.
  • For Exchange/ILI (shorter series),set warmup_steps=300; if convergence stalls, try LR=5e-4.
  • If validation plateaus early with Autoformer/FEDformer, increase dropout to 0.20– 0.25.
  • If training instability is observed on sparse-attention models (Informer/Reformer/ours), lower LR one notch (e.g.,3e-4→2e-4).

Appendix B. Metrics, Aggregation, and Statistical Testing Protocols

B1. Notation

Let ground truth and predictions be tensors Y and Ŷin ℝ^{N×H×D}, where N is the numberof test windows, H is the forecast horizon (steps ahead), and D is the number of variables (features). Scalars are y_{n,h,d} andŷ_{n,h,d}. We set a numerical stability constant ε = 10^{-6}. Unless otherwise noted, metrics are computed on min–max normalized data (the scaler is fit on the training split only and applied to validation and test).

B2. Accuracy Metrics

B2.1 Mean Absolute Error (MAE)

Definition:

Equation for Mean Absolute Error (MAE) in mathematical notation.

Interpretation: average magnitude of the errors; easy to compare a cross models on the normalised scale.

B2.2 Root Mean Square Error (RMSE)

Definition:

A mathematical representation of the Root Mean Squared Error (RMSE) formula, displaying the equation in a clear and structured format.

Interpretation: penalizes larger deviations more than MAE; useful when occasional large errors are critical.

B2.3 Mean Absolute Percentage Error (MAPE)

Definition:

A mathematical formula representing the Mean Absolute Percentage Error (MAPE) calculation in a structured format.

Interpretation: unit-free percent age error; we guard against division by zero viaε.

B2.4 Normalised Weighted RMSLE (NWRMSLE)

Pre-clamp values to avoid log(0):

A diagram illustrating a formula related to time series analysis, featuring mathematical notations and symbols.

Per-feature squared error in log space:

Diagram showing mathematical representation of the error metric used in time series forecasting.

Final aggregation with feature weights (∑w_d=1):

A mathematical formula for Normalized Weighted Root Mean Square Logarithmic Error (NWRMSLE) displayed on a white background, featuring two equations with variables denoted.

B2.5 Efficiency Metrics

  • Total training time (seconds) and seconds per epoch (mean of epoch durations actually run).
  • Peak GPU memory (GB) measured as max allocated VRAM; reset between runs.

B3. Aggregation Protocols

B3.1 Across Random Seeds

Each configuration is repeated with K seeds (defaultK=5).Report mean ± standard deviation: mean = (1/K) · Σ_k m^{(k)}; std = sqrt( (1/(K−1)) · Σ_k (m^{(k)} − mean)^2 ). Optional 95% confidence interval: mean ± t_{0.975, K−1} · std / sqrt(K).

B3.2 Across Horizons

Compute metrics separately for each forecast horizonH(e.g.,24, 48, 96, 168). If an overall horizon summary is needed, use an unweighted average across horizons unless a domain-specific weighting is pre-specified.

B3.3 Across Datasets

Primary reporting is per dataset. Optional global summaries may include average ranks (loweris better); ties receive average rank.

B4. Statistical Testing

Primary test: paired t-test

For models A and B, compute per-seed differencesΔ^{(k)}=m_A^{(k)}−m_B^{(k)}andruna two-sided paired t-test (α = 0.05). Check normality of Δ via Shapiro–Wilk; report the paired effect size Cohen’s d_z = mean(Δ) / std(Δ).

Nonparametric fall back: Wilcoxon signed-rank

Use Wilcoxon when normality is rejected or when K<5.Report the rank-biserial effectsizer= (W^{+}−W^{−})/(W^{+}+W^{−}).

Multiple comparisons

When comparing one model to many baselines within a table, control family-wise error using Holm–Bonferroni. False discovery rate control (Benjamini–Hochberg) is also acceptable if explicitly stated.

B5. Data Handling and Fairness Controls

  • Scaling : min–max per feature, fit on the training split; apply to validation and test with the same scaler.
  • Missing values: time‑ aware interpolation; forward/back fill if needed.
  • Splits: time‑ordered 70/15/15; indices identical across models.
  • Windowing: sliding windows with stride 1; document input length L and horizon H per dataset.
  • Determinism: set seeds for Python/Num Py/Py Torch; cudnn. Deterministic = True, cudnn. Benchmark = False.
  • Early stopping : patience = 10 epochs on validation MSE for all models.

B6. Rounding and Table Schema

  • Round error metrics to 3 decimals; time to 1 decimal. Do not round before statistical testing.
  • Use consistent decimal places with in a column for readability.
  • Recommended columns per dataset×horizon: Model;MAE(mean±std);RMSE(mean± std);MAPE(mean± std);NWRMSLE(mean± std);seconds per epoch (mean± std);Peak VRAM (GB, mean ± std).

B7. Reproduction Hooks in Code

  • Per‑run CSVs record: seed, device, epochsrun, total time, seconds per epoch, MAE,RMSE, MAPE, NWRMSLE, peak VRAM.
  • Statistical tests CLI: python-mts_benchmark. stats–csv_a…–csv_b…—metric rmse.
  • Store metrics as full‑precision floats in CSV; round only for presentation.


Premier Science
Publishing Science that inspires