An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (2025)

1. Introduction

Rapid urbanization enhances the importance of subway systems in managing city traffic and improving commuter experiences. Accurate subway passenger flow forecasting is crucial for distributing transportation resources efficiently, reducing congestion, and planning future traffic networks. Subway passenger flow predictions utilize three main approaches: mathematical statistics, machine learning, and deep learning [1]. Our method enhances prediction accuracy by integrating elements from these established techniques. Although diverse prediction techniques aim to improve the accuracy of subway passenger flow forecasts, the complex and dynamic nature of these data presents significant challenges. These techniques struggle to capture deep-level data, impacting the precision and reliability of forecasts.

Conventional mathematical statistical approaches like the Autoregressive Integrated Moving Average (ARIMA) [2] are popular in subway passenger flow forecasting due to their simplicity and ease of use. However, their effectiveness is limited when dealing with complex, non-smooth traffic flow data, often leading to poor outcomes. In response, researchers have turned to machine learning methods such as SVM, decision trees, and random forests [3,4,5], which offer improved data processing and pattern recognition capabilities, enhancing prediction accuracy and leading to advancements in predictive models. Despite their advancements, these machine learning models require extensive feature engineering and struggle with long-term dependencies and seasonal variations in time series data [6], limiting their effectiveness. Concurrently, more sophisticated deep learning techniques like Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) [7,8,9,10] have been broadly utilized for forecasting subway passenger flow. Recent advancements in subway passenger flow prediction have utilized a range of methods, from traditional ARIMA/SARIMA models to more complex deep learning approaches such as LSTM and Transformer-based architectures. These models have shown varying degrees of success in addressing the nonlinear and time-varying nature of passenger flow data. However, many face challenges such as scalability, interpretability, and computational cost [11,12].

To enhance the effective extraction of various types of information from raw traffic data, various methods for transforming time series data into frequency domain components have been widely applied to passenger flow data, road traffic flow data, and other types of time series data. Fourier transform is one such method, and empirical mode decomposition (EMD) proposed by NE Huang and others can decompose time series data into a few Intrinsic Mode Functions (IMFs) and a residue term [13]. Analyzing these components enables the extraction of the deep features of the sequence data at different time scales. Wu et al. employed the empirical mode decomposition and reinforcement learning techniques to predict short-term passenger flow in urban rail transit systems [14]. The model effectively enhanced the accuracy of the predictions. However, the process of EMD suffers from issues such as end effects and mode mixing, which can lead to the omission of valuable information and affect the decomposition accuracy. EEMD, CEEMDAN, and ICEEMDAN have addressed some of the limitations of the original EMD by reducing mode mixing and improving signal decomposition stability [15,16,17]. Sun et al. used a wavelet model to decompose subway passenger flow data into low-frequency and high-frequency components to provide more accurate information extraction, but the type of wavelet limits its accuracy. Classical frequency models, such as Fourier transforms, struggle to capture non-stationary data characteristics. Wavelet transforms, while effective in decomposing non-stationary data, have limitations in selecting an appropriate wavelet basis and in capturing complex nonlinear relationships [18]. While EWT is more versatile due to its ability to customize the transformation based on signal characteristics, VMD was selected in this study because it offers better control over the decomposition of specific frequency components. This makes it highly effective for handling complex and noisy subway passenger flow data, where the accurate separation of frequency components is critical [19,20,21,22,23].

Overcoming the shortcomings of current methods in the research process poses a significant challenge in the area of subway passenger flow forecasting. The latest research trend involves the effective improvement and proper integration of mathematical statistics methods, machine learning, and deep learning approaches, exploiting the advantages of each model while avoiding their risks and drawbacks [24]. The primary benefit of enhancing diverse models for subway passenger flow data is their targeted approach; they can analyze and forecast from various perspectives, considering the “non-stationary” and “seasonal” attributes of subway passenger flow data [25,26,27,28,29]. For instance, the SARIMA model and TimesNet model, although they are mathematical statistics and deep learning models, respectively, can model time series targeting the “seasonality”, “smoothness”, and “intra-period” and “inter-period” characteristics of passenger flow data. In light of the above models’ characteristics, this paper proposes an improved TimesNet subway passenger flow prediction model based on Multi-scale Variational Mode Decomposition (VMD) and the incorporation of a seasonal module. The model decomposes subway passenger flow data via VMD into IMFs of various frequencies, then feeds these decomposed IMFs into the improved TimesNet model with a seasonal module for prediction and, finally, reconstructs the predicted outcome. In our proposed method, judging from the seasonal detection module, the SARIMA component is primarily responsible for capturing the linear trends and seasonal variations in subway passenger flow data, while the improved TimesNet parallel is tasked with capturing and simulating the nonlinear patterns and complex “within-period” and “across-period” dependencies of subway passenger flow data [30]. Such targeted improvements, tailored to the characteristics of subway passenger flow data, can effectively enhance the accuracy of subway passenger flow predictions. Meanwhile, it is important to clarify that the “seasonal variations” referred to in this study specifically address short-term seasonality. In the context of subway passenger flow data, this refers to recurring patterns observed within a single day, such as the morning and evening rush hours, as well as the differences between weekdays and weekends. These short-term cycles are critical for understanding and optimizing subway operations, as they provide valuable insights into passenger flow fluctuations throughout the day and week.

The experiments conducted in this paper selected real swipe card data from the Hangzhou subway in 2019 (a total of 70 million records) and processed them to produce subway passenger flow data with a time step of five minutes. In time series analysis, especially for complex urban transportation systems, it is crucial to account for both linear and nonlinear dynamics. While linear models such as SARIMA are effective at capturing the seasonal and trend components of the data, many real-world systems, including metro passenger flow, exhibit nonlinear and chaotic behavior. Nonlinear dynamics can be identified through various metrics such as the Lyapunov exponent, correlation dimension, and pseudo-attractor analysis [31,32]. In this paper, we verify the nonlinearity and chaotic nature of the dataset time series by analyzing them as described above.

The results demonstrate that our method shows significant performance improvements over other baseline models (e.g., Autoformer, TimesNet, Crossformer) in terms of the Mean Absolute Error (MAE) and Mean Squared Error (MSE) evaluation metrics. In the model rationalization in our method section, the presented results reveal the impact of incorporating the improved TimesNet model, the SARIMA module, and the VMD module on the final model metrics, further validating the rational scientific basis of our prediction model. While the primary focus of this study is on predicting subway passenger flow, the proposed hybrid SARIMA-TimesNet model is designed as a versatile time series prediction framework. Due to its ability to effectively capture both linear and nonlinear components, the model can be applied to a wide range of domains that exhibit complex temporal patterns. For instance, in the energy sector, the model could be used to forecast consumption trends or electricity demand by accounting for both seasonal fluctuations and unexpected surges. In financial markets, it could help predict stock prices, exchange rates, or market indices by modeling inherent volatility and cyclical behavior. Moreover, in meteorology, the model could assist in predicting weather patterns by capturing recurring seasonal shifts and abrupt changes. These examples underscore the adaptability and robustness of the SARIMA-TimesNet model, making it suitable for any field where time series data involve a combination of predictable cycles and irregular variations.

The structure of this paper is outlined as follows: The passenger group prediction model for subway flow data, tailored to their characteristics, will be described in detail in Section 2; Section 3 will compare the experimental results of various models and analyze the functionality of each block in our proposed method; Section 4 will provide a conclusive summary; and the final section will discuss future research directions.

2. Methodology

2.1. Overall Model Framework

Figure 1 shows the model framework described in this paper. As illustrated in Figure 1, the overall operation process of this prediction model follows these 4 steps:

(1) Data Preprocessing: initially, the raw data undergo zero-padding operations and normalization to yield a cleaned dataset, which serves as the original input sequence.

(2) Variational Mode Decomposition (VMD): The cleaned dataset is then subjected to VMD to extract the corresponding Intrinsic Mode Functions (IMFs), obtaining a set of n IMFs. The VMD process involves iterative parameter initialization, optimization, and the update and check mechanism to produce the final IMFs.

(3) IMF Prediction: the obtained IMFs are standardized using the following formula:

I M F i = I M F i I M F i m i n I M F i m a x I M F i m i n

Each of the resulting IMFs is then fed into a seasonality detection module. This module utilizes the autocorrelation function (ACF) and partial autocorrelation function (PACF) to identify IMFs with significant seasonal characteristics [33,34]. The IMFs exhibiting strong seasonality are input into the SARIMA model, while those without strong seasonality, which exhibit stronger nonlinear and non-periodic patterns, are directed to an improved TimesNet model.

(4) Reconstruction: Finally, after each IMF has been predicted through either SARIMA or TimesNet, all the predicted components are reconstructed to generate the final prediction. This reconstruction step is essential because the VMD decomposes the original sequence into several IMFs, and reconstructing these predicted IMFs allows for the restoration of the overall predicted sequence [35]. This comprehensive framework captures both strong seasonal components and nonlinear features in the data, thus effectively utilizing all the available information for accurate prediction outcomes.

In this paper, a dataset constructed from real passenger flow data of the Hangzhou subway is used, and the sliding window technique is applied for forecasting. The data are segmented into five-minute intervals, with a total of 232 time steps in a day (covering the data within the statistics time), which serves as the length of the forecasting window (the sliding window length). Additionally, the sliding window can be adjusted based on experimental requirements. For each day forecasted, the window slides by 232 time steps, and the actual values from the previous day are used to update the dataset.

As shown in Figure 2, we utilized sliding windows of 232 time steps (a full day) to capture passenger flow patterns over different time horizons.

Similarly, the prediction of each component obtained after Variational Mode Decomposition (VMD) also employs the sliding window concept. A detailed description of the VMD module and improved TimesNet model as well as the SARIMA model will be provided in the subsequent sections.

2.2. VMD Module

Variational Mode Decomposition (VMD), first introduced in 2014 by Dragomiretskiy and Zosso, is a technique used to decompose nonlinear, non-stationary data into multiple Intrinsic Mode Functions (IMFs), separating the local characteristics of the data [36]. VMD decomposes the time series data into Intrinsic Mode Functions (IMFs), each representing a different frequency component. High-frequency IMFs capture short-term fluctuations, while low-frequency IMFs reveal long-term trends. This decomposition helps to isolate noise and enhance the detection of underlying periodic patterns, improving the accuracy of passenger flow predictions. The criteria for identifying an IMF can be summarized as follows:

(1) The number of extrema and the number of zero-crossings must differ at most by one throughout the entire sequence;

(2) At any point, the mean of the envelope defined by the local maxima and the envelope defined by the local minima is zero.

VMD serves as a key step in the subway passenger flow prediction model proposed in this paper. The VMD of a signal X can be represented as follows:

X = k = 1 K u k t

where u k t are the IMFs, and k represents the number of Intrinsic Mode Functions.

The purpose of VMD is to decompose complex subway passenger flow data into a finite set of IMFs, which each represent different frequency components. These IMFs are then used to reveal deeper characteristics of the subway passenger flow data. The objective of VMD is to minimize the bandwidth of the IMFs by solving the following optimization problem:

m i n u k k = 1 K t u k t u ^ k t 2

where u ^ k t are the Hilbert transforms of the IMF u ^ k t , and the operator t refers to the partial derivate with respect to time. The Hilbert transform is used for the instantaneous frequency analysis of the decomposed Intrinsic Mode Functions (IMFs), allowing for a more accurate capture of the nonlinear characteristics in the data. The optimization problem aims to minimize the total bandwidth of the IMFs by solving this equation iteratively. This ensures that the IMFs generated are as narrowband as possible, capturing distinct frequency components from the signal.

In VMD, each IMF is modeled as a mode function u k t that is the output of a bandpass filter with a fixed center frequency ω k . The mode function can be expressed as follows:

u k t = A k t c o s ϕ k t

where A k t is the amplitude envelope, and ϕ k t is the instantaneous phase.

VMD is a non-recursive, adaptive, and data-driven method that can decompose nonlinear and non-stationary data into IMFs with limited bandwidth. The center frequencies ω k and bandwidths are iteratively updated by solving the constrained optimization problem to ensure accurate decomposition.

The parameters of these mode functions, including their center frequencies and bandwidths, are determined by solving a constrained optimization problem that minimizes the bandwidth of each IMF and makes their sum as close as possible to the original signal.

As shown in Figure 3, the VMD module in this paper executes the following steps:

Step 1: Data Preprocessing. The original metro passenger flow data are normalized to ensure consistent scaling, reducing the influence of different data scales on the analysis and aiding convergence in subsequent optimization. The preprocessed data X t are input into the VMD module, where they are first transformed into a variational optimization problem. The original time series X t can be represented as follows:

X t = k = 1 K u k t

Step2: Transformation to an Unconstrained Optimization Problem. The augmented Lagrangian function for VMD is formulated to minimize the bandwidth of the Intrinsic Mode Functions (IMFs) while constraining the sum of all IMFs to

L u k , ω k , λ = α k = 1 K t δ t + j t π u k t e j ω k t 2 2 + | | X t k = 1 K u k t | | 2 2 + λ t , X t k = 1 K u k t

Here, α represents the penalty parameter, and λ t is the Lagrange multiplier used to ensure that the sum of all IMFs matches the original time series.

Step 3: Iterative Optimization Process. A dual optimization strategy is used to iteratively update the estimation of modes u k their center frequencies ω k , and the Lagrange multiplier λ until the solution converges.

Step 3.1: Initialization. u k , ω k , λ is initialized, and the iteration counter n = 0 is set.

Step 3.2: Iterative Updates and Convergence Check. For each iteration n , the following updates are performed:

1. Mode u k n + 1 ω is updated:

u k n + 1 ω = f ω i k u i n + 1 ω λ n ω 2 1 + 2 α ω ω k n 2

Here, f ω represents the Fourier transform of the original signal.

2. The center frequency ω k n + 1 is updated:

ω k n + 1 = 0 ω u k n + 1 ω 2 d ω 0 u k n + 1 ω 2 d ω

This update helps estimate the central frequency of each mode to ensure that it accurately captures the frequency characteristics of the IMF.

Step 3.3: The Lagrange multiplier λ n + 1 is updated:

λ n + 1 ω = λ n ω + γ f ω k = 1 K u k n + 1 ω

The parameter γ controls the adjustment of the Lagrange multiplier, aiding in the convergence of the solution. Convergence check: after each iteration, it is checked whether the solution has converged by evaluating the relative change:

k = 1 K u k n + 1 u k n 2 2 u k n 2 2 < ϵ

If the convergence criterion is met, the process terminates; otherwise, n is incremented, and the process continues.

Step 4: Extraction of IMFs. The output consists of K IMFs u k t that collectively represent the original signal. These IMFs are the core components that highlight different intrinsic frequency bands of the time series data, making the signal analysis more interpretable and effective for further modeling.

The VMD algorithm effectively deconstructs complex signals into simpler components that can be analyzed and modeled more easily. In the context of subway passenger flow forecasting, this decomposition allows for a refined analysis of the data’s inherent features, such as daily patterns, peak periods, and anomalies, thereby enhancing the forecasting model’s ability to capture and predict future trends based on historical data.

2.3. Seasonal Detection Module

In the previous section, this paper provided a comprehensive introduction to the VMD process. In the following section, we will elaborate on the seasonal detection model tailored for choosing the appropriate IMFs in subway passenger flow prediction.

Therefore, the model proposed in this article will first input the decomposed sequence data into the “seasonal detection” module to determine if they have strong seasonal characteristics (the specific algorithm in the seasonal detection module is described in detail in the Supplementary Materials).

As illustrated in Figure 4 the seasonality detection module introduced in this study initially receives sequences of Intrinsic Mode Functions (IMFs) as input. It then calculates the autocorrelation function (ACF) to assess the correlation between the time series.

The autocorrelation function (ACF) measures the correlation between observations of a time series that are separated by a specific time lag. The ACF also helps to identify the direct impact of a specific lag on the current observation, independent of the influence of intermediate lags. It provides insights into how past values influence future values.

Series X and its lagged versions are described by the following equation:

A C F k = t = 1 n k X t X ¯ X t + k X ¯ t = 1 n X t X ¯ 2

Equation (10) quantifies the autocorrelation between the time series X and its lagged version. Specifically, it calculates the relationship between the series value X t and the lagged series X t + k for a given lag k . This equation helps determine whether significant seasonality exists, providing a basis for seasonal analysis in time series.

The ACF values obtained are compared with a predetermined threshold, set at 0.9 for this study. This threshold was selected as a balance between detecting significant correlations and minimizing false positives. A higher threshold, such as 0.95, might be overly restrictive and risk missing relevant seasonal patterns, while a lower threshold could introduce noise, falsely identifying non-seasonal fluctuations as significant. By choosing 0.9, we ensure that only meaningful seasonal trends are detected, allowing for reliable analysis. However, we recognize that the choice of threshold may influence the detection of seasonality, and further exploration of optimal threshold values will be an important focus in future research. If the ACF values exceed this threshold, the module proceeds with further analysis by calculating the partial autocorrelation function (PACF) of the series to confirm the presence of seasonality. The partial autocorrelation function (PACF) measures the correlation between an observation and its lagged value while controlling for the correlations at all shorter lags. This helps to identify the direct impact of a specific lag on the current observation, independent of the influence of intermediate lags.

Specifically, the calculation of PACF values can be expressed as follows:

P A C F k = C o v X t , X t + k X t + 1 , , X t + k + 1

Here, C o v denotes the covariance given the intermediate variables. The PACF is computed using statistical software to measure the correlation between the series and its lagged versions, excluding the effects of intermediate lags. If the PACF value surpasses the 90% confidence interval, it is interpreted as an indication of significant seasonality. Should the series not exhibit significant seasonality in either the ACF or PACF assessments, the module will conclude the absence of seasonality in the series.

The framework of the seasonal detection model is illustrated in Figure 4.

Then, adaptively, the appropriate IMFs (Intrinsic Mode Functions) are inputted into the corresponding improved TimesNet and SARIMA (Seasonal Autoregressive Integrated Moving Average) modules.

2.4. Improved TimesNet Module

Subway passenger flow prediction inherently requires time series forecasting. Recent studies have seen a surge of prediction methodologies based on Transformer model variants, such as the Informer model composed of multi-head self-attention mechanisms and the Autoformer model that involves decomposing and then forecasting sequences before recombining data. These models have provided theoretical support for modeling subway passenger flow data prediction in this paper.

In ICLR 2023 (a prominent academic conference in the field of machine learning and AI), the TimesNet model composed of multiple “Timesblock” modules emerged as a state-of-the-art (SOTA) model across various domains of time series modeling. TimesNet is capable of predicting chaotic signals by leveraging its unique design, which focuses on multi-periodicity modeling. Chaotic signals are often characterized by their complexity and sensitivity to initial conditions, making them difficult to predict using conventional time series models. However, TimesNet overcomes this challenge by transforming 1D temporal data into 2D tensors, which capture both intra-period (short-term) and inter-period (long-term) variations. This structure allows the model to disentangle complex patterns and better capture the chaotic nature of the signal.

In the original TimesNet model, the TimesBlock module, which inputs sequence data, simply combines them in parallel to generate the final prediction data. In dealing with different patterns “within the cycle” of the data, the processing of each component is very uniform and standardized, lacking the ability to extract and predict specific characteristics of the data.

The main techniques contained in our I-TimesNet module used in this paper are as follows:

First and foremost, the TimesBlock is the central component of our I-TimesNet model, with its primary function being the identification of multi-periodicity within sequences and the capture of corresponding temporal changes. TimesBlock accomplishes this objective through the following three principal steps: Initially, it mines the data cycles. TimesBlock utilizes the Fast Fourier Transform (FFT) to detect multi-periodicity within the time series. FFT is a robust tool for analyzing the spectral composition of time series data, revealing periodic patterns therein. Specifically, for one-dimensional time series data X 1 D R L × C , where L is the sequence length and C is the number of variables, the periodic amplitude can be expressed by the following formula:

A = A v g A m p F F T X 1 D

In Equation (12), the formula represents the computation of the periodic amplitude using the Fast Fourier Transform (FFT) on the 1D time series data X 1 D . The result of the FFT is averaged to obtain the overall periodic amplitude A, which helps in detecting the multi-periodic patterns within the data. To avoid contamination by high-frequency noise, only the first k amplitude values (frequencies) are selected. The specific selection method is as follows:

f 1 , , f k = a r g T o p k A

f k 1 , , T 2

Here, T o p k refers to the selection of the top k frequencies from the amplitude values. This selection process ensures that only the most significant frequencies, which are relevant for modeling the multi-periodicity, are considered. The combination of the aforementioned formula can be summarized as follows:

A , f 1 , , f k , p 1 , p k = P e r i o d X 1 D

Continuing, the second step is the transformation from 1D to 2D sequences: After detecting the multi-periodicity within the time series, TimesBlock reshapes the one-dimensional time series into a set of two-dimensional tensors based on these periods. The specific procedural formula is as follows:

X 2 D i = r e s h a p e p i , f i P a d d i n g X 1 D , i 1 , , k

In the above equation, Padding involves appending zeros to the end of the time series to make it compatible with the reshaping process. p i and f i denote the number of rows and columns in the transformed 2D tensors, respectively.

Continuing, the following formula can express the input and output process within the TimesBlock module:

X 1 D l = T i m e s B l o c k X 1 D l 1 + X 1 D l 1

Specifically, when l = 0, the following holds true:

X 1 D 0 = E m b e d X 1 D 0

In the TimesBlock module, the softmax function is applied after the FFT for periods to determine the relative importance or weight of each detected frequency (or period) in the input time series. This helps the model to focus on the most relevant periods for further analysis and prediction, ensuring that less significant frequencies do not dominate the output. Specifically, the softmax function ensures that the selected frequencies are normalized, and their values sum to 1, providing a clear probabilistic interpretation of how important each period is in the model’s analysis. This probabilistic weighting aids in the final aggregation and prediction process. The various temporal two-dimensional variants from the k reshaped tensors are captured efficiently utilizing parameter-efficient Inception Blocks within the two-dimensional space. These are then fused based on normalized amplitude values (shown in Figure 5).

Within the 2D tensors, each column contains time points within a single period, while each row involves time points at the same phase across different periods. This allows for the representation of both intra-period and inter-period variations concurrently in the 2D space.

In the Inception Block module of the TimesNet model, the TimesBlock uses a parameter-efficient Inception Block within the 2D space to capture both intra-period and inter-period changes. Using the Inception V1 convolution, Inception Block is a structure from deep learning for visual models that process multi-scale information in parallel, allowing it to effectively capture and learn 2D features.

In metro passenger flow prediction tasks, models for processing time series data need to capture complex trends and periodic changes. The traditional Inception V1 module uses convolutional kernels of different sizes to capture features at various scales, thereby improving the model’s representation ability [37]. However, this design has limitations, such as a large number of parameters that may lead to overfitting and insufficient adaptability to dynamic changes in features. To overcome these issues, dynamic convolution is introduced, which adapts to the features of the input data by learning a weighted combination of different convolutional kernels, offering advantages in adaptive feature extraction and parameter efficiency [38]. Consequently, the design that concatenates dynamic convolution with the Inception V1 module not only inherits the advantages of both but also more effectively processes time series data.

Specifically, the dynamic convolution layer first performs adaptive feature extraction on the input data, adjusting the weights of the convolutional kernels based on the dynamic changes in the data, thereby better capturing the trends and periodic changes in the time series data. Then, the data processed by dynamic convolution pass through the Inception V1 module, which further extracts and fuses features at different scales, enhancing the model’s representation ability. Finally, the GeLU activation function introduces nonlinearity, enhancing the model’s expressive capability, enabling it to capture more complex feature relationships. In metro passenger flow prediction tasks, this design effectively captures the dynamic features of time series data and performs multi-scale feature fusion, thereby improving the accuracy of predictions.

Our improved TimesNet model incorporates dynamic convolution in the Inception Block for enhanced feature extraction. The dynamic convolution can be represented as follows:

D C o n v X = i = 1 N ω i · C o n v X , F i

where X is the input, ω i are the learned weights, C o n v is the convolution operation, and F i are the filters. The dynamic convolution allows our model to adaptively focus on different features by adjusting the weight ω i .

In detail, the dynamic convolution equations relevant to our improved TimesNet model are as follows:

y = g W ^ x · x + b ^ x

where y is the output of the dynamic convolution layer, and g · is an activation function (GeLU). W ^ x and b ^ x are the aggregated convolutional kernel and bias, respectively, which are functions of the input x . Equation (20) is the result of applying the dynamic convolution to the input data X , where the activation function GeLU (Gaussian Error Linear Unit) is applied to the output introducing nonlinearity. The aggregated convolutional kernel W ^ x and bias b ^ x are computed through kernel aggregation.

Furthermore, the kernel aggregation can be expressed as follows:

W ^ x = k = 1 K π k x · W k

b ^ x = k = 1 K π k x · b k

Equations (21) and (22) define the kernel aggregation process, where multiple convolution kernels W k and biases b k are combined based on the attention weights π k x , which are computed for each kernel. K is the number of parallel convolution kernels, and π k x is the attention weight for the k-th kernel, computed based on the input x .

Thus, our brand-new Inception Block can be expressed as follows. Initially, the dynamic convolution layer performs feature extraction on the input data X by adaptively adjusting the weights of the convolutional kernels, capturing the dynamic changes in the time series data, which can be expressed as follows:

Y d y n = i = 1 N ω i · C o n v X , F i

where N is the number of convolutional kernels, ω i are the learned weights, and C o n v X , F i represents the convolution operation using the kernel F i on the input X . Subsequently, the processed data are fed into the Inception V1 module, which further extracts and fuses features at different scales, expressed as follows:

Y i n c = 1 n i = 0 N 1 C o n v 2 i + 1 Y d y n

where C o n v 2 i + 1 Y d y n represents the convolution operation using a kernel of size 2 i + 1 on the output Y d y n of the dynamic convolution layer. Finally, the GeLU activation function introduces a nonlinear transformation, enhancing the model’s expressive capability, expressed as follows:

Y o u t = G e L U Y i n c

where G e L U Y i n c represents the GeLU activation function. The GeLU (Gaussian Error Linear Unit) activation function is a nonlinear function commonly used in modern neural networks, especially in Transformer-based models like BERT. The GeLU is defined as follows:

G e L U x = x · Φ x

where Φ x is the cumulative distribution function of a standard Gaussian distribution. Unlike the commonly used ReLU (Rectified Linear Unit) activation function, the GeLU introduces a smooth, probabilistic element into the activation. It computes the expected value of x under the assumption that x follows a Gaussian distribution, which allows the model to make smoother transitions between activated and non-activated states. This makes the GeLU particularly effective in handling more complex, nonlinear data. Through this concatenated convolution design, our processed model can more effectively extract details in time series data, improving the accuracy of metro passenger flow predictions. The framework of our improved TimesNet can be found in Figure 6.

The system of Equations (19)–(26) corresponds to the operations shown in Figure 6. Dynamic convolution (Equation (19)) is represented by the parallel convolution layers, followed by aggregation (Equations (20)–(22)) and further processing in the Inception V1 block (Equations (23)–(26)).

Following our improved Inception Block in our I-TimesNet model, the subsequent steps can be described as follows:

The output from the Inception Block, denoted as X ^ 2 D l , i , is reshaped to match the original sequence length. This is achieved using a truncation method, represented as follows:

X ^ 2 D l , i = T r u n c r e s h a p e 1 , p i × f i X ^ 2 D l , i , i 1 , , k

where p i and f i are the dimensions of the reshaped tensor, and T r u n c is the truncation operation to adjust the sequence back to its original length.

The reshaped sequence is then weighted using amplitude values derived from a softmax operation:

A ^ f 1 l i , , A ^ f k l i = S o f t m a x A f 1 l 1 , , A f k l 1

The final one-dimensional sequence, X 1 D l , is obtained by aggregating the weighted sequences:

X 1 D l = i = 1 k A ^ f 1 l i × X ^ 1 D l , i

The processed sequence data undergo a final stage of transformation before prediction. Initially, the data are reshaped to align with the requirements of the model’s subsequent components, ensuring the preservation of its temporal structure. This reshaping is achieved through a truncation method that adjusts the sequence length back to its original dimensions. Following this, the sequence is weighted using amplitude values derived from a softmax operation, which balances the contributions of different components within the sequence. This weighted sequence is then subjected to a softmax function to generate the predicted outcomes. This combined step of weighting and prediction encapsulates the final processing stage, enabling our improved TimesNet model to produce accurate predictions by effectively leveraging the temporal features of the metro passenger flow data.

2.5. SARIMA Module

In the context of our proposed method, the SARIMA module can be defined by a mathematical formula that represents its components: autoregressive (AR), differencing (I), moving average (MA), and the seasonal components of each. The generic form of a SARIMA model is typically denoted as follows:

S A R I M A p , d , q × P , D , Q s

where p is the number of autoregressive terms, d is the number of non-seasonal differences needed for stationarity, q is the number of lagged forecast errors in the prediction equation, P is the number of seasonal autoregressive terms, D is the number of seasonal differences, Q is the number of seasonal moving average terms, and s is the number of periods in a season. Even in seasonally inactive scenarios, Equation (30) provides a general framework that ensures the model is robust across various datasets, including those where seasonality is not a major factor. It serves as a baseline for comparing more complex, seasonally active systems.

Specifically, within our model, the formula for the SARIMA module can be expressed as follows:

ϕ B 1 B d 1 B s D X t = θ B Θ B s Z t

where ϕ B represents the non-seasonal autoregressive operator, θ B presents the non-seasonal moving average operator, Θ B s is the seasonal moving average operator, X t is the time series being modeled, and Z t is the white noise series.

The above equation can be further decomposed into the following:

ϕ B 1 B d 1 B s D X t = θ B ε t

Θ B s ε t = Z t

For the SARIMA module proposed in this paper, its establishment mainly includes the following steps:

(1) Model identification: According to the training dataset, the appropriate parameters for the corresponding IMFs are automatically determined. This process is automatically completed by fitting various combinations of ARIMA models for different orders p , d , q and seasonal orders P , D , Q s . The models are evaluated based on a given information criterion (the Akaike Information Criterion), and the model with the lowest criterion value is selected.

(2) Using the parameters determined in the previous step, future values of the time series X are predicted with the SARIMA model. The prediction is conducted in a rolling forecast manner, that is, predicting a certain number of future steps, with the model being updated with new data at each step.

(3) The model is continuously predicted and updated upon the receipt of new data. This iterative process continues until predictions have been made for the entire forecast range.

We also considered several boundary conditions during the application of the SARIMA model to ensure the accuracy and relevance of the predictions. These boundary conditions include the following:

(1)

Stationarity: The SARIMA model assumes that the time series data are stationary. To satisfy this condition, we employed the auto arima function, which automatically determines the differencing required to achieve stationarity if the raw data exhibit non-stationary behavior.

(2)

Autocorrelation: The autoregressive component of the SARIMA model requires that the time series data exhibit significant autocorrelation. Through the model selection process, auto arima tests various parameter combinations to ensure that the chosen model has sufficient autocorrelation for effective forecasting.

(3)

Short-term seasonality: We explicitly modeled short-term seasonality by specifying a weekly cycle (m = 7). This reflects the recurring patterns in subway passenger flow that occur within a week, such as weekday versus weekend differences. The seasonal component of the SARIMA model captures these variations to improve forecast accuracy.

In summary, in this paper, the original subway passenger flow data are first decomposed into multiple Intrinsic Mode Functions (IMFs) using the VMD method. In the seasonal detection module, IMFs exhibiting significant seasonal variations are selectively input into the SARIMA model for forecasting. Meanwhile, the remaining IMFs are input into the Timesblock module and eventually reshaped to obtain the final forecast results. The processing of the original data by VMD allows for the deeper information of the subway passenger flow data to be further mined; the introduction of the SARIMA module enables the forecasting model to better handle the IMFs of subway passenger flow data with seasonal variations, thereby enhancing the accuracy of the forecast. The improved TimesNet model will enhance its predictive performance, allowing for the more precise forecasting of subway passenger flow operations.

2.6. Evaluation Metrics

In order to assess the performance of the proposed subway passenger flow forecasting model, this paper employs two standard forecasting accuracy metrics: Mean Absolute Error (MAE) and Root Mean Squared Error (RMSE). These two metrics can comprehensively and intuitively reflect the accuracy and stability of the forecasting model.

2.6.1. Mean Absolute Error (MAE)

The Mean Absolute Error (MAE) is the average of the absolute errors between the actual observations and the forecasted values. The formula for its calculation is as follows:

M A E = 1 n i = 1 n y i y i ^

where y i is the actual observation value of the i instance, y i ^ is the corresponding forecasted value, and n is the number of forecasting periods. The smaller the value of the MAE, the higher the accuracy of the forecast model.

2.6.2. Root Mean Squared Error (RMSE)

The Root Mean Squared Error (RMSE) is the square root of the average of the squared differences between actual observations and forecasted values. The formula for its calculation is as follows:

RMSE = 1 n i = 1 n y i y i ^ 2

where y i is the actual observation value of the i instance, y i ^ is the corresponding forecasted value, and n is the number of forecasting periods. The smaller the value of the RMSE, the better the performance of the forecast model, and since the RMSE gives greater weight to larger errors, it can also reflect the stability of the forecast model.

By comparing and contrasting the magnitudes of these evaluation metrics, this paper can comprehensively assess the performance of the proposed model. After comparing with other baseline forecasting models, the superiority of the model proposed in this paper can be further highlighted.

3. Experimental Analysis

3.1. Dataset Introduction

The dataset used in this paper is derived from the Hangzhou subway passenger flow swipe card record data publicly available on the Alibaba Cloud Tianchi platform. The dataset contains data from 1 January 2019 to 25 January 2019, totaling 25 days, with a total data volume reaching 70 million records. From these, data not including the impact of transfer stations were selected, and using a time step of five minutes, a total of 5816 subway passenger flow datasets from a single station were obtained. The original data were divided into a training set and a test set, and the specific dataset table obtained is shown in Table 1(the data have been normalized).

3.2. Nonlinear Time Series Analysis of Datasets

In order to analyze the nonlinear and chaotic characteristics of the time series data of metro passenger flow data, this study delves into their autocorrelation function plot, lag relationship, Lyapunov exponent, etc., and comes up with the following conclusions. We analyze the autocorrelation function (ACF) of the metro passenger flow data, as shown in Figure 7. The ACF plot shows a high correlation at the initial lag, which gradually decays as the lag increases and tends to zero around lag 30, indicating strong short-term dependencies but an absence of long-term persistence. This observation suggests that the system’s dynamics are primarily governed by short-term influences, which aligns with the characteristics of chaotic systems where predictability declines rapidly over time. To quantify the chaotic nature of the system, we computed the largest Lyapunov exponent, which was found to be 2.18. A positive Lyapunov exponent indicates the presence of chaos, as it signifies sensitive dependence on initial conditions. The value of 2.18 suggests moderate chaotic behavior, implying that while there are complex, unpredictable dynamics, the system retains some structure that can be used for short-term predictions. The lag plot (Figure 8) demonstrates the relationship between the current value of the data and the previous value, forming a pseudo-attractor structure, further supporting the nonlinear and deterministic chaotic character of the data. The presence of pseudo-attractors indicates that the system is governed by complex interactions, and traditional linear models struggle to adequately portray these characteristics. The above results reveal the chaotic nature of the metro passenger flow dataset. Therefore, it is necessary to construct a nonlinear, chaotic prediction model around the metro passenger flow datasets.

3.3. VMD Results

In the application of VMD processing, this study anticipated the extraction of 12 IMFs, with the bandwidth constraint factor alpha set to 800, indicating a medium-strength bandwidth setting. The direct current component was set to 0 (meaning that the decomposition modes would not explicitly contain a constant term), and the frequency initialization parameter init was set to 100 (representing a uniform initialization strategy). The detailed results of the IMFs obtained following these settings are shown in the table below.

As shown in Table 2, the mean variance and average values of the decomposed data are significantly reduced compared to the original data. Meanwhile, the IMFs resulting from the VMD exhibit increasingly smooth and simplified periodic characteristics with the progression of decomposition. The IMFs (Intrinsic Mode Functions) represent different frequency components of the original time series data. Higher-indexed IMFs capture finer, higher-frequency patterns, while lower-indexed IMFs capture broader, low-frequency trends. IMF1 corresponds to the highest-frequency component, which captures the finest details or rapid fluctuations in the time series; IMF2-IMF12 gradually capture lower-frequency components, representing more long-term patterns in the data; the average of IMFs represents the aggregation of all IMFs to approximate the original time series data.

At the same time, the spectral images after decomposition are shown in Figure 8: it is evident from the figure that modal decomposition reveals the trends and patterns of passenger flow on different time scales, with each mode reflecting different characteristics of passenger flow variations.

As shown in Figure 9, the X-axis represents time steps, and the Y-axis shows the amplitude of the signal. The IMFs are sorted by frequency, with IMF1 representing the highest-frequency component and IMF12 capturing the lowest-frequency component. As you move from IMF1 to IMF12, you will notice that the oscillations (frequency) decrease, representing the long-term trends in the data. Each IMF reveals different aspects of the underlying data: IMF1-IMF4 generally capture short-term fluctuations, while IMF5 and above capture lower-frequency variations and broader trends.

In summary, the modes decomposed effectively extract the impact of various factors on subway passenger flow.

3.4. Proposed Model Results

3.4.1. Comparative Model Experimental Results

To evaluate the performance of the subway passenger flow prediction model based on the proposed improved TimesNet, we chose the MAE (Mean Absolute Error) and RMSE (Root Mean Square Error) as performance metrics. The processed data from our model were compared with other existing mainstream prediction models. Table 3 presents the performance comparison results of our model and other models in the task of subway passenger flow prediction.

All the models we use for comparison in this study (including TimesNet) are advanced deep learning models designed specifically for the task of time series forecasting. These models are powerful benchmark models in the field of time series forecasting with architectures optimized to capture complex temporal patterns such as long-term dependencies and chaotic behavior.

The comparisons we make in our results focus on models with similar goals, i.e., predicting complex time series data to ensure that baseline model comparisons are reasonable. At the same time, we ensure that the baseline models are relevant and competitive in terms of architecture, input data, and complexity of the tasks addressed.

To ensure a fair comparison, we used consistent datasets, preprocessing steps, and evaluation metrics for all models. Thus, the performance differences reflect the model’s capabilities, not the settings.

It can be seen from Table 3 that our proposed improved TimesNet model outperforms other models in both the MAE and RMSE performance metrics. This demonstrates the superiority of our model in predicting subway passenger flow. Additionally, in terms of fitting effect, as shown in Figure 10 (the rest of the result in the Supplementary Materials), our work shows more accurate prediction results compared to other baseline models.

3.4.2. Model Rationalization Experiment in Our Method

In order to rationalize the contribution of each module within our proposed method to the overall performance, this study conducted a Model Rationalization Experiment considering the use of VMD, the presence or exclusive use of the SARIMA module, and the use of the TimesNet model. The results are shown in Table 4.

From the data in the above table, we obtained the following three points:

(1) The proposed model significantly outperforms the SARIMA-only model, achieving a 49.61% improvement in the MAE and a 52.34% improvement in the RMSE. This indicates that the integration of additional components beyond SARIMA provides a substantial enhancement in prediction accuracy.

(2) When comparing the proposed model to the I-TimesNet model without the SARIMA module, we observe a 49.64% improvement in the MAE and a 52.37% improvement in the RMSE. This suggests that the SARIMA module plays a key role in enhancing the time series modeling capabilities of the I-TimesNet architecture.

(3) Without VMD, the proposed method’s performance drops by 19.63% in the MAE and 24.15% in the RMSE, indicating that VMD is crucial for effectively handling the different frequency components in the data. In order to further investigate the impact of each module on different time periods, we refined the time nodes and categorized the multiple time periods of the day into morning peak hours, evening peak hours, and normal hours. Specifically, the morning peak hour was defined as 7 a.m.–9 a.m., the evening peak hour was defined as 5 p.m.–7 p.m., and the rest of the day was considered as the regular hours. The results of the ablation experiments can be further obtained from the detailed ablation experiment table, as shown in Table 5 (best performance is marked in red, and second-best performance is underlined).

As shown in Table 5, the SARIMA model performs relatively poorly for the morning and evening peak hours, with MAE values of 1.4445 and 1.1110, respectively, which may be attributed to the model’s limited ability to capture sudden changes in passenger flow, which are common during peak hours. In contrast, the SARIMA model performs better during normal hours with a lower MAE value of 0.3353, but its linear nature may not be able to account for the nonlinear and sudden fluctuations that are typical of peak hours.

When using the I-TimesNet model alone, it has limited predictive power during the peak hours, with MAE values of 0.5666 and 0.8192 for the morning and evening peaks, respectively; however, when combined with the VMD (VMD-I_TimesNet), its performance improves dramatically, with the MAE values dropping to 0.6389 and 0.4078, respectively, which suggests that the VMD module helps the I-TimesNet model to better capture sudden changes in passenger flow.

Similarly, after adding the VMD module to the SARIMA model (VMD-SARIMA), the prediction accuracies of all time periods are significantly improved. During the morning and evening peaks, the MAE values drop to 0.4207 and 0.4303, respectively, which indicates that the VMD effectively improves the model performance by extracting the relevant sequence information.

Overall, the combination of the SARIMA model, I-TimesNet model, and VMD consistently outperforms the other models in almost all time periods except for the evening peak, with the lowest MAE value of 0.3626 in the morning peak, which suggests that the integration of these components can effectively capture the complex patterns of the metro passenger flow in the peak hours.

4. Conclusions

This paper presents a novel subway passenger flow prediction model based on the Variational Mode Decomposition (VMD) method, an improved TimesNet model, and the SARIMA model. The model addresses the shortcomings of traditional mathematical approaches, which struggle to capture the complex features of subway passenger flow, especially outliers and seasonal components. Additionally, the model proposes solutions to the challenges deep learning models face when predicting seasonal components and handling off-peak flow data.

The model’s structure starts with the VMD method, which decomposes the passenger flow data into multiple frequency components. After the seasonality judgment module, components are directed either to the SARIMA module or the TimesBlock for separate prediction. Finally, the predictions are synthesized to reconstruct the overall passenger flow. This hybrid approach utilizes real-world data from the Hangzhou subway system to validate the model’s performance.

In terms of the results, the proposed model achieves strong prediction performance with MAE and RMSE values of 0.2253 and 0.3178, respectively. Compared to baseline models such as Informer, Crossformer, and Autoformer, our model provides the best prediction accuracy and stability. The proposed approach demonstrates superior performance, especially during peak and off-peak hours, making it a valuable tool for urban transportation planning and optimization. In conclusion, this hybrid subway passenger flow prediction model offers a practical solution for urban transportation systems, contributing to the intelligent and data-driven management of urban mobility.

5. Future Research Prospects

Subway passenger flow prediction is inherently challenging due to the influence of multiple dynamic factors. Building on the limitations discussed earlier, future research will prioritize expanding the scope and depth of the model to enhance its predictive accuracy and robustness.

A key focus will be integrating additional data sources, such as weather conditions, holidays, large-scale events, and detailed subway network information, to capture a more comprehensive range of influencing factors. These diverse data sources will be incorporated as covariates in the model, enabling it to account for dynamic environmental changes and improving its capacity to predict passenger flow effectively.

Another limitation of the current study is the fixed use of a 5 min time interval for data aggregation. While this interval was chosen to balance data granularity and noise, it is critical to explore the impact of alternative time intervals, such as 4 or 6 min, on model performance. This exploration will determine whether varying time intervals can reveal additional insights into system dynamics, particularly in terms of chaotic behavior and short-term seasonal patterns. Specifically, we aim to investigate how these intervals affect the detection of chaotic attractors and prediction accuracy during critical periods like rush hours.

Given that our study addresses a chaotic system characterized by a positive Lyapunov exponent, it is essential to clarify how accurate predictions are achieved despite the inherent unpredictability of chaos. A positive Lyapunov exponent indicates a sensitive dependence on initial conditions, leading to the exponential divergence of trajectories and making long-term predictions particularly challenging. However, the model’s strong short-term predictive performance can be attributed to the limited predictability retained by chaotic systems over short time frames. To leverage this, we employ machine learning models capable of capturing short-term patterns and correlations in the data before chaotic divergence becomes significant. By utilizing local stability, the model extracts meaningful yet transient insights from the system’s dynamic behavior.

Future research will further refine the conditions under which these short-term predictions remain reliable and extend the predictive horizon wherever possible. This will involve enhancing preprocessing techniques, refining decomposition methods, and developing chaos-aware model architectures capable of managing chaotic divergence more effectively. Additionally, hybrid approaches that combine traditional statistical models with machine learning techniques will be explored to deepen the understanding and mitigate the limitations imposed by chaotic dynamics.

The computational complexity of the proposed hybrid model presents another challenge. To address this, future research will focus on optimizing the model architecture and exploring more efficient machine learning algorithms to reduce complexity without compromising performance. Furthermore, the current dataset, limited to a single city’s subway network, raises concerns about the model’s generalizability. To validate the approach across diverse urban contexts, we plan to extend the model to metro networks with different structural and passenger flow characteristics, incorporating data from major systems such as those in Beijing, London, and Shanghai.

Ultimately, the goal is to develop a comprehensive, dynamic prediction framework that integrates various models. By leveraging the strengths of these models, we aim to provide a holistic approach to managing urban metro systems, enabling operators to make real-time operational adjustments and informed long-term strategic decisions. This integrated approach is expected to enhance resource allocation, optimize passenger transfers, and reduce congestion, contributing to more efficient, reliable, and sustainable urban transportation systems.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/app15062874/s1. The Supplementary Materials consist of two parts. The first part, “Seasonal Detection Module Algorithm Pseudocode”, provides detailed information about the specific algorithms and presents the basic settings of the method proposed in this paper. The second part, “Supplementary Figures for Fitting Results”, comprehensively displays the fitting effects of various baseline models compared to the model proposed in this paper, using the remaining parts of the test dataset.

Author Contributions

Conceptualization, T.Z.; methodology, T.Z.; software, T.Z.; validation, T.Z.; formal analysis, T.Z.; investigation, S.T.; resources, T.Z.; data curation, T.Z.; writing—original draft preparation, T.Z.; writing—review and editing, T.Z.; visualization, H.K., H.S. and P.L.; supervision, S.T. and L.Z.; project administration, T.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by following projects: the China National Key R&D Program (Grant No. 2021YFB1715700) and the R&D Program of Beijing Municipal Education Commission (KM202111417003).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Since we used publicly available data, the following is a link to the data used in the experiment: https://tianchi.aliyun.com/competition/entrance/231712/information.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Medina-Salgado, B.; Sánchez-DelaCruz, E.; Pozos-Parra, P.; Sierra, J.E. Urban traffic flow prediction techniques: A review. Sustain. Comput. Inform. Syst. 2022, 35, 100739. [Google Scholar] [CrossRef]
  2. Feng, S.; Cai, G. Passenger flow forecast of metro station based on the ARIMA model. In Proceedings of the 2015 International Conference on Electrical and Information Technologies for Rail Transportation; Springer: Berlin/Heidelberg, Germany, 2016; pp. 463–470. [Google Scholar]
  3. Sun, Y.; Leng, B.; Guan, W. A novel wavelet-SVM short-time passenger flow prediction in Beijing subway system. Neurocomputing 2015, 166, 109–121. [Google Scholar] [CrossRef]
  4. Wu, W.; Xia, Y.; Jin, W. Predicting bus passenger flow and prioritizing influential factors using multi-source data: Scaled stacking gradient boosting decision trees. IEEE Trans. Intell. Transp. Syst. 2020, 22, 2510–2523. [Google Scholar] [CrossRef]
  5. Niu, Q.; Wang, G.; Liu, B.; Zhang, R.; Lei, J.; Wang, H.; Liu, M. Selection and prediction of metro station sites based on spatial data and random forest: A study of Lanzhou, China. Sci. Rep. 2023, 13, 22542. [Google Scholar] [CrossRef]
  6. Han, Y.; Wang, S.; Ren, Y.; Wang, C.; Gao, P.; Chen, G. Predicting station-level short-term passenger flow in a citywide metro network using spatiotemporal graph convolutional neural networks. ISPRS Int. J. Geo.-Inf. 2019, 8, 243. [Google Scholar] [CrossRef]
  7. Wang, J.; Zhang, Y.; Wei, Y.; Hu, Y.; Piao, X.; Yin, B. Metro passenger flow prediction via dynamic hypergraph convolution networks. IEEE Trans. Intell. Transp. Syst. 2021, 22, 7891–7903. [Google Scholar] [CrossRef]
  8. Lv, S.; Wang, K.; Yang, H.; Wang, P. An origin–destination passenger flow prediction system based on convolutional neural network and passenger source-based attention mechanism. Expert Syst. Appl. 2024, 238, 121989. [Google Scholar] [CrossRef]
  9. Liu, D.; Wu, Z.; Sun, S. Study on Subway passenger flow prediction based on deep recurrent neural network. Multimed. Tools Appl. 2022, 81, 18979–18992. [Google Scholar] [CrossRef]
  10. Yang, J.; Dong, X.; Jin, S. Metro passenger flow prediction model using attention-based neural network. IEEE Access 2020, 8, 30953–30959. [Google Scholar] [CrossRef]
  11. Sun, Y.; Liao, K. A hybrid model for metro passengers flow prediction. Syst. Sci. Control. Eng. 2023, 11, 2191632. [Google Scholar] [CrossRef]
  12. Lai, Y.; Wang, Y.; Xu, X.; Easa, S.M.; Zhou, X. Hybrid models of subway passenger flow prediction based on convolutional neural network. IET Intell. Transp. Syst. 2023, 17, 716–729. [Google Scholar] [CrossRef]
  13. Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, H.H.; Zheng, Q.; Yen, N.C.; Tung, C.C.; Liu, H.H. The empirical mode decomposition and the Hilbert spectrum for non-linear and non-stationary time series analysis. Proceedings of the Royal Society of London. Ser. A Math. Phys. Eng. Sci. 1998, 454, 903–995. [Google Scholar] [CrossRef]
  14. Wu, J.X.; He, D.Q.; Li, X.W.; He, S.Q.; Li, Q.; Ren, C.H. A Time Series Decomposition and Reinforcement Learning Ensemble Method for Short-Term Passenger Flow Prediction in Urban Rail Transit. Urban Rail Transit 2023, 9, 323–351. [Google Scholar] [CrossRef]
  15. Rhif, M.; Ben Abbes, A.; Farah, I.R.; Martínez, B.; Sang, Y. Wavelet Transform Application for/in Non-Stationary Time-Series Analysis: A Review. Appl. Sci. 2019, 9, 1345. [Google Scholar] [CrossRef]
  16. Kang, Y.-Z.; Yao, Y.-K.; Dong, R.-L.; Jia, Y.-S.; Xie, Q.-M.; Wang, J.-N. Improved complete ensemble empirical mode decomposition with adaptive noise and composite multiscale permutation entropy for denoising blast vibration signal. Heliyon 2024, 10, e37339. [Google Scholar] [CrossRef]
  17. Poongadan, S.; Lineesh, M.C. Non-linear Time Series Prediction using Improved CEEMDAN, SVD and LSTM. Neural Process. Lett. 2024, 56, 164. [Google Scholar] [CrossRef]
  18. Zhu, J.; Xu, W.X.; Jin, H.T.; Sun, H. Prediction of urban rail traffic flow based on multiply Wavelet-ARIMA model. In Proceedings of the 7th International Conference on Green Intelligent Transportation System and Safety, Nanjing, China, 1–4 July 2016; Springer: Singapore, 2018; pp. 561–576. [Google Scholar]
  19. Zhang, Y.; Zhu, C.; Wang, Q. LightGBM-based model for metro passenger volume forecasting. IET Intell. Transp. Syst. 2020, 14, 1815–1823. [Google Scholar] [CrossRef]
  20. Li, H.; Jin, K.; Sun, S.; Jia, X.; Li, Y. Metro passenger flow forecasting through multi-source time-series fusion: An ensemble deep learning approach. Appl. Soft Comput. 2022, 120, 108644. [Google Scholar] [CrossRef]
  21. Kim, E.J.; Park, H.C.; Kho, S.Y.; Kim, D.K. A hybrid approach based on variational mode decomposition for analyzing and predicting urban travel speed. J. Adv. Transp. 2019, 2019, 3958127. [Google Scholar] [CrossRef]
  22. Yang, H.; Cheng, Y.; Li, G. A new traffic flow prediction model based on cosine similarity variational mode decomposition, extreme learning machine and iterative error compensation strategy. Eng. Appl. Artif. Intell. 2022, 115, 105234. [Google Scholar] [CrossRef]
  23. Elouaham, S.; Nassiri, B.; Dliou, A.; Zougagh, H.; El Kamoun, N.; El Khadiri, K.; Said, S. Combination time frequency and empirical wavelet transform methods for removal of composite noise in EMG signals. TELKOMNIKA Telecommun. Comput. Control. 2023, 21, 1373–1381. [Google Scholar] [CrossRef]
  24. Hao, S.; Lee, D.-H.; Zhao, D. Sequence to sequence learning with attention mechanism for short-term passenger flow prediction in large-scale metro system. Transp. Res. Part C Emerg. Technol. 2019, 107, 287–300. [Google Scholar] [CrossRef]
  25. Milenković, M.; Švadlenka, L.; Melichar, V.; Bojović, N.; Avramović, Z. SARIMA modelling approach for railway passenger flow forecasting. Transport 2018, 33, 1113–1120. [Google Scholar] [CrossRef]
  26. Wu, H.; Hu, T.; Liu, Y. Timesnet: Temporal 2D-variation modeling for general time series analysis. arXiv 2022, arXiv:2210.02186. [Google Scholar]
  27. Zuo, C.; Wang, J.; Liu, M.; Deng, S.; Wang, Q. An Ensemble Framework for Short-Term Load Forecasting Based on TimesNet and TCN. Energies 2023, 16, 5330. [Google Scholar] [CrossRef]
  28. Zhao, H.; Huang, X.; Xiao, Z.; Shi, H. Weekly solar irradiation forecasting method based on ICCEMDAN and TimesNet networks. Renew. Energy 2023, 220, 119706. [Google Scholar] [CrossRef]
  29. Liu, J.X.; Jiang, R.; Zhu, D.; Zhao, J.D. Short-Term Subway Inbound Passenger Flow Prediction Based on AFC Data and PSO-LSTM Optimized Model. Urban Rail Transit 2022, 8, 56–66. [Google Scholar] [CrossRef]
  30. Huang, Y.; Zhou, C.; Cui, K.; Lu, X. A multi-agent reinforcement learning framework for optimizing financial trading strategies based on TimesNet. Expert Syst. Appl. 2023, 237, 121502. [Google Scholar] [CrossRef]
  31. Hu, S.; Chen, J.; Zhang, W.; Liu, G.; Chang, X. Graph transformer embedded deep learning for short-term passenger flow prediction in urban rail transit systems: A multi-gate mixture-of-experts model. Inf. Sci. 2024, 679, 121095. [Google Scholar] [CrossRef]
  32. Rosenstein, M.T.; Collins, J.J.; De Luca, C.J. A practical method for calculating largest Lyapunov exponents from small data sets. Phys. D Nonlinear Phenom. 1993, 65, 117–134. [Google Scholar] [CrossRef]
  33. Sirisha, U.M.; Belavagi, M.C.; Attigeri, G. Profit prediction using ARIMA, SARIMA and LSTM models in time series forecasting: A Comparison. IEEE Access 2022, 10, 124715–124727. [Google Scholar] [CrossRef]
  34. Rezaie-Balf, M.; Zahmatkesh, Z.; Kim, S. Forecasting the monthly incidence of scarlet fever in Chongqing, China using the SARIMA model. Epidemiol. Infect. 2022, 150, e90. [Google Scholar]
  35. Rezaie-Balf, M.; Zahmatkesh, Z.; Kim, S. Soft computing techniques for rainfall-runoff simulation: Local non-parametric paradigm vs. model classification methods. Water Resour. Manag. 2017, 31, 3843–3865. [Google Scholar] [CrossRef]
  36. Dragomiretskiy, K.; Zosso, D. Variational mode decomposition. IEEE Trans. Signal Process 2014, 62, 531–544. [Google Scholar] [CrossRef]
  37. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 7–12 June 2015; IEEE Computer Society: Piscataway, NJ, USA, 2015; pp. 1–9. [Google Scholar]
  38. Chen, Y.; Dai, X.; Liu, M.; Chen, D.; Yuan, L.; Liu, Z. Dynamic convolution: Attention over convolution kernels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; IEEE Computer Society: Piscataway, NJ, USA, 2020; pp. 11030–11039. [Google Scholar]

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (1)

Figure 1. The overall model framework.

Figure 1. The overall model framework.

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (2)

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (3)

Figure 2. Diagram of sliding window.

Figure 2. Diagram of sliding window.

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (4)

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (5)

Figure 3. The procedure of VMD.

Figure 3. The procedure of VMD.

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (6)

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (7)

Figure 4. Our seasonal detection model.

Figure 4. Our seasonal detection model.

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (8)

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (9)

Figure 5. Diagram of the TimesNet model.

Figure 5. Diagram of the TimesNet model.

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (10)

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (11)

Figure 6. Our improved TimesNet model.

Figure 6. Our improved TimesNet model.

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (12)

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (13)

Figure 7. Autocorrelation of the dataset.

Figure 7. Autocorrelation of the dataset.

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (14)

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (15)

Figure 8. Lag plot of the dataset.

Figure 8. Lag plot of the dataset.

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (16)

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (17)

Figure 9. VMD results.

Figure 9. VMD results.

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (18)

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (19)

Figure 10. January 25th forecasting performance fit graph.

Figure 10. January 25th forecasting performance fit graph.

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (20)

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (21)

Table 1. Dataset division information.

Table 1. Dataset division information.

DatasetSizeMeanVariance
Training4652−0.00571.0114
Testing11640.02290.9535

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (22)

Table 2. Components of each IMF in this study.

Table 2. Components of each IMF in this study.

IndexMeanMinMaxVariance
IMF11.71 × 10−17−1.35501.61550.5204
IMF2−2.29 × 10−18−0.86131.11680.1755
IMF32.29 × 10−18−0.67000.75140.0823
IMF41.03 × 10−18−0.57830.65020.0224
IMF55.92 × 10−19−0.37670.38460.0054
IMF68.26× 10−19−0.56080.53670.0076
IMF7−3.33 × 10−19−0.47270.46970.0055
IMF81.77 × 10−19−0.35630.36210.0045
IMF93.38 × 10−19−0.27850.29570.0034
IMF10−3.23 × 10−19−0.35150.35070.0035
IMF11−1.21 × 10−19−0.24960.24780.0035
IMF122.86 × 10−19−0.2890.27660.0032
Average of IMFs7.64 × 10−19−0.09560.31770.0069
Original data4.52 × 10−17−1.14653.81191.0000

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (23)

Table 3. Module performance evaluation index compared to baseline.

Table 3. Module performance evaluation index compared to baseline.

MODELMAERMSE
Informer0.79901.0194
Crossformer0.81101.1263
Autoformer0.81391.1215
Reformer0.68730.9475
PatchTST0.38560.5953
iTransformer0.66760.8793
TimesNet0.33990.4818
FiLM0.51800.7434
Proposed0.22530.3178

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (24)

Table 4. Model Rationalization Experiment result.

Table 4. Model Rationalization Experiment result.

PartMAERMSE
Proposed0.22530.3178
SARIMA0.44790.6667
I_TimesNet0.28030.4181
VMD-I_TimesNet0.31060.4320
VMD-SARIMA0.28060.3822

An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (25)

Table 5. Model rationalization result in detailed time period.

Table 5. Model rationalization result in detailed time period.

Time PeriodMorning Rush HourEvening Rush HourNormal Time
MetricsMAERMSEMAERMSEMAERMSE
Proposed0.36260.45130.54000.67740.18970.2522
Only SARIMA1.44451.68751.11101.3110.33530.3927
Only I_TimesNet0.56660.68470.81920.99220.28650.3368
VMD-I_TimesNet0.63890.69980.40780.50460.28190.3903
VMD-SARIMA0.42070.58250.43030.53580.25930.3001

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.


© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
An Enhanced TimesNet-SARIMA Model for Predicting Outbound Subway Passenger Flow with Decomposition Techniques (2025)
Top Articles
Latest Posts
Recommended Articles
Article information

Author: Sen. Ignacio Ratke

Last Updated:

Views: 5816

Rating: 4.6 / 5 (56 voted)

Reviews: 95% of readers found this page helpful

Author information

Name: Sen. Ignacio Ratke

Birthday: 1999-05-27

Address: Apt. 171 8116 Bailey Via, Roberthaven, GA 58289

Phone: +2585395768220

Job: Lead Liaison

Hobby: Lockpicking, LARPing, Lego building, Lapidary, Macrame, Book restoration, Bodybuilding

Introduction: My name is Sen. Ignacio Ratke, I am a adventurous, zealous, outstanding, agreeable, precious, excited, gifted person who loves writing and wants to share my knowledge and understanding with you.