[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
Physical Layer Security in RIS-NOMA-Assisted IoV Systems with Uncertain RIS Deployment
Previous Article in Journal
A Cybernetic Delay Analysis of the Energy–Economy–Emission Nexus in India via a Bistage Operational Amplifier Network
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Spatio-Temporal Hypergraph Convolutional Network for Traffic Flow Forecasting

1
School of Computer Science, Hubei University of Technology, Wuhan 430068, China
2
Department of Automation, Lublin University of Technology, Nadbystrzycka 38D, 20-618 Lublin, Poland
3
Vilnius Gediminas Technical University, Sauletekio al. 11, LT-10223 Vilnius, Lithuania
4
Department of Automatics and Metrology, Lublin University of Technology, Nadbystrzycka 38D, 20-618 Lublin, Poland
5
Department of Measuring-Information Technologies, Lviv Polytechnic National University, 79013 Lviv, Ukraine
*
Authors to whom correspondence should be addressed.
Electronics 2024, 13(22), 4435; https://doi.org/10.3390/electronics13224435
Submission received: 7 October 2024 / Revised: 31 October 2024 / Accepted: 4 November 2024 / Published: 12 November 2024

Abstract

:
Graph convolutional networks (GCN) are an important research method for intelligent transportation systems (ITS), but they also face the challenge of how to describe the complex spatio-temporal relationships between traffic objects (nodes) more effectively. Although most predictive models are designed based on graph convolutional structures and have achieved effective results, they have certain limitations in describing the high-order relationships between real data. The emergence of hypergraphs breaks this limitation. A dynamic spatio-temporal hypergraph convolutional network (DSTHGCN) model is proposed in this paper. It models the dynamic characteristics of traffic flow graph nodes and the hyperedge features of hypergraphs simultaneously, achieving collaborative convolution between graph convolution and hypergraph convolution (HGCN). On this basis, a hyperedge outlier removal mechanism (HOR) is introduced during the process of node information propagation to hyper-edges, effectively removing outliers and optimizing the hypergraph structure while reducing complexity. Through in-depth experimental analysis on real-world datasets, this method has better performance compared to other methods.

1. Introduction

Driven by continuous economic growth, self-driving has become the primary mode of transportation, with traffic congestion becoming increasingly severe. Consequently, there is an urgent need for an efficient intelligent transportation system to enhance road usage efficiency, optimize traffic flow, and meet the rising demand for mobility. In real traffic networks, excessive traffic flow may lead to congestion or accidental traffic accidents; therefore, accurate traffic flow forecasting can not only save travel time but also reduce the probability of occurring dangerous accidents.
Domestic and foreign scholars have conducted in-depth research on traffic flow data, fully considering external factors such as time, period, and weather, and have achieved a series of prominent research results in traffic flow prediction accuracy. We can categorize these methods into statistical prediction methods, machine learning-based prediction methods, neural network-based prediction methods, and graph convolution-based prediction methods. Among them, prediction methods based on statistics rely on the statistical analysis of historical traffic data, which makes it difficult to capture the complex dynamic characteristics of traffic flow data. Machine learning-based prediction methods predict future traffic flow by learning from historical traffic flow data using machine learning algorithms. As the volume of traffic flow data samples continues to increase, the process of finding the optimal parameters may significantly reduce the efficiency and accuracy of the model learning. In recent years, with the continuous development of deep learning, prediction methods based on neural networks and graph convolution have overcome the shortcomings of the first two methods and can better handle the complexity and nonlinearity of traffic flow data.
Among neural network-based prediction methods, the most commonly used are convolutional neural networks (CNN) [1,2], recurrent neural networks (RNN) [3], and long short-term memory (LSTM) [4,5,6]. Although these methods have significant advantages in feature extraction and prediction accuracy, the process of training neural networks requires a significant amount of computational resources and storage space, presenting numerous issues; hence, neural network-based prediction methods face certain difficulties.
Urban traffic networks are highly complex. Although we can convert data into grid image data, we can only capture local spatial features within the network. Capturing the global spatial and temporal features of the network is still a significant challenge [7,8,9,10]. To effectively capture the spatio-temporal features within the transportation network, researchers have combined graph convolutional networks with recurrent network models in their predictive methods to capture spatio-temporal features. Li et al. [11] considered the relationship of spatio-temporal features and proposed the STGCN model, which captures the road network’s spatio-temporal features by stacking spatio-temporal convolutional blocks using graph convolution and gated linear units. Based on the STGCN model, Guo et al. [12] introduced the ASTGCN model incorporating spatio-temporal attention mechanisms to capture features at different levels. The adaptive graph convolutional recurrent network (AGCR) is different from the models previously discussed [13]. It employs an innovative data-adaptive graph construction mechanism that dynamically generates graphs based on the inherent structure and characteristics of the data. Li et al. [14] proposed the DGCRN model, which integrates dynamic features of (velocity and timestamps) to construct a dynamic adjacency matrix. There are also many researchers [15,16,17,18,19] who combine graph neural networks with other models to construct spatio-temporal models for traffic flow forecasting.
Although these methods have achieved commendable research outcomes, the complexity and long-term spatio-temporal dependencies within traffic networks make spatio-temporal modeling of traffic flow more challenging. Based on the predictive outcomes of previous researchers, models based on graph convolutional networks have significantly enhanced forecasting accuracy. However, models based on graph structures still have some issues: (1) most graph-structure-based models are highly dependent on the node-to-node graph structure, and the static graph structure generated by the physical connections of the road network cannot explore the potential dynamic spatial correlations in traffic flow data or adaptively establish nonlinear temporal correlations; (2) simple graph structures only describe the pairwise interactions between entities and are incapable of depicting higher-order relationships among multiple entities. In real traffic networks, the traffic flow on a road segment is not only affected by the directly connected road segments but may also be influenced by other indirectly connected road segments.
For instance, Figure 1 illustrates two traffic network structures, each reflecting the complex relationships between multiple nodes of traffic flow input and output. Simple graph structures are incapable of effectively representing these complex relationships.
The major contributions are as follows:
(1)
Proposed a new dynamic spatio-temporal hypergraph convolutional network (DSTHGCN) framework for traffic flow prediction, which effectively captures more accurate traffic flow information by combining traffic flow graphs and hypergraphs through collaborative convolution.
(2)
Unlike most traffic flow prediction models, this paper proposes a feature extraction module that can extract dynamic features of nodes and edges separately. These features are used to update the hyperedges and graph node information in the hypergraph, thereby revealing more complex underlying relationships in the dynamic traffic system.
(3)
Introduced a hyperedge outlier removal mechanism to identify and remove outliers in the hyperedges, thereby optimizing the hypergraph structure and better capturing the higher-order relationships within the data.

2. Related Work

In recent research, traffic flow prediction models developed using graph neural network technology have shown a rapid pace of development. These models have transitioned from traditional methods such as time series analysis and regression models to more sophisticated deep learning approaches, and their predictive accuracy has greatly improved. For instance, [12] simulates the complex spatio-temporal correlations between different locations through spatial attention. Ref. [20] proposes a novel learning component for graph structures that optimally obtains the graph’s adjacency matrix from a micro and macro viewpoint and enhances the model’s predictive performance through a two-stage training process. Ref. [21] has developed an innovative graph learning module to capture the hidden relationships between variables. Furthermore, the study introduces a comprehensive framework that not only models multivariate time series data effectively but also learns and infers the graphical structure of the data. Ref. [22] proposed the TSGDC model that combines the graph neural networks and transformers into the spatio-temporal graph diffusion convolutional network, which can better understand and utilize the complex spatio-temporal relationships in traffic data.
Although these graph-based models have achieved good performance, they have neglected the interactions between two or more nodes in practical situations and cannot model the higher-order relationships among nodes. To break through the limitations of traditional graph structures, more and more scholars have begun to consider hypergraph structures [23].
In addition, hypergraph convolutional networks are increasingly becoming a research hotspot due to their unique advantages in handling complex higher-order interaction relationships. These networks can more naturally express and learn multi-element relationships within data, demonstrating strong application potential across various fields. As research progresses, the design of models based on hypergraph convolutional networks continues to innovate and develop. For instance, [24] proposes a spatio-temporal hyper-relation to fully model complex local spatio-temporal relationships. Ref. [25] incorporates hypergraph convolution and attention mechanisms as end-to-end trainable operators in graph neural networks to acquire in-depth embeddings of higher-order graph-structured data. Ref. [26] explores hidden hyperedges by employing a local hypergraph attention mechanism across different time spans and then optimizes the hypergraph using a global attention mechanism. In [27], a novel high-order multi-modal and multi-type data association modeling framework called HGNN+ is introduced. This framework is designed to capture and learn the optimal representation of data through a single hypergraph structure, thereby enhancing the efficiency and accuracy of data processing and analysis. Ref. [28] proposes a multi-level graph neural network capable of handling hypergraphs and paired graphs. Ref. [29] proposed a new multi-task hypergraph learning framework, which is composed of a primary task and an associated task. The latent features of hidden layers are shared between the tasks through a feature compression unit during training. In recent years, research on hypergraph convolutional networks has revealed some new trends. For instance, [30] proposed a unified method to capture local correlations and cross-network isomorphism by using the K-means clustering algorithm and the connectivity characteristics of the physical road network. Ref. [31] proposed a new network structure that can handle both temporal and spatial information, introducing a temporal hypergraph to capture action dependencies in time series. Ref. [32] proposed a new network structure that can effectively model dynamic higher-order relationships between nodes across multiple time scales, thereby enhancing the model’s predictive capabilities. Ref. [33] proposed a new multi-task spatio-temporal network model MT-STNet, which adopts an encoder-decoder structure, builds the encoder and decoder through spatio-temporal blocks, and integrates information about the physical structure into model of the spatio-temporal dependencies of the highway network. Ref. [34] proposed the Res-HGCN model, which combines residual blocks with hypergraph convolutional networks for a multi-sensor data-driven fault diagnosis method.
In the existing hypergraph model structure, this paper proposes a new hypergraph network structure that fully considers the spatio-temporal features of transportation networks. We propose to extract more comprehensive spatio-temporal feature information using co-convolution of traffic flow graphs and hypergraphs. Additionally, we introduce a hyperedge outlier removal mechanism to optimize the hypergraph structure, thereby enhancing the model’s predictive performance.

3. Methodology

3.1. Preliminaries

In the traffic flow prediction task, historical data are analyzed, and a prediction model is constructed to predict the traffic flow data of a specific road section for a future period of time, including time series analysis and spatial distribution characteristics of the traffic flow. In this section, we define the graph, hypergraph, GCN, and HGCN related to the predictive model.
1.
Graph: The traffic network G is defined as G = ( V , E , A ) , where V ( | V | = N ) and E correspond to the number of nodes (the number of sensors) and the set of edges in the traffic network, respectively. A R N × N denotes the adjacency matrix of the graph, indicating the proximity between two nodes. The left hand side half of Figure 2 shows a simple graph converted into the adjacency matrix, where the elements of the matrix indicate whether there is an edge between the nodes in the graph. If there is an edge between node i and j, then the element in the i-th row and j-th column of the matrix is 1; otherwise, it is 0. Given the adjacency matrix and historical information over T time steps, learning a function f that uses historical data from T time steps to predict traffic information for the next T time steps:
[ X t + 1 , X t + 2 , X t + T ] = f [ A , ( X t T + 1 , X t T + 2 , , X ) ] .
2.
GCN: This is a graph-based convolution that captures the interrelationships between nodes through graph convolution operations, thereby updating node features. We can describe the convolution process as follows:
G C N ( X ) = σ ( A ^ X Θ ) .
where A ^ R N × N refers to the adjacency matrix after normalization, Θ represents the learnable parameters, and σ ( · ) denotes the nonlinear activation function.
3.
Hypergraph: The hypergraph is defined as G h = ( V h , E h , H ) , where V h and E h denote the node set and the set of nodes within hyperedges, respectively; H R E × N is the incidence matrix of ϵ E h and ϵ E h . When ϵ is associated with v i , H i ϵ = 1 ; otherwise, it is 0. Then, the degrees of nodes and hyperedges are represented by D v R N × N and D e R E × E , respectively. The diagonal matrix W corresponds to hyperedge weights. The right hand side half of Figure 2 shows a hypergraph converted into a matrix. The rows of the matrix represent nodes, and the columns represent edges. If a node is connected to an edge, the corresponding element is 1; otherwise, it is 0. The specific representation is shown as follows:
D v = ϵ = 1 E W ϵ ϵ H i ϵ ,
D e = i = 1 N H i ϵ .
4.
HGCN: The convolution process of a hypergraph is defined as follows:
H G C N ( X h ) = σ ( H W H T X h Θ ) .
As it lacks a constrained spectral radius, we have normalized it, and the reformulation is as follows:
H G C N ( X h ) = σ ( D v 1 H W D E 1 H T X h Θ ) .

3.2. Framework of DSTHGCN

The structure of the model is shown in Figure 3. It consists of three main modules: the input layer, output layer, and dynamic spatio-temporal convolutional layer. The input part transforms traffic data into a high-dimensional representation. The output part performs skip connection operations for each output result. The most important dynamic spatio-temporal convolutional layer consists of three parts: dynamic feature extraction, dynamic hypergraph convolution (DHGCN), and dynamic graph convolution (DGCN). The following sections will provide a comprehensive discussion of the content related to these three parts.

3.2.1. Dynamic Feature Extraction

In order to capture more complex and dynamic correlations of traffic data, this paper designs a dynamic feature extraction module to extract the dynamic features of the graph nodes of traffic flow as inputs for the DHGCN, thereby updating the corresponding edges of the hypergraph. Similarly, by extracting the edge features in the hypergraph as inputs for the DGCN, the node features of the traffic flow graph are updated. Currently, modules similar to feature extraction rarely appear in other models. It has the function of fusing information between the traffic flow graph and the hypergraph, thereby enabling the model to achieve more effective prediction results.
(1)
Extraction of dynamic features of traffic flow graph nodes
First, the aggregate operation is used to aggregate node features along the temporal dimension, followed by the use of diffusion GCN to capture spatial correlations, with specific operations as shown below:
X d = A g g r e g a t e ( X ) R N × F ,
G C N d ( X d ) = σ ( n = 0 N P f n X d Θ n , f + P b n X d Θ n , b ) ,
where X represents the node features of the traffic flow graph; σ ( · ) denotes a nonlinear activation function like ReLU; P f = A / r o w s u m ( A ) and P b = A T / r o w s u m ( A T ) represent the forward and backward transition matrices of the directed graph, respectively; P and N represent the state transition matrix and diffusion coefficient, respectively; and Θ signifies learnable parameters. The output of G C N d ( X d ) will be used as the input for DHGCN.
(2)
Extraction of dynamic features of the graph edges of traffic flow
First, the extraction operation is used to obtain node-related information of the directed edges in the traffic flow graph as initial edge features; then, convolutional operations are used to update the edge features.
X h = [ ( W 1 X ) [ i n d s r c , : ] ; ( W 2 X ) [ i n d d s t , : ] ] R E × 2 F ,
X h d = C o n v 1 × 1 ( X h ) ,
where W 1 , W 2 represent trainable parameters; ( ) [ i n d , : ] is a tensor composed of slices of i n d s r c and i n d d s t , selected from the source and target nodes of the directed edges; and [ , ] performs the concatenation operation.
The process of implementation of dynamic edge features is as follows:
W h = D v 1 H W a d p D e 1 H T R E × E ,
H G C N Θ d ( X h d ) = σ ( n = 0 N 1 W h n X h d Θ n ) R E ,
where W a d p represents the diagonal matrix of the adaptively learned hyperedge weight vectors. The output of H G C N Θ d ( X h d ) will be used as the input for DGCN.

3.2.2. Dynamic Graph Convolution

First, the temporal convolutional network (TCN) combines one-dimensional causal convolution and dilated convolution to extract temporal correlations within traffic data along the time dimension.
F ( X ) = tanh ( Θ 1 X + b ) s i g m o i d ( Θ 1 X + c )
where Θ 1 , Θ 2 , b and c are trainable parameters; denotes the product of elements; tanh ( ) is an activation function acting as a filter; and s i g m o i d ( ) is an activation function that controls the ratio of information transfer.
Subsequently, the proposed DGCN is used to capture the spatio-temporal features of the traffic data, with the output of H G C N Θ d ( X h d ) serving as the input. The specific implementation process is shown in Equation (14):
D G C N ( X ) = σ n = 0 N 1 ( P f D f ) n X Θ n , f + ( P b D b ) n X Θ n , b + A a d p n X Θ n , a d p
where D f = r e s h a p e ( H G C N Θ f d ( X h d ) ) and D b = r e s h a p e ( H G C N Θ b d ( X h d ) ) T utilize r e s h a p e ( · ) to reshape H G C N Θ d ( X h d ) into a sparse matrix; Θ f , Θ b , and Θ a d p are all trainable parameters.

3.2.3. Dynamic Hypergraph Convolution

Dynamic hypergraph convolution utilizes the TCN from (13) to extract the features along the temporal dimension of traffic flow data and then employs DHGCN to capture the spatio-temporal features.
W h d = D v 1 H D w D e 1 H T R E × E ,
D H G C N ( X h ) = σ ( n = 0 N 1 W h d n X h Θ n ) ,
where D w represents the diagonalization of output results of G C N d ( X d ) , and Θ is a trainable parameter.
Finally, the node features of the traffic flow graph from (14) are connected with the hypernode features of the hypergraph from (16) to obtain the output of the spatio-temporal convolutional layer. DSTHGCN can be summarized as Algorithm 1.
Algorithm 1: Training Algorithm for DSTHGCN.
Input: Node features of the traffic flow graph  X ;
Output: New node features  X n e w  generated by the collaborative convolutions on the traffic flow graph and its dual hypergraph;
1: The dynamic features X d G C N d ( X d ) of the nodes are obtained through Equation (8);
2: The dynamic edge features X h d H G C N Θ d ( X h d ) are obtained through Equation (12);
3: Update X F ( X ) by Equation (13);
4: Update X D G C N ( X ) by Equation (14);
5: Transform the node features of the traffic flow graph into the hyperedge features of the hypergraph;
6: Update X h F ( X ) by Equation (13);
7: Update X h D H G C N ( X h ) by Equation (16);
8: Map the hyperedge features from Equation (16) back to the node features of the traffic flow graph to obtain X ;
9: Finally, a concatenation operation is performed to obtain X n e w = [ X ; X ] ;

3.2.4. Hyperedge Outlier Removal Mechanism

Hypergraph theory offers an innovative approach to modeling traffic flow, effectively capturing complex higher-order relationships in real datasets by transferring information from nodes to hyperedges and then feedback from hyperedges to nodes. In this model, each hyperedge can connect multiple nodes, thereby containing richer traffic flow information, which allows the model to more accurately reflect the dynamic characteristics of the traffic network. However, in practical applications, outlier nodes in traffic networks are inevitable. These outlier nodes may have a negative impact on the model’s training process, thereby causing deviations in the prediction results. To address this issue, we propose a mechanism that removes outliers on hyperedges during the process of information propagation between nodes and hyperedges. The core of this mechanism lies in dynamically adjusting the structure of the hypergraph to reduce the impact of outlier nodes on model performance. Specifically, the mechanism identifies and removes outlier nodes that have a negative impact on model training and prediction performance by learning embeddings. This dynamic adjustment strategy effectively removes outliers and enhances the model’s ability to handle noise and irrelevant information.
As shown in Figure 4, e i and v i j represent hyperedges and nodes, respectively. We believe that important nodes should have a high degree of similarity with their associated hyperedges. Therefore, this paper uses cosine similarity (Equation (17)) to calculate the similarity between each node v i j and its corresponding hyperedge e i . Then, based on the value of similarity to identify outliers, we sort by the magnitude of the similarity value and consider the bottom 10% of nodes as outliers, removing these outlier nodes from the hypergraph. As shown in Figure 4, if nodes v i 1 and v i 8 have low similarity with hyperedge e i , and their similarity ranking is in the bottom 10% of all nodes, then these two nodes will be recognized as outliers. Subsequently, these outlier nodes will be removed from hyperedge e i . This mechanism helps to optimize the structure of the hypergraph because it reduces the complexity of the model by removing outlier nodes, thereby improving the model’s efficiency and accuracy.
cos i j = x i y i x i y i ,
where x i is the representation of node v i j ; y i is the representation of hyperedge e i ; and | | refers to the norm operation on vectors. Typically, we assess the influence of a node on a hyperedge based on the similarity between the node and hyperedge features; the higher the similarity, the greater the influence; and vice versa: the lower the similarity, the less influence.

3.2.5. Evaluation Metrics

In this study, we chose the Huber loss function as the primary loss function, with the aim to enhance the accuracy and reliability of model predictions. The Huber loss function is a very effective statistical tool. In comparison to the commonly used the mean squared error (MSE) loss function, the Huber loss function can more robustly handle outliers in the data. We assume that y and y ^ represent the true values and predicted values, respectively. The Huber loss function is defined as follows:
L δ ( y , y ^ ) = 1 2 ( y y ^ ) 2 , f o r   | y y ^ | δ δ | y y ^ | 1 2 δ 2 , o t h e r w i s e .
The mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE) will be used for the overall evaluation of DSTHGCN and the baselines. The optimal values for these three metrics are obtained after running 200 epochs on the validation data.
M A E = 1 | Ω | i Ω | y i y ^ i |
R M S E = 1 | Ω | i Ω ( y i y ^ i ) 2
M A P E = 1 | Ω | i Ω | y i y ^ i y i | × 100 %

4. Experiments

4.1. Datasets

To comprehensively validate the performance and accuracy of the model, we employed a multidimensional evaluation approach. We selected two widely recognized real-world traffic datasets, PeMSD4 [4] and PeMSD8 [35], to conduct experiments on the model proposed in this paper and the baseline models. These datasets include information on traffic flow, velocity, and occupancy rates. This paper mainly focuses on the study of traffic flow [13,36]. The PeMSD4 dataset, commencing on 1 January 2018, consists of traffic flow data collected continuously over 59 days by 307 sensors, with data gathered every five minutes. The PeMSD8 dataset, starting from 1 July 2016, comprises traffic flow data collected continuously over 62 days by 170 sensors, with data being gathered every five minutes. More detailed information about these two datasets is shown in Table 1. To gain a deeper understanding of the geographical context of the data and the interrelationships among the data, Figure 5 illustrates the sensor distribution maps of the PeMSD4 and PeMSD8 datasets. These maps not only intuitively display the specific geographic locations where the data are collected but also delineate the layout of the traffic network in detail.

4.2. Experimental Setups

This paper divides the entire dataset into three parts based on chronological order: training set, validation set, and test set. The ratio of this division is 6:2:2, with the training set accounting for 60% and the validation set and test set each accounting for 20%. The traffic flow prediction task takes historical traffic flow information from 12 time steps as input to forecast the traffic flow for the next 12 time steps, with each time step being five minutes long. This experiment was conducted on NVIDIA GeForce RTX 3090 GPU using the Adam optimizer for training. The initial learning rate is set to 0.001, the batch size is 16, the number of stacked layers in the model is B = 2, and the number of filters is F = 40.

4.3. Baselines

We compared the method proposed in this paper with the following six traffic flow prediction methods. These baseline models are either widely used or are new models proposed in the last two years. A brief description of each baseline model is provided as follows:
(1)
HA: The historical average method uses the mean traffic flow from the same time point over the past few days or weeks as the predicted value.
(2)
ARIMA: Autoregressive integrated moving average model, a classic method for time series forecasting.
(3)
FC-LSTM: This model is a combination of fully connected layers and Long Short-Term Memory networks, often used for time series problems.
(4)
ASTGCN (r) [12]: This model captures the dynamic spatio-temporal features of traffic flow data using a spatio-temporal attention mechanism while considering the periodicity of the spatio-temporal network.
(5)
STGODE [37]: This model utilizes ordinary differential equations on graphs to model the dynamics of spatio-temporal data.
(6)
Graph WaveNet [38]: It combines graph convolution with causal convolutional networks to capture the spatio-temporal dependencies in the data.
(7)
DSTAGNN [39]: This model employs an improved multi-head attention mechanism to capture the dynamic spatial dependencies between nodes.

4.4. Experimental Results and Analysis

We compared our model with the seven abovementioned baseline models on two real-life datasets. To ensure the fairness of the experiment, the parameters of each baseline were adjusted.
As can be seen from Table 2, the model we proposed significantly outperforms the existing baseline models in terms of predictive accuracy. DSTHGCN outperforms all seven baseline models in all three-evaluation metrics. On the PeMSD4 dataset, DSTHGCN improved the evaluation metrics of MAE, RMSE, and MAPE by 10.6%, 6.8%, and 9.74% respectively, with respect to the most comparable baseline model STGODE. On the PeMSD8 dataset, the evaluation metrics of MAE, RMSE, and MAPE improved by 11.26%, 7.98%, and 14.19%, respectively, with respect to the baseline model STGODE.
Table 3 provides an in-depth analysis of the predictive performance of the five models, such as ASTGCN, Graph WaveNet, STGODE, DSTAGNN, and DSTHGCN at different time steps (3, 6, and 12 time steps). Through this analysis, we can gain a more comprehensive understanding of the capabilities and limitations of these models in handling prediction tasks across various time spans. It can be observed that the longer time of prediction, the greater the forecasting error. We can see from Table 3 that the DSTHGCN model outperforms the other five baseline models at every forecasting horizon. To more intuitively see the outstanding forecasting level of the models, a line chart is used to display the forecasting level of each time step for these five models. As shown in Figure 6, the DSTHGCN model performs better on all metrics (MAE, RMSE, MAPE), and its error growth trend is slower with the increase in time steps. The DSTAGNN and STGODE models perform second only to DSTHGCN, while ASTGCN and Graph WaveNet models perform poorly, especially the Graph WaveNet model, mainly because as time progresses, the model needs to consider more variables and uncertainties, leading to increased complexity in forecasting.
To conduct an in-depth analysis of the model’s performance, we selected four nodes from the PeMSD4 dataset and visualized the traffic flow data for these nodes over a day. We also conducted a comparative analysis with ground truth and Graph WaveNet, and the comparative curve chart is shown in Figure 7. The forecast curves for the DSTHGCN model on all four nodes are very close to the actual data, demonstrating its high accuracy in traffic flow prediction. The forecast curves of the Graph WaveNet model are close to the actual data for most time steps, but there are some deviations at key points (such as traffic flow peaks). Overall, the DSTHGCN model outperforms the Graph WaveNet model in terms of prediction performance on these nodes.

4.5. Ablation Study

The experimental results mentioned above demonstrate that the DSTHGCN model outperforms most baseline models across various forecasting levels, with this significant advantage attributed to the innovative integration of hypergraph convolution technology and hyperedge outlier processing mechanisms. Therefore, this section focuses on verifying whether these two modules have a positive effect on the final prediction results. We have developed two variants, DSTHGCN-1 and DSTHGCN-2, and conducted ablation studies on these variants to explore the contributions of their different components to the overall forecasting capability. Under the same datasets and evaluation metrics, the ablation study results of the DSTHGCN model are shown in Table 4:
In DSTHGCN-1, the dynamic hypergraph convolution module is removed. Then, TCN is combined with DGCN to capture the complex relationships within the traffic network. It is evident that the accuracy of DSTHGCN-1, which lacks the hypergraph convolutional module, significantly decreased. This is because hypergraphs are an extension of traditional graphs, capable of representing non-pairwise relationships between nodes, thereby simulating the intrinsic connections between high-order data and capturing more complex traffic flow information, which, in turn, enhances the prediction accuracy. In DSTHGCN-2, the hyperedge outlier removal mechanism is removed, and the dynamic graph convolution and dynamic hypergraph convolution modules work in tandem to capture node features through a cooperative convolution approach. As shown in Table 4, the accuracy of DSTHGCN-2, which lacks the hyperedge outlier removal mechanism, also decreased. This is because although hypergraphs contain abundant traffic flow information, there may be outliers in these data that could negatively impact the model’s predictive performance. The mechanism of hyperedge outlier removal is capable of identifying and removing these outliers, optimizing the hypergraph structure, and enhancing prediction accuracy.

4.6. Hyperparameter Sensitivity Analysis

Parameters are very important for the performance of the model. By analyzing the model under different hyperparameter settings, it can help us better understand the model’s sensitivity to different parameters and how these hyperparameters affect the model’s learning process and final performance. Therefore, this section analyzes different values of two important hyperparameters, B and F.
B: In graph convolutional networks, the number of stacked layers of the model is crucial for capturing complex relationships between nodes. In this paper, we explored the range of values for B, from 1 to 5, and incrementally increased the number of layers in experiments to determine the value of B that optimizes model performance. As shown in Figure 8a, when B = 3, the model’s MAE metric achieved the best performance on both datasets. When B is greater than 3, the excessive number of stacked layers causes a waste of computational resources, so the performance decreases. Therefore, the optimal number of stacked layers is B = 3.
F: As for the number of filters, it has a significant impact on the effectiveness of feature extraction. In experiments, we start with an initial value of F at 8 and incrementally increase its value, doubling the previous value each time, to find the optimal number of filters F for enhancing the model’s feature extraction capability. As shown in Figure 8b, when F = 40, the model’s MAE metric achieved the best performance on both datasets. When F is greater than 40, the performance decreases due to the excessive number of filters leading to overfitting. Therefore, the optimal number of filters is F = 40.

4.7. Computational Time Analysis

To better assess the complexity of the DSTHGCN model, Table 5 compares the training and inference times of the baseline ASTGCN, Graph WaveNet, STGODE, DSTAGNN, and DSTHGCN models.
Table 5 clearly shows that the STGODE and DSTACNN models have the shortest training and inference times on the PeMSD4 dataset. The DSTHGCN model has the second shortest training time after STGODE. On the PeMSD4 dataset, Graph WaveNet and DSTHGCN have the shortest training and inference times, respectively, while the inference time of the DSTHGCN model is second only to Graph WaveNet. Due to the complexity of the DSTHGCN model, although it is not optimal overall, it is within an acceptable range.
From the table, we can also observe that an increase in dataset complexity leads to an extension of training and inference time, which implies that processing large-scale datasets in real-world applications may consume more computational resources, potentially limiting the feasibility of the model in practical applications. To overcome this challenge, we plan to introduce advanced attention mechanisms and combine them with optimized dynamic graph structures in future research. Attention mechanisms can help the model focus on key features in the data, thereby improving processing efficiency. At the same time, the optimization of dynamic graph structures can further reduce the computational burden on the model during training and inference. By employing these strategies, our goal is to achieve a reduction in the model’s demand for computational resources while maintaining existing predictive accuracy.

5. Conclusions

This paper proposes a dynamic spatio-temporal hypergraph attention convolutional network that combines graph convolutional networks and hypergraph convolutional networks to form a predictive model capable of capturing more comprehensive spatio-temporal features in the task of traffic flow prediction. Furthermore, we have introduced a hyperedge outlier removal mechanism that can effectively eliminate abnormal nodes in the hypergraph, thereby optimizing the hypergraph structure and reducing the complexity of the model. To thoroughly assess the model’s performance, we meticulously designed an extensive series of experiments that covered two representative real-world datasets, and the results strongly demonstrated the model’s excellent capabilities.
At the same time, we also consider that the paper may have the following limitations: there may be data compatibility and integration issues in the process of integrating with existing traffic management systems, insufficient computational resources, and issues with the model’s generalization ability. When dealing with large-scale data and needing to expand the prediction range, the model’s response time and resource consumption requirements are higher. Therefore, to reduce the impact of these limitations on the model, our future research directions will include the following aspects:
(1) Develop advanced data fusion and preprocessing techniques to solve problems in data integration, ensuring data consistency and accuracy. (2) Incorporate attention mechanisms to help the model better focus on key time steps and spatial locations, thereby improving the accuracy of long-term predictions. (3) Explore methods to decompose long-term forecasting tasks into multiple short-term predictions to reduce the burden on individual forecasts. (4) Optimize the model structure, such as using shallower network layers or optimized graph convolution methods, which can not only reduce the demand for computational resources but also reduce the model’s latency.

Author Contributions

Conceptualization, Z.Y. and K.P.; methodology, J.M.; software, H.W.; validation, Z.Y., K.P., and N.H.; formal analysis, H.W.; investigation, N.H.; resources, J.S.; data curation, J.S.; writing—original draft preparation, Z.Y., J.S., and K.P.; writing—review and editing, H.W. and J.M.; visualization, H.W.; supervision, N.H.; funding acquisition, J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was financed as part of the Lublin University of Technology projects FD-24/IM-5/087 and FD-24/EE-2/801. This research was supported by the National Natural Science Foundation of China (Grant Nos. 62376089, 62302153, 62302154), the Key Research and Development Program of Hubei Province, China (Grant No. 2023BEB024), the Young and Middle-aged Scientific and Technological Innovation Team Plan in Higher Education Institutions in Hubei Province, China (Grant No. T2023007), and the National Natural Science Foundation of China (Grant No. U23A20318).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors would like to thank Orest Kochan for his assistance in preparing this paper.

Conflicts of Interest

Authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  2. Tian, H.; Su, J.; Kochan, O. Research on Traffic Flow Prediction Based on ISMA-CNN-GRU Model. In Proceedings of the COLINS, Kharkiv, Ukraine, 20–21 April 2023; Volume 1, pp. 40–50. [Google Scholar]
  3. Cho, K.; van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN Encoder-Decoder for statistical machine translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1724–1734. [Google Scholar]
  4. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  5. Yao, Y.; Ye, Z.; Bai, W.; Kochan, O.; Mokhun, S. Time Series Prediction Based on LSTM and Modified Hybrid Breeding Optimization Algorithm. In Proceedings of the 2023 13th International Conference on Advanced Computer Information Technologies (ACIT), Wrocław, Poland, 21–23 September 2023; pp. 584–590. [Google Scholar] [CrossRef]
  6. Sun, L.; Qin, H.; Przystupa, K.; Majka, M.; Kochan, O. Individualized Short-Term Electric Load Forecasting Using Data-Driven Meta-Heuristic Method Based on LSTM Network. Sensors 2022, 22, 7900. [Google Scholar] [CrossRef] [PubMed]
  7. Guo, S.; Lin, Y.; Wan, H.; Li, X.; Cong, G. Learning dynamics and heterogeneity of spatial-temporal graph data for traffic forecasting. IEEE Trans. Knowl. Data Eng. 2021, 34, 5415–5428. [Google Scholar] [CrossRef]
  8. Zheng, C.; Fan, X.; Pan, S.; Jin, H.; Peng, Z.; Wu, Z.; Wang, C.; Yu, P.S. Spatio-temporal joint graph convolutional networks for traffic forecasting. IEEE Trans. Knowl. Data Eng. 2023, 36, 372–385. [Google Scholar] [CrossRef]
  9. Jiang, R.; Cai, Z.; Wang, Z.; Yang, C.; Fan, Z.; Chen, Q.; Tsubouchi, K.; Song, X.; Shibasaki, R. DeepCrowd: A deep model for large-scale citywide crowd density and flow prediction. IEEE Trans. Knowl. Data Eng. 2021, 35, 276–290. [Google Scholar] [CrossRef]
  10. Gu, J.; Zhou, Q.; Yang, J.; Liu, Y.; Zhuang, F.; Zhao, Y.; Xiong, H. Exploiting interpretable patterns for flow prediction in dockless bike sharing systems. IEEE Trans. Knowl. Data Eng. 2020, 34, 640–652. [Google Scholar] [CrossRef]
  11. Yu, B.; Yin, H.; Zhu, Z. Spatio-Temporal Graph Convolutional Networks: A Deep Learning Framework for Traffic Forecasting. arXiv 2017, arXiv:1709.04875. [Google Scholar]
  12. Guo, S.; Lin, Y.; Feng, N.; Song, C.; Wan, H. Attention based spatial-temporal graph convolutional networks for traffic flow forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January 2019. [Google Scholar]
  13. Bai, L.; Yao, L.; Li, C.; Wang, X.; Wang, C. Adaptive graph convolutional recurrent network for traffic forecasting. Adv. Neural Inf. Process. Syst. 2020, 33, 17804–17815. [Google Scholar]
  14. Li, F.; Feng, J.; Yan, H.; Jin, G.; Yang, F.; Sun, F.; Jin, D.; Li, Y. Dynamic graph convolutional recurrent network for traffic forecasting: Benchmark and solution. ACM Trans. Knowl. Discov. Data 2021, 17, 1–21. [Google Scholar]
  15. Sun, X.; Cheng, H.; Liu, B.; Li, J.; Chen, H.; Xu, G.; Yin, H. Self-supervised hypergraph representation learning for sociological analysis. IEEE Trans. Knowl. Data Eng. 2023, 35, 11860–11871. [Google Scholar] [CrossRef]
  16. Liu, S.; Chen, H.; Chen, X.Y.; He, J.J. A dual-branch spatio-temporal graph convolutional neural network for traffic flow prediction. Inf. Control. 2022, 1–14. [Google Scholar] [CrossRef]
  17. Zeng, Y.C.; Shao, M.H.; Sun, L.J.; Lu, C. Traffic prediction and congestion control based on directed graph convolutional neural network. China J. Highw. Transp. 2021, 34, 239–248. [Google Scholar]
  18. Feng, Y.; You, H.; Zhang, Z.; Ji, R.; Gao, Y. Hypergraph neural networks. In Proceedings of the 33rd AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 3558–3565. [Google Scholar]
  19. Pedronette, D.C.G.; Valem, L.P.; Almeida, J.; Torres, R.D.S. Multimedia retrieval through unsupervised hypergraph-based manifold ranking. IEEE Trans. Image Process. 2019, 28, 5824–5838. [Google Scholar] [CrossRef]
  20. Ta, X.; Liu, Z.; Hu, X.; Yu, L.; Sun, L.; Du, B. Adaptive spatio-temporal graph neural network for traffic forecasting. Knowl.-Based Syst. 2022, 242, 108199. [Google Scholar] [CrossRef]
  21. Wu, Z.; Pan, S.; Long, G.; Jiang, J.; Chang, X.; Zhang, C. Connecting the dots: Multivariate time series forecasting with graph neural networks. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual Event, CA, USA, 6–10 July 2020. [Google Scholar]
  22. Wei, S.; Yang, Y.; Liu, D.; Deng, K.; Wang, C. Transformer-Based Spatiotemporal Graph Diffusion Convolution Network for Traffic Flow Forecasting. Electronics 2024, 13, 3151. [Google Scholar] [CrossRef]
  23. Naganand, Y.; Madhav, N.; Prateek, Y.; Vikram, N.; Anand, L.; Partha, T. HyperGCN: A new method of training graph convolutional networks on hypergraphs. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 1511–1522. [Google Scholar]
  24. Zhang, Y.K.; Wu, Z.H.; Lin, Y.F.; Zhao, Y.J. Spatio-temporal hypergraph convolutional network for traffic flow prediction. J. Comput. Appl. 2021, 41, 3578–3584. [Google Scholar]
  25. Bai, S.; Zhang, F.; Torr, P.H.S. Hypergraph convolution and hypergraph attention. Pattern Recognit. 2021, 110, 107637. [Google Scholar] [CrossRef]
  26. Wang, J.; Zhang, Y.; Hu, Y.; Yin, B. Metro flow prediction with hierarchical hypergraph attention networks. IEEE Trans. Artif. Intell. 2021, 5, 3012–3021. [Google Scholar] [CrossRef]
  27. Gao, Y.; Feng, Y.; Ji, J.R. HGNN+: General hypergraph neural networks. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 3181–3199. [Google Scholar] [CrossRef]
  28. Chen, H.; Yin, H.; Sun, X.; Chen, T.; Gabrys, B.; Musial, K. Multi-level graph convolutional networks for cross-platform anchor link prediction. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual Event, 6–10 July 2020. [Google Scholar]
  29. Wang, J.; Zhang, Y.; Wang, L.; Hu, Y.; Piao, X.; Yin, B. Multitask Hypergraph Convolutional Networks: A Heterogeneous Traffic Prediction Framework. IEEE Trans. Intell. Transp. Syst. 2022, 23, 18557–18567. [Google Scholar] [CrossRef]
  30. Zhao, Z.Z.; Shen, G.J.; Zhou, J.J.; Jin, J.C.; Kong, X.J. Spatial-temporal hypergraph convolutional network for traffic forecasting. PeerJ Comput. Sci. 2023, 9, 341–345. [Google Scholar] [CrossRef] [PubMed]
  31. Wang, S.; Zhang, Y.; Qi, H.; Zhao, M.; Jiang, Y. Dynamic spatial-temporalhypergraph convolutional network for skeleton-based action recognition. In Proceedings of the 2023 IEEE International Conference on Multimedia and Expo (ICME), Brisbane, Australia, 10–14 July 2023; pp. 2147–2152. [Google Scholar]
  32. Dong, Z.; Yu, S.; Shen, Y. Multi-scale dynamic hypergraph convolutional network for traffic flow forecasting. J. Shanghai Jiaotong Univ. (Sci.) 2024. [Google Scholar] [CrossRef]
  33. Zou, G.; Lai, Z.; Wang, T.; Liu, Z.; Li, Y. MT-STNet: A novel multi-task spatio-temporal network for highway traffic flow prediction. IEEE Trans. Intell. Transp. Syst. 2024, 25, 8221–8236. [Google Scholar] [CrossRef]
  34. Xia, L.; Liang, Y.; Zheng, P.; Huang, X. Residual-hypergraph convolution network: A model-based and data-driven integrated approach for fault diagnosis in complex equipment. IEEE Trans. Instrum. Meas. 2023, 72, 1–11. [Google Scholar] [CrossRef]
  35. Song, C.; Lin, Y.; Guo, S.; Wan, H. Spatial-temporal synchronous graph convolutional networks: A new framework for spatial-temporal network data forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 914–921. [Google Scholar]
  36. Chen, Y.; Segovia-Dominguez, I.; Gel, Y.R. Z-gcnets: Time zigzags at graph convolutional networks for time series forecasting. In Proceedings of the International Conference on Machine Learning, Online, 18–24 July 2021. [Google Scholar]
  37. Fang, Z.; Long, Q.; Song, G.; Xie, K. Spatial-temporal graph ODE networks for traffic flow forecasting. arXiv 2021, arXiv:2106.12931. [Google Scholar]
  38. Wu, Z.; Pan, S.; Long, G.; Jiang, J.; Zhang, C. Graph WaveNet for deep spatial-temporal graph modeling. arXiv 2019, arXiv:1906.00121. [Google Scholar]
  39. Lan, S.; Ma, Y.; Huang, W.; Wang, W.; Yang, H.; Li, P. DSTAGNN: Dynamic spatial-temporal aware graph neural network for traffic flow forecasting. In Proceedings of the International Conference on Machine Learning, Baltimore, ML, USA, 17–23 July 2022; pp. 11906–11917. [Google Scholar]
Figure 1. Schematic diagram of two types of intersection structures in a traffic network.
Figure 1. Schematic diagram of two types of intersection structures in a traffic network.
Electronics 13 04435 g001
Figure 2. Graph structure and hypergraph structure.
Figure 2. Graph structure and hypergraph structure.
Electronics 13 04435 g002
Figure 3. The overall framework of the model.
Figure 3. The overall framework of the model.
Electronics 13 04435 g003
Figure 4. Hyperedge outlier mechanism.
Figure 4. Hyperedge outlier mechanism.
Electronics 13 04435 g004
Figure 5. PeMSD4 and PeMSD8 datasets sensor distribution maps.
Figure 5. PeMSD4 and PeMSD8 datasets sensor distribution maps.
Electronics 13 04435 g005
Figure 6. Performance comparison of different forecasting models across multiple datasets at various time steps.
Figure 6. Performance comparison of different forecasting models across multiple datasets at various time steps.
Electronics 13 04435 g006
Figure 7. Comparison of predictive curves for DSTHGCN with ground truth and Graph WaveNet on PeMSD4 at Nodes #1, #69, #123, and #196.
Figure 7. Comparison of predictive curves for DSTHGCN with ground truth and Graph WaveNet on PeMSD4 at Nodes #1, #69, #123, and #196.
Electronics 13 04435 g007
Figure 8. Hyperparameter study on PeMSD4 and PeMSD8.
Figure 8. Hyperparameter study on PeMSD4 and PeMSD8.
Electronics 13 04435 g008
Table 1. Summary statistics for PeMSD4 and PeMSD8.
Table 1. Summary statistics for PeMSD4 and PeMSD8.
Datasets#NodesEdges#TimeStepsTime IntervalTime Range
PeMSD430734016,9925 min1/1/2018–2/28/2018
PeMSD817029517,8565 min7/1/2016–8/31/2016
Table 2. Analysis of prediction performance of traffic forecasting models on different datasets.
Table 2. Analysis of prediction performance of traffic forecasting models on different datasets.
ModelsPeMSD4PeMSD8
MAERMSEMAPEMAERMSEMAPE
HA38.0359.2427.8834.8659.2427.88
ARIMA33.7348.8024.1831.0944.3222.73
FC-LSTM26.7740.6518.2323.0935.1714.99
ASTGCN(r)22.2034.6915.017.3126.9311.0
Graph WaveNet25.3239.0417.5119.1630.5912.24
STGODE20.7732.5014.0616.7825.8011.27
DSTAGNN19.5831.9113.2815.7525.0810.54
DSTHGCN18.9530.2912.6914.8923.749.67
Table 3. Comparison predictive performance of traffic forecasting models at 15 min, 30 min, and 60 min horizons (3, 6, and 12 time steps).
Table 3. Comparison predictive performance of traffic forecasting models at 15 min, 30 min, and 60 min horizons (3, 6, and 12 time steps).
DatasetsModels15 Min30 Min60 Min
MAERMSEMAPEMAERMSEMAPEMAERMSEMAPE
PeMSD4ASTGCN (r)20.0031.5913.021.9634.315.026.0439.6318.0
Graph WaveNet20.9732.9214.6724.5838.1116.6732.6649.1623.09
STGODE18.8930.8513.3120.4432.6214.0423.6335.4215.32
DSTAGNN18.6330.0812.6419.5531.8913.2121.3434.7514.41
DSTHGCN (ours)18.1229.1112.1219.0530.5312.6820.6332.7213.69
PeMSD8ASTGCN (r)15.8624.6910.017.1726.7511.020.1530.8412.0
Graph WaveNet15.8725.349.9818.6730.1011.7324.8039.0916.21
STGODE15.7523.9410.2416.7825.8711.1918.428.6412.64
DSTAGNN14.823.739.7715.5524.9810.4617.7527.3811.61
DSTHGCN (ours)14.0722.099.0415.0223.879.6316.7026.4210.69
Table 4. Results of ablation studies on PeMSD4 and PeMSD8.
Table 4. Results of ablation studies on PeMSD4 and PeMSD8.
ModelsPeMSD4PeMSD8
MAERMSEMAPEMAERMSEMAPE
DSTHGCN-119.2630.7013.2815.4024.2610.53
DSTHGCN-219.0830.6013.2315.2624.1610.45
DSTHGCN18.9530.2912.6914.8923.749.67
Table 5. Comparison of training and inference times of various models on PeMSD4 and PeMSD8 datasets.
Table 5. Comparison of training and inference times of various models on PeMSD4 and PeMSD8 datasets.
DatasetsComputation Time (Training (s/epoch)/Inference (s))
ASTGCNGraph WaveNetSTGODEDSTAGNNDSTHGCN
PeMSD484.395/9.45580.31/7.5232.04/6.0145.58/4.2637.66/7.44
PeMSD845.58/4.2624.61/1.6727.69/5.1328.51/3.1921.58/2.43
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ye, Z.; Wang, H.; Przystupa, K.; Majewski, J.; Hots, N.; Su, J. Dynamic Spatio-Temporal Hypergraph Convolutional Network for Traffic Flow Forecasting. Electronics 2024, 13, 4435. https://doi.org/10.3390/electronics13224435

AMA Style

Ye Z, Wang H, Przystupa K, Majewski J, Hots N, Su J. Dynamic Spatio-Temporal Hypergraph Convolutional Network for Traffic Flow Forecasting. Electronics. 2024; 13(22):4435. https://doi.org/10.3390/electronics13224435

Chicago/Turabian Style

Ye, Zhiwei, Hairu Wang, Krzysztof Przystupa, Jacek Majewski, Nataliya Hots, and Jun Su. 2024. "Dynamic Spatio-Temporal Hypergraph Convolutional Network for Traffic Flow Forecasting" Electronics 13, no. 22: 4435. https://doi.org/10.3390/electronics13224435

APA Style

Ye, Z., Wang, H., Przystupa, K., Majewski, J., Hots, N., & Su, J. (2024). Dynamic Spatio-Temporal Hypergraph Convolutional Network for Traffic Flow Forecasting. Electronics, 13(22), 4435. https://doi.org/10.3390/electronics13224435

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop