Academic Journal
An indoor positioning method based on bluetooth array/PDR fusion using the SVD-EKF
Title: | An indoor positioning method based on bluetooth array/PDR fusion using the SVD-EKF |
---|---|
Authors: | Chenhui Li, Jie Zhen, Jianxin Wu |
Source: | Scientific Reports, Vol 15, Iss 1, Pp 1-21 (2025) |
Publisher Information: | Nature Portfolio, 2025. |
Publication Year: | 2025 |
Collection: | LCC:Medicine LCC:Science |
Subject Terms: | Medicine, Science |
More Details: | Abstract With the development of mobile internet and artificial intelligence, the core of current surveying and mapping science and technology is no longer confined solely to outdoor applications. High-precision indoor positioning technology is now one of the core technologies in the era of artificial intelligence. An angle measurement and positioning system based on wireless signal array antennas can achieve high accuracy in indoor unblocked conditions. However, indoor environment is unpredictable, and the user's behavior is also random. Inevitably, some signal reflection and other factors will affect the positioning accuracy. Considering all the aforementioned, this study analyzes the principles and characteristics of a Bluetooth signal based array antenna angle measurement and positioning system. In addition, aiming at the multi-phase problem caused by antenna switching, the frequency estimation method based on FFT (Fast Fourier Transform) is studied in this paper, achieving high-precision angle measurement and positioning. Aiming at the problem that the system positioning error increase in the complex and variable indoor environment, a fusion positioning method of Bluetooth array/PDR (Pedestrian Dead Reckoning) based on the SVD-EKF (Singular Value Decomposition–Extended Kalman Filter) is proposed. This study introduces a few improvements to the EKF (Extended Kalman Filter). First, the predicted state covariance matrix is decomposed by singular value decomposition, which improves the robustness of the EKF. Second, a Bluetooth array self-evaluation factor is introduced and combined with the Huber function to construct an adaptive factor, thus further enhancing filtering accuracy and environmental adaptability. Through static and dynamic data collected in indoor environment, the feasibility of the algorithm is verified. The results of static and dynamic experimental results show that the array angle measurement and positioning system can achieve high accuracy, the near point positioning maximum error is 0.3m and the far point positioning accuracy is 0.6m. The dynamic test results in the room show that after SVD-EKF algorithm optimization, the positioning error is reduced by 0.05m, which is equivalent to the EKF algorithm. However, in corridor areas, the improvement of accuracy of SVD-EKF algorithm is better than EKF, achieving an improvement of 0.373 m and a smoother positioning result compared to the traditional EKF algorithm. This study provides a new practical technology with a high precision and easy deployment for indoor positioning. |
Document Type: | article |
File Description: | electronic resource |
Language: | English |
ISSN: | 2045-2322 28144066 |
Relation: | https://doaj.org/toc/2045-2322 |
DOI: | 10.1038/s41598-025-88860-1 |
Access URL: | https://doaj.org/article/1caa281440664c1e87264a56660404da |
Accession Number: | edsdoj.1caa281440664c1e87264a56660404da |
Database: | Directory of Open Access Journals |
Full text is not displayed to guests. | Login for full access. |
FullText | Links: – Type: pdflink Url: https://content.ebscohost.com/cds/retrieve?content=AQICAHjPtM4BHU3ZchRwgzYmadcigk49r9CVlbU7V5F6lgH7WwGxwvD3NioAhtEqp86KGPr4AAAA4jCB3wYJKoZIhvcNAQcGoIHRMIHOAgEAMIHIBgkqhkiG9w0BBwEwHgYJYIZIAWUDBAEuMBEEDE0ZhbUCuUvvBb3u2QIBEICBmqU7hWH4Ry9apzBjaH87BTZjaOpuAph_kQM393MtB-khIqILbWd5thsfZMr-fQUerhAhkLjp9llzEEbdB42T3acD7UxcYAumv2XITAGJ4bKOLUKnx3GvCnLFEuCznVjKin5ahYsij2LFZ97gPR82eC_cC0qBwE-TnWKBBN6mKhU3Q3r1WHSm--Nem0UrxpxekY6C2wV5vbr923c= Text: Availability: 1 Value: <anid>AN0183073501;[fkqs]15feb.25;2025Feb19.04:14;v2.2.500</anid> <title id="AN0183073501-1">An urban road traffic flow prediction method based on multi-information fusion </title> <sbt id="AN0183073501-2">Introduction</sbt> <p>Accurate traffic flow prediction not only relies on historical traffic flow information, but also needs to take into account the influence of a variety of external factors such as weather conditions and the distribution of neighbouring POIs. However, most of the existing studies have used historical data to predict future traffic flows for short periods of time. Spatio-Temporal Graph Neural Networks (STGNN) solves the problem of combining temporal properties and spatial dependence, but does not extract long-term trends and cyclical features of historical data. Therefore, this paper proposes a MIFPN (Multi information fusion prediction network) traffic flow prediction method based on the long and short-term features in the historical traffic flow data and combining with external information. First, a subsequence converter is utilised to allow the model to learn the temporal relationships of contextual subsequences from long historical sequences that incorporate external information. Then, a superimposed one-dimensional inflated convolutional layer is used to extract long-term trends, a dynamic graph convolutional layer to extract periodic features, and a short-term trend extractor to learn short-term temporal features. Finally, long-term trends, cyclical features and short-term features are fused to obtain forecasts. Experiments on real datasets show that the MIFPN model improves by an average of 11.2% over the baseline model in long term predictions up to 60 min ago.</p> <p>The rapid growth of urban traffic demand has further aggravated the problems of traffic congestion, air pollution and traffic accidents. As an important part of intelligent transportation system, accurate and efficient traffic flow prediction method can not only support more scientific road design and land use planning, but also effectively reduce congestion and carbon emissions and promote the realization of green travel. In addition, the results of traffic flow prediction can optimize the operation of urban logistics and public transportation, reduce costs and improve resource utilization efficiency. Therefore, accurate prediction and effective management of traffic flow has become an urgent challenge for urban traffic managers and researchers.</p> <p>With the development of deep learning, convolutional neural network-based traffic flow prediction algorithm introduces spatial features, but only realizes grid-based road network representation. Then, the graph neural network is introduced to improve the spatial feature extraction significantly, and the prediction accuracy is also greatly improved. However, current traffic flow prediction algorithms still face some limitations. On the one hand, they still have shortcomings in learning nonlinear transformation features, mainly focusing on short-term features of traffic flow data, and lacking long-term trend extraction and cycle analysis. On the other hand, the future state of traffic depends not only on the historical state, but also on various external factors, including static factors such as street restaurants, schools, bus stops, and dynamic factors such as weather conditions and traffic control. To solve these problems, this paper proposes a multi-information fusion prediction network (MIFPN) for traffic flow prediction. This method takes into account the short-term characteristics, long-term trends and periodic characteristics of historical traffic flow data, and combines external static and dynamic information. The main innovations of this paper include:</p> <p></p> <ulist> <item> The proposed MIFPN model integrates external road information. Static information is used to learn node traffic features under different spatial attributes, while dynamic information captures flow transformation features under various weather conditions. A feature-enhancement unit is designed to collect the dynamic <emph>D</emph> and static <emph>S</emph> attributes of road sections, which fuses the traffic feature matrix <emph>X</emph> and the attribute matrix into an expanded matrix .</item> <p></p> <item> A long-term trend extractor is designed based on a subsequence transformer, enabling the MIFPN model to capture the long-term characteristics of traffic flow data. By leveraging a masked subsequence Transformer, compressed and contextually rich subsequences are generated, and a long-term trend extractor is obtained using stacked one-dimensional dilated convolution layers on these subsequences.</item> <p></p> <item> A cycle extractor is designed to capture the periodic characteristics within a week and a day. A graph convolutional module is employed to combine the spatial dependencies of the traffic graph and the hidden spatial dependencies within the graph, resulting in a cycle extractor.</item> </ulist> <p>The proposed multi-information fusion traffic flow forecasting method synthesizes external road information, static and dynamic characteristics, and temporal and spatial relations, and has a positive impact on urban planning, environmental sustainability and economic benefits. By learning spatial attribute characteristics and traffic changes under dynamic conditions, combined with long-term trend and periodic feature extractor, the method enhances the predictability of traffic flow prediction, optimizes intelligent traffic signal control and route planning, and strongly supports the development of smart cities.</p> <p>The remainder of the paper is structured as follows: the second section reviews related work and trends in traffic flow prediction. Section III presents the details of the proposed method in this paper. In Section IV, experiments based on real datasets are conducted to evaluate the performance of the proposed method in comparison with the baseline method, and perturbation analysis is performed to test the robustness of the model. Section V summarises the work of this paper and gives an outlook.</p> <hd id="AN0183073501-3">Related work</hd> <p>Earlier, traffic flow prediction[<reflink idref="bib1" id="ref1">1</reflink>], [<reflink idref="bib2" id="ref2">2</reflink>]–[<reflink idref="bib3" id="ref3">3</reflink>] was mainly based on mathematical and statistical models. Autoregressive Integral Moving Average (ARIMA)[<reflink idref="bib4" id="ref4">4</reflink>] and its variants are classical methods[<reflink idref="bib5" id="ref5">5</reflink>] and are widely used in traffic prediction problems. However, these methods are mainly for small datasets and are not suitable for dealing with complex and dynamic time series data. With the development of machine learning[<reflink idref="bib6" id="ref6">6</reflink>], [<reflink idref="bib7" id="ref7">7</reflink>]–[<reflink idref="bib8" id="ref8">8</reflink>], it is possible to model more complex time series data, including feature models[<reflink idref="bib9" id="ref9">9</reflink>], Gaussian process models and state space models Such models are capable of handling non-linear time series features but are poor for modelling complex road networks and dynamic traffic data. Deep learning models with more feature layers and more complex architectures have better results for modelling spatio-temporal correlation of traffic flow with large sample data processing For example, Convolutional Neural Networks (CNNs) are effective in modelling spatial relationships of traffic data constructed in the form of road networks, but lack the construction of road topological relationships. However, road networks can be naturally represented as graphs, where distances between roads can be expressed as weights of edges, and the modelling problem of non-Euclidean data structures can be effectively solved by graph neural networks (GNN)[<reflink idref="bib10" id="ref10">10</reflink>], [<reflink idref="bib11" id="ref11">11</reflink>]–[<reflink idref="bib12" id="ref12">12</reflink>].</p> <p>GNNs can be classified into two categories, spectrum-based GNNs and space-based GNNs.Bruna et al. first proposed spectral CNNs to generalise CNNs to non-Euclidean spaces, but the spectral decomposition process for Laplace matrices has excessive computational complexity. ChebNet[<reflink idref="bib13" id="ref13">13</reflink>] applies Chebyshev polynomials to approximate complex spectral convolutions, which effectively reduces the complexity of spectral convolutions and reduces the number of parameters, and is therefore less likely to overfit the data. GCN[<reflink idref="bib14" id="ref14">14</reflink>],[<reflink idref="bib15" id="ref15">15</reflink>] can be regarded as a first-order ChebNet with regularisation, which is computationally efficient and easy to stack in multiple layers. Space-based GNN describes the process of graph convolution as an aggregation of information from the central node and its neighbouring nodes from the point of view of local spatial associations between nodes. Glimer[<reflink idref="bib16" id="ref16">16</reflink>] summarised the operations of aggregation, readout and so on for spatially-based GNNs and generalised one-class message-passing neural networks (MPNNs). The GCN can also be viewed as a spatial-based message-passing neural network, where the weights of messages from neighbouring nodes during aggregation are fixed and determined by the degree of the node. GAT[<reflink idref="bib17" id="ref17">17</reflink>] introduced a multi-head self-attention mechanism in graph neural networks to dynamically assign weights to messages from neighbouring nodes during aggregation. Therefore, GAT can better focus on information from important nodes compared to GCN.</p> <p>Based on GNNs, modelling of temporal correlation in traffic flow has been further introduced and many spatio-temporal graphical neural network models have been proposed for traffic prediction tasks[<reflink idref="bib18" id="ref18">18</reflink>], [<reflink idref="bib19" id="ref19">19</reflink>]–[<reflink idref="bib20" id="ref20">20</reflink>]. Based on the time-dependent modelling, STGNNs are classified into RNN-based, CNN-based and Transformer-based STGNNs. The RNN-based STGNN is represented by the DCRNN[<reflink idref="bib21" id="ref21">21</reflink>], [<reflink idref="bib22" id="ref22">22</reflink>]–[<reflink idref="bib23" id="ref23">23</reflink>], which uses gated recurrent units (GRUs), extracts temporal features from the data, and employs a diffusive convolutional neural network (DCNN) to simulate the effects of spatio-temporal variations from one node to the others in the traffic network. The CNN-based STGNN is ASTGCN[<reflink idref="bib24" id="ref24">24</reflink>], [<reflink idref="bib25" id="ref25">25</reflink>]–[<reflink idref="bib26" id="ref26">26</reflink>], which introduces a spatio-temporal attention mechanism to capture long-range temporal dependencies and implicit spatial connectivity in road networks and combines it with spatial graph convolution and temporal one-dimensional convolution in order to model the spatio-temporal dynamics in traffic data. The transformer-based STGNN is STTN[<reflink idref="bib27" id="ref27">27</reflink>], [<reflink idref="bib28" id="ref28">28</reflink>]–[<reflink idref="bib29" id="ref29">29</reflink>], which integrates a graphical convolution process into a spatial Transformer, and together with a temporal Transformer, STTN can capture long-term dependencies from traffic data.</p> <p>However, the traffic forecasting task not only relies on historical traffic information and spatial relationships, but is also influenced by various external factors, such as weather conditions and the distribution of surrounding POIs. How to integrate information from external influences into the model is a major issue in current transport work. For example, Liaoetal integrated an LSTM-based encoder to encode external information and modelled multimodal data as sequential inputs. Zhangetal implemented a traffic prediction task with external weather information by feature fusion of input features and weather information, mainly based on the GRU model.In summary, this paper exploits the long-term trend and periodicity of historical traffic data and incorporates information external to the road to improve the accuracy and robustness of traffic flow prediction.</p> <hd id="AN0183073501-4">Methodology</hd> <p></p> <hd id="AN0183073501-5">Relevant definition</hd> <p></p> <hd id="AN0183073501-6">Definition 1</hd> <p>Road network to represent the connectivity about road segments. denotes the set of road segments, <emph>n</emph> is the number of road segments, is the set of edges indicating connectivity between two road segments, and <emph>m</emph> denotes the number of edges. In general, the adjacency matrix <emph>A</emph> is used to illustrate the connectivity of a road network. When <emph>G</emph> is an unweighted network, <emph>A</emph> is a matrix of 0 and 1, which 1 indicates a connection to the corresponding road segment, and 0 otherwise.</p> <hd id="AN0183073501-7">Definition 2</hd> <p>Traffic speed is regarded as an intrinsic attribute of each node on the urban road network, which is represented by a traffic characterization matrix <emph>X</emph>, and denotes the traffic speed on the i-th road segment at time <emph>t</emph> as matrix .</p> <hd id="AN0183073501-8">Definition 3</hd> <p>External factors affecting traffic conditions are used as auxiliary attributes of online sections of urban roads, which can form the attribute matrix , as <emph>l</emph> is the category number of the auxiliary information. The set of auxiliary information of type <emph>j</emph>, is denoted as , which is the j-th auxiliary information of the i-th road segment at time <emph>t</emph>.</p> <p>In summary, the traffic prediction problem can be seem as learning a function <emph>f</emph> on the basic road network <emph>G</emph>, feature matrix <emph>X</emph> and attribute matrix <emph>K</emph> to obtain the traffic information for the future time period <emph>T</emph>, as shown in Eq. (<reflink idref="bib1" id="ref30">1</reflink>):</p> <olist> <item></item> </olist> <p>Graph</p> <hd id="AN0183073501-9">MIFPN framework</hd> <p>The entire MIFPN framework is shown in Fig. 1. Firstly, external information is fused into the traffic data. Secondly, a subsequence level time series representation is extracted from the long term series based on a subsequence learner, and long term and short term time features are obtained from the subsequence by means of a trend extractor, a period extractor and a feature fusion. Finally, the obtained temporal features are combined with short-term feature data fused with external information for traffic flow prediction.</p> <p>Graph: Fig. 1 MIFPN model diagram.</p> <hd id="AN0183073501-10">External information</hd> <p>This paper analyzes the influence of external factors on traffic states from both static and dynamic perspectives. The external factors are defined as dynamic <emph>D</emph> and static <emph>S</emph> attributes of road segments in the road network. Then, the traffic characteristic matrix <emph>X</emph> and the attribute matrix are synthesized into an expansion matrix .</p> <hd id="AN0183073501-11">Static factors</hd> <p>Primarily refers to static geographic information that does not change over time, but still has an impact on the state of traffic. For example, the distribution of POIs around a road section can determine people's access patterns and the attractiveness of the road section, which in turn is reflected in its traffic state.</p> <p>Graph: Fig. 2 POI speed distribution graph.</p> <p>Figure 2 shows the average traffic performance for Shenzhen taxis at different types of nodes over the course of a day. After analysing the data, different types of nodes have different behavioural characteristics, and the surrounding buildings and environment of a road node can have an impact on traffic flow. For example, morning and evening flows would be higher in residential areas than at other times of the day, while midday and evening flows would be significantly higher in catering areas. Therefore, classifying the architectural attributes of the road nodes themselves as points of interest and incorporating them into the data stream will help to improve the accuracy of the traffic flow data.</p> <p>is a collection of <emph>p</emph> distinct static attributes . As the attribute values don't change by time, the matrix <emph>S</emph> is always used, and only the corresponding columns of the feature matrix <emph>X</emph> is extracted during the generation of the augmented matrix at each timestamp. The extended matrix with static properties is formed at time<emph>t</emph>, as shown in Eq. (<reflink idref="bib2" id="ref31">2</reflink>):</p> <p>2</p> <p>Graph</p> <hd id="AN0183073501-12">Dynamic factor</hd> <p>Weather information is an important factor affecting traffic flow and is selected as a dynamic feature. Traffic peaks and flow characteristics have different manifestations under varying weather conditions. Figure 3 shows 10 nodes selected from the Shenzhen taxi dataset and analysed for their speed averages under different weather conditions. It can be clearly seen that the speed behaviour of the vehicles changes significantly in different weather conditions, with the lowest speeds in foggy weather and the highest mean speeds on sunny days.</p> <p>Graph: Fig. 3 Speed conditions in different weather.</p> <p>, unlike <emph>S</emph>, is <emph>W</emph> different dynamic properties. It is noteworthy to consider that traffic states are cumulatively affected by dynamic factors over time. We extend the size of the selection window to when forming , So choose . Finally, through the Attribute Enhancement Cell (A-Cell), the augmented matrix containing information about static and dynamic external attributes and traffic characteristics at time t is formed as . shown in Eq. (<reflink idref="bib3" id="ref32">3</reflink>)</p> <p>3</p> <p>Graph</p> <hd id="AN0183073501-13">Subsequence learner</hd> <p>Masked Subsequence Transformer (MST) is to infer the contents of masked subsequences from a small number of subsequences and their temporal contexts, allowing the model to efficiently learn compressed, contextually informative subsequence representations from long time sequences. The design of the MST consists of two fundamental issues: (<reflink idref="bib1" id="ref33">1</reflink>) masking strategies and (<reflink idref="bib2" id="ref34">2</reflink>) models for learning the representations.</p> <p>(<reflink idref="bib1" id="ref35">1</reflink>) Masking strategy. There are two important factors to be considered in the MST masking strategy, the basic unit of masking and the masking ratio. Existing methods usually use a 5-minute time step as the basic unit of input data, which does not capture the trend of long time series well. Inspired by[<reflink idref="bib27" id="ref36">27</reflink>], long sequences are divided into equal-length subsequences containing multiple time steps, and use these subsequences as the basic unit of model input. Both BERT[<reflink idref="bib30" id="ref37">30</reflink>] and MAE use template reconstruction to learn the basic semantic information in the data. The information density of image data is relatively low and the pixel points have spatial continuity, so even if the MAE masks out 75% of the pixels, the main content of the image can still be inferred. Long-term traffic flow data is similar to images with temporal continuity and low information density. Thus requires a relatively high masking rate 75% random masking.</p> <p>(<reflink idref="bib2" id="ref38">2</reflink>) Model for learning the representation. For time series, the difference between Transformer and temporal models such as RNN and 1DCNN is that the inputs of each time step in Transformer are directly connected to each other. Regardless of the increase in time step length, Transformer considers the representation of previous temporal features. In this paper, the Transformer encoder is used as the STRL, as shown in Fig. 4. The MST consists of two parts, the STRL and the self-supervised task head. The STRL learns the temporal representations of the subsequences, and the self-supervised task head reconstructs the complete long sequence based on the temporal representations of the unmasked subsequences and the masking tokens.</p> <p>Graph: Fig. 4 Schematic diagram of the mask subsequence model.</p> <p>Specifically, long history sequences are partitioned into non-overlapping subsequences . Then, randomly mask 75% of the subsequences, and marked as the masked subsequence. The remaining unmasked subsequences serve as input to <emph>STRL</emph>, as shown in Eq. (<reflink idref="bib4" id="ref39">4</reflink>):</p> <p>4</p> <p>Graph</p> <p>where, represents the <emph>STRL</emph> processed output of . The self-supervised task header consists of a Transformer layer and a linear output layer, which can reconstruct the given unmasked mask and the complete long sequence of learnable mask tokens, as shown in Eq. (<reflink idref="bib5" id="ref40">5</reflink>):</p> <p>5</p> <p>Graph</p> <p>The goal of pre-training is to minimize the error between the reconstructed mask value and the mask true value. Hence, only masked subsequences are considered in calculating losses, as shown in Eq. (<reflink idref="bib6" id="ref41">6</reflink>):</p> <p>6</p> <p>Graph</p> <p>where, is the learnable parameters of the whole Transformer.</p> <hd id="AN0183073501-14">Trend extractor</hd> <p>The relatively small amount of information in the short-term historical series is insufficient to infer complex future traffic flow changes, whereas the long-term historical series can help the model to determine traffic flow fluctuations at future moments. For this purpose, a long-term trend extractor is designed to extract the long-term trend characteristics of the traffic flow from the temporal representation of the subsequence.</p> <p>Commonly used basic structures for temporal feature extraction include RNNs and 1D CNNs. However, it is difficult for RNNs to handle long sequences because they cannot process features at each time step in parallel and are prone to the problems of gradient vanishing and gradient explosion. Ordinary one-dimensional CNNs have a limited receptive field, and increasing that receptive field requires stacking multiple CNN layers, which leads to a significant increase in the number of model parameters as the depth of the model increases. As shown in Fig. 5, a stacked one-dimensional dilated convolutional layer is used as the long-term trend extractor.</p> <p>Graph: Fig. 5 Schematic diagram of the trend extractor model.</p> <p>The sensory field of this module grows exponentially with the number of 1-dimensional dilated convolutional layers, which allows for efficient capture of trending features while avoiding problems such as gradient vanishing. The dilation convolution operation is expressed as shown in Eq. (<reflink idref="bib7" id="ref42">7</reflink>):</p> <p>7</p> <p>Graph</p> <p>where, <emph>m</emph> denotes the m-th element in the sequence <emph>x</emph>, denotes the convolution kernel, and <emph>d</emph> denotes the expansion rate. In this paper, the convolutional layer can be represented as follows in Eq. (<reflink idref="bib8" id="ref43">8</reflink>):</p> <p>8</p> <p>Graph</p> <p>where the maximum pooling operation is used to reduce the dimensionality. The dilated rate <emph>d</emph> of i-th layer is set to . When , the input to the module is the set of subsequence time <emph>S</emph>, . The output of the last convolutional layer is considered as a long-term trend feature .</p> <hd id="AN0183073501-15">Cycle extractor</hd> <p>Traffic flows are usually cyclical, with similar spatial and temporal patterns for the same time periods on different dates and days of the week. In this paper, a periodicity extractor is built, a module that extracts the spatial dependence of input features across nodes with different time steps while preserving the temporal information of the subsequence. Suppose that the duration of a day corresponds to a time period, denoted by <emph>l</emph>. Then, the representation of the corresponding moments of the previous week and the previous day can be expressed as respectively: and (<emph>N</emph> denotes the number of subsequences and is equal to ). As shown in Fig. 6, passing and to the spatial-based graphical convolution module to obtain and for periodic temporal features. This module is similar to the one proposed in [<reflink idref="bib30" id="ref44">30</reflink>], where the graph convolution module combines the spatial dependencies of the flow graph and the spatial dependencies hidden in the graph.</p> <p>Graph: Fig. 6 Schematic diagram of the cycle extractor.</p> <p>Specifically as shown in Eqs. (<reflink idref="bib9" id="ref45">9</reflink>) and (<reflink idref="bib10" id="ref46">10</reflink>).</p> <p>9</p> <p>Graph</p> <p>10</p> <p>Graph</p> <p>where, and matrices correspond to the forward and backward diffusion of the graphical signals, respectively, and denote the corresponding transfer matrices. The power <emph>k</emph> of the matrix represents the number of steps in the diffusion process. <emph>W</emph> denotes the weight matrix and <emph>Aadp</emph> is an adaptive neighbourhood matrix which is considered as the transfer matrix for the hidden diffusion process. , and denote the source and target node embeddings, respectively, and the spatial dependency weights between the source and target nodes are derived by multiplying and together. is used to remove weak dependencies and is used for normalisation.</p> <hd id="AN0183073501-16">Fusion module</hd> <p>There is a strong temporal correlation between future short-term traffic flows and historical short-term traffic flows, so short-term trends need to be modelled separately.</p> <p>It has been widely demonstrated that STGNN excels at capturing fine-grained features from short-term sequences [<reflink idref="bib5" id="ref47">5</reflink>, 34]. Firstly, spatial and temporal features are learnt through spatial and temporal learning networks respectively. Then, the two features are fused by a certain spatio-temporal fusion neural network structure. In this paper, an existing STGNN such as Graph WaveNet is used as a short-term trend extractor to obtain a finer-grained short-term trend map, as shown in Eq. (<reflink idref="bib11" id="ref48">11</reflink>):</p> <olist> <item>1</item> </olist> <p>Graph</p> <p>where the short sequence denotes the last subsequence in the long sequence, and <emph>A</emph> is the neighbourhood matrix. denotes the STGNN model used.</p> <p>In order to comprehensively consider the long and short-term features in the long historical series, the previously obtained long-term trend features, cyclical features and short-term trend features are fused to obtain the final prediction results <emph>Y</emph>, as shown in Eq. (<reflink idref="bib12" id="ref49">12</reflink>):</p> <olist> <item>2</item> </olist> <p>Graph</p> <p>where the symbol || denotes a join operation. The goal of the traffic flow prediction task is to make the output of the model as close as possible to the true value, so <emph>L</emph>1 loss is chosen as the objective function. This is expressed in the following Eq. (<reflink idref="bib13" id="ref50">13</reflink>), whereindicates the actual value.</p> <p>13</p> <p>Graph</p> <hd id="AN0183073501-17">Experiments and analyses</hd> <p></p> <hd id="AN0183073501-18">Experimental setup</hd> <p>The runtime environment used in the experiments of this paper is shown in Table 1.</p> <p>Table 1 Experimental environment.</p> <p> <ephtml> &lt;table frame="hsides" rules="groups"&gt;&lt;thead&gt;&lt;tr&gt;&lt;th align="left"&gt;&lt;p&gt;Appliances&lt;/p&gt;&lt;/th&gt;&lt;th align="left"&gt;&lt;p&gt;Model number&lt;/p&gt;&lt;/th&gt;&lt;/tr&gt;&lt;/thead&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;CPU&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;11thGenIntel(R)Core(TM)i7-11700@2.50 GHz&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;GPU&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;NvidiaGeForceRTX3060&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;System&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;Linux&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;GraphIics Memory&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;12GB&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;CUDA&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;11.7.101&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;Torch&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.8.2&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt; </ephtml> </p> <p>Before model training, the original data set needs to be preprocessed. Firstly, the graph structure representation of the road network is constructed according to the address information of the sensor, which is mainly used to calculate the adjacency matrix. In this paper, it is defined by the sensor distance and the connectivity property between nodes. Then, the data set is divided into the training set, the validation set, and the test set in a ratio of 7:2:1, and the data is shred in a way that predicts the data of the next hour based on the historical data of 1 h. Based on the premise of external information fusion, and facilitate data management, this paper splices time, POI and weather attributes on the basis of historical sequences and corresponds to each other in the time dimension. The usable data for training and testing is finally obtained.</p> <p>In the experiments, each Batch is set to have a size of 64, with a total of 100 epochs, and the decay rate of course learning is updated every 2000 rounds. The length of data input and output is 12, the data dimension is 4, and the hidden dimension of long-term trend extractor and periodic extractor is set to 4. The STGNN used in the experimental part of this paper is Graph WaveNet [<reflink idref="bib33" id="ref51">33</reflink>]. During the training process, the initial learning rate is set to 0.01, the decay rate is 0.1, and the optimizer is trained using Adam, and the training model is validated every 5 epoch. The whole model training time is about 6 h, and each epoch takes about 4 min.</p> <hd id="AN0183073501-19">Data set</hd> <p>SZ-taxi dataset: It was collected from the taxi operation data system in Shenzhen and first used in the T-GCN network by Ling et<sups>[<reflink idref="bib16" id="ref52">16</reflink>]</sups>. The time span of the dataset is from 1st January 2015 to 31st January 2015, with a 15-minute interval, and includes speed values for 156 road nodes, as well as an adjacency matrix representing connections between nodes.</p> <p>SZ_POI: This dataset provides information about POIs around the selected road section. POI categories can be classified into 9 types, food and beverage services, businesses, shopping services, transport facilities, educational services, living services, medical services, accommodation and others. Calculating the distribution of POIs on each road segment, the POI type with the largest percentage is used as the characteristic of the road segment. Thus, the size of the obtained static attribute matrix is 156*1.</p> <p>SZ_Weather: The auxiliary information contains the weather conditions of the study area recorded in January 2015 every 15 min. The weather conditions were classified into five categories, sunny, cloudy, foggy, light rain and heavy rain. Using the time-varying weather information, a 156*2976 dynamic attribute matrix was constructed.</p> <hd id="AN0183073501-20">Evaluation indicators</hd> <p>The traffic flow prediction task is essentially a large-scale, characterised data regression simulation problem. In this paper, mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE) are used to evaluate the accuracy of prediction results. As shown in Eqs. (14–16)</p> <p>14</p> <p>Graph</p> <p>15</p> <p>Graph</p> <p>16</p> <p>Graph</p> <p>where <emph>n</emph> is the number of samples, the and are the predicted and true values of the i-th sample, respectively.</p> <hd id="AN0183073501-21">Analysis of experimental results</hd> <p>In order to test the effectiveness of the algorithm, nine representative algorithms were selected to compare the accuracy of the prediction results under different time steps. Among them, GWnet[<reflink idref="bib31" id="ref53">31</reflink>], STSGCN[<reflink idref="bib32" id="ref54">32</reflink>], AGCRN[<reflink idref="bib33" id="ref55">33</reflink>] and DSTET[<reflink idref="bib34" id="ref56">34</reflink>] are the newly published prediction methods, which have achieved better prediction accuracy in the original paper.</p> <p>Table 2 Experimental results of each model on the SZ-TAXI dataset.</p> <p> <ephtml> &lt;table frame="hsides" rules="groups"&gt;&lt;thead&gt;&lt;tr&gt;&lt;th align="left" /&gt;&lt;th align="left" colspan="3"&gt;&lt;p&gt;15 min&lt;/p&gt;&lt;/th&gt;&lt;th align="left" colspan="3"&gt;&lt;p&gt;30 min&lt;/p&gt;&lt;/th&gt;&lt;th align="left" colspan="3"&gt;&lt;p&gt;60 min&lt;/p&gt;&lt;/th&gt;&lt;/tr&gt;&lt;/thead&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td align="left" /&gt;&lt;td align="left"&gt;&lt;p&gt;MAE&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;RMSE&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;MAPE(%)&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;MAE&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;RMSE&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;MAPE(%)&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;MAE&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;RMSE&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;MAPE(%)&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;ARIMA&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.62&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.30&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.50&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.33&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.76&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;5.40&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.38&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;6.50&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;8.30&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;LSTM(FC)&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.05&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.19&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.80&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.20&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.55&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;5.20&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.37&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.96&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;5.70&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;DCRNN&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.38&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.95&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.90&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.74&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.97&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.90&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.07&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.74&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.90&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;STGCN&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.36&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.96&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.90&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.81&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.27&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.17&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.49&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;5.69&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;5.79&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;ASTGCN&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.52&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.13&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.22&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.01&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.27&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.48&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.61&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;5.42&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;6.00&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;GWnet&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.30&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.74&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.93&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.63&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;&lt;bold&gt;3.70&lt;/bold&gt;&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.67&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.95&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;&lt;bold&gt;4.52&lt;/bold&gt;&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.63&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;STSGCN&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.44&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.01&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.04&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.83&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.18&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.17&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.26&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;5.21&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;5.40&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;AGCRN&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.37&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.87&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.94&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.69&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.85&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.87&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.96&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.54&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.64&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;DSTET&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.31&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.84&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;2.92&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.62&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.82&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.65&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;1.95&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.55&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.62&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;Ours&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;&lt;bold&gt;1.28&lt;/bold&gt;&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;&lt;bold&gt;2.76&lt;/bold&gt;&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;&lt;bold&gt;2.83&lt;/bold&gt;&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;&lt;bold&gt;1.58&lt;/bold&gt;&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;3.72&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;&lt;bold&gt;3.55&lt;/bold&gt;&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;&lt;bold&gt;1.93&lt;/bold&gt;&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;4.62&lt;/p&gt;&lt;/td&gt;&lt;td align="left"&gt;&lt;p&gt;&lt;bold&gt;4.58&lt;/bold&gt;&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt; </ephtml> </p> <p>As can be seen from the Table 2 data, the model proposed in this paper performs well on most of the indicators in the Shenzhen taxi dataset. Compared with the second best performing model, the prediction accuracy is improved in all the three prediction steps, and the MAE, RMSE, and MAPE are improved by 28%, 19.7%, and 28.5% on average, respectively. The effectiveness of the algorithm design proposed in this paper in reducing the relative error and the strong long-term prediction ability are demonstrated.</p> <p>The STGCN and ASTGCN models perform well on the dataset, which is closely related to the characteristics of the dataset itself.The SZ dataset is collected in an area with high complexity of urban roads, which has a high degree of influence from external factors. The STGCN and ASTGCN models use multinomial Chebyshev map convolution, which can obtain richer feature information from the space. Although the algorithm designed in this paper does not use Chebyshev polynomials, it is better than the above models because it obtains the features of the high-order graph signal flow through the combination of graph convolution modules, and, at the same time, introduces the POI attribute, which enables the graph convolution process to make use of the hidden spatial features of the nodes.</p> <p>Another noteworthy point is that out of the nine baseline models, DSTET is the only one that takes periodicity into account when predicting traffic flow and shows the best performance. This result proves the importance of periodicity.One of the main differences between DSTET and MIFPN models is that MIFPN models have the ability to capture long term trends, and the advantage of MIFPN models over DSTET may be attributed to their introduction of long term trends for forecasting.</p> <p>DCRNN's diffusion convolution and temporal codec structure based on graph network has a simpler structure, but adopting the principle of end-to-end sequence prediction makes the network structure enough to learn the transformed features of the traffic, so the effect is slightly better. In 30 min and 60 min, the RMSE value of MIFPN is slightly higher than that of GWnet model, and the reason for this is that GWnet model adopts the non-fixed structure of graph network, which can learn the hidden edge relationship between the nodes in the whole graph, which indicates that this model has something special in learning the relationship between the nodes. The subsequent algorithm design can pay more attention in this aspect.</p> <p>In summary, the model proposed by the algorithm has improved its effect at most of the time points of the dataset. On the other hand, ASTGCN, due to the complexity of the model and the many parameters, although the attention network is used in both space and time, its prediction effect depends on the complexity of the road conditions, and the results are not obvious.The effect of adaptive neighbour matrix used by the two networks, Graph Wavenet and AGCRN, is also approaching the best in the dataset, which shows from the side that the relationship between the nodes is not static. How to extract the dynamically changing correlation between nodes, so that the model can improve the learning ability is also the main research direction of the subsequent research.</p> <hd id="AN0183073501-22">Ablation experiment</hd> <p>In order to verify the role of each module in the model in predicting the results, this section conducts an ablation experiment of the model on the SZ-taxi dataset. The degree of enhancement of each module on the prediction results and its theoretical basis is analysed by comparing the experiment with the complete model after module elimination and without elimination.</p> <p>Firstly, all external information (POI &amp; Weather) is ablated. The specific way of ablation is to use only the speed as an input based on the original model and will add only the vehicle speed as an input to the graph convolution layer for prediction.</p> <p>Next, the components added to the model are ablated to validate the model. The trend extractor, specifically ablated by removing the trend extractor module from the base of the model and using external information and extracted cycle information as inputs to get the prediction results. The cycle extractor, specifically ablated by removing the cycle extractor module on top of the model and using only external information and the trend extractor as inputs to generate prediction results. The experimental results are shown in Table 3.</p> <p>Table 3 Results of ablation experiments.</p> <p> <ephtml> &lt;table frame="hsides" rules="groups"&gt;&lt;thead&gt;&lt;tr&gt;&lt;th align="left"&gt;&lt;p&gt;Models&lt;/p&gt;&lt;/th&gt;&lt;th align="left"&gt;&lt;p&gt;MAE&lt;/p&gt;&lt;/th&gt;&lt;th align="left"&gt;&lt;p&gt;RMSE&lt;/p&gt;&lt;/th&gt;&lt;th align="left"&gt;&lt;p&gt;MAPE&lt;/p&gt;&lt;/th&gt;&lt;/tr&gt;&lt;/thead&gt;&lt;tbody&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;POI&amp;Weather&lt;/p&gt;&lt;/td&gt;&lt;td char="." align="char"&gt;&lt;p&gt;2.91&lt;/p&gt;&lt;/td&gt;&lt;td char="." align="char"&gt;&lt;p&gt;4.24&lt;/p&gt;&lt;/td&gt;&lt;td char="." align="char"&gt;&lt;p&gt;9.28&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;timeinday/week&lt;/p&gt;&lt;/td&gt;&lt;td char="." align="char"&gt;&lt;p&gt;2.88&lt;/p&gt;&lt;/td&gt;&lt;td char="." align="char"&gt;&lt;p&gt;4.26&lt;/p&gt;&lt;/td&gt;&lt;td char="." align="char"&gt;&lt;p&gt;9.31&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;subsequece&lt;/p&gt;&lt;/td&gt;&lt;td char="." align="char"&gt;&lt;p&gt;2.85&lt;/p&gt;&lt;/td&gt;&lt;td char="." align="char"&gt;&lt;p&gt;4.19&lt;/p&gt;&lt;/td&gt;&lt;td char="." align="char"&gt;&lt;p&gt;9.22&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;tr&gt;&lt;td align="left"&gt;&lt;p&gt;Full model&lt;/p&gt;&lt;/td&gt;&lt;td char="." align="char"&gt;&lt;p&gt;2.77&lt;/p&gt;&lt;/td&gt;&lt;td char="." align="char"&gt;&lt;p&gt;4.10&lt;/p&gt;&lt;/td&gt;&lt;td char="." align="char"&gt;&lt;p&gt;9.14&lt;/p&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/tbody&gt;&lt;/table&gt; </ephtml> </p> <p>The external information fusion model proposed by the algorithm improves the MAE metric by 4.81%, the RMSE by 3.3% and the MAPE by 1.51% on the SZ dataset, proving that the predictive ability of the model with the addition of the road information is by significantly improved. Long-term trend characteristics also improve the prediction effect on the dataset, in which MAE improves by 2.8%, RMSE improves by 2.15%, and MAPE improves by 0.86%. From the experimental results, it can be seen that the long-term trend characteristic effectively improves the prediction ability of the model. This is because in the long term, the data is showing a smooth upward and downward trend as well as staying constant. Cyclic characteristics are very obvious to improve the model prediction results, MAE improved by 3.8%, RMSE improved by 3.75%, and MAPE improved by 1.82%. Within urban roads, people and cars share lanes, resulting in a drastic degree of flow variation over time, and traffic flow information is highly cyclical within days and weeks.</p> <p>Through the ablation experiments on the dataset, it can be concluded that in traffic flow prediction, the flow information of the road network itself is affected by other factors outside the road network, and by incorporating these factors into the prediction model, the effect of prediction and the speed of convergence are effectively improved. In addition, the cycle module and the long-term trend extraction module designed in this paper both help to improve the model, in which the cycle module is more useful than the long-term trend extraction module. In addition, through experimental validation of this paper, it is found that the cycle module is helpful for the traffic flow prediction algorithm when data is missing.</p> <p>In order to demonstrate the role of the period module for the traffic flow prediction algorithm when data are missing, this paper selects the traffic flow prediction results of two nodes in the same time period in the SZ-taxi dataset for visualisation, as shown in Fig. Figure 7, (a) demonstrates the prediction of this paper's traffic flow prediction algorithm in the normal situation, and (b) represents the prediction task under the premise of partial data missing, i.e., part of the short-term traffic flow data is missing. As can be seen from the comparison of the data results, even in the case of partial data missing reality. The model can also predict the values of traffic flow attributes based on the periodic characteristics of traffic flow data. It shows that the periodicity extractor can effectively enhance the robustness of the model.</p> <p>Graph: Fig. 7 Results comparison chart.</p> <hd id="AN0183073501-23">Summary and outlook</hd> <p>The MIFPN traffic flow prediction algorithm proposed in this paper, the model model incorporates external information on the one hand, and on the other hand, the cycle extractor as well as the long-term trend extractor are designed. Through experimental verification, it is found that the proposed algorithm in this paper is greatly superior to machine learning algorithms, and compared with similar deep learning algorithms, there is also a great improvement. On the basis of enhancing the performance of the algorithm, this paper also explores the use of the period extractor to greatly improve the robustness of the model when part of the data is missing. For the case of missing data this paper explores the strength and depth of this paper is still to be explored, in addition to the long-term prediction of this paper is also to be studied in depth.</p> <hd id="AN0183073501-24">Acknowledgements</hd> <p>This work was supported by the Natural Science Foundation of Chongqing, China (Grant No. 2024NSCQ-MSX3950), the Natural Science Foundation of Chongqing Science &amp; Technology Commission (Grant No. cstc2021jcyj-msxmX0532), the Science and Technology Research Program of Chongqing Municipal Education Commission (Grant No. KJZD-K202303405, KJZD-M202203401, KJQN202303423, KJQN202403401, and KJQN202103101), and the Program for Innovation Research Groups at Institutions of Higher Education in Chongqing (Grant No. CXQT21032).</p> <hd id="AN0183073501-25">Author contributions</hd> <p>W. Z. and H.T wrote the main manuscript, W designed the experiments, Z participated in the experiments, H collected and organised the experimental data, T collected the relevant information and co-edited the final manuscript. W made the pictures of the experimental results, and W touched up the manuscript. All authors reviewed the manuscript.</p> <hd id="AN0183073501-26">Data availability</hd> <p>Due to the privacy implications of the team's work, such as patent applications, the datasets and codes generated and/or analyzed during the current research period are not publicly available, but are available from the corresponding authors upon reasonable request.</p> <hd id="AN0183073501-27">Declarations</hd> <p></p> <hd id="AN0183073501-28">Competing interests</hd> <p>The authors declare no competing interests.</p> <hd id="AN0183073501-29">Publisher's note</hd> <p>Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.</p> <ref id="AN0183073501-30"> <title> References </title> <blist> <bibl id="bib1" idref="ref1" type="bt">1</bibl> <bibtext> Wu C-H, Ho J-H, Lee D-T. Travel-time prediction with support vector regression. IEEE Trans. Intell. Transp. Syst. 2004; 5; 4: 276-281. 10.1109/TITS.2004.837813. 1466.62130</bibtext> </blist> <blist> <bibl id="bib2" idref="ref2" type="bt">2</bibl> <bibtext> Yuan, S. et al. The application of nonparametric regressive algorithm for short-term traffic flow forecast. 2009 First International Workshop on Education Technology and Computer Science3, 767–770 (2009).</bibtext> </blist> <blist> <bibl id="bib3" idref="ref3" type="bt">3</bibl> <bibtext> Lippi M, Bertini M, Frasconi P. Short-term traffic flow forecasting: an experimental comparison of time-series analysis and supervised learning. IEEE Trans. Intell. Transp. Syst. 2013; 14; 2: 871-882. 10.1109/TITS.2013.2247040. 1334.68186</bibtext> </blist> <blist> <bibl id="bib4" idref="ref4" type="bt">4</bibl> <bibtext> Shumway, R. H. et al. ARIMA models. Time series analysis and its applications: with R examples 75–163 (2017).</bibtext> </blist> <blist> <bibl id="bib5" idref="ref5" type="bt">5</bibl> <bibtext> Wu J. Traffic flow anomaly detection based on wavelet denoising and support vector regression. J. Algorithms Comput. Technol. 2013; 7; 2: 209-225. 10.1260/1748-3018.7.2.209. 1289.15032</bibtext> </blist> <blist> <bibl id="bib6" idref="ref6" type="bt">6</bibl> <bibtext> Zhang, A. Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. Association for Computing Machinery, (2022).</bibtext> </blist> <blist> <bibl id="bib7" idref="ref7" type="bt">7</bibl> <bibtext> Bai D. Spatial-temporal graph neural network based on gated convolution and topological attention for traffic flow prediction. Appl. Intell. 2023; 53; 24: 30843-30864. 10.1007/s10489-023-05053-x</bibtext> </blist> <blist> <bibl id="bib8" idref="ref8" type="bt">8</bibl> <bibtext> Guo, Y. &amp; ACM Digital Library. Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &amp; Data Mining (ACM, 2018).</bibtext> </blist> <blist> <bibl id="bib9" idref="ref9" type="bt">9</bibl> <bibtext> Wu, Z. et al. Graph wavenet for deep spatial-temporal graph modeling. arXiv preprint arXiv:1906.00121 (2019).</bibtext> </blist> <blist> <bibtext> He, S. et al. STGC-GNNs: A GNN-based Traffic Prediction Framework with a spatial–temporal Granger Causality Graph 623128913 (Statistical Mechanics and its Applications, 2023).</bibtext> </blist> <blist> <bibtext> Sharma A. A graph neural network (GNN)-based approach for real-time estimation of traffic speed in sustainable smart cities. Sustainability. 2023; 15; 15: 11893. 10.3390/su151511893. 1485.74015</bibtext> </blist> <blist> <bibtext> Zhang D, Kabuka MR. Combining weather condition data to predict traffic flow: a GRU-based deep learning approach. IET Intel. Transport Syst. 2018; 12; 7: 578-585. 10.1049/iet-its.2017.0313. 0989.93095</bibtext> </blist> <blist> <bibtext> Defferrard, M, Bresson, X. &amp; Vandergheynst, P. Convolutional neural networks on graphs with fast localized spectral filtering. Adv. Neural. Inf. Process. Syst. 29 (2016).</bibtext> </blist> <blist> <bibtext> Zhao L. T-GCN: a temporal graph convolutional network for traffic prediction. IEEE Trans. Intell. Transp. Syst. 2019; 21; 9: 3848-3858. 10.1109/TITS.2019.2935152. 1429.65210</bibtext> </blist> <blist> <bibtext> Bai J. A3t-gcn: attention temporal graph convolutional network for traffic forecasting. ISPRS Int. J. Geo-Information. 2021; 10; 7: 485. 2021IJGI.10.485B. 10.3390/ijgi10070485. 1026.62042</bibtext> </blist> <blist> <bibtext> Gilmer, J. et al. Neural message passing for quantum chemistry. International conference on machine learning. PMLR 1263–1272 (2017).</bibtext> </blist> <blist> <bibtext> Velickovic P. Graph Atten. Networks stat. 2017; 1050; 20: 10-48550</bibtext> </blist> <blist> <bibtext> Wang Z. Spatiotemporal Fusion Transformer for large-scale traffic forecasting. Inform. Fusion. 2024; 107: 102293. 10.1016/j.inffus.2024.102293</bibtext> </blist> <blist> <bibtext> Ahmed SF. Insights into internet of medical things (IoMT): data fusion, security issues and potential solutions. Inform. Fusion. 2024; 102: 102060. 10.1016/j.inffus.2023.102060</bibtext> </blist> <blist> <bibtext> Zhang, J. et al. Spatio-temporal pre-training enhanced fast pure tansformer network for traffic flow forecasting. 2023 International Joint Conference on Neural Networks (IJCNN) 1–8 (IEEE, 2023).</bibtext> </blist> <blist> <bibtext> Cai L. Traffic transformer: capturing the continuity and periodicity of time series for traffic forecasting. Trans. GIS. 2020; 24; 3: 736-755. 10.1111/tgis.12644. 07887911</bibtext> </blist> <blist> <bibtext> Park, C. et al. ST-GRAT: A novel spatio-temporal graph attention networks for accurately forecasting dynamically changing road speed. Proceedings of the 29th ACM international conference on information &amp; knowledge management 1215–1224 (2020).</bibtext> </blist> <blist> <bibtext> Shao, H. Deep Learning Approaches for Traffic Prediction (Nanyang Technological University, 2020).</bibtext> </blist> <blist> <bibtext> Roy, A. et al. Unified spatio-temporal modeling for traffic forecasting using graph neural network. 2021 International Joint Conference on Neural Networks (IJCNN) 1–8 (IEEE, 2021).</bibtext> </blist> <blist> <bibtext> He, K. et al. Proceedings of the IEEE conference on computer vision and pattern recognition. Going deeper with convolutions 1–9 (2016).</bibtext> </blist> <blist> <bibtext> Nie, Y. et al. A time series is worth 64 words: long-term forecasting with transformers. arXiv preprint arXiv:2211.14730 (2022).</bibtext> </blist> <blist> <bibtext> Luo Q. LSTTN: a long-short term transformer-based spatiotemporal neural network for traffic flow forecasting. Knowl. Based Syst. 2024; 293: 111637. 10.1016/j.knosys.2024.111637. 1537.93040</bibtext> </blist> <blist> <bibtext> Li, Z. et al. Ti-mae: self-supervised masked time series autoencoders. arXiv Preprint arXiv:2301.08871 (2023).</bibtext> </blist> <blist> <bibtext> Liu J. STGHTN: spatial-temporal gated hybrid transformer network for traffic flow forecasting. Appl. Intell. 2023; 53; 10: 12472-12488. 10.1007/s10489-022-04122-x. 1454.93065</bibtext> </blist> <blist> <bibtext> Devlin, J. &amp; Bert Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).</bibtext> </blist> <blist> <bibtext> Chen S. Gate-based GWNet for process quality filter and multioutput prediction. Expert Syst. Appl. 2025; 264: 125921. 10.1016/j.eswa.2024.125921. 1508.93286</bibtext> </blist> <blist> <bibtext> Yu X, Bao Y, Shi Q. Spatial-temporal synchronous graphsage for traffic prediction. Appl. Intell. 2025; 55; 1: 1-17. 10.1007/s10489-024-05970-5. 1205.92025</bibtext> </blist> <blist> <bibtext> Gao X. An AGCRN Algorithm for pressure prediction in an Ultra-long Mining Face in a medium–thick coal Seam in the Northern Shaanxi Area, China. Appl. Sci. 2023; 13; 20: 11369. 1:CAS:528:DC%2BB3sXit1elu7fP. 10.3390/app132011369. 1280.33016</bibtext> </blist> <blist> <bibtext> Sun W. Transformer network with decoupled spatial–temporal embedding for traffic flow forecasting. Appl. Intell. 2023; 53; 24: 30148-30168. 10.1007/s10489-023-05126-x</bibtext> </blist> </ref> <aug> <p>By Xiao Wu; Hua Huang; Tong Zhou; Yudan Tian; Shisen Wang and Jingting Wang</p> <p>Reported by Author; Author; Author; Author; Author; Author</p> </aug> <nolink nlid="nl1" bibid="bib10" firstref="ref10"></nolink> <nolink nlid="nl2" bibid="bib11" firstref="ref11"></nolink> <nolink nlid="nl3" bibid="bib12" firstref="ref12"></nolink> <nolink nlid="nl4" bibid="bib13" firstref="ref13"></nolink> <nolink nlid="nl5" bibid="bib14" firstref="ref14"></nolink> <nolink nlid="nl6" bibid="bib15" firstref="ref15"></nolink> <nolink nlid="nl7" bibid="bib16" firstref="ref16"></nolink> <nolink nlid="nl8" bibid="bib17" firstref="ref17"></nolink> <nolink nlid="nl9" bibid="bib18" firstref="ref18"></nolink> <nolink nlid="nl10" bibid="bib19" firstref="ref19"></nolink> <nolink nlid="nl11" bibid="bib20" firstref="ref20"></nolink> <nolink nlid="nl12" bibid="bib21" firstref="ref21"></nolink> <nolink nlid="nl13" bibid="bib22" firstref="ref22"></nolink> <nolink nlid="nl14" bibid="bib23" firstref="ref23"></nolink> <nolink nlid="nl15" bibid="bib24" firstref="ref24"></nolink> <nolink nlid="nl16" bibid="bib25" firstref="ref25"></nolink> <nolink nlid="nl17" bibid="bib26" firstref="ref26"></nolink> <nolink nlid="nl18" bibid="bib27" firstref="ref27"></nolink> <nolink nlid="nl19" bibid="bib28" firstref="ref28"></nolink> <nolink nlid="nl20" bibid="bib29" firstref="ref29"></nolink> <nolink nlid="nl21" bibid="bib30" firstref="ref37"></nolink> <nolink nlid="nl22" bibid="bib33" firstref="ref51"></nolink> <nolink nlid="nl23" bibid="bib31" firstref="ref53"></nolink> <nolink nlid="nl24" bibid="bib32" firstref="ref54"></nolink> <nolink nlid="nl25" bibid="bib34" firstref="ref56"></nolink> CustomLinks: – Url: https://resolver.ebsco.com/c/xy5jbn/result?sid=EBSCO:edsdoj&genre=article&issn=20452322&ISBN=&volume=15&issue=1&date=20250201&spage=1&pages=1-21&title=Scientific Reports&atitle=An%20indoor%20positioning%20method%20based%20on%20bluetooth%20array%2FPDR%20fusion%20using%20the%20SVD-EKF&aulast=Chenhui%20Li&id=DOI:10.1038/s41598-025-88860-1 Name: Full Text Finder (for New FTF UI) (s8985755) Category: fullText Text: Find It @ SCU Libraries MouseOverText: Find It @ SCU Libraries – Url: https://doaj.org/article/1caa281440664c1e87264a56660404da Name: EDS - DOAJ (s8985755) Category: fullText Text: View record from DOAJ MouseOverText: View record from DOAJ |
---|---|
Header | DbId: edsdoj DbLabel: Directory of Open Access Journals An: edsdoj.1caa281440664c1e87264a56660404da RelevancyScore: 1082 AccessLevel: 3 PubType: Academic Journal PubTypeId: academicJournal PreciseRelevancyScore: 1081.56311035156 |
IllustrationInfo | |
Items | – Name: Title Label: Title Group: Ti Data: An indoor positioning method based on bluetooth array/PDR fusion using the SVD-EKF – Name: Author Label: Authors Group: Au Data: <searchLink fieldCode="AR" term="%22Chenhui+Li%22">Chenhui Li</searchLink><br /><searchLink fieldCode="AR" term="%22Jie+Zhen%22">Jie Zhen</searchLink><br /><searchLink fieldCode="AR" term="%22Jianxin+Wu%22">Jianxin Wu</searchLink> – Name: TitleSource Label: Source Group: Src Data: Scientific Reports, Vol 15, Iss 1, Pp 1-21 (2025) – Name: Publisher Label: Publisher Information Group: PubInfo Data: Nature Portfolio, 2025. – Name: DatePubCY Label: Publication Year Group: Date Data: 2025 – Name: Subset Label: Collection Group: HoldingsInfo Data: LCC:Medicine<br />LCC:Science – Name: Subject Label: Subject Terms Group: Su Data: <searchLink fieldCode="DE" term="%22Medicine%22">Medicine</searchLink><br /><searchLink fieldCode="DE" term="%22Science%22">Science</searchLink> – Name: Abstract Label: Description Group: Ab Data: Abstract With the development of mobile internet and artificial intelligence, the core of current surveying and mapping science and technology is no longer confined solely to outdoor applications. High-precision indoor positioning technology is now one of the core technologies in the era of artificial intelligence. An angle measurement and positioning system based on wireless signal array antennas can achieve high accuracy in indoor unblocked conditions. However, indoor environment is unpredictable, and the user's behavior is also random. Inevitably, some signal reflection and other factors will affect the positioning accuracy. Considering all the aforementioned, this study analyzes the principles and characteristics of a Bluetooth signal based array antenna angle measurement and positioning system. In addition, aiming at the multi-phase problem caused by antenna switching, the frequency estimation method based on FFT (Fast Fourier Transform) is studied in this paper, achieving high-precision angle measurement and positioning. Aiming at the problem that the system positioning error increase in the complex and variable indoor environment, a fusion positioning method of Bluetooth array/PDR (Pedestrian Dead Reckoning) based on the SVD-EKF (Singular Value Decomposition–Extended Kalman Filter) is proposed. This study introduces a few improvements to the EKF (Extended Kalman Filter). First, the predicted state covariance matrix is decomposed by singular value decomposition, which improves the robustness of the EKF. Second, a Bluetooth array self-evaluation factor is introduced and combined with the Huber function to construct an adaptive factor, thus further enhancing filtering accuracy and environmental adaptability. Through static and dynamic data collected in indoor environment, the feasibility of the algorithm is verified. The results of static and dynamic experimental results show that the array angle measurement and positioning system can achieve high accuracy, the near point positioning maximum error is 0.3m and the far point positioning accuracy is 0.6m. The dynamic test results in the room show that after SVD-EKF algorithm optimization, the positioning error is reduced by 0.05m, which is equivalent to the EKF algorithm. However, in corridor areas, the improvement of accuracy of SVD-EKF algorithm is better than EKF, achieving an improvement of 0.373 m and a smoother positioning result compared to the traditional EKF algorithm. This study provides a new practical technology with a high precision and easy deployment for indoor positioning. – Name: TypeDocument Label: Document Type Group: TypDoc Data: article – Name: Format Label: File Description Group: SrcInfo Data: electronic resource – Name: Language Label: Language Group: Lang Data: English – Name: ISSN Label: ISSN Group: ISSN Data: 2045-2322<br />28144066 – Name: NoteTitleSource Label: Relation Group: SrcInfo Data: https://doaj.org/toc/2045-2322 – Name: DOI Label: DOI Group: ID Data: 10.1038/s41598-025-88860-1 – Name: URL Label: Access URL Group: URL Data: <link linkTarget="URL" linkTerm="https://doaj.org/article/1caa281440664c1e87264a56660404da" linkWindow="_blank">https://doaj.org/article/1caa281440664c1e87264a56660404da</link> – Name: AN Label: Accession Number Group: ID Data: edsdoj.1caa281440664c1e87264a56660404da |
PLink | https://login.libproxy.scu.edu/login?url=https://search.ebscohost.com/login.aspx?direct=true&site=eds-live&scope=site&db=edsdoj&AN=edsdoj.1caa281440664c1e87264a56660404da |
RecordInfo | BibRecord: BibEntity: Identifiers: – Type: doi Value: 10.1038/s41598-025-88860-1 Languages: – Text: English PhysicalDescription: Pagination: PageCount: 21 StartPage: 1 Subjects: – SubjectFull: Medicine Type: general – SubjectFull: Science Type: general Titles: – TitleFull: An indoor positioning method based on bluetooth array/PDR fusion using the SVD-EKF Type: main BibRelationships: HasContributorRelationships: – PersonEntity: Name: NameFull: Chenhui Li – PersonEntity: Name: NameFull: Jie Zhen – PersonEntity: Name: NameFull: Jianxin Wu IsPartOfRelationships: – BibEntity: Dates: – D: 01 M: 02 Type: published Y: 2025 Identifiers: – Type: issn-print Value: 20452322 – Type: issn-print Value: 28144066 Numbering: – Type: volume Value: 15 – Type: issue Value: 1 Titles: – TitleFull: Scientific Reports Type: main |
ResultId | 1 |