CN114493014A - Multivariate time series prediction method, multivariate time series prediction system, computer product and storage medium - Google Patents

Multivariate time series prediction method, multivariate time series prediction system, computer product and storage medium Download PDF

Info

Publication number
CN114493014A
CN114493014A CN202210107028.XA CN202210107028A CN114493014A CN 114493014 A CN114493014 A CN 114493014A CN 202210107028 A CN202210107028 A CN 202210107028A CN 114493014 A CN114493014 A CN 114493014A
Authority
CN
China
Prior art keywords
time series
vector
space
term
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210107028.XA
Other languages
Chinese (zh)
Inventor
谢鲲
刘丹
陈小迪
张大方
文吉刚
李肯立
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hunan University
Original Assignee
Hunan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hunan University filed Critical Hunan University
Priority to CN202210107028.XA priority Critical patent/CN114493014A/en
Publication of CN114493014A publication Critical patent/CN114493014A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biophysics (AREA)
  • Strategic Management (AREA)
  • Economics (AREA)
  • Human Resources & Organizations (AREA)
  • Game Theory and Decision Science (AREA)
  • Development Economics (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Marketing (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Tourism & Hospitality (AREA)
  • General Business, Economics & Management (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a multivariate time sequence prediction method, a multivariate time sequence prediction system, a computer product and a storage medium, wherein two feature extraction codes are used for respectively extracting space-time feature vectors of long-term and short-term historical data matrixes, the historical time sequence matrixes are input into a space feature extraction coder to generate weighted attention space feature vectors, and the weighted space feature vectors are input into a gating circulation unit to generate the space-time feature vectors; inputting the space-time feature vector extracted by the long-term historical data matrix into an interactive attention module to generate a weighted feature vector; inputting the short-term historical data matrix into an autoregressive layer to generate a linear prediction result of the short-term historical time sequence data; and combining the weighted feature vector and the coding feature vector, inputting the weighted feature vector and the coding feature vector into a full-link layer to generate a neural network prediction result, and adding the neural network prediction result and an autoregressive layer linear prediction result to obtain a final prediction result. The invention realizes the accurate prediction of the multivariate time series data.

Description

Multivariate time series prediction method, multivariate time series prediction system, computer product and storage medium
Technical Field
The invention relates to the field of multivariate time series data prediction, in particular to a multivariate time series prediction method and a multivariate time series prediction system based on a feature extraction coding and interaction attention module.
Background
With the development of big data technology and the rapid increase of data, there are wide application scenarios for predicting the future state of the big data technology by using time series data, such as traffic flow prediction on traffic lines, stock price prediction on stock markets, and air quality index prediction in different cities. The accurate prediction of new trends or potential events is often the content of real interest to the user, and provides powerful support for future decision making and planning, which is helpful for the implementation of advanced applications. However, problems with complex periodic patterns and dependencies in time series data are not well modeled. The industry has conducted a deep study of the above problems.
Research shows that time series data have complex periodic patterns and time and space dependencies, so that the short-term and long-term periodic patterns are accurately mined, the time and space dependencies are learned and information from other variables is effectively combined by extracting the internal relation among historical time series, and the accurate time series prediction is still a challenging task.
In order to solve the problem of time series prediction, many researches and methods for predicting a sequence have been conducted. In summary, there are two main categories:
the first category of methods is the traditional time series prediction model. Generally referred to as statistical models for time series analysis/prediction[1,2,3,4]For example, there are common methods such as mean regression, autoregressive moving average, and exponential smoothing prediction. The method mainly extracts the trend of historical data through modeling analysis of the historical data, and finally obtains the change of the demand in the future period of time through prediction of the trend. The method has the advantages of low complexity and high calculation speed, can provide a reasonable range of a prediction result, and helps to carry out post calibration when the black box model finally outputs the result by using the reasonable range, so that the prediction system is more stable.
However, these traditional regression-based models typically assume that the time series conform to some distribution or functional form, and then analyze and predict based thereon. This analysis method cannot capture complex non-linear relationships in the sequence and many datasets do not perform satisfactorily. Furthermore, conventional models based on functional distributions are not suitable for multivariate sequence analysis. Models like autoregressive integrated moving average (ARIMA) can only capture the distribution information of single sequence history data, and cannot model complex patterns or dependencies in real data and capture and integrate relationships between sequences.
With the improvement of computing power and the development of machine learning theory, the deep neural network gets more and more attention in various fields, and the second category of research direction of extracting features and learning data correlation for prediction by using a machine learning method becomes popular[5]. A common processing method is to use a self-encoder[6]Convolutional neural network[7]Adding a recurrent neural network[8,9](e.g., gated cycle units), etc. The self-encoder and the convolutional neural network are used for capturing the internal spatial relationship of the time sequence data at each moment, the gating cycle unit is used for learning the time relationship of the time sequence data, and finally the time-space relationship is integrated for prediction.
However, conventional recurrent neural networks vanish due to the presence of gradients[10]The problem of (2) is not well dealing with long sequences and tasks that require long-term historical information. To address the problem of gradient disappearance, researchers have proposed short-term and long-term memory network (LSTM) models and gated cyclic units (GRU) models. The latter has similar effects but is computationally less intensive and is therefore widely used in research today. An attention mechanism is introduced in subsequent studies to allow the model to selectively focus on a portion of all information while ignoring other visible information.
Hybrid neural networks, as compared to single neural networks[11,12,13]Has better performance. Finally, and most importantly, we note that there are complex periodic patterns and spatio-temporal correlations in time series data, employing multitaskingThe learning method can effectively improve the generalization and the accuracy of the model. The existing prediction algorithm is usually insufficient in the capturing capability of trend features and multi-period modes among multivariate time sequences, for example, in a TCN multivariate time sequence prediction method based on a parallel space-time attention mechanism, in a module for obtaining multivariate time sequence data space-time features, complex space-time correlation among historical time sequence data cannot be obtained only by performing simple linear transformation and normalization on original input, and a TCN needs a complete time sequence to predict.
[1]Elvin Isufifi,Andreas Loukas,Nathanael Perraudin,and Geert Leus.Forecasting time series with varma recursions on graphs.IEEE Transactions on Signal Processing,67(18):4870–4885,2019.
[2]Helmut Lutkepohl.New introduction to multiple time series analysis.Springer Science&Business Media,2005.
[3]Jiahan Li and Weiye Chen.Forecasting macroeconomic time series:Lasso-based approaches and their forecast combinations with dynamic factor models.International Journal of Forecasting,30(4):996–1015,2014.
[4]George EP Box,Gwilym M Jenkins,Gregory C Reinsel,and Greta M Ljung.Time series analysis:forecasting and control.John Wiley&Sons,2015.
[5]Liu,Hanpeng,et al.“CoSTCo:A Neural Tensor Completion Model for Sparse Tensors.”Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery&Data Mining,2019,pp.324–334.
[6]Fan,Jicong,and Tommy W.S.Chow.“Deep Learning Based Matrix Completion.”Neurocomputing,vol.266,2017,pp.540–549.
[7]Lecun Y,Bottou L.Gradient-based learning applied to document recognition[J].Proceedings of the IEEE,1998,86(11):2278-2324.
[8]Zhuoning Yuan,Xun Zhou,and Tianbao Yang.Hetero-convlstm:A deep learning approach to traffiffiffic accident prediction on heterogeneous spatio-temporal data.In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery&Data Mining,pages 984–992,2018.
[9]Xu Zhang,Furao Shen,Jinxi Zhao,and GuoHai Yang.Time series forecasting using gru neural network with multi-lag after decomposition.In International Conference on Neural Information Processing,pages 523–532.Springer,2017.
[10]Yoshua Bengio,Patrice Simard,and Paolo Frasconi.Learning long-term dependencies with gradient descent is difficult.IEEE transactions on neural networks,5(2):157–166,1994.
[11]Guokun Lai,Wei-Cheng Chang,Yiming Yang,and Hanxiao Liu.Modeling long-and short-term temporal patterns with deep neural networks.In The 41st International ACM SIGIR Conference on Research&Development in Information Retrieval,pages 95–104,2018.
[12]Huaxiu Yao,Xianfeng Tang,Hua Wei,Guanjie Zheng,and Zhenhui Li.Revisiting spatial-temporal similarity:A deep learning framework for traffiffiffic prediction.In Proceedings of the AAAI conference on artifificial intelligence,volume 33,pages 5668–5675,2019.
[13]Zhenxiong Yan,Kun Xie,Xin Wang,Dafang Zhang,Gaogang Xie,KenliLi,and Jigang Wen.Multivariate time series forecasting exploitingtensor projection embedding and gated memory network.In29thIEEE/ACM International Symposium on Quality of Service,IWQOS2021,Tokyo,Japan,June 25-28,2021,pages 1–6.IEEE,2021.
Disclosure of Invention
The technical problem to be solved by the invention is to provide a multivariate time series prediction method, a multivariate time series prediction system, a computer product and a storage medium aiming at the defects of the prior art, overcome the problems that the prior art cannot better learn complex periodic patterns, time and space dependencies and has low prediction precision, and realize the accurate prediction of multivariate time series data with dynamically changing time-space relationship.
In order to solve the technical problems, the technical scheme adopted by the invention is as follows: a multivariate time series prediction method comprises the following steps:
s1, historical time-series data [ X ]k-nT,...,Xk-1]Partitioning into long-term historical time series data [ X ]k -nT,...,Xk-T-1]And short-term historical time series data [ X ]k-T,...,Xk-1](ii) a T is the length of the long-term historical time sequence data or the short-term historical time sequence data;
s2, using the long-term historical time-series data [ X ]k-nT,...,Xk-T-1]And short-term historical time series data [ X ]k-T,...,Xk-1]Obtaining a weighted attention space feature vector of a long-term historical time series { [ z ]k-nT,...,zk -(n-1)T-1],...,[zk-2T,...,zk-T-1]Weighted attention space feature vector [ z ] of short-term historical time seriesk -T,...,zk-1](ii) a Short-term historical time-series data [ X ]k-T,...,Xk-1]Inputting the autoregressive layer to obtain a prediction matrix
Figure BDA0003493749250000041
S3, utilizing the acquired long-term historical time series weighted attention space feature vector { [ z ]k-nT,...,zk -(n-1)T-1],...,[zk-2T,...,zk-T-1]Weighted attention space feature vector [ z ] of short-term historical time seriesk -T,...,zk-1]Respectively obtaining a feature code vector sequence { m) containing space-time correlationiAnd a eigen-code vector u containing spatio-temporal correlations;
s4, mixing { miInputting an interactive attention network with u to obtain attention weight distribution directionQuantity piBy calculating piAnd { miThe product of which yields a weighted feature vector oi
S5, weighting characteristic vector o with size of T multiplied by diCombining the feature coding vector u with the size of 1 × d and containing space-time correlation into a new vector with the size of (T +1) × d, inputting the combined new vector into a generation model, and generating a nonlinear prediction matrix
Figure BDA0003493749250000042
S6, generating a prediction matrix of model output
Figure BDA0003493749250000043
Prediction matrix with autoregressive layer output
Figure BDA0003493749250000044
And adding to obtain the final prediction result.
The invention uses the characteristic extraction encoder to obviously improve the characteristic learning speed; the relevance among the space-time characteristic vectors can be captured as much as possible through the interactive attention network, effective information contained in historical data is greatly reserved, and accurate prediction of the multivariate time sequence data can be achieved. The method solves the problems that the prior art cannot better learn complex periodic patterns, time and space dependencies and has low prediction precision, and realizes accurate prediction of the multi-element time sequence data with dynamically changing time-space relationship. The specific implementation process of step S2 includes: long-term historical time series data [ X ]k-nT,...,Xk-T-1]And short-term historical time series data [ X ]k-T,...,Xk-1]Obtaining a space feature vector [ e ] of a long-term historical time sequence through a first space feature extractor and a second space feature extractor respectivelyk-nT,...,ek-T-1]And a short-term historical time series of spatial feature vectors [ ek-T,...,ek-1]Will [ e ]k-nT,...,ek-T-1]And [ ek-T,...,ek-1]Inputting the first attention layer and the second attention layer respectively to obtain the long termWeighted attention space feature vector of historical time series { [ z ]k-nT,...,zk -(n-1)T-1],...,[zk-2T,...,zk-T-1]Weighted attention space feature vector [ z ] of short-term historical time seriesk -T,...,zk-1]. Because the matrix sequences contain complex space-time correlation, the spatial relation of the adjacent positions of the matrix can be extracted through the spatial feature extractor, and the influence between different position data can be further acquired by adopting an attention mechanism, so that the weighted attention spatial feature vector is acquired.
The specific implementation process of step S3 includes: weighted attention space feature vector of long-term history time series { [ z ] to be acquiredk-nT,...,zk-(n-1)T-1],...,[zk-2T,...,zk-T-1]Weighted attention space feature vector [ z ] of short-term historical time seriesk-T,...,zk-1]Sequentially inputting a first gating circulation unit and a second gating circulation unit according to the time sequence, and respectively outputting a characteristic coding vector sequence { m } containing space-time correlationiAnd a feature code vector u containing spatio-temporal correlations. The gate control circulation unit can capture a nonlinear time relation, and the obtained space-time characteristic vectors retain hidden space-time information in historical data.
oi=pi×mi;pi=Softmax(uTmi)。
The obtained characteristic coding vector sequence containing space-time correlation { miAnd inputting the feature coding vector u containing the space-time correlation into an interactive attention network, and better capturing the relation between the space-time feature vectors output by the prediction encoder. Obtaining an attention weight distribution vector p by calculating an inner product of feature vectors and Softmax ()iAnd a weighted feature vector Oi. According to the method, the mutual dependency relationship between the characteristic vector matrixes is further mined and dynamically captured through the interactive attention network, so that the finally obtained space-time characteristic vector can greatly retain effective information contained in historical data.
In step S5, the generative model includes a plurality of cascaded fully connected layers. And performing weighted mixing on the historical time sequence data through the weight matrix, introducing an activation function and adding a nonlinear relation, and finally extracting the relation between different elements of the matrix. Theoretically, the more full connection layers, the stronger the extraction capability of the feature extraction network.
The invention also provides a multivariate time series prediction system, which comprises:
a first feature extraction encoder for inputting the long-term history time-series data [ X ]k-nT,...,Xk-T-1]The output is a characteristic code vector sequence { m) containing space-time correlationi};
A second feature extraction encoder for inputting a weighted attention space feature vector [ z ] of the short-term historical time seriesk -T,...,zk-1]Outputting a characteristic coding vector u containing space-time correlation;
autoregressive layer, input as short-term historical time series data [ X ]k-T,...,Xk-1]The output is a prediction matrix
Figure BDA0003493749250000061
An interactive attention network, the input is a characteristic coding vector sequence { m ] containing space-time correlationiAnd a feature coded vector u containing a spatio-temporal correlation, outputting a new vector of (T +1) × d;
generating a model, inputting a new vector of (T +1) x d, and outputting a nonlinear prediction matrix
Figure BDA0003493749250000062
A prediction module for generating a prediction matrix of the model output
Figure BDA0003493749250000063
Prediction matrix with autoregressive layer output
Figure BDA0003493749250000064
Adding to obtain a final prediction result;
wherein the long-term history time-series data [ Xk-nT,...,Xk-T-1]And short-term historical time series data [ X ]k -T,...,Xk-1]From historical time series data [ X ]k-nT,...,Xk-1]Dividing to obtain; and T is the length of the long-term historical time sequence data or the short-term historical time sequence data.
According to the multivariate time sequence prediction system provided by the invention, firstly, the spatial relationship of the adjacent positions of the matrix can be extracted through the space-time information contained in the matrix sequence by the characteristic extraction encoder, the incidence relationship of data among different time dimensions can be further captured, the nonlinear time relationship can be captured, and the obtained space-time characteristic vectors retain the space-time information hidden in the historical data; and further mining and dynamically capturing the interdependence relation between the characteristic vector matrixes through an interactive attention network, so that the finally obtained space-time characteristic vector can greatly keep effective information contained in historical data. According to the method, the model effect is generated by integrating the space-time characteristic vector further captured and the space characteristic vector information at the current moment and training through the loss function, and the prediction result at the future moment is finally generated by combining the linear prediction result output by the autoregressive layer, so that the problems that the prior art cannot better learn complex periodic patterns, time and space dependence and low prediction precision are solved, and the accurate prediction of the multi-element time sequence data with the dynamic change of the space-time relationship is realized. .
The first feature extraction encoder and the second feature extraction encoder have the same structure; wherein the first feature extraction encoder includes:
a first spatial feature extractor for inputting the long-term historical time-series data [ X ]k-nT,...,Xk-T-1]The output is a space characteristic vector [ e ] of the long-term historical time sequencek-nT,...,ek-T-1];
First attention level, input of [ e ]k-nT,...,ek-T-1]And outputting a weighted attention space feature vector { [ z ] of the long-term historical time seriesk-nT,...,zk-(n-1)T-1],...,[zk-2T,...,zk-T-1]};
A first gated cyclic unit with the input of weighted attention space feature vector { [ z ] of long-term history time seriesk -nT,...,zk-(n-1)T-1],...,[zk-2T,...,zk-T-1]The output is a characteristic coding vector sequence { m) containing space-time correlationi}。
The adoption of the feature extraction encoder can obviously improve the learning speed of the model on the space-time correlation of the sequence data. The spatiotemporal relationship of the adjacent positions of the matrix can be extracted through spatiotemporal information contained in the matrix sequence by the characteristic extraction encoder and the gating circulation unit, and the incidence relationship of data among different time dimensions can be further captured by adopting the attention layer. The gate control circulation unit can capture a nonlinear time relation, and the obtained space-time characteristic vectors retain hidden space-time information in historical data.
The interactive attention network is configured to perform operations comprising: will { miUsing the u and the u as the input of the self-attention layer to obtain the attention weight value distribution vector piBy calculating piAnd { miThe product of which yields a weighted feature vector oiWeighted feature vector o of size T × diAnd a feature code vector u of size 1 × d containing spatio-temporal correlation is combined into a new vector of size (T +1) × d.
As one inventive concept, the present invention also provides a computer-readable storage medium including a program running in a processor; the program is configured for performing the steps of the method of the invention.
As an inventive concept, the present invention also provides a computer program product comprising computer programs/instructions; which when executed by a processor implement the steps of the method of the present invention.
In the invention, the long-term time sequence historical data sequence can be divided into a plurality of groups, for example, the long-term time sequence historical data sequence can be divided into 7 groups, and a sliding window can be set for the long-term time sequence historical data sequence, so that the long-term and short-term time sequences are all T according to the unit length of a specified window (because the input feature extraction encoders have the same structure, in order to ensure that the input lengths are all T, different groups of data can be input by setting the sliding window).
Compared with the prior art, the invention has the beneficial effects that: the invention arranges the historical time sequence data matrix into a sequence one by one according to the time sequence, and the sequence contains the spatial correlation and the time correlation among the data. The historical time series data is divided into a long-term time series historical data sequence and a short-term time series historical data sequence, and the long-term time series historical data sequence and the short-term time series historical data sequence are arranged in time sequence to be used as input of the feature extraction codes. A multivariate time sequence prediction method based on feature extraction coding and an interactive attention module firstly extracts space-time information contained in a matrix sequence through a feature extraction coder and a gating circulation unit. In the process, the feature extraction encoder can extract the spatial relationship of the adjacent positions of the matrix, the attention layer can further capture the incidence relationship of data among different time dimensions, the gating circulation unit can capture the nonlinear time relationship, and the obtained space-time feature vectors retain the hidden space-time information in the historical data; according to the method, the interdependency relation between the characteristic vector matrixes is further mined and dynamically captured through an interactive attention network, so that the finally obtained space-time characteristic vector can greatly keep effective information contained in historical data; then, the method generates a model effect by integrating the space-time characteristic vector further captured and the space characteristic vector information at the current moment and training through a composite loss function, and finally generates a prediction result at the future moment by combining the linear prediction result input by the autoregressive layer; in the process, compared with other algorithms, the representation learning speed can be remarkably improved by using the space-time predictive coding, and the convergence speed of the generated model can be remarkably accelerated by adopting the relative root mean square error loss function and the empirical correlation coefficient combined training; the invention can input long-term historical data into the characteristic extraction encoder from the memory component, and can find that when new moment data comes temporarily, the data filling can be completed only by executing a plurality of steps of the method, rather than training the model parameters from the beginning like other algorithms, so the time complexity and the calculation complexity of the method are lower; by introducing the interactive attention network, the relevance among the space-time characteristic vectors can be captured as much as possible, effective information contained in historical data is greatly reserved, and accurate prediction of the multivariate time series data can be realized. The method adopts two independent feature extraction encoders to accurately and efficiently acquire the complex space-time correlation among historical time sequence data, firstly extracts the spatial features among the data through a convolutional neural network, weights the acquired spatial features through an attention layer, and further captures the spatial correlation; the weighted attention space-time characteristics of the data are further acquired through the gate control circulation unit, so that the problem of long dependence in the RNN is solved; the long-term historical time sequence data fully acquire the influence among different time steps through an interactive attention network, and meanwhile, the short-term historical time data further improve the prediction accuracy and the experimental robustness through an autoregressive component.
Drawings
FIG. 1 is a network model for predicting future time data of a plurality of time-series data according to an embodiment of the present invention;
FIG. 2 is a schematic diagram of extracting spatial information from historical time-series data and outputting a spatial feature vector according to an embodiment of the present invention;
FIG. 3 is a schematic diagram illustrating an embodiment of extracting temporal correlations by inputting spatial feature vectors into a gated cyclic unit one by one;
FIG. 4 is a diagram of the prediction of the neural network portion by combining weighted feature vectors and short-term time-series space-time features in accordance with an embodiment of the present invention.
Detailed Description
The embodiment of the invention provides a multivariate time sequence prediction method based on a feature extraction coding and interactive attention module, which comprises the following steps:
step B1, historical time series data [ X ]k-nT,...,Xk-1]Partitioning into long-term historical time series data [ X ]k -nT,...,Xk-T-1]And short term historical time seriesAccording to [ X ]k-T,...,Xk-1]In the embodiment of the invention, the value of n is 8, and the lengths of the input long-term and short-term historical time sequences are all T;
the time series data matrix mentioned in B1 has rows representing time nodes and columns representing variable dimensions, and the element values in the matrix represent the values of different indexes at corresponding time instants.
Step B2, inputting a long-term historical time sequence [ X ] with the length of Tk-nT,...,Xk-T-1]And short-term time series historical data series [ X ]k-T,...,Xk-1]Respectively obtaining a space feature vector [ e ] of a long-term history time sequence through a space feature extractor consisting of a convolutional layer and a plurality of fully-connected layersk-nT,...,ek-T-1]And a short-term historical time series of spatial feature vectors [ ek-T,...,ek-1]Obtaining a weighted attention space feature vector of a long-term historical time series immediately after passing through an attention layer { [ z ]k-nT,...,zk-(n-1)T-1],...,[zk-2T,...,zk-T-1]Weighted attention space feature vector [ z ] of short-term historical time seriesk-T-1,...,zk]。
The spatial feature extractor reasonably constructed by the convolutional layer and the multilayer full-connection layer is mainly used for extracting the spatial correlation of the historical time data sequence and outputting the spatial correlation in the form of a spatial feature vector. Taking the full-connection layer in the short-term historical time sequence matrix characteristic extraction code as an example, the original input [ Xk-T,...,Xk-1]The layer output weighted attention space feature vector is obtained by:
[zk-T,...,zk-1]=Softmax(f(W·[Xk-T,...,Xk-1]+b))
where W represents the weight matrix, b is the bias term, and f () is the activation function.
And performing weighted mixing on the historical time sequence data through the weight matrix, introducing an activation function and adding a nonlinear relation, and finally extracting the relation between different elements of the matrix. Theoretically, the more full connection layers, the stronger the extraction capability of the feature extraction network. The step can obtain the spatial correlation contained in each matrix surface and output the spatial correlation respectively in the form of vector sequences.
Step B3, obtaining the weighted attention space feature vector of the long-term history time series acquired in the step B2 { [ z { [k -nT,...,zk-(n-1)T-1],...,[zk-2T,...,zk-T-1]Weighted attention space feature vector [ z ] of short-term historical time seriesk-T,...,zk-1]Sequentially inputting the time-sequence into a gating cycle unit to learn the time correlation of the sequence, and respectively outputting a characteristic coding vector sequence { m } containing the space-time correlationiAnd a feature code vector u containing spatio-temporal correlations. A loop layer with gated loop units (GRUs) uses the ReLU function as a hidden activation function. the hidden state of the cyclic unit at time t can be expressed as,
Figure BDA0003493749250000101
the weighted attention space feature vectors are input into a gated round robin unit (GRU) one by one to obtain context vectors. Thus, the short-term history matrix sequence [ z ]k-T,...,zk-1]Outputting a feature-coded vector u containing a spatio-temporal correlation, where u ═ ck-1K is a time value; long-term history matrix sequence { [ z ]k-nT,...,zk-(n-1)T-1],...,[zk-2T,...,zk-T-1]Outputs a feature code vector sequence (m) containing space-time correlationiIn which m isi}=[ck-(n-1)T,...,ck-T-1]. In fact GRU inputs a weighted attention space feature vector z each timek-n(1. ltoreq. n. ltoreq.T) results in a corresponding context vector ck-nAnd a hidden state hk-n. In which the hidden state hk-nInput c for passing history information to the next momentk-n+1. It is the transfer of the hidden state that enables the learning of the temporal correlation of the input sequence.
Finally, we succeed in extracting the encoder module through the characteristicTime series of long-term histories [ X ]k-nT,...,Xk-T-1]Conversion into a sequence of eigenvectors { m } containing spatio-temporal correlationsi}. Short term time series historical data series [ X ]k-T,...,Xk-1]And obtaining a feature coding vector u containing space-time correlation by another feature extraction coder, wherein the expression formula is as follows,
mi=Encoderl([Xk-nT,...,Xk-T-1])
u=Encoders([Xk-T,...,Xk-1]);
wherein m isi∈Rd,u∈Rd,miConstructing a sequence of eigen-coded vectors { m } containing spatio-temporal correlationsiU is a feature code vector containing space-time correlation.
Step B4, the feature coding vector sequence { m } containing space-time correlation obtained in step B3 is codediAnd inputting the feature coding vector u containing the space-time correlation into an interactive attention network, so that the relationship between the space-time feature vectors output by the predictive encoder is better captured, the calculation speed is accelerated, and the defects of a recurrent neural network are avoided. Computing a feature-coded vector u containing spatio-temporal correlation and a plurality of sequences of feature-coded vectors m containing spatio-temporal correlation by computing the inner product of the feature vectors and a Softmax ()iAttention weight distribution vector piAnd a weighted feature vector OiThe calculation formula is as follows:
pi=SOftmax(uTmi);
oi=pi×mi
wherein Softmax () is
Figure BDA0003493749250000111
Each of the inputs miAll have a corresponding weighted feature vector oi
Figure BDA0003493749250000112
d is the variable dimension of the weighted feature vector.
Step B5, weighting feature vector o containing space-time feature after trainingiCombined with a feature code vector u containing spatio-temporal correlations to combine the spatio-temporal correlations with information of the data of the recent time instant.
Specifically, a weighted feature vector o of size T × diAnd a feature encoding vector u having a size of 1d and including a space-time correlation are combined into a new vector having a size of (T +1) × d, and the combined new vector is input to a generation model composed of a fully-connected layer to generate a nonlinear prediction matrix. The output value of the fully-connected layer is calculated as,
Figure BDA0003493749250000113
wherein b is a paranoid, [ u; o1;o2;…;oT]Is a feature vector u and a set of weighted feature vectors { O }iThe connection of the electronic component is realized,
Figure BDA0003493749250000114
is a prediction of the neural network portion.
Step B6, generating the prediction matrix of the model output in the step B5
Figure BDA0003493749250000115
Prediction matrix with autoregressive layer output
Figure BDA0003493749250000116
And adding to obtain a final prediction result, and fully combining the space-time correlation and the short-term historical time sequence data. Using a classical Autoregressive (AR) model as the linear component, all dimensional variables share the same set of linear parameters. The expression of the AR model is as follows,
Figure BDA0003493749250000117
wherein the prediction result of the AR component at the time stamp t is represented as
Figure BDA0003493749250000118
SarFor the size of the input window, the correlation coefficient of the AR model is
Figure BDA00034937492500001111
And bar∈R。
The final prediction of the method is obtained by integrating the neural network part and the output of the AR model,
Figure BDA0003493749250000119
in the training process, we use the mean absolute error and the objective function as follows,
Figure BDA00034937492500001110
where N is the number of training samples and D is the dimension of the target data. All neural models were trained using Adam optimizer.
And generating a model training model by jointly training a relative root mean square error RRSE loss function and an empirical correlation coefficient CORR. The model can have faster convergence speed during training, and relative errors have higher precision.
And step B7, performing iterative training from steps B1 to B6 based on the existing historical data loop until convergence. And at the moment, the model is trained in the early stage, and long-term data is input into the feature extraction module according to groups at the subsequent moment.
The multivariate time series prediction algorithm provided by the invention takes a group of long-term time series historical data matrixes and short-term time series historical data matrixes as input by arranging multivariate time series data of space correlation and time correlation according to the time sequence. Firstly, spatial relation of adjacent positions of a matrix can be extracted through space-time information contained in a matrix sequence by a feature extraction encoder and a gating circulation unit, incidence relation of data among different time dimensions can be further captured by adopting an attention layer, the gating circulation unit can capture nonlinear time relation, and obtained space-time feature vectors retain the space-time information hidden in historical data; and then, the method further excavates and dynamically captures the interdependence relation between the characteristic vector matrixes through an interactive attention network, so that the finally obtained space-time characteristic vector can greatly keep effective information contained in historical data. The method generates a model effect by integrating the space-time characteristic vector further captured and the current time space characteristic vector information and training through a loss function, and finally generates a prediction result of the future time by combining a linear prediction result output by an autoregressive layer.
In the process, compared with other algorithms, the characteristic extraction coding module can be used for remarkably improving the characteristic learning speed, and the convergence speed of the generated model can be remarkably accelerated by adopting the relative root mean square error loss function and the empirical correlation coefficient combined training, so that the time complexity is lower, and the calculation complexity is lower; by introducing the interactive attention network, the relevance among the space-time characteristic vectors can be captured as much as possible, effective information contained in historical data is greatly reserved, and accurate prediction of the multivariate time series data can be realized.
The method provided by the invention adopts the traditional autoregressive linear model and the nonlinear neural network in parallel, and provides reliable guarantee for the robustness and accuracy of the prediction model.
The results of experiments using the indices RRSE and CORR were obtained by summarizing 4 multivariate data sets Traffic, electric, Exchange-Rate and Solar-Energy with 6 baseline algorithms, see Table 1. Wherein the smaller the RRSE value is, the better the experimental result is; the larger the CORR value, the better the experimental results. We set Horizon ═ 3,6, the larger the field of view, the more difficult the prediction. Comparing neural networks with statistical algorithms, we can also directly see that algorithms based on neural networks (i.e. RNN-GRU, DA-RNN, DASNet, MTNet, LSTNet, Our Model) have a clear advantage over statistical algorithms based on autoregressive models (i.e. VAR) in overall results, which has been confirmed by many studies. Due to the limited modeling capability of the traditional time series prediction method based on autoregression, the prediction result is not good. Comparing the recurrent neural network variant with our method, the RNN-GRU and DA-RNN models are only suitable for extracting temporal features, are not suitable for processing spatial features, and have no long-term memory capability, so that they have obvious defects in processing multivariate time series prediction tasks and have poor prediction results. Comparing other hybrid neural networks with our method, LSTNet provides a corresponding solution to extract long-term and long-term complex patterns and spatial correlations, MTNet learns long-term patterns using an end-to-end memory network with a single memory layer, DSANet exploits self-attention mechanisms to mine relationships between multivariate time sequences, but the experimental effect is poor due to the inability to comprehensively capture spatio-temporal correlations and complex periodic patterns between historical time sequences. The method comprises the steps of firstly converting latest short-term historical time sequence data into space-time characteristic vectors, calculating a weighted attention distribution matrix by using the long-term historical time sequence data, and finally obtaining a weighted space-time characteristic vector sequence. Intuitively, the method of the present invention knows which time supports the prediction, and can thus produce a better prediction for a data set with a periodic pattern. Experimental results show that the method is superior to other prediction methods in most cases. To demonstrate the effectiveness of the model design of the present invention, we performed a careful ablation study, with the results shown in table 2. Specifically, the present invention deletes one module of a model at a time and implements the remaining modules using the same environment and data set. A withoutEncoder experiment is carried out by deleting a spatial feature extractor in the feature extraction code, and the experiment effect is poor because the spatial correlation among the historical time sequence data cannot be extracted; the method has the advantages that a withoutInteractivetention experiment is carried out by deleting an interactive attention network, and the experiment effect is poor because the time-space correlation among historical time sequence data cannot be fully obtained; the Without experiment is carried out by deleting the autoregressive layer, and the experiment effect is poor because a linear prediction result of short-term historical time sequence data cannot be obtained. Experimental results show that deleting a certain module of the model directly influences the accuracy of prediction, so that the effectiveness of each experimental module in the model is further verified.
Table 1 prediction accuracy of four different data sets at different future times for six different methods in the experiment
Figure BDA0003493749250000141
Table 2 experimental data of ablation experiments performed on four different data sets according to the invention
Figure BDA0003493749250000142

Claims (10)

1. A multivariate time series prediction method is characterized by comprising the following steps:
s1, historical time-series data [ X ]k-nT,...,Xk-1]Partitioning into long-term historical time series data [ X ]k-nT,...,Xk -T-1]And short-term historical time series data [ X ]k-T,...,Xk-1](ii) a T is the length of the long-term historical time series data or the short-term historical time series data;
s2, using the long-term historical time-series data [ X ]k-nT,...,Xk-T-1]And short-term historical time series data [ X ]k -T,...,Xk-1]Obtaining a weighted attention space feature vector of a long-term historical time series { [ z ]k-nT,...,zk -(n-1)T-1],...,[zk-2T,...,zk-T-1]Weighted attention space feature vector [ z ] of short-term historical time seriesk -T,...,zk-1](ii) a Short-term historical time-series data [ X ]k-T,...,Xk-1]Inputting the autoregressive layer to obtain a prediction matrix
Figure FDA0003493749240000011
S3, weighted attention space feature direction using acquired long-term historical time seriesQuantity { [ z ]k-nT,...,zk -(n-1)T-1],...,[zk-2T,...,zk-T-1]Weighted attention space feature vector [ z ] of short-term historical time seriesk -T,...,zk-1]Respectively obtaining a feature code vector sequence { m) containing space-time correlationiAnd a eigen-code vector u containing spatio-temporal correlations;
s4, mixing { miInputting an interactive attention network with u to obtain an attention weight value distribution vector piBy calculating piAnd { miThe product of (f) and (f) obtains a weighted feature vector oi
S5, weighting characteristic vector o with size of T multiplied by diCombining the feature coding vector u with the size of 1 × d and containing space-time correlation into a new vector with the size of (T +1) × d, inputting the combined new vector into a generation model, and generating a nonlinear prediction matrix
Figure FDA0003493749240000012
S6, generating a prediction matrix of model output
Figure FDA0003493749240000013
Prediction matrix with autoregressive layer output
Figure FDA0003493749240000014
And adding to obtain the final prediction result.
2. The multivariate time series prediction method as defined in claim 1, wherein the step S2 is implemented by: time series data [ X ] of long-term historyk-nT,...,Xk-T-1]And short-term historical time series data [ X ]k-T,...,Xk-1]Respectively obtaining a space feature vector [ e ] of the long-term historical time sequence through a first space feature extractor and a second space feature extractork-nT,...,ek-T-1]And a short-term historical time series of spatial feature vectors [ ek-T,...,ek-1]Will [ e ]k-nT,...,ek-T-1]And [ ek-T,...,ek-1]Respectively inputting a first attention layer and a second attention layer to obtain a weighted attention space feature vector { [ z ] of a long-term history time seriesk-nT,...,zk-(n-1)T-1],...,[zk-2T,...,zk-T-1]Weighted attention space feature vector [ z ] of short-term historical time seriesk-T,...,zk-1]。
3. The multivariate time series prediction method as defined in claim 1, wherein the step S3 is implemented by: weighted attention space feature vector of long-term history time series { [ z ] to be acquiredk-nT,...,zk-(n-1)T-1],...,[zk-2T,...,zk-T-1]Weighted attention space feature vector [ z ] of short-term historical time seriesk-T,...,zk-1]Sequentially inputting a first gating circulation unit and a second gating circulation unit according to the time sequence, and respectively outputting a characteristic coding vector sequence { m } containing space-time correlationiAnd a feature code vector u containing spatio-temporal correlations.
4. The multivariate time series prediction method as defined in claim 1, wherein in step S4, oi=pi×mi;pi=Softmax(uTmi)。
5. The multivariate time series prediction method as defined in claim 1, wherein the generative model comprises a plurality of cascaded fully-connected layers in step S5.
6. A multivariate time series prediction system, comprising:
a first feature extraction encoder for inputting the long-term history time-series data [ X ]k-nT,...,Xk-T-1]The output is a characteristic code vector sequence { m) containing space-time correlationi};
A second feature extraction encoder for inputting a weighted attention space feature vector [ z ] of the short-term historical time seriesk -T,...,zk-1]Outputting a characteristic coding vector u containing space-time correlation;
autoregressive layer, input as short-term historical time series data [ X ]k-T,...,Xk-1]The output is a prediction matrix
Figure FDA0003493749240000021
An interactive attention network, the input is a characteristic coding vector sequence { m ] containing space-time correlationiAnd a feature coding vector u containing space-time correlation, and outputting a new vector of (T +1) x d;
generating a model, inputting a new vector of (T +1) x d, and outputting a nonlinear prediction matrix
Figure FDA0003493749240000031
A prediction module for generating a prediction matrix of the model output
Figure FDA0003493749240000032
Prediction matrix with autoregressive layer output
Figure FDA0003493749240000033
Adding to obtain a final prediction result;
wherein the long-term history time-series data [ Xk-nT,...,Xk-T-1]And short-term historical time series data [ X ]k -T,...,Xk-1]From historical time series data [ X ]k-nT,...,Xk-1]Dividing to obtain; and T is the length of the long-term historical time sequence data or the short-term historical time sequence data.
7. The multivariate time series prediction system as defined in claim 6, wherein the first and second feature extraction encoders are identical in structure; wherein the first feature extraction encoder includes:
a first spatial feature extractor for inputting the long-term historical time-series data [ X ]k-nT,...,Xk-T-1]The output is a space characteristic vector [ e ] of the long-term historical time sequencek-nT,...,ek-T-1];
First attention tier, input of [ e ]k-nT,...,ek-T-1]And outputting a weighted attention space feature vector { [ z ] of the long-term historical time seriesk-nT,...,zk-(n-1)T-1],...,[zk-2T,...,zk-T-1]};
A first gated cyclic unit with the input of weighted attention space feature vector { [ z ] of long-term history time seriesk-nT,...,zk-(n-1)T-1],...,[zk-2T,...,zk-T-1]The output is a characteristic code vector sequence (m) containing space-time correlationi}。
8. The multivariate time series prediction system as defined in claim 6 or 7, wherein the interactive attention network is configured to perform operations comprising: will { miUsing the u and the u as the input of the self-attention layer to obtain the attention weight value distribution vector piBy calculating piAnd { m }iThe product of which yields a weighted feature vector oiWeighted eigenvector o of size T × diAnd the eigen-code vector u containing the spatio-temporal correlation of size 1 x d are combined into a new vector of size (T +1) x d.
9. A computer-readable storage medium characterized by; comprising a program running in a processor; the program is configured for carrying out the steps of the method according to one of claims 1 to 5.
10. A computer program product comprising a computer program/instructions; characterized in that the computer program/instructions, when executed by a processor, performs the steps of the method according to one of claims 1 to 5.
CN202210107028.XA 2022-01-28 2022-01-28 Multivariate time series prediction method, multivariate time series prediction system, computer product and storage medium Pending CN114493014A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210107028.XA CN114493014A (en) 2022-01-28 2022-01-28 Multivariate time series prediction method, multivariate time series prediction system, computer product and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210107028.XA CN114493014A (en) 2022-01-28 2022-01-28 Multivariate time series prediction method, multivariate time series prediction system, computer product and storage medium

Publications (1)

Publication Number Publication Date
CN114493014A true CN114493014A (en) 2022-05-13

Family

ID=81476715

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210107028.XA Pending CN114493014A (en) 2022-01-28 2022-01-28 Multivariate time series prediction method, multivariate time series prediction system, computer product and storage medium

Country Status (1)

Country Link
CN (1) CN114493014A (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115081586A (en) * 2022-05-19 2022-09-20 中国科学院计算机网络信息中心 Short-term time sequence prediction method and system based on time and space attention
CN115146842A (en) * 2022-06-24 2022-10-04 沈阳建筑大学 Multivariate time series trend prediction method and system based on deep learning
CN115204535A (en) * 2022-09-16 2022-10-18 湖北信通通信有限公司 Purchasing business volume prediction method based on dynamic multivariate time sequence and electronic equipment
CN115796407A (en) * 2023-02-13 2023-03-14 中建科技集团有限公司 Production line fault prediction method and related equipment
CN115938112A (en) * 2022-11-23 2023-04-07 华侨大学 Traffic demand prediction method, system, electronic device, and computer storage medium
CN117688453A (en) * 2024-02-02 2024-03-12 山东科技大学 Traffic flow prediction method based on space-time embedded attention network

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111651504A (en) * 2020-06-03 2020-09-11 湖南大学 Multi-element time sequence multilayer space-time dependence modeling method based on deep learning
CN112183862A (en) * 2020-09-29 2021-01-05 长春理工大学 Traffic flow prediction method and system for urban road network
CN112508173A (en) * 2020-12-02 2021-03-16 中南大学 Traffic space-time sequence multi-step prediction method, system and storage medium

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111651504A (en) * 2020-06-03 2020-09-11 湖南大学 Multi-element time sequence multilayer space-time dependence modeling method based on deep learning
CN112183862A (en) * 2020-09-29 2021-01-05 长春理工大学 Traffic flow prediction method and system for urban road network
CN112508173A (en) * 2020-12-02 2021-03-16 中南大学 Traffic space-time sequence multi-step prediction method, system and storage medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
ZHENXIONG YAN ET.AL: "Multivariate Time Series Forecasting exploiting Tensor Projection Embedding and Gated Memory Network", 《2021 IEEE/ACM 29TH INTERNATIONAL SYMPOSIUM ON QUALITY OF SERVICE (IWQOS)》, 26 August 2021 (2021-08-26) *
张建晋;王韫博;龙明盛;***;王海峰;: "面向季节性时空数据的预测式循环网络及其在城市计算中的应用", 计算机学报, vol. 43, no. 02, 16 October 2019 (2019-10-16) *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115081586A (en) * 2022-05-19 2022-09-20 中国科学院计算机网络信息中心 Short-term time sequence prediction method and system based on time and space attention
CN115081586B (en) * 2022-05-19 2023-03-31 中国科学院计算机网络信息中心 Photovoltaic power generation time sequence prediction method and system based on time and space attention
CN115146842A (en) * 2022-06-24 2022-10-04 沈阳建筑大学 Multivariate time series trend prediction method and system based on deep learning
CN115146842B (en) * 2022-06-24 2023-07-18 沈阳建筑大学 Multi-element time sequence trend prediction method and system based on deep learning
CN115204535A (en) * 2022-09-16 2022-10-18 湖北信通通信有限公司 Purchasing business volume prediction method based on dynamic multivariate time sequence and electronic equipment
CN115938112A (en) * 2022-11-23 2023-04-07 华侨大学 Traffic demand prediction method, system, electronic device, and computer storage medium
CN115796407A (en) * 2023-02-13 2023-03-14 中建科技集团有限公司 Production line fault prediction method and related equipment
CN115796407B (en) * 2023-02-13 2023-05-23 中建科技集团有限公司 Production line fault prediction method and related equipment
CN117688453A (en) * 2024-02-02 2024-03-12 山东科技大学 Traffic flow prediction method based on space-time embedded attention network
CN117688453B (en) * 2024-02-02 2024-04-30 山东科技大学 Traffic flow prediction method based on space-time embedded attention network

Similar Documents

Publication Publication Date Title
CN114493014A (en) Multivariate time series prediction method, multivariate time series prediction system, computer product and storage medium
Xuan et al. Multi-model fusion short-term load forecasting based on random forest feature selection and hybrid neural network
Wang et al. Res2Fusion: Infrared and visible image fusion based on dense Res2net and double nonlocal attention models
Ma et al. A hybrid attention-based deep learning approach for wind power prediction
JP2021060992A (en) Machine learning system and method
Zhou et al. Exploiting operation importance for differentiable neural architecture search
Fang et al. A new sequential image prediction method based on LSTM and DCGAN
Zheng et al. Prompt vision transformer for domain generalization
Khan et al. A survey of the vision transformers and their CNN-transformer based variants
Zheng et al. Model compression based on differentiable network channel pruning
Xue et al. Deep Correlated Predictive Subspace Learning for Incomplete Multi-View Semi-Supervised Classification.
Tan et al. Openstl: A comprehensive benchmark of spatio-temporal predictive learning
An et al. Object recognition algorithm based on optimized nonlinear activation function-global convolutional neural network
CN115659254A (en) Power quality disturbance analysis method for power distribution network with bimodal feature fusion
Zhou et al. IF2CNN: Towards non-stationary time series feature extraction by integrating iterative filtering and convolutional neural networks
Yao et al. ModeRNN: Harnessing spatiotemporal mode collapse in unsupervised predictive learning
Yi et al. An Effective Lightweight Crowd Counting Method Based on an Encoder-Decoder Network for the Internet of Video Things
CN116680456A (en) User preference prediction method based on graph neural network session recommendation system
CN116502774A (en) Time sequence prediction method based on time sequence decomposition and Legend projection
CN113835964B (en) Cloud data center server energy consumption prediction method based on small sample learning
Liu et al. Diverse hyperspectral remote sensing image synthesis with diffusion models
Li et al. Zero-Shot Neural Architecture Search: Challenges, Solutions, and Opportunities
Kashyap et al. Quantum convolutional neural network architecture for multi-class classification
CN115348182A (en) Long-term spectrum prediction method based on depth stack self-encoder
Shang et al. Mshyper: Multi-scale hypergraph transformer for long-range time series forecasting

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination