CN111475987B - SAE and ON-LSTM-based gear residual life prediction method - Google Patents

SAE and ON-LSTM-based gear residual life prediction method Download PDF

Info

Publication number
CN111475987B
CN111475987B CN202010255897.8A CN202010255897A CN111475987B CN 111475987 B CN111475987 B CN 111475987B CN 202010255897 A CN202010255897 A CN 202010255897A CN 111475987 B CN111475987 B CN 111475987B
Authority
CN
China
Prior art keywords
hidden layer
output
representing
layer
input
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010255897.8A
Other languages
Chinese (zh)
Other versions
CN111475987A (en
Inventor
秦毅
阎昊冉
项盛
陈定粮
奚德君
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing University
Original Assignee
Chongqing University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing University filed Critical Chongqing University
Priority to CN202010255897.8A priority Critical patent/CN111475987B/en
Publication of CN111475987A publication Critical patent/CN111475987A/en
Application granted granted Critical
Publication of CN111475987B publication Critical patent/CN111475987B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Testing Of Devices, Machine Parts, Or Other Structures Thereof (AREA)

Abstract

The invention relates to a gear residual life prediction method based ON SAE and ON-LSTM, belonging to the field of big data and intelligent manufacturing. The method utilizes SAE to extract gear health indexes, and uses a novel ON-LSTM neural network to predict the residual life of gears, and the hierarchical structures of input data and historical data are divided by a hierarchy divider in the ON-LSTM. And defining the positions of the largest elements in the output vectors of the main forgetting gate and the main input gate as hierarchical positions, so that the recurrent neural network is updated hierarchically. The invention predicts the residual service life of the gear through SAE feature extraction and ON-LSTM neural network, greatly reduces the network calculation amount, reduces the calculation time, and improves the prediction speed and accuracy.

Description

SAE and ON-LSTM-based gear residual life prediction method
Technical Field
The invention belongs to the field of big data and intelligent manufacturing, and relates to a gear residual life prediction method based ON SAE and ON-LSTM.
Background
Gears are widely used in mechanical equipment and are one of the most widely used mechanical parts. The gear has the unique advantages of high transmission efficiency, compact structure, good transmission smoothness, large bearing capacity, long service life and the like, so that the gear has strong and durable vitality. Under complex operating conditions and circumstances, gears are prone to failure, potentially leading to disasters in machine operation and even compromising personal safety. This is especially true for large or ultra-large equipment such as hydraulic generators, mine conveyor machines, helicopter power transmission systems, heavy machine tools, and the like. The service life of the in-service gear is predicted, the maintenance time of equipment can be effectively determined, the continuous and efficient production is ensured, the production efficiency is improved, the accident rate is reduced, and the occurrence of sudden accidents is prevented, so that the failure of the gear is predicted in advance, and the method has great significance for engineering production.
The commonly used life prediction methods of mechanical equipment are mainly divided into the following three types: 1) A model-based prediction method; 2) Data driving; 3) The first two methods are mixed. The model-based approach builds a physical model that describes the component degradation process. This method requires specific mechanical knowledge and is therefore less applicable. The data driven method derives a predictive model from conventionally collected monitored data. It is mainly based on statistical and machine learning methods, with the aim of finding the behaviour of the system. Thus, these methods provide a compromise between accuracy, complexity and applicability. The hybrid approach combines model-based and data-driven techniques. And establishing a model by using physical knowledge of the monitoring system, and learning and updating parameters by using a data driving technology. The combination of model-based and data-driven techniques makes the method accurate, but still requires specific physical knowledge and is computationally expensive. The machine learning-based method in the data driving method can overcome the problem that the degradation model is unknown, and meanwhile, the input of the built model is not limited to the state monitoring data, and can be various different types of data. The residual life prediction method based on the RNN can integrate the original learning sample and the new learning mode to realize sample retraining, not only can improve the accuracy of residual life prediction, but also has the characteristics of high convergence speed, high stability and the like, and plays an important role in the fields of reliability evaluation and residual life prediction. However, when long-term dependent degradation data is processed, the conventional RNN method suffers from the problem of gradient extinction or explosion, and the residual life prediction accuracy is seriously affected.
To solve this problem, long short-term memory (LTSM) networks have been developed, however LSTM has insufficient capabilities to handle long-term dependent degradation data. In the method for predicting the residual life of the gear, the actually measured gear degradation signal contains rich useful information, and the time domain and frequency domain characteristics of the gear vibration signal reflect the descending trend of the state and the health condition of the gear vibration signal to a certain extent. In order to comprehensively and accurately express the degradation process of the gear, it is necessary to calculate all the characteristics of the gear. However, the conventional LTSM network does not perform comprehensive analysis on the gear degradation data and does not perform mining on the sequence information of the neural network, so that the life prediction capability is poor. Therefore, in order to consider the influence of sequence information in the overall analysis of gear degradation data to thereby improve prediction accuracy, it is necessary to design a new neural network to replace the conventional LTSM neural network to predict the remaining life of gears.
Disclosure of Invention
In view of the above, the present invention aims to provide a gear residual life prediction method based ON SAE and ON-LSTM, which aims at the unordered characteristics of neurons of the conventional LSTM neural network, and digs the sequence information of the neurons by embedding a tree structure between the neurons. And then carrying out hierarchical updating on different sequence information, thereby reducing the calculated amount of the traditional LSTM neural network.
In order to achieve the above purpose, the present invention provides the following technical solutions:
a gear residual life prediction method based ON SAE and ON-LSTM specifically comprises the following steps:
S1: collecting gear vibration signals with the time length of T at intervals of delta T until the gears fail, wherein the number of sampled gear vibration signal segments is n;
s2: processing the gear vibration signal by using a sparse self-encoder (sparse autoencoder, SAE) to obtain an n-dimensional characteristic vector V;
s3: selecting a feature matrix V1 consisting of the n1 sampling points as a training matrix;
S4: using least squares to function the target And (3) carrying out minimum, normalizing all elements in the vector V with the vector V1 through a formula V' i=avi +b, wherein a and b respectively represent the weight and bias of a function used in the process of unifying all element vectors V1 in the vector V, and the specific size is determined through the minimization of the objective function;
S5: normalizing the vector V1 to obtain a normalized vector w= (W 1,w2,…,wn1)T;
S6: reconstruction matrix Wherein p is the number of units of the input layer of the neural network;
S7: taking the p rows in front of the matrix U as the input of an ON-LSTM (ON-LSTM) of a long-term memory neural network of ordered neurons, and taking the last row as the output of the ON-LSTM neural network to train the network;
s8: after training the matrix U, taking p outputs of the reciprocal as the network input of the ON-LSTM to obtain the output at the next moment;
S9: and repeating the step S8 for a certain number of times, comparing the output denormalized values with an actual characteristic value V '= (V' p+1,v'p+2,…,v'n)T), and when the output denormalized values exceed a set threshold value, multiplying the predicted sampling point number by the sum delta t+T of the interval time and the sampling time of the gear vibration signal to obtain the residual service life of the gear.
Further, in step S2, the SAE sparse automatic encoder has a structure of: let Y t=[y1 y2 … yn be the original vibration signal for the entire time dimension, where Y t=[yt,1 yt,2 … yt,m represents the input data at time t, where m represents the number of signal points acquired during each time period. Through the first layer network, a vector z t with 50 dimensions can be obtained; through the second layer network, a 1-dimensional health index x t can be obtained; then through the third layer network, a 50-dimensional vector z' t can be obtained; through the last layer of network, an m-dimensional vector y t' can be obtained. The least square method is used as an objective function to make the difference between y t and y t' as small as possible, and x t is used as a health index at the time t to be applied to a subsequent prediction network.
Wherein z t denotes an output of the first layer network, x t denotes an output of the second layer network, z t denotes an output of the hidden third layer network, y' t denotes an output of the fourth layer network, σ denotes a sigmod activation function, W 1 denotes a weight of the first layer network, W 2 denotes a weight of the second layer network, W 3 denotes a weight of the third layer network, W 4 denotes a weight of the fourth layer network, b 1 denotes a bias of the first layer network, b 2 denotes a bias of the second layer network, b 3 denotes a bias of the third layer network, and b 4 denotes a bias of the fourth layer network.
Further, the structure of the ON-LSTM neural network is as follows: assuming that the input information of the whole time dimension is X t=[x1x2 … xt ], wherein X t=[xt,1 xt,2 ... xt,n represents input data at the time t, and h t-1=[ht-1,1 ht-1,2 ... ht-1,m represents recursive data of the neural network at the time t-1; first, defining a hierarchy divider, determining a hierarchy position L1 of the input information x t through a constructor F 1, and determining a hierarchy position L2 of the history information h t-1 through a constructor F 2; with the division of hierarchical information, the structure of an ON-LSTM neural network is expressed as:
Where f t denotes the output of the forget gate in the hidden layer, i t denotes the output of the input gate in the hidden layer, o t denotes the output of the output gate in the hidden layer, Representing the output of the hyperbolic tangent function in the hidden layer,/>Output representing main forgetting gate in hidden layer,/>Representing the output of a main input gate in the hidden layer, σ representing sigmod activation functions, tanh representing hyperbolic tangent functions, W f representing the weight between the input layer and a forgetting gate in the hidden layer, W i representing the weight between the input layer and an input gate in the hidden layer, W o representing the weight between the input layer and an output gate in the hidden layer, W c representing the weight between the input layer and the hyperbolic tangent functions in the hidden layer,/>Representing the weight between the main forgetting gate in the input layer and hidden layer,/>Representing the weight between the input gate and the main output gate in the hidden layer, U f representing the weight between the forgetting gate in the hidden layer at the current time and the hidden layer output at the last time, U i representing the weight between the input gate in the hidden layer at the current time and the hidden layer output at the last time, U o representing the weight between the output gate in the hidden layer at the current time and the hidden layer output at the last time, U c representing the weight between the hyperbolic tangent function in the hidden layer at the current time and the hidden layer output at the last time,/>Representing the weight between the main forgetting gate in the hidden layer at the current moment and the output of the hidden layer at the last moment,/>Representing the weight between the main input gate in the hidden layer at the current moment and the output of the hidden layer at the previous moment, b f representing the bias of the linear combination in the forgetting gate in the hidden layer, b i representing the bias of the linear combination in the input gate in the hidden layer, b o representing the bias of the linear combination in the output gate in the hidden layer, b c representing the bias of the linear combination in the hyperbolic tangent function in the hidden layer,/>Bias representing linear combinations in the main forgetting gate in hidden layer,/>Representing the bias of the linear combination in the main input gate in the hidden layer,/>Representing the multiplication operation symbol of the corresponding element, c t-1 represents the output of the memory cell memory unit at time t-1, function cumax (…) = cumsum (softmax (…)), cumsum is the accumulation function,/>Representing middle level,/>Representing a low-level of the hierarchy,Representing a high level.
Further, in the structure of the ON-LSTM neural network, the hierarchical position L1 of the input information x t is determined by the constructor F 1, and the hierarchical position L2 of the history information h t-1 is also determined by the constructor F 2, where the expression is as follows:
L1=F1(xt,ht-1)
L2=F2(xt,ht-1)
Wherein F 1 and F 2 are defined as:
And outputting values of positions of maximum elements in the vector after nonlinear combination of the input data and the historical data. During the network training process, x t+1,n is set to the query vector q m; in the prediction phase, x t,n is set to q m.
Further, in the structure of the ON-LSTM neural network, the neural network hierarchy is divided by ordering the cell units by the required L1 and L2. When L2 is greater than or equal to L1, the historical data level is lower than the current input level, and the overlapped part of the levels isThe cell memory unit c t is updated by the following rule:
When L2 < L1, the current input level is lower than the history level, and the cell memory unit c t is updated by the following rule:
The invention has the beneficial effects that: the SAE adopted by the invention is different from the traditional time-frequency domain characteristic value, the health index can be extracted by utilizing a digital model, and the adopted ON-LSTM neural network aims at the unordered characteristic of the interlayer neurons of the common neural network, the sequence information of the interlayer neurons is mined by embedding a tree structure among the neurons, and then the hierarchical updating is carried out ON different level information, so that the high level information is not easy to replace, the updating rule of the middle level information is the same as that of the traditional long-short-term memory network (LSTM), and the low level information is easier to be influenced by the input data at the current moment. Through the use of sequence information ignored by other neural networks, the ON-LSTM has advantages over the traditional LSTM and variants thereof in the field of predicting the residual service life of gears, and the prediction accuracy of the network is greatly improved.
Additional advantages, objects, and features of the invention will be set forth in part in the description which follows and in part will become apparent to those having ordinary skill in the art upon examination of the following or may be learned from practice of the invention. The objects and other advantages of the invention may be realized and obtained by means of the instrumentalities and combinations particularly pointed out in the specification.
Drawings
For the purpose of making the objects, technical solutions and advantages of the present invention more apparent, the present invention will be described in the following preferred detail with reference to the accompanying drawings, in which:
FIG. 1 is a block diagram of an SAE sparse automatic encoder;
FIG. 2 is a block diagram of an ON-LSTM neural network;
FIG. 3 is a schematic diagram of an hidden layer of an ON-LSTM neural network;
FIG. 4 is a schematic diagram showing a hierarchical cell unit update process based on layer sequence information, wherein (a) is a first update mode and (b) is a second update mode;
FIG. 5 is a flow chart of a method for predicting the remaining life of a gear based ON SAE and ON-LSTM according to the present invention;
FIG. 6 is a simulation diagram of failure threshold, training value, predicted value and actual value for 380 known sampling points;
FIG. 7 is a graph of predictive power versus known 380 fused eigenvalues for different models.
Detailed Description
Other advantages and effects of the present invention will become apparent to those skilled in the art from the following disclosure, which describes the embodiments of the present invention with reference to specific examples. The invention may be practiced or carried out in other embodiments that depart from the specific details, and the details of the present description may be modified or varied from the spirit and scope of the present invention. It should be noted that the illustrations provided in the following embodiments merely illustrate the basic idea of the present invention by way of illustration, and the following embodiments and features in the embodiments may be combined with each other without conflict.
Referring to fig. 1 to 7, fig. 5 is a flowchart of a method for predicting the remaining life of a gear based ON SAE and ON-LSTM, wherein ON-LSTM aims at the disorder characteristic of inter-layer neurons of a general neural network, a tree structure is embedded between neurons to mine sequence information of the inter-layer neurons, and then hierarchical updating is performed ON different level information, so that high level information is not easy to replace, update rules of the middle level information are the same as those of a traditional long short term memory network (LSTM), and low level information is more easily affected by input data at the current moment. The ON-LSTM has advantages over conventional LSTM and its variants in the field of gear life prediction by using other neural networks that ignore sequence information. The prediction method based ON the ON-LSTM comprises the following steps:
S1: collecting gear vibration signals with the time length of T at intervals of delta T until the gears fail, wherein the number of sampled gear vibration signal segments is n;
s2: processing the gear vibration signal by using SAE to obtain an n-dimensional characteristic vector V;
s3: selecting a feature matrix V1 consisting of the n1 sampling points as a training matrix;
S4: using least squares to function the target And (3) carrying out minimum, normalizing all elements in the vector V with the vector V1 through a formula V' i=avi +b, wherein a and b respectively represent the weight and bias of a function used in the process of unifying all element vectors V1 in the vector V, and the specific size is determined through the minimization of the objective function;
S5: normalizing the vector V1 to obtain a normalized vector w= (W 1,w2,…,wn1)T;
S6: reconstruction matrix Wherein p is the number of units of the input layer of the neural network;
S7: taking the p rows in front of the matrix U as the input of an ON-LSTM (ON-LSTM) of a long-term memory neural network of ordered neurons, and taking the last row as the output of the ON-LSTM neural network to train the network;
s8: after training the matrix U, taking p outputs of the reciprocal as the network input of the ON-LSTM to obtain the output at the next moment;
S9: and repeating the step S8 for a certain number of times, comparing the output denormalized values with an actual characteristic value V '= (V' p+1,v'p+2,…,v'n)T), and when the output denormalized values exceed a set threshold value, multiplying the predicted sampling point number by the sum delta t+T of the interval time and the sampling time of the gear vibration signal to obtain the residual service life of the gear.
As shown in fig. 1, the SAE sparse automatic encoder has the structure: let Y t=[y1 y2 … yn be the original vibration signal for the entire time dimension, where Y t=[yt,1 yt,2 … yt,m represents the input data at time t, where m represents the number of signal points acquired during each time period. Through the first layer network, a vector z t with 50 dimensions can be obtained; through the second layer network, a 1-dimensional health index x t can be obtained; then through the third layer network, a 50-dimensional vector z' t can be obtained; through the last layer of network, an m-dimensional vector y t' can be obtained. The least square method is used as an objective function to make the difference between y t and y t' as small as possible, and x t is used as a health index at the time t to be applied to a subsequent prediction network.
Wherein z t denotes an output of the first layer network, x t denotes an output of the second layer network, z t denotes an output of the hidden third layer network, y' t denotes an output of the fourth layer network, σ denotes a sigmod activation function, W 1 denotes a weight of the first layer network, W 2 denotes a weight of the second layer network, W 3 denotes a weight of the third layer network, W 4 denotes a weight of the fourth layer network, b 1 denotes a bias of the first layer network, b 2 denotes a bias of the second layer network, b 3 denotes a bias of the third layer network, and b 4 denotes a bias of the fourth layer network.
As shown in fig. 2 to 4, the hierarchical structure of the input data and the history data is divided by a learnable hierarchy definer in the ON-LSTM. The position of the largest element in the output vector of the main forgetting gate and the main input gate is defined as a hierarchical position, and the division of the hierarchical structure is determined by the largest element, so that the tree hierarchical structure is embedded into the recurrent neural network. This means that the low-level information is more easily replaced by the input data at the next time; while high-level information is not easily replaced; the update rules of the middle level information follow the original LSTM. The deducing process of the ON-LSTM structure is as follows:
Assuming that the input information of the whole time dimension is X t=[x1 x2 … xt ], wherein X t=[xt,1 xt,2 … xt,n represents input data at the time t, and h t-1=[ht-1,1 ht-1,2 … ht-1,m represents recursive data of the neural network at the time t-1; first, a hierarchy divider is defined, whose hierarchy position L1 is determined by the constructor F 1 for the input information x t, and whose hierarchy position L2 is also determined by the constructor F 2 for the history information h t-1.
L1=F1(xt,ht-1) (2)
L2=F2(xt,ht-1) (3)
Wherein F 1 and F 2 are defined as:
And outputting values of positions of maximum elements in the vector after nonlinear combination of the input data and the historical data. During the network training process, x t+1,n is set to the query vector q m; in the prediction phase, x t,n is set to q m.
In the structure of the ON-LSTM neural network, the neural network hierarchy is divided by the required L1 and L2 ordering of the cell units. When L2 is greater than or equal to L1, the historical data level is lower than the current input level, and the overlapped part of the levels isThe cell memory unit c t is updated by the following rule:
When L2 < L1, the current input level is lower than the history level, and the cell memory unit c t is updated by the following rule:
with the division of hierarchical information, a variant structure of an LSTM neural network is derived as follows:
Where f t denotes the output of the forget gate in the hidden layer, i t denotes the output of the input gate in the hidden layer, o t denotes the output of the output gate in the hidden layer, Representing the output of the hyperbolic tangent function in the hidden layer,/>Output representing main forgetting gate in hidden layer,/>Representing the output of a main input gate in the hidden layer, σ representing sigmod activation functions, tanh representing hyperbolic tangent functions, W f representing the weight between the input layer and a forgetting gate in the hidden layer, W i representing the weight between the input layer and an input gate in the hidden layer, W o representing the weight between the input layer and an output gate in the hidden layer, W c representing the weight between the input layer and the hyperbolic tangent functions in the hidden layer,/>Representing the weight between the main forgetting gate in the input layer and hidden layer,/>Representing the weight between the input gate and the main output gate in the hidden layer, U f representing the weight between the forgetting gate in the hidden layer at the current time and the hidden layer output at the last time, U i representing the weight between the input gate in the hidden layer at the current time and the hidden layer output at the last time, U o representing the weight between the output gate in the hidden layer at the current time and the hidden layer output at the last time, U c representing the weight between the hyperbolic tangent function in the hidden layer at the current time and the hidden layer output at the last time,/>Representing the weight between the main forgetting gate in the hidden layer at the current moment and the output of the hidden layer at the last moment,/>Representing the weight between the main input gate in the hidden layer at the current moment and the output of the hidden layer at the previous moment, b f representing the bias of the linear combination in the forgetting gate in the hidden layer, b i representing the bias of the linear combination in the input gate in the hidden layer, b o representing the bias of the linear combination in the output gate in the hidden layer, b c representing the bias of the linear combination in the hyperbolic tangent function in the hidden layer,/>Bias representing linear combinations in the main forgetting gate in hidden layer,/>Representing the bias of the linear combination in the main input gate in the hidden layer,/>Representing the multiplication operation symbol of the corresponding element, c t-1 represents the output of the memory cell memory unit at time t-1, function cumax (.+ -.) = cumsum (softmax (.+ -.)), cumsum is the summation function,/>Representing middle level,/>Representing a low-level of the hierarchy,Representing a high level.
And (3) experimental verification:
The experiment adopts a mode of first-stage transmission acceleration and second-stage transmission deceleration, so that the transmission ratio of the experiment gearbox is just 1:1. The experimental gear is made of 40Cr, the machining precision is 5 grades, the surface hardness is 55HRC, and the modulus is 5. In particular, the number of teeth of the large gear is 31, the number of teeth of the small gear is 25, and the width of the first stage transmission gear is 21mm. In the experiment, the torque is 1400 N.m, the rotation speed of the large gear is 500r/min, the lubricating oil quantity of the experimental gearbox is 4L/h, and the cooling temperature is 70 ℃. The mode of collecting data selects all data in the collection process. Because of the high torque, the first stage drive gearwheel fails to break after 814 minutes of operation.
On the premise of knowing 380 fusion characteristic points, the gear residual life is predicted, and the predicted result is shown in fig. 6. The present invention provides an ON-LSTM with higher prediction accuracy than conventional LSTMs, see fig. 7, where MAE (mean absolute error), NRMSE (standard root mean square error), score (scoring function of american society of electrical and electronic engineers) for life prediction performance) are three evaluation criteria. As can be easily seen from fig. 7, the ON-LSTM can fully grasp the health information contained in the input data due to the excavation of the layer sequence information, so that local optimization is easier to obtain, and high-precision prediction of the remaining service life of the gear is easier to realize.
Finally, it is noted that the above embodiments are only for illustrating the technical solution of the present invention and not for limiting the same, and although the present invention has been described in detail with reference to the preferred embodiments, it should be understood by those skilled in the art that modifications and equivalents may be made thereto without departing from the spirit and scope of the present invention, which is intended to be covered by the claims of the present invention.

Claims (5)

1. The method for predicting the residual life of the gear based ON SAE and ON-LSTM is characterized by comprising the following steps:
S1: collecting gear vibration signals with the time length of T at intervals of delta T until the gears fail, wherein the number of sampled gear vibration signal segments is n;
s2: processing the gear vibration signal by using SAE to obtain an n-dimensional characteristic vector V; wherein SAE represents a sparse self-encoder;
s3: selecting a feature matrix V1 consisting of the n1 sampling points as a training matrix;
S4: using least squares to function the target The minimum value is then normalized to the vector V1 by the formula V' i=avi +b, where a, b represent the weight and bias of the function used in the unification of the element vector V1 in the vector V ,V1=(v11,v12,…,v1n1)T,V=(v1,v2,...,vn)T;
S5: normalizing the vector V1 to obtain a normalized vector w= (W 1,w2,...,wn1)T;
S6: reconstruction matrix Wherein p is the number of units of the input layer of the neural network;
S7: taking the p rows in front of the matrix U as the input of an ON-LSTM (ON-LSTM) of a long-term memory neural network of ordered neurons, and taking the last row as the output of the ON-LSTM neural network to train the network;
s8: after training the matrix U, taking p outputs of the reciprocal as the network input of the ON-LSTM to obtain the output at the next moment;
S9: and repeating the step S8 for a certain number of times, comparing the output denormalized values with an actual characteristic value V '= (V' p+1,v'p+2,...,v'n)T), and when the output denormalized values exceed a set threshold value, multiplying the predicted sampling point number by the sum delta t+T of the interval time and the sampling time of the gear vibration signal to obtain the residual service life of the gear.
2. The method for predicting the remaining life of a gear based ON SAE and ON-LSTM as set forth in claim 1, wherein in step S2, the SAE is structured as follows: setting an original vibration signal of the whole time dimension as Y t=[y1 y2 … yn, wherein Y t=[yt,1 yt,2 … yt,m represents input data at the moment t, and m represents the number of signal points acquired in each time period; obtaining a vector z t with 50 dimensions through a first layer network; obtaining a 1-dimensional health index x t through a second-layer network; then, a 50-dimensional vector z' t is obtained through a third layer network; obtaining an m-dimensional vector y' t through a last layer of network; using a least square method as an objective function to make y t and y' t small in difference, and applying x t as a health index at the time t to a subsequent prediction network;
Where z t denotes the output of the first layer network, x t denotes the output of the second layer network, z 't denotes the output of the hidden third layer network, y' t denotes the output of the fourth layer network, σ denotes sigmod the activation function, W 1、W2、W3、W4 denotes the weights of the first, second, third and fourth layer networks, respectively, and b 1、b2、b3、b4 denotes the bias of the first, second, third and fourth layer networks, respectively.
3. The method for predicting the remaining life of a gear based ON SAE and ON-LSTM as set forth in claim 1, wherein the ON-LSTM neural network has a structure of: setting the input information of the whole time dimension as X t=[x1 x2 … xt ], wherein X t=[xt,1 xt,2 … xt,n represents input data at the time t, and h t-1=[ht-1,1 ht-1,2 … ht-1,m represents recursive data of the neural network at the time t-1; first, defining a hierarchy divider, determining a hierarchy position L1 of the input information x t through a constructor F 1, and determining a hierarchy position L2 of the history information h t-1 through a constructor F 2; with the division of hierarchical information, the structure of an ON-LSTM neural network is expressed as:
Where f t denotes the output of the forget gate in the hidden layer, i t denotes the output of the input gate in the hidden layer, o t denotes the output of the output gate in the hidden layer, Representing the output of the hyperbolic tangent function in the hidden layer,/>Output representing main forgetting gate in hidden layer,/>Representing the output of the main input gate in the hidden layer, σ represents sigmod the activation function, tanh represents the hyperbolic tangent function, W f represents the weight between the input layer and the forgetting gate in the hidden layer, W i represents the weight between the input layer and the input gate in the hidden layer, W o represents the weight between the input layer and the output gate in the hidden layer, W c represents the weight between the hyperbolic tangent function in the input layer and the hidden layer,Representing the weight between the main forgetting gate in the input layer and hidden layer,/>Representing the weight between the input gate and the main output gate in the hidden layer, U f representing the weight between the forgetting gate in the hidden layer at the current time and the hidden layer output at the last time, U i representing the weight between the input gate in the hidden layer at the current time and the hidden layer output at the last time, U o representing the weight between the output gate in the hidden layer at the current time and the hidden layer output at the last time, U c representing the weight between the hyperbolic tangent function in the hidden layer at the current time and the hidden layer output at the last time,/>Representing the weight between the main forgetting gate in the hidden layer at the current moment and the output of the hidden layer at the last moment,/>Representing the weight between the main input gate in the hidden layer at the current moment and the output of the hidden layer at the previous moment, b f representing the bias of the linear combination in the forgetting gate in the hidden layer, b i representing the bias of the linear combination in the input gate in the hidden layer, b o representing the bias of the linear combination in the output gate in the hidden layer, b c representing the bias of the linear combination in the hyperbolic tangent function in the hidden layer,/>Bias representing linear combinations in the main forgetting gate in hidden layer,/>Representing the bias of the linear combination in the main input gate in the hidden layer,/>Representing the multiplication operation symbol of the corresponding element, c t-1 represents the output of the memory cell memory unit at time t-1, function cumax (.+ -.) = cumsum (softmax (.+ -.)), cumsum is the summation function,/>Representing middle level,/>Representing a low level,/>Representing a high level.
4. A method for predicting the remaining life of a gear based ON SAE and ON-LSTM according to claim 3, characterized in that in the structure of the ON-LSTM neural network, the level position L1 is determined by the constructor F 1 for the input information x t, and the level position L2 is also determined by the constructor F 2 for the history information h t-1, expressed as follows:
L1=F1(xt,ht-1)
L2=F2(xt,ht-1)
Wherein F 1 and F 2 are defined as:
Outputting values of positions of maximum elements in the vector after nonlinear combination of the input data and the historical data; during the network training process, x t+1,n is set to the query vector q m; in the prediction phase, x t,n is set to q m.
5. The method for predicting gear residual life based ON SAE and ON-LSTM according to claim 4, wherein in the structure of the ON-LSTM neural network, the neural network hierarchy is divided by ordering the cell units by the L1 and L2; when L2 is greater than or equal to L1, the historical data level is lower than the current input level, and the overlapped part of the levels isThe cell memory unit c t is updated by the following rule:
When L2 < L1, the current input level is lower than the history level, and the cell memory unit c t is updated by the following rule:
CN202010255897.8A 2020-04-02 2020-04-02 SAE and ON-LSTM-based gear residual life prediction method Active CN111475987B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010255897.8A CN111475987B (en) 2020-04-02 2020-04-02 SAE and ON-LSTM-based gear residual life prediction method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010255897.8A CN111475987B (en) 2020-04-02 2020-04-02 SAE and ON-LSTM-based gear residual life prediction method

Publications (2)

Publication Number Publication Date
CN111475987A CN111475987A (en) 2020-07-31
CN111475987B true CN111475987B (en) 2024-05-24

Family

ID=71749576

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010255897.8A Active CN111475987B (en) 2020-04-02 2020-04-02 SAE and ON-LSTM-based gear residual life prediction method

Country Status (1)

Country Link
CN (1) CN111475987B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112926505B (en) * 2021-03-24 2022-11-11 重庆大学 Rotating machine health index construction method based on DTC-VAE neural network
CN113642248B (en) * 2021-08-30 2023-11-07 平安国际融资租赁有限公司 Method and device for evaluating residual use time of positioning equipment

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109343505A (en) * 2018-09-19 2019-02-15 太原科技大学 Gear method for predicting residual useful life based on shot and long term memory network
CN110516833A (en) * 2019-07-03 2019-11-29 浙江工业大学 A method of the Bi-LSTM based on feature extraction predicts road traffic state
CN110941928A (en) * 2019-11-26 2020-03-31 哈尔滨理工大学 Rolling bearing residual life prediction method based on dropout-SAE and Bi-LSTM

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109343505A (en) * 2018-09-19 2019-02-15 太原科技大学 Gear method for predicting residual useful life based on shot and long term memory network
CN110516833A (en) * 2019-07-03 2019-11-29 浙江工业大学 A method of the Bi-LSTM based on feature extraction predicts road traffic state
CN110941928A (en) * 2019-11-26 2020-03-31 哈尔滨理工大学 Rolling bearing residual life prediction method based on dropout-SAE and Bi-LSTM

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
A deep learning approach for anomaly detection based on SAE and LSTM in mechanical equipment;Zhe Li等;《The International Journal of Advanced Manufacturing Technology》;第103卷;499–510 *
基于ANFIS的外啮合齿轮泵寿命预测研究;郭锐等;《仪器仪表学报》;第41卷(第01期);223-232 *
基于奇异值分解与深度递归神经网络的齿轮剩余寿命预测方法研究;张清亮;《中国优秀硕士学位论文全文数据库 (工程科技Ⅱ辑)》(第04期);C035-114 *
基于深度学习的航空发动机剩余使用寿命预测研究;温海茹;;内燃机与配件(第03期);41-42 *

Also Published As

Publication number Publication date
CN111475987A (en) 2020-07-31

Similar Documents

Publication Publication Date Title
CN111274737A (en) Method and system for predicting remaining service life of mechanical equipment
CN107145634B (en) Multi-state dynamic reliability assessment method for shield cutter head and driving system
CN109766583A (en) Based on no label, unbalanced, initial value uncertain data aero-engine service life prediction technique
CN111475987B (en) SAE and ON-LSTM-based gear residual life prediction method
CN111723527B (en) Method for predicting residual life of gear based on cocktail long-short-term memory neural network
CN110175425B (en) Prediction method of residual life of gear based on MMALSTM
CN108520320A (en) A kind of equipment life prediction technique based on multiple shot and long term memory network and Empirical Bayes
CN110210126B (en) LSTMPP-based gear residual life prediction method
CN113188794B (en) Gearbox fault diagnosis method and device based on improved PSO-BP neural network
CN114218872B (en) DBN-LSTM semi-supervised joint model-based residual service life prediction method
CN114692993B (en) Water conservancy facility deformation prediction method integrating ARIMA and BiLSTM
CN114282443B (en) Residual service life prediction method based on MLP-LSTM supervised joint model
CN111475986B (en) LSTM-AON-based gear residual life prediction method
CN112149953B (en) Electromechanical equipment operation safety assessment method based on multimode linkage and multistage cooperation
CN115758290A (en) Fan gearbox high-speed shaft temperature trend early warning method based on LSTM
Xue et al. An improved generic hybrid prognostic method for RUL prediction based on PF-LSTM learning
CN115329853B (en) Equipment parameter prediction and knowledge transfer method based on multi-source domain migration
CN114492642A (en) Mechanical fault online diagnosis method for multi-scale element depth residual shrinkage network
CN114757365A (en) High-speed railway roadbed settlement prediction and early warning method based on deep learning
CN113029619A (en) Underground scraper fault diagnosis method based on C4.5 decision tree algorithm
Liu et al. A novel health prognosis method for system based on improved degenerated Hidden Markov model
Du et al. Stacked convolutional LSTM models for prognosis of bearing performance degradation
CN116628444A (en) Water quality early warning method based on improved meta-learning
Wang et al. Similarity-based echo state network for remaining useful life prediction
CN114818500A (en) Method for predicting soil bin pressure based on LSTM algorithm

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant