CN117272782A - Hot rolled steel mechanical property prediction method based on self-adaptive multi-branch depth separation - Google Patents
Hot rolled steel mechanical property prediction method based on self-adaptive multi-branch depth separation Download PDFInfo
- Publication number
- CN117272782A CN117272782A CN202310986868.2A CN202310986868A CN117272782A CN 117272782 A CN117272782 A CN 117272782A CN 202310986868 A CN202310986868 A CN 202310986868A CN 117272782 A CN117272782 A CN 117272782A
- Authority
- CN
- China
- Prior art keywords
- convolution
- channel
- data
- model
- bottleneck
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 54
- 229910000831 Steel Inorganic materials 0.000 title claims abstract description 47
- 239000010959 steel Substances 0.000 title claims abstract description 47
- 238000000926 separation method Methods 0.000 title claims description 8
- 238000012549 training Methods 0.000 claims abstract description 22
- 239000011159 matrix material Substances 0.000 claims abstract description 11
- 238000013527 convolutional neural network Methods 0.000 claims description 23
- 230000003044 adaptive effect Effects 0.000 claims description 18
- 238000005096 rolling process Methods 0.000 claims description 13
- 230000006870 function Effects 0.000 claims description 11
- 238000010438 heat treatment Methods 0.000 claims description 10
- OKTJSMMVPCPJKN-UHFFFAOYSA-N Carbon Chemical compound [C] OKTJSMMVPCPJKN-UHFFFAOYSA-N 0.000 claims description 9
- 229910052799 carbon Inorganic materials 0.000 claims description 9
- 238000011156 evaluation Methods 0.000 claims description 9
- 238000010606 normalization Methods 0.000 claims description 9
- 238000011176 pooling Methods 0.000 claims description 9
- 238000005457 optimization Methods 0.000 claims description 8
- 239000000126 substance Substances 0.000 claims description 8
- 238000012360 testing method Methods 0.000 claims description 8
- 238000004364 calculation method Methods 0.000 claims description 7
- 238000010586 diagram Methods 0.000 claims description 7
- OAICVXFJPJFONN-UHFFFAOYSA-N Phosphorus Chemical compound [P] OAICVXFJPJFONN-UHFFFAOYSA-N 0.000 claims description 6
- 230000000694 effects Effects 0.000 claims description 6
- 229910052698 phosphorus Inorganic materials 0.000 claims description 6
- 239000011574 phosphorus Substances 0.000 claims description 6
- 238000012545 processing Methods 0.000 claims description 6
- XUIMIQQOPSSXEZ-UHFFFAOYSA-N Silicon Chemical compound [Si] XUIMIQQOPSSXEZ-UHFFFAOYSA-N 0.000 claims description 5
- 229910052710 silicon Inorganic materials 0.000 claims description 5
- 239000010703 silicon Substances 0.000 claims description 5
- NINIDFKCEFEMDL-UHFFFAOYSA-N Sulfur Chemical compound [S] NINIDFKCEFEMDL-UHFFFAOYSA-N 0.000 claims description 4
- 230000004913 activation Effects 0.000 claims description 4
- 238000013528 artificial neural network Methods 0.000 claims description 4
- 229910052717 sulfur Inorganic materials 0.000 claims description 4
- 239000011593 sulfur Substances 0.000 claims description 4
- NAWXUBYGYWOOIX-SFHVURJKSA-N (2s)-2-[[4-[2-(2,4-diaminoquinazolin-6-yl)ethyl]benzoyl]amino]-4-methylidenepentanedioic acid Chemical compound C1=CC2=NC(N)=NC(N)=C2C=C1CCC1=CC=C(C(=O)N[C@@H](CC(=C)C(O)=O)C(O)=O)C=C1 NAWXUBYGYWOOIX-SFHVURJKSA-N 0.000 claims description 3
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 claims description 3
- 238000012935 Averaging Methods 0.000 claims description 3
- 238000006243 chemical reaction Methods 0.000 claims description 3
- WPBNNNQJVZRUHP-UHFFFAOYSA-L manganese(2+);methyl n-[[2-(methoxycarbonylcarbamothioylamino)phenyl]carbamothioyl]carbamate;n-[2-(sulfidocarbothioylamino)ethyl]carbamodithioate Chemical compound [Mn+2].[S-]C(=S)NCCNC([S-])=S.COC(=O)NC(=S)NC1=CC=CC=C1NC(=S)NC(=O)OC WPBNNNQJVZRUHP-UHFFFAOYSA-L 0.000 claims description 2
- 230000006872 improvement Effects 0.000 abstract description 10
- 238000000605 extraction Methods 0.000 abstract description 7
- 230000008569 process Effects 0.000 description 13
- 238000005098 hot rolling Methods 0.000 description 6
- 238000004519 manufacturing process Methods 0.000 description 5
- 230000006835 compression Effects 0.000 description 4
- 238000007906 compression Methods 0.000 description 4
- 239000011572 manganese Substances 0.000 description 4
- 239000000243 solution Substances 0.000 description 4
- PWHULOQIROXLJO-UHFFFAOYSA-N Manganese Chemical compound [Mn] PWHULOQIROXLJO-UHFFFAOYSA-N 0.000 description 3
- 230000014509 gene expression Effects 0.000 description 3
- 230000035882 stress Effects 0.000 description 3
- 230000006399 behavior Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 150000001875 compounds Chemical class 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 238000009826 distribution Methods 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 229910052748 manganese Inorganic materials 0.000 description 2
- 238000001953 recrystallisation Methods 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 239000007787 solid Substances 0.000 description 2
- 239000006104 solid solution Substances 0.000 description 2
- 229910000975 Carbon steel Inorganic materials 0.000 description 1
- 238000003723 Smelting Methods 0.000 description 1
- UCKMPCXJQFINFW-UHFFFAOYSA-N Sulphide Chemical compound [S-2] UCKMPCXJQFINFW-UHFFFAOYSA-N 0.000 description 1
- 238000002679 ablation Methods 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 239000010962 carbon steel Substances 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000001816 cooling Methods 0.000 description 1
- 239000013078 crystal Substances 0.000 description 1
- 230000003811 curling process Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000013210 evaluation model Methods 0.000 description 1
- 230000006355 external stress Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 239000007769 metal material Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000001556 precipitation Methods 0.000 description 1
- 238000002360 preparation method Methods 0.000 description 1
- 238000004886 process control Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 238000007670 refining Methods 0.000 description 1
- 230000001105 regulatory effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 238000003860 storage Methods 0.000 description 1
- 238000005728 strengthening Methods 0.000 description 1
- 239000000758 substrate Substances 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
- 239000013585 weight reducing agent Substances 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
- G06F30/20—Design optimisation, verification or simulation
- G06F30/27—Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2119/00—Details relating to the type or aim of the analysis or the optimisation
- G06F2119/14—Force analysis or force optimisation, e.g. static or dynamic forces
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02P—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
- Y02P90/00—Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
- Y02P90/30—Computing systems specially adapted for manufacturing
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Evolutionary Computation (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Computing Systems (AREA)
- Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- General Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Geometry (AREA)
- Medical Informatics (AREA)
- Computer Hardware Design (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a method for predicting mechanical properties of hot rolled steel based on self-adaptive multi-branch depth separable, and relates to the technical field of prediction of mechanical properties of hot rolled steel. On one hand, the invention aims at the problem of weaker original data characteristic representation capability, and provides a Gram Angle Field (GAF) method, which is used for converting original one-dimensional data into a two-dimensional matrix and enriching the correlation characteristics among the data, so that the model can extract more effective characteristic information and obtain better training, thereby realizing the improvement of performance. On the other hand, in order to improve model precision when the high real-time requirement of task compactness is solved, a combination module of multi-branch depth separable convolution (MB-DSC) and a self-adaptive bottleneck layer (AB) is provided, the characteristic extraction capacity is improved, intuitionistic is introduced on the basis of depth separable, the model performance is improved while the model is lightened, and the relationship between the model and the model is balanced, so that a lightweight high-precision model is constructed.
Description
Technical Field
The invention relates to the technical field of prediction of mechanical properties of hot rolled steel, in particular to a method for predicting the mechanical properties of hot rolled steel based on self-adaptive multi-branch depth separable.
Background
With the development and progress of the times, the requirements of people on the mechanical properties of the hot rolled steel are increasingly strict, and industrial products produced in industries such as construction, bridges, traffic and the like are not separated from the excellent mechanical properties of the hot rolled steel.
At present, in the smelting process of the hot-rolled steel, the microstructure of the hot-rolled steel crystal phase can be changed differently due to the influence of different technological parameters and chemical component contents, and the microstructure directly determines the mechanical properties of the hot-rolled steel. The process of hot rolling is shown in fig. 1. Important mechanical properties of hot rolled steel include Yield Strength (YS), tensile Strength (TS) and Elongation (EL). YS refers to the stress magnitude when the metal material is subjected to external stress and is subjected to plastic deformation, and is also called the yield limit when the yield phenomenon occurs. TS refers to the maximum tensile stress to which the steel is subjected before breaking under tensile stress. EL means that when a steel sheet is stretched by an external force until it breaks, the percentage of the stretched length to the length of the original sample is called elongation.
The influencing factors on the performance index are divided into two types, namely process parameters and chemical components. In terms of process parameters, the furnace heating temperature (FT) has a certain influence on the Start Rolling Temperature (SRT). Higher heating temperatures can increase the plasticity and heat deformability of the steel, resulting in a relatively low initial rolling temperature. Lower initial rolling temperatures can reduce recrystallization behavior during hot deformation, thereby refining the grains and improving the strength and toughness of the steel. During the process of reducing the initial rolling temperature to the final rolling temperature, the mechanical properties may be reduced due to the influence of time and speed, and the excessively high or excessively low Final Rolling Temperature (FRT) may further influence the Coiling Temperature (CT), and the cooling rate, the phase change behavior and the curling process of the steel are influenced by the coiling temperature. In addition, the degree of compression (Redn) also has an effect on the finishing temperature. Higher degrees of compression will generate more heat, resulting in a relatively higher finishing temperature. Therefore, in selecting the finishing temperature, the limitation of the degree of compression needs to be considered in order to avoid the decrease in strength and toughness of the steel caused by excessive recrystallization and grain growth. In terms of chemical content, the interrelationship has a non-negligible effect on the mechanical properties. Increasing the content of carbon (C) and silicon (Si) can increase the yield strength and tensile strength of the steel, but can decrease the elongation and cold workability. However, too high a carbon and silicon content may result in increased brittleness of the steel. The proper amount of manganese (Mn) content can promote solid solution strengthening of carbon, thereby improving the yield strength and the tensile strength of the hot rolled steel. Phosphorus (P) is also present in solid solution in steel to strengthen the steel rolling, but when manganese forms a compound with phosphorus, the solid solubility of phosphorus is reduced, reducing its content. When sulfur (S) forms a compound with carbon, the solid solubility of carbon is lowered, resulting in precipitation of carbon and formation of brittle sulfide, thereby further lowering the performance of the steel.
In the past, a random sampling method or a physical metallurgical model method is adopted to measure the mechanical properties of each batch of products, so that the nonlinear relation in the production process is difficult to describe, a great deal of time and labor cost are also spent, and most importantly, the method depends on personal experience of an operator, so that a correct and effective conclusion cannot be obtained.
The prior application (publication No. CN 110472349A) provides a one-dimensional CNN model for mechanical property prediction aiming at the field. And the prior invention application (publication No. CN 116307195A) uses a sequential interpolation method to enhance the data of the field data, and performs rolling width prediction by combining a CNN model. Conventional convolution is a basic local operation, as shown in fig. 2. The core idea is to obtain an output feature map by performing an operation while sliding a small window called a convolution kernel on the input data. This convolution kernel is typically a small matrix that is much smaller in size than the input data. The convolution kernel is operated with the input data by element-wise multiplication, the multiplication results are summed to obtain a scalar value, and the scalar value is used as an element of the output feature map. The moving window repeats this process until the entire input data is traversed, resulting in an output profile. Typically, the model will convolve the input features using multiple convolution checks to achieve feature rich objectives.
Convolution kernels play a key role in the convolution process. Convolution kernels of different sizes and shapes can be used to extract different features, which provides the possibility for the network to learn more advanced feature expressions, inspiring the idea of subsequent optimization improvement. However, compared with other image data, the hot rolled steel data still has unsatisfactory feature performance, which is shown in unreasonable way of the data enhancement method, and the primary purpose is to make the original one-dimensional data more in line with the feature extraction process of the convolutional neural network in the process of converting the original one-dimensional data into two-dimensional data, and meanwhile, more relevant feature is introduced to assist the learning training of the network. And for the hot rolling prediction problems of wide development and compact tasks, how to construct a lightweight model with excellent performance has important value for the production and utilization of hot rolled steel.
However, the prior art proposes to use sequential interpolation to convert the original data into two-dimensional data for data enhancement. The method has the defects that correlation characteristics among data are not fully considered, simple sequential arrangement is only performed, original data are placed in a space matrix with a specified size one by one, dimensional change is realized, and the method has little help to model performance improvement. Meanwhile, the prior art provides a CNN model for the hot rolling performance prediction problem, and the model performance is improved by applying the strong characteristic extraction capability of the CNN model. However, in actual production, the hot rolling process has high real-time requirements, front-end equipment resources are limited, and the traditional CNN model is thick and complicated.
Disclosure of Invention
The invention aims at two problems existing in the prior art: on the one hand, in the field of hot-rolled steel mechanical property prediction, the hot-rolled steel data is complex and multi-coupled, and the correlation characteristics among influence factors can be ignored when the hot-rolled steel data is directly used as the input of a network, so that the model prediction precision is insufficient; on the other hand, in the field, the hot rolling process is a widely developed and compact process, has a plurality of prediction index tasks, has real-time requirements, and needs a lightweight model under the condition that the computing resources of front-end equipment are limited. Therefore, the self-adaptive multi-branch depth separable hot-rolled steel mechanical property prediction method is provided, the original one-dimensional data is converted into two-dimensional data by adopting a gram angle field method, and the data enhancement is realized by introducing data correlation characteristics, so that more effective characteristic information can be extracted from a model, and the performance is improved; and a depth separable convolution module combining a multi-branch enhanced convolution kernel and a self-adaptive bottleneck layer is provided to balance the relation between model precision and light weight, so that a light weight high-precision model is constructed.
In order to achieve the above object, the present invention provides the following technical solutions:
the invention provides a method for predicting mechanical properties of hot rolled steel based on self-adaptive multi-branch depth separable, wherein input parameters comprise 5 heat treatment parameters and 5 chemical components, the 5 heat treatment parameters comprise heating temperature in a furnace, start rolling temperature, finish rolling temperature, coiling temperature and compressibility, and the 5 chemical components comprise carbon, silicon, manganese, phosphorus and sulfur, and the method comprises the following steps:
s1, carrying out data enhancement on original one-dimensional data to obtain a two-dimensional matrix, and dividing a data set;
s2, constructing an optimized and improved convolutional neural network model; the optimized and improved convolutional neural network model uses depth separable convolutional operations to carry out channel-by-channel convolution and point-by-point convolution; the channel-by-channel convolution uses a single channel as a feature map, and local information of each channel is extracted by using each convolution kernel; the point-by-point convolution uses a convolution kernel of the same channel dimension as the channel-by-channel convolution, and integrates the convolution kernel into the same number of output feature maps as the convolution kernel of the point-by-point convolution; in a channel-by-channel convolution stage, constructing a multi-branch convolution neural network, wherein each branch represents convolution kernels with different sizes, and summing the multi-branch outputs; in the point-by-point convolution stage, combining the self-adaptive bottleneck layer with the self-adaptive bottleneck layer, and adopting global average pooling operation to self-adaptively adjust the bottleneck coefficient through weight size;
s3, setting super parameters, learning, training, optimizing and improving a convolutional neural network model on a training set, and judging through a loss function to obtain a trained model;
s4, predicting the test set data by using the trained model to obtain a hot rolled steel mechanical property prediction result.
Further, in step S1, the data enhancement is performed by using a glamer angle field method, which includes the steps of:
s11, performing normalization operation on the original one-dimensional data, and assuming that one-dimensional data X= { X exists 1 ,x 2 ,...,x N Normalized one-dimensional data to [ -1,1 } length N]As shown in formula (1):
wherein x is i Data for the ith position, x i The value of the position after normalization is min (X) and max (X) respectively represent the minimum value and the maximum value in X;
s12, representing the normalized data by using polar coordinates, wherein the value of the one-dimensional data is taken as a polar angle theta, and the value of the position where the corresponding value is located is taken as a polar axis r, as shown in the formula (2):
where i is a position value, L is a position length, r i For its polar axis, θ i For its polar angle, x after normalization i Is within the range of [ -1,1]So theta is i The value range of (2) is [0, pi ]]And cos θ is monotonic in this interval;
s13, summing different position angles to obtain a gram and an angle field, wherein the gram and the angle field are shown in a formula (3):
further, the method is characterized in that in step S1, the enhanced data is represented by 9: the ratio of 1 is divided into a training set and a test set.
Further, the step S2 of constructing an optimized and improved convolutional neural network model includes the steps of:
s21, performing two convolution processing steps of channel-by-channel convolution and point-by-point convolution by using depth separable convolution operation on the basis of CNN; let the input feature map size be D in ×D in X M, output feature map size D out ×D out The dimension of the x N corresponds to the height, width and channel number of the feature map, and the dimension of the convolution kernel is set as D k ×D k The method comprises the steps of carrying out a first treatment on the surface of the The number of convolutions per channel is M and the size is D k ×D k Processing by the x 1 convolution kernel to produce D out ×D out Feature map of x M; the point-by-point convolution uses N number and 1×1×M size convolution to check the intermediate feature map to perform channel conversion to generate D size out ×D out Feature map of x N;
s22, constructing a multi-branch convolution neural network from the aspect of channel-by-channel convolution on the basis of depth separable convolution, wherein each branch represents convolution kernels with different sizes, namely Kx K, K x 1 and 1 xK, and summing the outputs of the three branches;
s23, combining the self-adaptive bottleneck layer with the self-adaptive bottleneck layer from the aspect of point-by-point convolution, adopting global average pooling operation to adaptively adjust the size of the bottleneck coefficient through the weight size so as to realize self-adaptive bottleneck layer optimization, enabling the bottleneck coefficient to set the kernel number to be a value smaller than the number of input channels in a ratio mode, adopting a point convolution mode to perform restoration operation after the bottleneck, and recovering the number of channels;
s24, a global averaging layer (Global Average Pooling, GAP) is adopted to replace a full-connection layer (Fully Connected layers, FC) at the tail end, the feature map is converted into a vector with a fixed length, weight parameters in the FC layer are replaced, the number of weight parameters needing to be learned in a network is obviously reduced, and the model is lighter. Further, in step S23, it is assumed that the number of channels of the input feature map is C in The number of the output characteristic diagram channels is C out The channel number of the bottleneck layer is C b If the bottleneck coefficient is B and the bottleneck layer is not set, the point convolution calculation parameters are shown as formula (4), if the bottleneck layer is setThen the formula (5) is shown, wherein the bottleneck layer channel number is shown as formula (6), and the formula (7) is obtained according to the formulas (5) and (6), when P 2 Less than P 1 When present, formula (8) is present, which yields about formula (9):
P 1 =C in ×C out (4)
P 2 =C in ×C b +C b ×C out (5)
C b =B×C in (6)
further, step S3 includes the steps of:
s31, placing two-dimensional matrix data enhanced by GAF data into multi-branch depth separable convolution, performing enhanced convolution by using three branch paths in a channel-by-channel stage of the depth separable convolution, respectively using convolution kernels with the sizes of 3 multiplied by 3, 3 multiplied by 1 and 1 multiplied by 3, wherein the convolution kernels are increased by 64, 128, 256 and 512 along with the increase of the number of convolution layers, and performing total four-layer depth separable convolution; batch size is set to 220, reLU is used as a nonlinear activation function, model parameters are updated by adopting an Adam optimizer with a learning rate of 0.001, and iteration is carried out for 30 times;
s32, in a point-by-point convolution stage, adopting global average pooling operation to adaptively adjust the size of a bottleneck coefficient through weight magnitude so as to realize adaptive bottleneck layer optimization, and setting the number of cores to be a value smaller than the number of input channels in a ratio mode by the bottleneck coefficient;
s33, the loss function adopts a mean square error MSE shown in the formula (10), the weight bias parameter is updated through back propagation, the prediction error is minimized, and meanwhile, different evaluation indexes are used for describing the influence of the model on the prediction precision;
and S34, after training, saving weight bias to obtain a trained model structure.
6. The method for predicting mechanical properties of hot-rolled steel based on the separable depth of the adaptive multi-branch according to claim 1, wherein in step S33, different evaluation indexes are used, including mean absolute error MSE, root mean square error RMSE and determination coefficient R2, as shown in formulas (11) - (13), the closer the error is to 0, the closer the determination coefficient is to 1, and the better the fitting effect of the model is;
wherein y is i In order to predict the data,for the corresponding actual data +.>Is the average of the corresponding actual data.
Compared with the prior art, the invention has the beneficial effects that:
according to the self-adaptive multi-branch depth separable-based hot rolled steel mechanical property prediction method, on one hand, a Gram Angle Field (GAF) method is adopted to convert original one-dimensional data into a two-dimensional matrix aiming at the problem of weak original data characteristic representation capability, and correlation characteristics among data are enriched, so that the model can extract more effective characteristic information, better training is obtained, and further performance improvement is realized. On the other hand, in order to improve model precision when the high real-time requirement of task compactness is solved, a combination module of multi-branch depth separable convolution (Multi Branch Depthwise Separable Convolution, MB-DSC) and a self-adaptive bottleneck layer (AB) is provided, the feature extraction capability is improved, visual intuition is introduced on the basis of depth separation, the model performance is improved while the model is lightened, and the relationship between the model and the model is balanced, so that a lightweight high-precision model is constructed.
Drawings
In order to more clearly illustrate the embodiments of the present application or the technical solutions in the prior art, the drawings that are needed in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments described in the present invention, and other drawings may be obtained according to these drawings for a person having ordinary skill in the art.
FIG. 1 is a flow chart of a hot rolling process.
Fig. 2 is a diagram of a standard convolution process.
Fig. 3 is a view of GAF data enhancement provided by an embodiment of the present invention, where a is original one-dimensional data and b is GAF two-dimensional data.
Fig. 4 is a diagram of a multi-branch depth separable convolution process according to an embodiment of the present invention.
Fig. 5 is a schematic diagram of an adaptive bottleneck structure according to an embodiment of the present invention.
Fig. 6 is a schematic structural diagram of a multi-branch enhanced depth separable convolution model based on an adaptive bottleneck layer according to an embodiment of the present invention.
Detailed Description
In order to better understand the technical solution, the technical solution in the embodiment of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiment of the present invention. It will be apparent that the described examples are only some embodiments, but not all embodiments, of the present invention. Based on the embodiments of the present invention, those of ordinary skill in the art will be able to devise all other embodiments that are obtained based on this application and are within the scope of the present invention.
The invention provides a hot-rolled steel mechanical property prediction method based on self-adaptive multi-branch depth separable, a data-enhanced gram angle field method and a multi-branch depth separable model based on a self-adaptive bottleneck layer, so as to obtain a hot-rolled steel mechanical property prediction result with better performance in a light-weight state. Which comprises the following steps:
s1, carrying out data enhancement on original one-dimensional data to obtain a two-dimensional matrix with rich characteristics, and dividing a data set.
S2, constructing an optimized and improved convolutional neural network model,
s3, setting super parameters, learning a training model on the training set, and judging through a loss function to obtain a trained model. The superiority and the light weight of the model are proved by a plurality of evaluation index formulas.
S4, predicting the test set data by using the trained model. Obtaining better results.
Specifically, step S1 includes:
s11, data preparation. The data used in the experiment are hot rolled steel production data in a wine steel group carbon steel sheet factory based on a real industrial environment, accord with international standards and have practical significance, and the obtained experimental result can be used in a real life production environment. The original data is 7898 pieces in total. The input parameters include 5 heat treatment parameters including in-furnace heating temperature (FT), start Rolling Temperature (SRT), finish Rolling Temperature (FRT), coiling Temperature (CT), compressibility (Redn), and 5 chemical components including carbon (C), silicon (Si), manganese (Mn), phosphorus (P), sulfur (S). The output parameters include three important mechanical properties, namely Yield Strength (YS), tensile Strength (TS) and Elongation (EL). The data distribution statistics of the minimum, maximum, mean and standard deviation of all input parameters in the original dataset are shown in table 1.
TABLE 1 Hot Rolling data distribution
S12, data enhancement using the glamer angle field (Gramian Angular Field, GAF) method, comprising the steps of:
s121, performing normalization operation on the original one-dimensional data. Assume that there is one-dimensional data x= { X 1 ,x 2 ,...,x N Normalized one-dimensional data to [ -1,1 } length N]As shown in formula (1), wherein x i Data for the ith position, x i Is the value of the position after normalization. min (X) and max (X) represent the minimum and maximum values of X, respectively.
S122, the normalized data is represented using polar coordinates. Wherein the value of the one-dimensional data is taken as a polar angle theta, and the value of the position where the corresponding value is located is taken as a polar axis r. As shown in formula (2), wherein i is a position value, L is a position length, r i For its polar axis, θ i For its polar angle, the numerical relationship is preserved. Converting the recalibrated raw data onto a polar coordinate system, the interaction relationship between the components of the raw vector can be identified from an angular perspective by the triangles and/or differences between each point.
After normalization, x i Is within the range of [ -1,1]So theta is i The value range of (2) is [0, pi ]]And cos θ is monotonic in this interval, a unique mapping result is produced in polar coordinates.
And S123, summing the different position angles to obtain the gram and the angle field (Gramian summation angular field, GASF) as shown in the formula (3).
As shown in fig. 3, the data enhancement is performed visually on the unobvious characteristic relationship between the original one-dimensional data, and 10 input data which are originally one-dimensional are converted into 10×10 matrix data, so that a rich set of original characteristics and correlation characteristics with obvious distinction is obtained. Provides a more accurate and comprehensive data basis for the subsequent learning training of CNN. It should be noted that this section of data enhancement process is performed only once during the data acquisition phase, and does not add to the complexity of the model. The application of the method can bring important help and improvement to the aspects of process control, quality optimization and the like of the hot rolled steel.
S13, training set test set segmentation is carried out on the enhanced data. 9:1 to the training set and the test set.
Further, step S2 constructs an optimized and improved convolutional neural network model, comprising the steps of:
s21, channel dimension and spatial position features in the feature map can be decoupled, so that depth separable convolution operation is used on the basis of CNN, and the purpose of feature extraction consistent with the traditional convolution is achieved through two convolution processing steps of channel-by-channel convolution and point-by-point convolution, and meanwhile calculation parameters are effectively reduced, so that the purpose of light weight is achieved.
The process of depth separable convolution is illustrated in detail in fig. 4. Wherein the channel-by-channel convolution uses a single channel as a feature map and uses each convolution kernel to extract local information of the respective channel. Unlike full-channel convolution in conventional convolution, channel-by-channel convolution greatly reduces the number of convolution parameters. However, since the characteristic information obtained by the channel-by-channel convolution is relatively independent, the characteristic information between different channels is not considered to be extracted. Thus, by using a 1×1 convolution kernel of the same channel dimension as the channel-wise convolution, the same number of output feature maps as the convolution kernel of the point-wise convolution are integrated, integrating its channel information.
Under the condition of ensuring that the sizes of the input and output characteristic diagrams are the same, the parameter quantity and the calculated quantity of the two convolution processing methods are compared. Let the input feature map size be D in ×D in X M, output feature map size D out ×D out X N. The dimension dimensions correspond to the height, width, and number of channels of the feature map, respectively. The convolution kernel is sized to D k ×D k . Thus, the first and second substrates are bonded together,the reference quantity occupied by the standard convolution is D k ×D k X M N, calculated as D k ×D k ×M×D out ×D out X N. For depth separable convolution, a number M and a size D are used for channel-by-channel convolution k ×D k Processing by the x 1 convolution kernel to produce D out ×D out Feature map of xM, point-by-point convolution uses N number, size 1 x 1 xM convolution check intermediate feature map to make channel conversion to generate size D out ×D out Feature map of x N. The total parameter of the method is D k ×D k X1 XM+1 X1 XM N, calculated as D k ×D k ×D in ×D in ×M+1×1×M×D out ×D out X N. Comparing the parameter quantity and the calculated quantity to obtain the ratio of the formula (4) to the formula (5). Since Dk and N are both values greater than 1, the depth separable convolution has fewer parameters and less computation than the standard convolution.
S22, on the basis of the depth separable convolution, in the aspect of channel-by-channel convolution, in order to achieve performance improvement under a lightweight model, multi-branch convolution kernel structure improvement is used, the characteristic relation of the separable convolution kernels is enriched, and the model prediction performance is improved. Due to the additivity of convolution operations, two-dimensional convolution kernels of compatible sizes are operated on the same input with the same operation to obtain output features of the same size, and convolutions of compatible sizes can be added at corresponding positions to obtain an equivalent convolution kernel of the same output features. Multi-branch depth separable convolutions (Multi Branch Depthwise Separable Convolution, MB-DSC) were constructed, each branch representing a different size convolution kernel, K x K, K x 1 and 1 x K, respectively, and the three branch outputs were summed to enrich the feature space. Because of its additivity, the model whose training is complete can remain the same response time as a one-way convolutional neural network if it has already been deployed. Therefore, the model performance can be improved on the basis of not affecting the depth-separable light model.
S23, although the depth separable convolution module significantly reduces parameters and calculation amount of the model, we hope to combine the requirement with practical situations, and hope that important features can be learned more, so that the model can obtain higher precision, and less important features can be learned in a small amount, so as to further reduce occupation of parameters and time loss. It is therefore proposed that an adaptive bottleneck layer (Adaptive Bottleneck, AB) is combined with a multi-branch depth separable convolution module. And (5) balancing the relationship between the weight reduction and the precision of the model.
The bottleneck layer is a bottleneck-like layer with a small number of channels of the intermediate feature map, and is used for reducing the computational complexity and the number of parameters, and meanwhile, a certain model expression capability is maintained. The method is effective for eliminating unimportant features by reducing the data through a narrower layer and reducing the data, reducing the compression calculation cost. The invention adopts global average pooling operation to adaptively adjust the size of the bottleneck coefficient (Bottleneck coefficient) through weight size so as to realize the optimization of the adaptive bottleneck layer. The bottleneck coefficient is made to set the number of cores to a value smaller than the number of input channels in the form of a ratio. In order to prevent the loss of detail information in the dimension reduction operation, the reduction operation is performed in a point convolution mode after the bottleneck, the channel number is recovered, more nonlinear activation functions are introduced, and richer characteristic representations can be captured, so that the method is beneficial to the performance and the expression capability of the network. As can be obtained from equations (4) (5), the point convolution integrates the channel information with more computation than the channel-by-channel convolution. The adaptive bottleneck layer is combined with the depth separable convolution from the point-wise convolution aspect on the basis of it.
P 1 =C in ×C out (6)
P 2 =C in ×C b +C b ×C out (7)
C b =B×C in (8)
Let Cin be the number of channels of the input feature map, cout be the number of channels of the output feature map, cb be the number of channels of the bottleneck layer, and B be the bottleneck coefficient. If the bottleneck layer is not set, the point convolution calculation parameters are shown in the formula (6), and if the bottleneck layer is set as shown in the formula (7), wherein the number of the bottleneck layer channels is shown in the formula (8). Then the formula (9) can be obtained, when P2 is smaller than P1, the formula (10) exists, the formula (11) is obtained by about division, and under the condition that the formula is established, a lighter model can be obtained by adding the point convolution module of the self-adaptive bottleneck layer, and otherwise, a more accurate model can be obtained.
The transfer of bottleneck coefficients in the adaptive bottleneck structure is illustrated in fig. 6. The bottleneck coefficients are adaptively regulated according to the weights, so that the characteristic bottleneck layers with larger weights perform further characteristic extraction learning, and the model performance is improved; and the layers with smaller weights reduce the bottleneck, reduce model parameters, accelerate the overall training speed and improve the generalization capability of the model. Thereby balancing the weight and performance of the model.
S24, a global averaging layer is adopted to replace a traditional full-connection layer at the tail end, and the characteristic map is converted into a vector with a fixed length to replace a large number of weight parameters in the full-connection layer. The number of weight parameters to be learned in the network model is obviously reduced, the risk of overfitting is reduced, the calculation and storage requirements of the model are reduced, and the model is light. Meanwhile, important spatial information can be reserved, so that the model is less sensitive to the position of the target, and the robustness of the model is enhanced. Finally, the multi-branch enhanced depth separable convolution model based on the adaptive bottleneck layer as shown in fig. 6 is obtained.
S3, setting super parameters, learning a training model on the training set, and judging through a loss function to obtain a trained model. The superiority and the light weight of the model are proved by a plurality of evaluation index formulas.
Next, step S3 sets super parameters, and performs model learning on the training set, including the following steps:
s31, two-dimensional matrix data enhanced by GAF data are put into multi-branch depth separable convolution, enhancement convolution is carried out by using three branch paths in a channel-by-channel stage of the depth separable convolution, convolution kernels with the sizes of 3 multiplied by 3, 3 multiplied by 1 and 1 multiplied by 3 are respectively used, the convolution kernels are increased by 64, 128, 256 and 512 along with the increase of the number of convolution layers, and the total four layers of depth separable convolution is carried out. Batch size was set to 220, and ReLU was iterated 30 times as a nonlinear activation function, with model parameters updated using an Adam optimizer with a learning rate of 0.001.
S32, in the point-by-point convolution stage, the size of the bottleneck coefficient (Bottleneck coefficient) is adaptively adjusted through weight size by adopting global average pooling operation, so that adaptive bottleneck layer optimization is realized. The bottleneck coefficient is made to set the number of cores to a value smaller than the number of input channels in the form of a ratio.
And S33, the loss function adopts a mean square error (Mean Square Error, MSE) shown in a formula (12), and the weight bias parameter is updated through back propagation to minimize the prediction error. Meanwhile, different evaluation indexes, such as average absolute errors (Mean Absolute Error, MAE) and root mean square errors (Root Mean Square Error, RMSE) shown in the formula (13) and the formula (14) and a determination coefficient (R2) shown in the formula (15), are used for better describing the influence of the model on the prediction precision, and the closer the error is to 0, the closer the determination coefficient is to 1 and the better the fitting effect of the model is. Wherein y is i In order to predict the data,for the corresponding actual data +.>Is the average of the corresponding actual data.
And S34, after training, saving weight bias to obtain a trained model structure for subsequent direct use.
S4, predicting the test set data by using the trained model. The evaluation index in step S33 again proves that a better result can be obtained in practical application.
To further verify the effectiveness of the model improvement, the effect of each component in the evaluation model on overall performance was studied. The different improvement modules are stepped up to determine the contribution of each component to the overall model performance.
The experimental results are shown in Table 2, and it can be seen from a plurality of evaluation indexes that the GAF method introduces more features than the one-dimensional CNN (1 d-CNN), but increases the parameter burden of the model. The multi-branch depth separable module (Multi Branch Depthwise Separable Convolution, MB-DSC) performs more effective feature extraction, so that the performance of the model is greatly improved on the basis of light weight. The adaptive bottleneck structure (AB) further balances the relationship between model performance and light weight, resulting in an acceptable model scale, making it easier to deploy mobile configurations while maintaining higher performance accuracy.
Table 2 ablation experiments
The foregoing is merely illustrative of the preferred embodiments and principles of the present invention, and not in limitation thereof. Any modification, equivalent replacement, improvement, etc. which are within the spirit and principle of the present invention, should be considered as the protection scope of the present invention, based on the ideas provided by the present invention, for those skilled in the art.
Claims (7)
1. The method for predicting the mechanical properties of the hot rolled steel based on the self-adaptive multi-branch depth separable is characterized in that the input parameters comprise 5 heat treatment parameters and 5 chemical components, wherein the 5 heat treatment parameters comprise in-furnace heating temperature, start rolling temperature, finish rolling temperature, coiling temperature and compressibility, and the 5 chemical components comprise carbon, silicon, manganese, phosphorus and sulfur, and the method comprises the following steps:
s1, carrying out data enhancement on original one-dimensional data to obtain a two-dimensional matrix, and dividing a data set;
s2, constructing an optimized and improved convolutional neural network model; the optimized and improved convolutional neural network model uses depth separable convolutional operations to carry out channel-by-channel convolution and point-by-point convolution; the channel-by-channel convolution uses a single channel as a feature map, and local information of each channel is extracted by using each convolution kernel; the point-by-point convolution uses a convolution kernel of the same channel dimension as the channel-by-channel convolution, and integrates the convolution kernel into the same number of output feature maps as the convolution kernel of the point-by-point convolution; in a channel-by-channel convolution stage, constructing a multi-branch convolution neural network, wherein each branch represents convolution kernels with different sizes, and summing the multi-branch outputs; in the point-by-point convolution stage, combining the self-adaptive bottleneck layer with the self-adaptive bottleneck layer, and adopting global average pooling operation to self-adaptively adjust the bottleneck coefficient through weight size;
s3, setting super parameters, learning, training, optimizing and improving a convolutional neural network model on a training set, and judging through a loss function to obtain a trained model;
s4, predicting the test set data by using the trained model to obtain a hot rolled steel mechanical property prediction result.
2. The method for predicting mechanical properties of hot rolled steel based on the self-adaptive multi-branch depth separable according to claim 1, wherein the step S1 of data enhancement by using a glamer angle field method comprises the steps of:
s11, performing normalization operation on the original one-dimensional data, and assuming that one-dimensional data X= { X exists 1 ,x 2 ,…,x N Normalized one-dimensional data to [ -1,1 } length N]As shown in formula (1):
wherein x is i Data for the ith position, x i The value of the position after normalization is min (X) and max (X) respectively represent the minimum value and the maximum value in X;
s12, representing the normalized data by using polar coordinates, wherein the value of the one-dimensional data is taken as a polar angle theta, and the value of the position where the corresponding value is located is taken as a polar axis r, as shown in the formula (2):
where i is a position value, L is a position length, r i For its polar axis, θ i For its polar angle, x after normalization i Is within the range of [ -1,1]So theta is i The value range of (2) is [0, pi ]]And cos θ is monotonic in this interval;
s13, summing different position angles to obtain a gram and an angle field, wherein the gram and the angle field are shown in a formula (3):
3. the method for predicting mechanical properties of hot rolled steel based on adaptive multi-branch depth separation according to claim 1, wherein in step S1, the reinforced data is obtained by using the method of 9: the ratio of 1 is divided into a training set and a test set.
4. The method for predicting mechanical properties of hot rolled steel based on adaptive multi-branch depth separation according to claim 1, wherein the step S2 of constructing an optimized and improved convolutional neural network model comprises the steps of:
s21, performing two convolution processing steps of channel-by-channel convolution and point-by-point convolution by using depth separable convolution operation on the basis of CNN; let the input feature map size be D in ×D in X M, output feature map size D out ×D out The dimension of the x N corresponds to the height, width and channel number of the feature map, and the dimension of the convolution kernel is set as D k ×D k The method comprises the steps of carrying out a first treatment on the surface of the The number of convolutions per channel is M and the size is D k ×D k Processing by the x 1 convolution kernel to produce D out ×D out Feature map of x M; the point-by-point convolution uses N number and 1×1×M size convolution to check the intermediate feature map to perform channel conversion to generate D size out ×D out Feature map of x N;
s22, constructing a multi-branch convolution neural network from the aspect of channel-by-channel convolution on the basis of depth separable convolution, wherein each branch represents convolution kernels with different sizes, namely Kx K, K x 1 and 1 xK, and summing the outputs of the three branches;
s23, combining the self-adaptive bottleneck layer with the self-adaptive bottleneck layer from the aspect of point-by-point convolution, adopting global average pooling operation to adaptively adjust the size of the bottleneck coefficient through the weight size so as to realize self-adaptive bottleneck layer optimization, enabling the bottleneck coefficient to set the kernel number to be a value smaller than the number of input channels in a ratio mode, adopting a point convolution mode to perform restoration operation after the bottleneck, and recovering the number of channels;
s24, a global averaging layer is adopted to replace a full-connection layer at the tail end, and the feature map is converted into a vector with fixed length to replace weight parameters in the full-connection layer.
5. The method for predicting mechanical properties of hot-rolled steel based on the adaptive multi-branch depth separation as claimed in claim 4, wherein in step S23, it is assumed that the number of channels of the input feature map is C in The number of the output characteristic diagram channels is C out The channel number of the bottleneck layer is C b If the bottleneck coefficient is B, the point convolution calculation parameters are shown as a formula (4) and a formula (5) if the bottleneck layer is not arranged, wherein the bottleneck layer channel number is shown as a formula (6), a formula (7) is obtained according to the formulas (5) and (6), and when P is the sum of the values 2 Less than P 1 When present, formula (8) is present, which yields about formula (9):
P 1 =C in ×C out (4)
P 2 =C in ×C b +C b ×C out (5)
C b =B×C in (6)
6. the method for predicting mechanical properties of hot rolled steel based on the adaptive multi-branch depth separation according to claim 1, wherein the step S3 comprises the following steps:
s31, placing two-dimensional matrix data enhanced by GAF data into multi-branch depth separable convolution, performing enhanced convolution by using three branch paths in a channel-by-channel stage of the depth separable convolution, respectively using convolution kernels with the sizes of 3 multiplied by 3, 3 multiplied by 1 and 1 multiplied by 3, wherein the convolution kernels are increased by 64, 128, 256 and 512 along with the increase of the number of convolution layers, and performing total four-layer depth separable convolution; batch size is set to 220, reLU is used as a nonlinear activation function, model parameters are updated by adopting an Adam optimizer with a learning rate of 0.001, and iteration is carried out for 30 times;
s32, in a point-by-point convolution stage, adopting global average pooling operation to adaptively adjust the size of a bottleneck coefficient through weight magnitude so as to realize adaptive bottleneck layer optimization, and setting the number of cores to be a value smaller than the number of input channels in a ratio mode by the bottleneck coefficient;
s33, the loss function adopts a mean square error MSE shown in the formula (10), the weight bias parameter is updated through back propagation, the prediction error is minimized, and meanwhile, different evaluation indexes are used for describing the influence of the model on the prediction precision;
and S34, after training, saving weight bias to obtain a trained model structure.
7. The method for predicting mechanical properties of hot-rolled steel based on the adaptive multi-branch depth separation according to claim 1, wherein in step S33, different evaluation indexes are used, including mean absolute error MAE, root mean square error RMSE and determination coefficient R2, as shown in formulas (11) - (13), the closer the error is to 0, the closer the determination coefficient is to 1, and the better the fitting effect of the model is;
wherein y is i In order to predict the data,to pair(s)Actual data of the response->Is the average of the corresponding actual data.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310986868.2A CN117272782B (en) | 2023-08-07 | Hot rolled steel mechanical property prediction method based on self-adaptive multi-branch depth separation |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310986868.2A CN117272782B (en) | 2023-08-07 | Hot rolled steel mechanical property prediction method based on self-adaptive multi-branch depth separation |
Publications (2)
Publication Number | Publication Date |
---|---|
CN117272782A true CN117272782A (en) | 2023-12-22 |
CN117272782B CN117272782B (en) | 2024-06-21 |
Family
ID=
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117892258A (en) * | 2024-03-12 | 2024-04-16 | 沃德传动(天津)股份有限公司 | Bearing migration diagnosis method based on data fusion, electronic equipment and storage medium |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110472349A (en) * | 2019-08-20 | 2019-11-19 | 武汉科技大学 | A kind of hot-rolled steel performance prediction method based on EEMD and depth convolutional network |
CN113033106A (en) * | 2021-04-06 | 2021-06-25 | 东北大学 | Steel material performance prediction method based on EBSD and deep learning method |
US20230004781A1 (en) * | 2020-12-04 | 2023-01-05 | Northeastern University | Lstm-based hot-rolling roll-bending force predicting method |
CN116307195A (en) * | 2023-03-22 | 2023-06-23 | 苏州大学 | Strip steel finish rolling expansion prediction method and system based on deep learning |
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110472349A (en) * | 2019-08-20 | 2019-11-19 | 武汉科技大学 | A kind of hot-rolled steel performance prediction method based on EEMD and depth convolutional network |
US20230004781A1 (en) * | 2020-12-04 | 2023-01-05 | Northeastern University | Lstm-based hot-rolling roll-bending force predicting method |
CN113033106A (en) * | 2021-04-06 | 2021-06-25 | 东北大学 | Steel material performance prediction method based on EBSD and deep learning method |
CN116307195A (en) * | 2023-03-22 | 2023-06-23 | 苏州大学 | Strip steel finish rolling expansion prediction method and system based on deep learning |
Non-Patent Citations (2)
Title |
---|
ZHI-WEI XU 等: "Mechanical Properties Prediction for Hot Rolled Alloy Steel Using Convolutional Neural Network", 《IEEE》, vol. 7, 18 April 2019 (2019-04-18) * |
胡石雄;李维刚;杨威;: "基于卷积神经网络的热轧带钢力学性能预报", 武汉科技大学学报, no. 05, 29 September 2018 (2018-09-29) * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117892258A (en) * | 2024-03-12 | 2024-04-16 | 沃德传动(天津)股份有限公司 | Bearing migration diagnosis method based on data fusion, electronic equipment and storage medium |
CN117892258B (en) * | 2024-03-12 | 2024-06-07 | 沃德传动(天津)股份有限公司 | Bearing migration diagnosis method based on data fusion, electronic equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Xu et al. | Mechanical properties prediction for hot rolled alloy steel using convolutional neural network | |
Gupta et al. | Prediction of flow stress in dynamic strain aging regime of austenitic stainless steel 316 using artificial neural network | |
CN111241952A (en) | Reinforced learning reward self-learning method in discrete manufacturing scene | |
WO2021004198A1 (en) | Plate performance prediction method and apparatus | |
CN108845501B (en) | Blast furnace molten iron quality self-adaptive optimization control method based on lazy learning | |
CN101391268B (en) | Reverse optimization method of steel plate rolling and cooling controlling-process temperature institution | |
CN104375478B (en) | A kind of method and device of Rolling production process product quality on-line prediction and optimization | |
US20230004781A1 (en) | Lstm-based hot-rolling roll-bending force predicting method | |
CN102284512B (en) | Control method for manufacturing process of large deformation pipeline steel pipe | |
CN112036081B (en) | Method for determining addition amount of silicon-manganese alloy in converter tapping based on yield prediction | |
CN111367181B (en) | Hybrid drive intelligent judgment control method for sewage treatment system | |
CN112668234B (en) | Intelligent control method for converter steelmaking end point | |
CN113987949B (en) | Data-driven-based plate strip steel deformation resistance prediction method | |
CN106019940B (en) | Converter steelmaking process cost control method and system based on UKF neural network | |
CN117272782B (en) | Hot rolled steel mechanical property prediction method based on self-adaptive multi-branch depth separation | |
CN113469891A (en) | Neural network architecture searching method, training method and image completion method | |
CN117272782A (en) | Hot rolled steel mechanical property prediction method based on self-adaptive multi-branch depth separation | |
CN116543228A (en) | Infrared image vehicle detection method based on single-stage network | |
Mi et al. | Prediction of accumulated temperature in vegetation period using artificial neural network | |
CN109214513B (en) | Solid-liquid waste intelligent coupling treatment method based on adaptive deep belief network | |
CN114792051A (en) | Cold-rolling and pickling parameter optimization method for strip steel | |
CN112951344B (en) | Material uniform elongation prediction method based on machine learning | |
CN115438570A (en) | GA-BP neural network-based vehicle dynamic fuel consumption prediction model method | |
Pengtao | Based on adam optimization algorithm: Neural network model for auto steel performance prediction | |
CN110989522B (en) | Multi-steel-coil-oriented optimal setting method for technological parameters in continuous annealing production process |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
CB03 | Change of inventor or designer information | ||
CB03 | Change of inventor or designer information |
Inventor after: Zhang Qiwen Inventor after: Wang Yichao Inventor after: Tang Xingchang Inventor after: Guo Xinxin Inventor after: Guo Rongping Inventor before: Zhang Qiwen Inventor before: Guo Rongping Inventor before: Tang Xingchang Inventor before: Guo Xinxin Inventor before: Wang Yichao |
|
GR01 | Patent grant |