CN112598684A - Open-pit area ground feature segmentation method based on semantic segmentation technology - Google Patents

Open-pit area ground feature segmentation method based on semantic segmentation technology Download PDF

Info

Publication number
CN112598684A
CN112598684A CN202011577613.3A CN202011577613A CN112598684A CN 112598684 A CN112598684 A CN 112598684A CN 202011577613 A CN202011577613 A CN 202011577613A CN 112598684 A CN112598684 A CN 112598684A
Authority
CN
China
Prior art keywords
convolution
layer
mining area
training set
convolution kernel
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011577613.3A
Other languages
Chinese (zh)
Inventor
罗霄
特日根
仪锋
韩宇
刘欣悦
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chang Guang Satellite Technology Co Ltd
Original Assignee
Chang Guang Satellite Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chang Guang Satellite Technology Co Ltd filed Critical Chang Guang Satellite Technology Co Ltd
Priority to CN202011577613.3A priority Critical patent/CN112598684A/en
Publication of CN112598684A publication Critical patent/CN112598684A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a method for segmenting surface mine land features based on a semantic segmentation technology, which comprises the following steps: acquiring an image data set containing RGB three-channel colors of a mining area; making a mining area training set according to the image data set; carrying out image preprocessing on the mining area training set to obtain a preprocessed mining area training set; respectively training different deep learning models by utilizing the preprocessed mining area training set to obtain at least two trained deep learning models; performing feature segmentation on the same target mining area image by using each trained deep learning model to obtain corresponding feature segmentation results; and fusing the ground feature segmentation results to obtain a final mining area ground feature segmentation result. The open-pit mine land feature segmentation method based on the semantic segmentation technology is high in precision, high in automation degree and simple in processing flow, and has very important significance in open-pit mine land feature classification.

Description

Open-pit area ground feature segmentation method based on semantic segmentation technology
Technical Field
The invention relates to the technical field of image segmentation, in particular to an open-pit mine ground feature segmentation method based on a semantic segmentation technology.
Background
Mineral resources are used as an important material basis for the development of socioeconomic development and have been the core and key of national resource safety for a long time. In recent years, with the rapid development of mining and utilization of mineral resources, mine areas around the country emerge like spring shoots after rain. However, the occurrence of some illegal behaviors in the mining area development process restricts the national environmental safety. Therefore, how to effectively monitor the mining area environment in real time has become a primary problem in the national mineral resource report in recent years.
With the rapid development of remote sensing satellites, the space-to-ground observation data acquisition capability is continuously improved, the time, space and spectral resolution of the open-pit mine images provided by the remote sensing images is continuously improved, and more real-time and more reliable data support is provided for mine image interpretation work.
The traditional image interpretation method needs to manually set features to segment the image, needs to have stronger professional knowledge in the parameter adjusting process, has poorer generalization capability and robustness, low recognition rate and lower automation degree, and cannot be suitable for large-scale commercial application. And the characteristics of the pits in the ground objects of the mining area are similar to those of the ground objects of the dumping site, the shapes of the tailings ponds of different mining areas are different, the characteristic difference is obvious, and the difficulty is increased for the classification of the ground objects of the mining area by using the traditional method. With the rapid development of deep learning and big data, the deep learning algorithm based on big data obtains a breakthrough result in the field of computer vision, so that the difficult problem which cannot be solved by the traditional image processing algorithm is solved successively. Therefore, it is an effective solution to try to train the deep learning algorithm by using the images of the mining areas in different regions and to classify the ground features of the mining areas by using the trained deep learning algorithm model.
Disclosure of Invention
In order to solve the problems of low recognition rate and low automation degree of the traditional mining area image interpretation method in the prior art, the invention improves the prior semantic segmentation algorithm, and provides an open-pit mining area ground object segmentation method based on the semantic segmentation technology by combining the mining area remote sensing data, thereby realizing the automatic mining area ground object classification. The invention provides technical support for mining area change detection, mining area ecological environment detection and the like through a deep learning means.
In order to achieve the purpose, the invention adopts the following technical scheme:
a method for segmenting surface features in an open-pit mine area based on a semantic segmentation technology comprises the following steps:
the method comprises the following steps: acquiring an image data set containing RGB three-channel colors of a mining area;
step two: making a mining area training set according to the image data set;
step three: carrying out image preprocessing on the mining area training set to obtain a preprocessed mining area training set;
step four: respectively training different deep learning models by utilizing the preprocessed mining area training set to obtain at least two trained deep learning models;
step five: performing feature segmentation on the same target mining area image by using each trained deep learning model to obtain corresponding feature segmentation results;
step six: and fusing the ground feature segmentation results to obtain a final mining area ground feature segmentation result.
Compared with the prior art, the invention has the following beneficial effects:
1) the open-pit mine area ground object segmentation method based on the semantic segmentation technology trains the deep learning model by utilizing different ground object image blocks of different open-pit mines, and segments the open-pit mine area ground objects by utilizing the trained deep learning model, so that the problem that the traditional image processing algorithm cannot accurately segment the mine area ground objects is solved;
2) the open-pit mine ground feature segmentation method based on the semantic segmentation technology can efficiently identify different ground features with the same type in different mine areas, utilizes a Google open source deep learning system TensorFlow for development, utilizes an English Kyoda GPU for accelerating a deep learning algorithm, and can meet the real-time requirement of industrial application;
3) the open-pit area ground feature segmentation method based on the semantic segmentation technology adopts the idea of voting by voting of each pixel point in a model fusion mode, can well remove some pixel points with obvious classification errors, and greatly improves the prediction capability of the model;
4) the open-pit mine land feature segmentation method based on the semantic segmentation technology is high in precision, high in automation degree and simple in processing flow, and has very important significance in open-pit mine land feature classification.
Drawings
FIG. 1 is a flow chart of a method for segmenting surface mine land features based on semantic segmentation technology according to the present invention;
FIG. 2 is a schematic diagram of a multi-layer convolutional neural network Unet in an embodiment of the present invention;
fig. 3 is a schematic diagram of a multilayered convolutional neural network depllabv 3 in the embodiment of the present invention.
Detailed Description
The technical solution of the present invention will be described in detail with reference to the accompanying drawings and preferred embodiments.
In one embodiment, as shown in fig. 1, the invention relates to a method for segmenting surface mine features based on a semantic segmentation technology, which comprises the following steps:
step one (S100): an image data set containing RGB three-channel colors of a mine area is obtained. The images in the image dataset include different features of different strip mines and the images are in RGB three channel color.
Step two (S200): a mine area training set is made from the image dataset.
Preferably, in the second step, the step of preparing the mining area training set according to the image data set comprises the following steps: and marking ground objects on each image in the image data set, and obtaining a mining area training set after marking, wherein the ground objects comprise a mining area, a tailing pond, a dumping site, a mine pile, an industrial area and the like.
Step three (S300): and carrying out image preprocessing on the mining area training set to obtain a preprocessed mining area training set.
Preferably, in step three, in the image preprocessing process, the data size of the mining area training set is increased by using a random transformation method, and the specific steps of the random transformation method are as follows:
step three, firstly: randomly rotating the mining area training set according to 90 degrees, 180 degrees and 270 degrees;
step three: randomly turning the mining area training set left and right;
step three: carrying out random gamma transformation on the mining area training set;
step three and four: carrying out fuzzy processing on the mining area training set;
step three and five: carrying out bilateral filtering and Gaussian filtering processing on the mining area training set;
step three and six: random noise point adding processing is carried out on the mining area training set;
step three, pseudo-ginseng: and carrying out image cutting on the mining area training set.
Step four (S400): and respectively training different deep learning models by utilizing the preprocessed mining area training set to obtain at least two trained deep learning models.
In order to meet the requirements of subsequent fusion models, a training set is utilized to train a deep learning method, and various deep learning algorithms are selected. Preferably, in step four, the different deep learning models include a multilayer convolutional neural network uet and a multilayer convolutional neural network deplab v 3. When the multi-layer convolutional neural network uet and the multi-layer convolutional neural network deplab v3 are selected as deep learning models, the multi-layer convolutional neural network uet and the multi-layer convolutional neural network deplab v3 are trained respectively, it should be noted that training processes of the multi-layer convolutional neural network uet and the multi-layer convolutional neural network deplab v3 are not sequential, and in this embodiment, the description is given by taking training of the multi-layer convolutional neural network uet first and training of the multi-layer convolutional neural network deplab v3 later as examples.
Further, when the deep learning model is the multilayer convolutional neural network Unet, training the multilayer convolutional neural network Unet by using the preprocessed mining area training set comprises the following steps:
step four, firstly: constructing a multi-layer convolutional neural network Unet
The specific structure of the multi-layer convolutional neural network uet is shown in fig. 2, and the size of an input image block (i.e., an original image) is 256 × 256;
the first layer is two convolution operations, the convolution kernel size is 3 x 3, and the number of convolution kernels is 64;
the second layer is maximum pooling operation;
the third layer is two convolution operations, the convolution kernel size is 3 x 3, and the number of convolution kernels is 128;
the fourth layer is maximum pooling operation;
the fifth layer is two convolution operations, the convolution kernel size is 3 x 3, and the convolution kernel number is 256;
the sixth layer is maximum pooling operation;
the seventh layer is two convolution operations, the convolution kernel size is 3 x 3, and the number of convolution kernels is 512;
the eighth layer is a maximum pooling operation.
A ninth layer of two convolution operations, the convolution kernel size is 3 x 3, and the number of convolution kernels is 1024;
the tenth layer is up-sampling operation, and the convolution results of the seventh layer are added to generate an image matrix with the size of 1024;
the eleventh layer is convolution operation, the convolution kernel size is 3 x 3, and the number of convolution kernels is 512;
the twelfth layer is an upsampling operation, and the fifth layer convolution results are added to generate an image matrix with the size of 512;
the thirteenth layer is convolution operation, the convolution kernel size is 3 x 3, and the convolution kernel number is 256;
the fourteenth layer is an upsampling operation, and the convolution results of the third layer are added to generate an image matrix with the size of 256;
the eleventh layer is a convolution operation, the convolution kernel size is 3 x 3, and the number of convolution kernels is 128;
the sixteenth layer is an upsampling operation, and the convolution results of the third layer are added to generate an image matrix with the size of 128;
the seventeenth layer is two convolution operations, the convolution kernel size is 3 x 3, and the number of convolution kernels is 64;
the eighteenth layer is a convolution operation, convolution kernel size 1 x 1, convolution kernel number 5.
In the constructed multilayer convolutional neural network Unet, convolution operation is used for extracting high-level features of an image, the input of maximum pooling operation generally comes from the last convolution operation, the main function is to provide strong robustness, the maximum in a small area is taken, if other values in the area slightly change or the image slightly translates, the result after pooling is still unchanged, the number of parameters is reduced, the occurrence of an overfitting phenomenon is prevented, and the pooling operation generally has no parameters, so that when the parameter is propagated reversely, only derivation is needed on the input parameters, and weight updating is not needed; the picture information can be complemented by the up-sampling operation, the subsequent semantic segmentation is convenient, and meanwhile, the convolved high-resolution picture is added, which is equivalent to making a compromise between high resolution and more abstract characteristics, so that the prediction result is more accurate.
Step four and step two: multilayer convolutional neural network Unet constructed by off-line training
And setting training parameters by utilizing a preprocessed mining area training set, performing steepest descent optimization on the error gradient of the constructed multilayer convolutional neural network Unet by adopting a random gradient descent algorithm, and training the constructed multilayer convolutional neural network Unet in an off-line manner to obtain at least one trained Unet deep learning model. For example, the training parameters set in this step are the training times of the model, the training times are set to 40000 times and 45000 times respectively, the constructed multi-layer convolutional neural network Unet is trained by using the preprocessed mining area training set and the random gradient descent algorithm, and two trained Unet deep learning models can be obtained.
Further, when the deep learning model is the multilayer convolutional neural network deplab v3, training the multilayer convolutional neural network deplab v3 by using the preprocessed mining area training set includes the following steps:
step four and step three: construction of a multilayer convolutional neural network Deeplab V3
The specific structure of the multi-layer convolutional neural network deplab v3 is shown in fig. 3, and the size of an input image block (i.e., an original image) is 256 × 256;
the first layer consists of three 3 x 3 convolution operations and one maximum pooling operation, with a number of convolution kernels of 128;
the second layer consists of 3 residual units, each residual unit containing a 1 × 1 convolution operation with a convolution kernel size of 64, a 3 × 3 convolution operation with a convolution kernel size of 64, and a 1 × 1 convolution operation with a convolution kernel step size of 1 and a convolution kernel size of 256;
the third layer consists of 4 residual units, each residual unit comprises 1 × 1 convolution operation with a convolution kernel size of 128, 3 × 3 convolution operation with a convolution kernel size of 128, and 1 × 1 convolution operation with a convolution kernel step size of 2 and a convolution kernel size of 512;
the fourth layer consists of 6 residual units, each of which contains 1 × 1 convolution operation with convolution kernel size 256, 3 × 3 hole convolution operation with convolution kernel size 256 and 1 × 1 convolution operation with rate 2 and convolution kernel size 1024;
the fifth layer consists of 3 residual units, each of which contains 1 × 1 convolution operation with convolution kernel size 512, 3 × 3 hole convolution operation with convolution kernel size 512, and 1 × 1 convolution operation with rate 4 and convolution kernel size 2048;
the sixth layer comprises five branches, wherein the 1 st branch is 1 × 1 convolution operation, the 2 nd, 3 rd and 4 th branches are 3 × 3 hole convolution operations, the corresponding rates are respectively 12, 24 and 36, the fifth branch is global average pooling operation plus 1 × 1 convolution operation, the number of convolution kernels of the five branches is 256, and finally the five branches are combined and used for outputting 256-dimensional new features by using the 1 × 1 convolution operation;
the seventh layer comprises a 3 x 3 convolution operation, a 1 x 1 convolution operation with a convolution kernel size of 5, and a bilinear downsampling operation;
step four: multilayer convolutional neural network Deeplab V3 constructed by offline training
And setting training parameters by utilizing a preprocessed mining area training set, performing steepest descent optimization on the error gradient of the constructed multilayer convolutional neural network Deeplab V3 by adopting a random gradient descent algorithm, and training the constructed multilayer convolutional neural network Deeplab V3 in an off-line manner to obtain at least one trained deep learning model of Deeplab V3. For example, the training parameters set in this step are the training times of the model, the training times are set to 40000 times, 45000 times and 50000 times respectively, and the constructed multilayer convolutional neural network deep learning model deplab v3 is trained by using the preprocessed mining area training set and the stochastic gradient descent algorithm, so that three trained deep learning models of the deplab v3 can be obtained.
Step five (S500): and performing feature segmentation on the same target mining area image by using each trained deep learning model to obtain a corresponding feature segmentation result.
And after at least two trained deep learning models are obtained in the fourth step, in the fifth step, performing feature segmentation on the same target mining area image by using each trained deep learning model, so as to obtain a corresponding feature segmentation result.
For example, for selecting a multi-layer convolutional neural network Unet and a multi-layer convolutional neural network Deeplab V3 as deep learning models, and obtaining two trained Unet deep learning models and three trained Deeplab V3 deep learning models by respectively training through adjusting training parameters, the step uses 5 models, namely the two trained Unet deep learning models and the three trained Deeplab V3 deep learning models, to respectively perform feature segmentation on the same target mine image, so as to obtain a feature segmentation result corresponding to each model.
Step six (S600): and fusing the ground feature segmentation results to obtain the final mining area ground feature segmentation result.
And in the step five, the ground feature segmentation results corresponding to all the trained deep learning models obtained in the step five are fused to obtain the final mining area ground feature segmentation result.
Preferably, in the sixth step, a fusion method of voting a plurality of prediction results is adopted when the ground feature segmentation results are fused, and the fusion method includes the following steps:
step six: classifying and predicting pixel points in the target mining area image by using each trained deep learning model, and selecting a prediction result with the most prediction categories as a classification prediction result of the pixel points;
step six and two: traversing all pixel points of the target mine area image according to the method of the sixth step to obtain a classification prediction result of all pixel points in the target mine area image;
step six and three: and drawing an image matrix according to the classification prediction results of all the pixel points to obtain a final mining area ground object segmentation result.
The model prediction of the final image semantic segmentation is to perform prediction classification on each pixel point of an image, each category corresponds to a pixel value, then the pixel value corresponding to the category of each pixel point is drawn into an image matrix, and finally a prediction result is obtained. When ground feature segmentation results are fused, each trained deep learning model is used for carrying out classification prediction on pixel points in the target mine area image, and the prediction result with the largest prediction category is selected as the classification prediction result of the pixel points, wherein the pixel points are any point in the target mine area image; after traversing all pixel points in the target mine area image according to the method of the sixth step, obtaining classification prediction results of all pixel points in the target mine area image; and finally, drawing an image matrix according to the classification prediction results of all the pixel points, and obtaining the final mining area ground object segmentation result. By adopting the idea of voting by voting of each pixel point in a model fusion mode, some pixel points with obvious classification errors can be well removed, and the prediction capability of the model is improved to a great extent.
Compared with the prior art, the invention has the following beneficial effects:
1) the open-pit mine area ground object segmentation method based on the semantic segmentation technology trains the deep learning model by utilizing different ground object image blocks of different open-pit mines, and segments the open-pit mine area ground objects by utilizing the trained deep learning model, so that the problem that the traditional image processing algorithm cannot accurately segment the mine area ground objects is solved;
2) the open-pit mine ground feature segmentation method based on the semantic segmentation technology can efficiently identify different ground features with the same type in different mine areas, utilizes a Google open source deep learning system TensorFlow for development, utilizes an English Kyoda GPU for accelerating a deep learning algorithm, and can meet the real-time requirement of industrial application;
3) the open-pit area ground feature segmentation method based on the semantic segmentation technology adopts the idea of voting by voting of each pixel point in a model fusion mode, can well remove some pixel points with obvious classification errors, and greatly improves the prediction capability of the model;
4) the open-pit mine land feature segmentation method based on the semantic segmentation technology is high in precision, high in automation degree and simple in processing flow, and has very important significance in open-pit mine land feature classification.
The technical features of the embodiments described above may be arbitrarily combined, and for the sake of brevity, all possible combinations of the technical features in the embodiments described above are not described, but should be considered as being within the scope of the present specification as long as there is no contradiction between the combinations of the technical features.
The above-mentioned embodiments only express several embodiments of the present invention, and the description thereof is more specific and detailed, but not construed as limiting the scope of the invention. It should be noted that, for a person skilled in the art, several variations and modifications can be made without departing from the inventive concept, which falls within the scope of the present invention. Therefore, the protection scope of the present patent shall be subject to the appended claims.

Claims (10)

1. A method for segmenting surface features in an open-pit mine area based on a semantic segmentation technology is characterized by comprising the following steps:
the method comprises the following steps: acquiring an image data set containing RGB three-channel colors of a mining area;
step two: making a mining area training set according to the image data set;
step three: carrying out image preprocessing on the mining area training set to obtain a preprocessed mining area training set;
step four: respectively training different deep learning models by utilizing the preprocessed mining area training set to obtain at least two trained deep learning models;
step five: performing feature segmentation on the same target mining area image by using each trained deep learning model to obtain corresponding feature segmentation results;
step six: and fusing the ground feature segmentation results to obtain a final mining area ground feature segmentation result.
2. The method for surface property segmentation based on semantic segmentation technology according to claim 1, wherein the process of image preprocessing the training set of the mine area comprises using a stochastic transformation method to increase the data size of the training set of the mine area, and the stochastic transformation method comprises the following steps:
step three, firstly: randomly rotating the mining area training set according to 90 degrees, 180 degrees and 270 degrees;
step three: randomly turning the mining area training set left and right;
step three: carrying out random gamma transformation on the mining area training set;
step three and four: carrying out fuzzy processing on the mining area training set;
step three and five: carrying out bilateral filtering and Gaussian filtering processing on the mining area training set;
step three and six: carrying out random noise point adding processing on the mining area training set;
step three, pseudo-ginseng: and carrying out image cutting on the mining area training set.
3. The method for segmenting the land features of the strip mine area based on the semantic segmentation technology according to claim 1 or 2, wherein a fusion method for voting on a plurality of prediction results is adopted when the land feature segmentation results are fused, and the fusion method comprises the following steps:
step six: classifying and predicting pixel points in the target mining area image by using each trained deep learning model, and selecting a prediction result with the most prediction categories as a classification prediction result of the pixel points;
step six and two: traversing all pixel points of the target mine area image according to the method of the sixth step to obtain a classification prediction result of all pixel points in the target mine area image;
step six and three: and drawing an image matrix according to the classification prediction results of all the pixel points to obtain the final mining area ground object segmentation result.
4. The method for segmenting the land features of the strip mine area based on the semantic segmentation technology as claimed in claim 3,
in step four, the different deep learning models include the multilayer convolutional neural network uet and the multilayer convolutional neural network depllabv 3.
5. The method for segmenting the surface features based on the semantic segmentation technology as claimed in claim 4, wherein when the deep learning model is a multi-layer convolutional neural network Unet, training the multi-layer convolutional neural network Unet by using the preprocessed mining area training set comprises the following steps:
step four, firstly: constructing a multi-layer convolutional neural network Unet
The first layer is two convolution operations, the convolution kernel size is 3 x 3, and the number of convolution kernels is 64;
the second layer is maximum pooling operation;
the third layer is two convolution operations, the convolution kernel size is 3 x 3, and the number of convolution kernels is 128;
the fourth layer is maximum pooling operation;
the fifth layer is two convolution operations, the convolution kernel size is 3 x 3, and the convolution kernel number is 256;
the sixth layer is maximum pooling operation;
the seventh layer is two convolution operations, the convolution kernel size is 3 x 3, and the number of convolution kernels is 512;
the eighth layer is a maximum pooling operation.
A ninth layer of two convolution operations, the convolution kernel size is 3 x 3, and the number of convolution kernels is 1024;
the tenth layer is up-sampling operation, and the convolution results of the seventh layer are added to generate an image matrix with the size of 1024;
the eleventh layer is convolution operation, the convolution kernel size is 3 x 3, and the number of convolution kernels is 512;
the twelfth layer is an upsampling operation, and the fifth layer convolution results are added to generate an image matrix with the size of 512;
the thirteenth layer is convolution operation, the convolution kernel size is 3 x 3, and the convolution kernel number is 256;
the fourteenth layer is an upsampling operation, and the convolution results of the third layer are added to generate an image matrix with the size of 256;
the eleventh layer is a convolution operation, the convolution kernel size is 3 x 3, and the number of convolution kernels is 128;
the sixteenth layer is an upsampling operation, and the convolution results of the third layer are added to generate an image matrix with the size of 128;
the seventeenth layer is two convolution operations, the convolution kernel size is 3 x 3, and the number of convolution kernels is 64;
the eighteenth layer is convolution operation, the convolution kernel size is 1 x 1, and the convolution kernel number is 5;
step four and step two: multilayer convolutional neural network Unet constructed by off-line training
And setting training parameters by utilizing a preprocessed mining area training set, performing steepest descent optimization on the error gradient of the constructed multilayer convolutional neural network Unet by adopting a random gradient descent algorithm, and training the constructed multilayer convolutional neural network Unet in an off-line manner to obtain at least one trained Unet deep learning model.
6. The method for segmenting the surface features of the strip mine based on the semantic segmentation technology, as claimed in claim 5, wherein when the deep learning model is a multi-layer convolutional neural network Deeplab V3, training the multi-layer convolutional neural network Deeplab V3 by using the preprocessed mine area training set comprises the following steps:
step four and step three: construction of a multilayer convolutional neural network Deeplab V3
The first layer consists of three 3 x 3 convolution operations and one maximum pooling operation, with a number of convolution kernels of 128;
the second layer consists of 3 residual units, each residual unit containing a 1 × 1 convolution operation with a convolution kernel size of 64, a 3 × 3 convolution operation with a convolution kernel size of 64, and a 1 × 1 convolution operation with a convolution kernel step size of 1 and a convolution kernel size of 256;
the third layer consists of 4 residual units, each residual unit comprises 1 × 1 convolution operation with a convolution kernel size of 128, 3 × 3 convolution operation with a convolution kernel size of 128, and 1 × 1 convolution operation with a convolution kernel step size of 2 and a convolution kernel size of 512;
the fourth layer consists of 6 residual units, each of which contains 1 × 1 convolution operation with convolution kernel size 256, 3 × 3 hole convolution operation with convolution kernel size 256 and 1 × 1 convolution operation with rate 2 and convolution kernel size 1024;
the fifth layer consists of 3 residual units, each of which contains 1 × 1 convolution operation with convolution kernel size 512, 3 × 3 hole convolution operation with convolution kernel size 512, and 1 × 1 convolution operation with rate 4 and convolution kernel size 2048;
the sixth layer comprises five branches, wherein the 1 st branch is 1 × 1 convolution operation, the 2 nd, 3 rd and 4 th branches are 3 × 3 hole convolution operations, the corresponding rates are respectively 12, 24 and 36, the fifth branch is global average pooling operation plus 1 × 1 convolution operation, the number of convolution kernels of the five branches is 256, and finally the five branches are combined and used for outputting 256-dimensional new features by using the 1 × 1 convolution operation;
the seventh layer comprises a 3 x 3 convolution operation, a 1 x 1 convolution operation with a convolution kernel size of 5, and a bilinear downsampling operation;
step four: multilayer convolutional neural network Deeplab V3 constructed by offline training
And setting training parameters by utilizing a preprocessed mining area training set, performing steepest descent optimization on the error gradient of the constructed multilayer convolutional neural network Deeplab V3 by adopting a random gradient descent algorithm, and training the constructed multilayer convolutional neural network Deeplab V3 in an off-line manner to obtain at least one trained deep learning model of Deeplab V3.
7. The method for segmenting the land features in the strip mine area based on the semantic segmentation technology as claimed in claim 6,
and adjusting the training parameters in the second step and the fourth step to respectively obtain two trained Unet deep learning models and three trained Deeplab V3 deep learning models.
8. The method for segmenting surface mine land features based on semantic segmentation technology according to claim 7,
the training parameter is the training times of the model.
9. The method for segmenting surface mine land features based on semantic segmentation technology according to claim 8,
the training times of the two trained Unet deep learning models are 40000 times and 45000 times respectively, and the training times of the three trained Deeplab V3 deep learning models are 40000 times, 45000 times and 50000 times respectively.
10. The method for segmenting surface features based on semantic segmentation technology according to claim 1 or 2, characterized in that the step of preparing a mining area training set according to the image data set comprises the following steps:
and marking ground objects on each image in the image data set, and obtaining the mining area training set after marking, wherein the ground objects comprise a mining area, a tailing pond, a refuse dump, a mine pile and an industrial area.
CN202011577613.3A 2020-12-28 2020-12-28 Open-pit area ground feature segmentation method based on semantic segmentation technology Pending CN112598684A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011577613.3A CN112598684A (en) 2020-12-28 2020-12-28 Open-pit area ground feature segmentation method based on semantic segmentation technology

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011577613.3A CN112598684A (en) 2020-12-28 2020-12-28 Open-pit area ground feature segmentation method based on semantic segmentation technology

Publications (1)

Publication Number Publication Date
CN112598684A true CN112598684A (en) 2021-04-02

Family

ID=75202811

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011577613.3A Pending CN112598684A (en) 2020-12-28 2020-12-28 Open-pit area ground feature segmentation method based on semantic segmentation technology

Country Status (1)

Country Link
CN (1) CN112598684A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113887459A (en) * 2021-10-12 2022-01-04 中国矿业大学(北京) Open-pit mining area stope change area detection method based on improved Unet +
CN113902765A (en) * 2021-12-10 2022-01-07 聚时科技(江苏)有限公司 Automatic semiconductor partitioning method based on panoramic segmentation
CN115862014A (en) * 2023-02-27 2023-03-28 中国地质大学(武汉) Mining area semantic segmentation model construction method and device and mining area semantic segmentation method

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113887459A (en) * 2021-10-12 2022-01-04 中国矿业大学(北京) Open-pit mining area stope change area detection method based on improved Unet +
CN113902765A (en) * 2021-12-10 2022-01-07 聚时科技(江苏)有限公司 Automatic semiconductor partitioning method based on panoramic segmentation
CN113902765B (en) * 2021-12-10 2022-04-12 聚时科技(江苏)有限公司 Automatic semiconductor partitioning method based on panoramic segmentation
CN115862014A (en) * 2023-02-27 2023-03-28 中国地质大学(武汉) Mining area semantic segmentation model construction method and device and mining area semantic segmentation method
CN115862014B (en) * 2023-02-27 2023-04-28 中国地质大学(武汉) Mining area semantic segmentation model construction method and device and mining area semantic segmentation method

Similar Documents

Publication Publication Date Title
Sony et al. A systematic review of convolutional neural network-based structural condition assessment techniques
CN109800736B (en) Road extraction method based on remote sensing image and deep learning
CN113688808B (en) Landslide mass identification method based on Laplacian pyramid remote sensing image fusion
CN111986099B (en) Tillage monitoring method and system based on convolutional neural network with residual error correction fused
CN112598684A (en) Open-pit area ground feature segmentation method based on semantic segmentation technology
CN109934153B (en) Building extraction method based on gating depth residual error optimization network
Abdollahi et al. Improving road semantic segmentation using generative adversarial network
Giang et al. U-Net convolutional networks for mining land cover classification based on high-resolution UAV imagery
Pi et al. Detection and semantic segmentation of disaster damage in UAV footage
CN110276402B (en) Salt body identification method based on deep learning semantic boundary enhancement
WO2020046213A1 (en) A method and apparatus for training a neural network to identify cracks
CN107273938B (en) Multi-source remote sensing image ground object classification method based on two-channel convolution ladder network
CN107564009B (en) Outdoor scene multi-target segmentation method based on deep convolutional neural network
CN112488025B (en) Double-temporal remote sensing image semantic change detection method based on multi-modal feature fusion
He et al. Mask R-CNN based automated identification and extraction of oil well sites
CN113887459A (en) Open-pit mining area stope change area detection method based on improved Unet +
CN111553869B (en) Method for complementing generated confrontation network image under space-based view angle
CN106910202B (en) Image segmentation method and system for ground object of remote sensing image
CN114943893B (en) Feature enhancement method for land coverage classification
Azmoon et al. Image-data-driven slope stability analysis for preventing landslides using deep learning
CN113887472A (en) Remote sensing image cloud detection method based on cascade color and texture feature attention
CN112950780A (en) Intelligent network map generation method and system based on remote sensing image
Guo et al. A novel transformer-based network with attention mechanism for automatic pavement crack detection
CN113673369A (en) Remote sensing image scene planning method and device, electronic equipment and storage medium
CN114022777A (en) Sample manufacturing method and device for ground feature elements of remote sensing images

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination