CN110443822B - Semantic edge-assisted high-resolution remote sensing target fine extraction method - Google Patents

Semantic edge-assisted high-resolution remote sensing target fine extraction method Download PDF

Info

Publication number
CN110443822B
CN110443822B CN201910638370.0A CN201910638370A CN110443822B CN 110443822 B CN110443822 B CN 110443822B CN 201910638370 A CN201910638370 A CN 201910638370A CN 110443822 B CN110443822 B CN 110443822B
Authority
CN
China
Prior art keywords
target
boundary
edge
image
detection model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910638370.0A
Other languages
Chinese (zh)
Other versions
CN110443822A (en
Inventor
夏列钢
张雄波
吴炜
杨海平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University of Technology ZJUT
Original Assignee
Zhejiang University of Technology ZJUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University of Technology ZJUT filed Critical Zhejiang University of Technology ZJUT
Priority to CN201910638370.0A priority Critical patent/CN110443822B/en
Publication of CN110443822A publication Critical patent/CN110443822A/en
Application granted granted Critical
Publication of CN110443822B publication Critical patent/CN110443822B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/13Satellite images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Astronomy & Astrophysics (AREA)
  • Remote Sensing (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

A high-resolution remote sensing target fine extraction method based on edge assistance comprises the steps of designing a deep convolutional neural network according to a remote sensing target extraction task, manufacturing a sample, and training to obtain a target detection model and an edge detection model. Then inputting the remote sensing image into a target detection model for extracting a surface feature outsourcing frame, thereby determining the type and the position range of a surface feature target; and inputting the remote sensing image into an edge detection model for ground object edge extraction, thereby obtaining a ground object target edge intensity image. And finally, aiming at the target in each outer frame, performing target boundary extraction operation by taking the edge strength graph of the corresponding position as an indication, directly refining the obvious edge with higher strength as a boundary, repairing the fuzzy, particularly interrupted edge by taking the complete boundary as the target, finally determining a target fine boundary, and vectorizing the target boundary into a polygon element according to task requirements. The invention can realize the fine extraction of the high-resolution remote sensing target.

Description

Semantic edge-assisted high-resolution remote sensing target fine extraction method
Technical Field
The invention belongs to the field of remote sensing and the field of target detection and edge extraction in computer vision, and provides a two-stage remote sensing target fine extraction method combining target detection and edge detection based on deep learning.
Technical Field
The target extraction is an important method for extracting information from remote sensing data, the observable target of the image and the extraction precision requirement of the observable target of the image are correspondingly and continuously improved along with the continuous improvement of the spatial resolution of the remote sensing image, although the example segmentation in general computer vision can be adopted for partial application, the requirements on the actual ground object boundary are far away, and the method is difficult to be directly applied to the fine target extraction of the high-resolution remote sensing image like the currently commonly adopted target detection-Mask segmentation two-stage method (such as Mask-RCNN and the like). In fact, target detection based on the deep learning method is generally high in precision because only positioning of a target and determination of an outer frame are required; the main actual influence on the target extraction effect lies in the fine determination of the actual boundary of the ground object in the mask segmentation stage, and due to the complex changes of the ground object target in the aspects of image spectral representation, spatial shape dimension and the like, the boundary identification of local data in a detection frame is limited by the current depth segmentation model, which is still difficult to obtain better precision.
The current mainstream target detection algorithm proposed by deep learning can be roughly divided into two types, namely a target detection convolutional neural network based on region suggestion and an end-to-end integrated target detection convolutional neural network. In the end-to-end target detection method, in order to pursue faster detection speed, the recognition result of single detection is adopted as the final target detection result, and the effect is not ideal compared with the multi-stage training target detection method based on the region suggestion. In the field of target detection based on region suggestion, R-CNN is an initiative proposal, changes the original complicated exhaustive sliding window form, extracts a small number of suitable region candidate frames through selective search, then normalizes the size of the candidate regions, extracts features by using a deep convolutional neural network, and finally uses an SVM for classification and recognition to obtain a final target detection result. Detection algorithms such as Fast-RCNN, FPN, Mask-RCNN and the like are all purposefully improved on the basis. As the target detection algorithm only requires positioning and classifying the target, the method has good detection precision.
The edge detection algorithm requires accurate determination of the target boundary, and has greater technical difficulty than simple target positioning. In the edge detection algorithm based on deep learning, RCF is a relatively effective convolutional neural network model. The method mainly includes the steps of utilizing a VGG model for fine adjustment, reducing dimensions of a plurality of convolution layers in 5 stages in the VGG to obtain a feature fusion graph, then respectively deconvoluting and independently calculating a loss value of each stage, and finally synthesizing the feature fusion graphs in the 5 stages in a multi-scale mode to obtain a final feature fusion graph and calculate the loss value. Utilizing the characteristics of all of the convolutional layers can provide an increase in efficiency over utilizing only the last convolutional layer characteristics of each stage in the HED. Under the condition of a strong boundary, the edge contains relatively rich semantic information, and then global information is obtained through convolution, so that a good edge detection effect is achieved. However, in the case of a weak boundary, it is difficult for the target edge to provide effective semantic information, and even due to the shadow occlusion of a tall object, the distinction between the target and the non-target edge is completely lost, so that the neural network cannot identify the target edge, and thus the edge extraction disconnection occurs. In practical requirements, the edge of the broken line cannot be regarded as complete boundary extraction, so that the boundary repairing work aiming at the integrity of the boundary can be regarded as very important content in the task of extracting the target boundary.
The method tries to integrate the advantages of target detection and edge extraction, finely positions the target boundary by using the edge detection result under the guidance of a target detection outer frame with higher precision, and actively speculates and completes the blocked or interfered edge based on the target detection result, thereby realizing the fine extraction of the remote sensing target.
Disclosure of Invention
The invention provides a high-resolution remote sensing target fine extraction method based on edge assistance, aiming at overcoming the defects of imprecise and incomplete extraction of the edges of ground objects caused by fuzzy and mutual shielding of the boundaries of the ground objects in high-resolution remote sensing images.
The invention can overcome the defect of inaccurate mask boundary in general case segmentation by edge assistance, simultaneously ensure the integrity of the target boundary by two-stage process under the assistance of the outer covering frame, overcome the defect of easy disconnection in general edge extraction, and finally realize the fine extraction of the high-resolution remote sensing target.
In order to achieve the purpose, the technical scheme provided by the invention is as follows:
a high-resolution remote sensing target fine extraction method based on edge assistance comprises the following steps:
step 1: making a remote sensing image target extraction sample, drawing a fine boundary of a target by contrasting an image and determining a type; generating two different labels for the same image sample, namely generating a target outer covering frame label according to the requirement of a target detection model, and generating a target boundary label according to the requirement of an edge detection model; the method specifically comprises the following steps:
step 1.1: obtaining a high-resolution remote sensing image: optical satellite remote sensing data with a visible light-near infrared sensor or aviation remote sensing data carrying a common optical camera are adopted, and multispectral images or fusion panchromatic images can be directly used according to resolution requirements;
step 1.2: cutting a remote sensing image: selecting a range of a typical target in a production area, and cutting the remote sensing image according to uniform pixels;
step 1.3: manufacturing a deep learning training sample: drawing a deep learning training sample by using ArcGIS or other GIS software, marking the boundary of each ground feature target to obtain a corresponding shp file, and generating a target outer frame mark according to the requirements of a target detection model; generating a target boundary label according to the requirement of the edge detection model;
step 1.4: collecting more than 200 images and corresponding labels as training samples according to task requirements, and independently preparing test samples for detection precision;
step 2: training a target detection model and an edge detection model by using the prepared image sample; training a Faster R-CNN neural network to obtain a target detection model, training an RCF neural network to obtain an edge detection model, and correspondingly replacing and modifying the network according to a production target; the method specifically comprises the following steps:
step 2.1: designing a deep convolutional neural network: in order to train a target detection model and an edge detection model, two neural networks of RCF and Faster R-CNN are selected, and the networks can be replaced and modified correspondingly according to a production target;
step 2.2: initializing the weight: initializing RCF network weights using a VGG pre-training model, initializing weights of Faster R-CNN using a pre-training model on a COCO data set;
step 2.3: setting a training hyper-parameter: configuring hyper-parameters, and specifically setting numerical values after model tuning;
training parameter setting of RCF: 8000 for iteration number, 4 for batch _ size, step for learning rate update strategy, 3200,4800,6400,8000 for learning rate update step, 0.001 for initial learning rate, 0.1 for learning rate update coefficient;
training parameter settings for Faster R-CNN: training phase number is 3, each phase iteration round number is [40,120, 40], each iteration round number is 1000, verification interval is 50, batch _ size is 4, learning rate updating strategy is step, learning rate updating step size is 1000, initial learning rate is 0.001, learning rate updating momentum is 0.9, and weight decay is 0.0001;
step 2.4: inputting a sample, training a model: inputting the training samples into the RCF model, and training according to the hyper-parameters in the step 2.3 to obtain an edge detection model capable of extracting the edge contour of the ground object target; inputting the training sample into the Faster R-CNN, training according to the hyperparameter in the step 2.3, and obtaining a target detection model capable of extracting a surface feature target outer wrapping frame;
and step 3: inputting an image sample for testing into a target detection model and an edge detection model, and acquiring a surface feature target outer frame and an edge intensity map; the object target edge intensity graph can show the possibility that the corresponding position in the remote sensing image is the target edge; the ground object target outer covering frame is a rectangular limiting frame for marking the position range of the target and the type of the target; the method specifically comprises the following steps:
step 3.1: inputting the high-resolution remote sensing image into a target detection model, and obtaining a rectangular outer frame of the ground object target; inputting the high-resolution remote sensing image into an edge detection model to obtain a surface feature target edge intensity map;
step 3.2: parameterizing an outer package frame: converting the rectangular outer frame parameters output by the target detection model into a left lower vertex coordinate and a right upper vertex coordinate of the rectangular outer frame parameters; specifically, the method comprises the following steps:
x1=x-w y1=y-h
x2=x+w y2=y+h
wherein x, y, w and h respectively represent the central abscissa, the central ordinate, the width and the height of the rectangular outer wrapping frame;
and 4, step 4: the edge intensity graph in the step 3 is refined into a single-pixel-wide boundary; the method specifically comprises the following steps:
step 4.1: and converting the edge intensity map of the ground object target from a gray scale map into a binary map according to a set threshold value. The edge judgment condition of the ground object at the image (x, y) is represented by binary _ image (x, y) (1 represents edge, 0 represents not edge), and can be represented as:
Figure GDA0002726304250000041
where threshold is a real number in the interval [0,1], which can be set by the user, and the initial default value is 0.5. And x and y are horizontal and vertical coordinates of the image.
Step 4.2: for the thick edge lines with a plurality of pixel widths in the binary image, the center of each edge line is continuously etched until all the edge lines are only one pixel width, so that the aim of skeleton extraction is fulfilled.
And 5: and (4) repairing the boundary with the single pixel width in the step (4) by taking the foreign object target outer covering frame in the step (3) as a constraint so as to obtain a complete and fine polygonal foreign object target boundary.
The method specifically comprises the following steps:
step 5.1: the method comprises the steps of finding out an unclosed part in a ground object target boundary line, and dividing the unclosed part into three types, namely a boundary broken line at the edge of an image (the boundary at the edge of the image is incomplete due to a skeleton extraction algorithm), a boundary broken line inside the image (the part which is not correctly identified by a boundary detection model) and a boundary line head inside the image (the target boundary has an excessive boundary line due to the skeleton extraction algorithm).
Step 5.2: three types of boundary breaks are handled in three different ways.
Step 5.2.1: and regarding the boundary broken line at the edge of the image, taking the edge strength image at the corresponding position as an indication, selecting pixel points with higher edge strength values to fill the target boundary broken line, and if a gap still exists between the pixel points and the image boundary, adopting a straight line perpendicular to the image boundary to connect the pixel points and the target boundary broken line and form a closed target boundary line by connecting the pixel points and the image boundary.
Step 5.2.2: and resetting the threshold for the boundary broken line in the image, mapping pixel points with the intensity values higher than the new threshold in the edge intensity graph into a binary graph by taking the edge intensity graph at the corresponding position as an indication, and connecting the two broken points by the original geometric characteristics if the target boundary broken line is not closed. The steps of repairing the break point with the original geometric characteristics are as follows.
Step 5.2.2.1: and taking out the innermost boundary line connected with the two break points.
Step 5.2.2.2: : the most inner circle boundary line is divided into a plurality of parts according to the change condition of the slope, and the approximate geometrical shape of the boundary line is determined.
Step 5.2.2.3: and (4) repairing the broken points into a closed graph according to the geometric shape, so that the repaired positions keep the original geometric characteristics.
Step 5.2.3: the isolated boundary lines within the image, i.e., the isolated boundary lines that are far from other broken lines, are deleted.
Step 5.3: and 5, performing skeleton extraction again, and refining the part filled in the step 5.2 into a target boundary line with the width of a single pixel.
Step 5.4: and traversing the image again, and deleting all unclosed target boundary lines to obtain a complete and fine polygonal ground object target boundary.
Due to the adoption of the technical scheme, the invention has the following advantages and beneficial effects:
1. the invention adopts a method of combining the target detection model and the edge detection model, thereby not only avoiding the defect of inaccurate mask boundary in the common example segmentation algorithm, but also solving the problems of easy line breakage and thick outline in the common edge extraction algorithm, and ensuring the integrity and the refinement of the target boundary.
2. Compared with the traditional method for manually identifying the remote sensing target, the method has higher identification speed, and for a remote sensing image with hundreds of targets, the method can identify the remote sensing image in less than one second, and the manual drawing usually needs several hours; the invention has higher accuracy, and avoids the condition of accuracy reduction caused by manual drawing misoperation due to the adoption of an end-to-end design method.
3. The invention adopts a deep neural network algorithm in machine learning, can identify various targets and extract target boundaries only by changing sample training, does not need to redesign the algorithm, and has strong reusability and robustness.
Drawings
FIG. 1 is a schematic diagram of the edge-assisted fine extraction method of high-resolution remote sensing targets of the invention;
FIG. 2 is a sample graph of a target detection model in an embodiment of the invention;
FIG. 3 is a sample graph of an edge detection model in an embodiment of the invention;
FIG. 4 is a diagram illustrating an example of a prediction result of an object detection model in an embodiment of the invention;
FIG. 5 is a diagram illustrating an example of the prediction results of the edge detection model in an embodiment of the invention;
FIG. 6 is a graph illustrating a comparison of final boundary results in an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the detailed description and specific examples, while indicating the preferred embodiment of the invention, are intended for purposes of illustration only and are not intended to limit the scope of the invention. The components of embodiments of the present invention generally described in the figures herein may be arranged and designed in a wide variety of different configurations.
Thus, the following detailed description of the embodiments of the present invention, presented in the figures, is not intended to limit the scope of the invention, as claimed, but is merely representative of selected embodiments of the invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present invention without making any creative effort, shall fall within the protection scope of the present invention.
FIG. 1 is a schematic diagram of a method for finely extracting a high-resolution remote sensing target with edge assistance.
Referring to fig. 1, a preferred embodiment of the present invention is provided, which includes the following steps:
step 1: making a remote sensing image target extraction sample, drawing a fine boundary of a target by contrasting an image and determining a type;
step 2: training a target detection model and an edge detection model by using the prepared image sample;
and step 3: inputting an image sample for testing into a target detection model and an edge detection model, and acquiring a surface feature target outer frame and an edge intensity map;
and 4, step 4: the edge intensity graph in the step 3 is refined into a single-pixel-wide boundary;
and 5: and (4) repairing the boundary with the single pixel width in the step (4) by taking the foreign object target outer covering frame in the step (3) as a constraint so as to obtain a complete and fine polygonal foreign object target boundary.
According to the above example, step 1 is detailed as follows:
step 1.1: obtaining a high-resolution remote sensing image: the multispectral image or the fusion panchromatic image can be directly used according to the resolution requirement by adopting optical satellite remote sensing data with a visible light-near infrared sensor or aviation remote sensing data carrying a common optical camera.
Step 1.2: cutting a remote sensing image: and uniformly cutting the remote sensing image into 512 by 512 pixels in the range of the typical target in the production area (the cutting range according to the size of the production area has certain randomness and is widely covered).
Step 1.3: manufacturing a deep learning training sample: using ArcGIS or other GIS software to draw a deep learning training sample, in the embodiment, the boundary of each ground feature target needs to be marked to obtain a corresponding shp file, and generating a target outer enclosure frame mark according to the requirements of a target detection model, as shown in FIG. 2; the target boundary label is generated according to the requirements of the edge detection model, as shown in fig. 3.
Step 1.4: more than 200 images and corresponding labels are generally collected as training samples according to task requirements, and test samples can be independently prepared if detection precision is required.
Step 2 detailed steps are as follows:
according to the above example, step 2.1: designing a deep convolutional neural network: in order to train the target detection model and the edge detection model, two neural networks, namely RCF and Faster R-CNN, are selected in the invention, and can be replaced and modified correspondingly according to the production target network.
Step 2.2: initializing the weight: the VGG pre-training model is used to initialize RCF network weights, and the pre-training model on the COCO data set is used to initialize the weights of Faster R-CNN.
Step 2.3: setting a training hyper-parameter: the hyper-parameters are configured, and the specific values are set as follows after the model is tuned and optimized.
Training parameter setting of RCF: the number of iterations is 8000, batch _ size is 4, the learning rate update strategy is step, the learning rate update step is 3200,4800,6400,8000, the initial learning rate is 0.001, and the learning rate update coefficient is 0.1.
Training parameter settings for Faster R-CNN: the training phase number is 3, the number of iteration rounds of each phase is [40,120, 40], the number of iterations of each round is 1000, the verification interval is 50, the batch _ size is 4, the learning rate updating strategy is step, the learning rate updating step size is 1000, the initial learning rate is 0.001, the learning rate updating momentum is 0.9, and the weight decay is 0.0001.
Step 2.4: inputting a sample, training a model: inputting the training samples into the RCF model, and training according to the hyper-parameters in the step 2.3 to obtain an edge detection model capable of extracting the edge contour of the ground object target; and (4) inputting the training sample into the Faster R-CNN, and training according to the hyperparameter in the step 2.3 to obtain a target detection model capable of extracting the surface feature target outer covering frame.
According to the above embodiment, step 3 is detailed as follows:
step 3.1: inputting the high-resolution remote sensing image into a target detection model, and obtaining a rectangular outer frame of the ground object target, as shown in FIG. 4; and inputting the high-resolution remote sensing image into an edge detection model to obtain a surface feature target edge intensity map, as shown in fig. 5.
Step 3.2: parameterizing an outer package frame: and converting the rectangular outer enclosure frame parameters output by the target detection model into the left lower vertex coordinate and the right upper vertex coordinate of the rectangular outer enclosure frame parameters. Specifically, the method comprises the following steps:
x1=x-w y1=y-h
x2=x+w y2=y+h
wherein x, y, w and h respectively represent the central abscissa, the central ordinate, the width and the height of the rectangular outer wrapping frame.
According to the above embodiment, step 4 is detailed as follows:
step 4.1: and converting the edge intensity map of the ground object target from a gray scale map into a binary map according to a set threshold value. The edge judgment condition of the ground object at the image (x, y) is represented by binary _ image (x, y) (1 represents edge, 0 represents not edge), and can be represented as:
Figure GDA0002726304250000081
where threshold is a real number in the interval [0,1], which can be set by the user, and the initial default value is 0.5. And x and y are horizontal and vertical coordinates of the image.
Step 4.2: for the thick edge lines with a plurality of pixel widths in the binary image, the center of each edge line is continuously etched until all the edge lines are only one pixel width, so that the aim of skeleton extraction is fulfilled.
According to the above embodiment, step 5 is detailed as follows:
step 5.1: the method comprises the steps of finding out an unclosed part in a ground object target boundary line, and dividing the unclosed part into three types, namely a boundary broken line at the edge of an image (the boundary at the edge of the image is incomplete due to a skeleton extraction algorithm), a boundary broken line inside the image (the part which is not correctly identified by a boundary detection model) and a boundary line head inside the image (the target boundary has an excessive boundary line due to the skeleton extraction algorithm).
Step 5.2: three types of boundary breaks are handled in three different ways.
Step 5.2.1: and regarding the boundary broken line at the edge of the image, taking the edge strength image at the corresponding position as an indication, selecting pixel points with higher edge strength values to fill the target boundary broken line, and if a gap still exists between the pixel points and the image boundary, adopting a straight line perpendicular to the image boundary to connect the pixel points and the target boundary broken line and form a closed target boundary line by connecting the pixel points and the image boundary.
Step 5.2.2: and resetting the threshold for the boundary broken line in the image, mapping pixel points with the intensity values higher than the new threshold in the edge intensity graph into a binary graph by taking the edge intensity graph at the corresponding position as an indication, and connecting the two broken points by the original geometric characteristics if the target boundary broken line is not closed. The steps of repairing the break point with the original geometric characteristics are as follows.
Step 5.2.2.1: and taking out the innermost boundary line connected with the two break points.
Step 5.2.2.2: : the most inner circle boundary line is divided into a plurality of parts according to the change condition of the slope, and the approximate geometrical shape of the boundary line is determined.
Step 5.2.2.3: and (4) repairing the broken points into a closed graph according to the geometric shape, so that the repaired positions keep the original geometric characteristics.
Step 5.2.3: the isolated boundary lines within the image, i.e., the isolated boundary lines that are far from other broken lines, are deleted.
Step 5.3: and 5, performing skeleton extraction again, and refining the part filled in the step 5.2 into a target boundary line with the width of a single pixel.
Step 5.4: and traversing the image again, and deleting all unclosed target boundary lines to obtain a complete and fine ground object target boundary shown in fig. 6-A, wherein fig. 6-B is a predicted boundary result of Mask-RCNN, and the comparison of the two results shows that the method has a relatively obvious precision advantage.
The embodiments described in this specification are merely illustrative of implementations of the inventive concept and the scope of the present invention should not be considered limited to the specific forms set forth in the embodiments but rather by the equivalents thereof as may occur to those skilled in the art upon consideration of the present inventive concept.

Claims (1)

1. A high-resolution remote sensing target fine extraction method based on edge assistance comprises the following steps:
step 1: making a remote sensing image target extraction sample, drawing a fine boundary of a target by contrasting an image and determining a type; generating two different labels for the same image sample, namely generating a target outer covering frame label according to the requirement of a target detection model, and generating a target boundary label according to the requirement of an edge detection model; the method specifically comprises the following steps:
step 1.1: obtaining a high-resolution remote sensing image: optical satellite remote sensing data with a visible light-near infrared sensor or aviation remote sensing data carrying a common optical camera are adopted, and multispectral images or fusion panchromatic images can be directly used according to resolution requirements;
step 1.2: cutting a remote sensing image: selecting a range of a typical target in a production area, and cutting the remote sensing image according to uniform pixels;
step 1.3: manufacturing a deep learning training sample: drawing a deep learning training sample by using ArcGIS or other GIS software, marking the boundary of each ground feature target to obtain a corresponding shp file, and generating a target outer frame mark according to the requirements of a target detection model; generating a target boundary label according to the requirement of the edge detection model;
step 1.4: collecting more than 200 images and corresponding labels as training samples according to task requirements, and independently preparing test samples for detection precision;
step 2: training a target detection model and an edge detection model by using the prepared image sample; training a Faster R-CNN neural network to obtain a target detection model, training an RCF neural network to obtain an edge detection model, and correspondingly replacing and modifying the network according to a production target; the method specifically comprises the following steps:
step 2.1: designing a deep convolutional neural network: in order to train a target detection model and an edge detection model, two neural networks of RCF and Faster R-CNN are selected, and the networks can be replaced and modified correspondingly according to a production target;
step 2.2: initializing the weight: initializing RCF network weights using a VGG pre-training model, initializing weights of Faster R-CNN using a pre-training model on a COCO data set;
step 2.3: setting a training hyper-parameter: configuring hyper-parameters, and specifically setting numerical values after model tuning;
training parameter setting of RCF: 8000 for iteration number, 4 for batch _ size, 3200,4800,6400,8000 for learning rate update step, 0.001 for initial learning rate, and 0.1 for learning rate update coefficient;
training parameter settings for Faster R-CNN: training phase number is 3, each phase iteration round number is [40,120, 40], each iteration round number is 1000, verification interval is 50, batch _ size is 4, learning rate update step size is 1000, initial learning rate is 0.001, learning rate update momentum is 0.9, and weight decay is 0.0001;
step 2.4: inputting a sample, training a model: inputting the training samples into the RCF model, and training according to the hyper-parameters in the step 2.3 to obtain an edge detection model capable of extracting the edge contour of the ground object target; inputting the training sample into the Faster R-CNN, training according to the hyperparameter in the step 2.3, and obtaining a target detection model capable of extracting a surface feature target outer wrapping frame;
and step 3: inputting an image sample for testing into a target detection model and an edge detection model, and acquiring a surface feature target outer frame and an edge intensity map; the surface feature target edge intensity chart indicates the possibility that the corresponding position in the remote sensing image is the target edge; the ground object target outer covering frame is a rectangular limiting frame for marking the position range of the target and the type of the target; the method specifically comprises the following steps:
step 3.1: inputting the high-resolution remote sensing image into a target detection model, and obtaining a rectangular outer frame of the ground object target; inputting the high-resolution remote sensing image into an edge detection model to obtain a surface feature target edge intensity map;
step 3.2: parameterizing an outer package frame: converting the rectangular outer frame parameters output by the target detection model into a left lower vertex coordinate and a right upper vertex coordinate of the rectangular outer frame parameters; specifically, the method comprises the following steps:
x1=x-w y1=y-h
x2=x+w y2=y+h
wherein x, y, w and h respectively represent the central abscissa, the central ordinate, the width and the height of the rectangular outer wrapping frame;
and 4, step 4: the edge intensity graph in the step 3 is refined into a single-pixel-wide boundary; the method specifically comprises the following steps:
step 4.1: converting the surface feature target edge intensity map into a binary map from a gray map according to a set threshold; the edge judgment condition of the ground object target at the image (x, y) is represented by binary _ image (x, y), and can be represented as:
Figure FDA0002726304240000021
wherein 1 represents edge, 0 represents not edge, threshold is real number in interval [0,1], which can be set by user, and initial default value is 0.5; x and y are horizontal and vertical coordinates of the image;
step 4.2: for the thick edge lines with a plurality of pixel widths in the binary image, continuously performing corrosion operation by using the center of each edge line until only one pixel width is left in all the edge lines, thereby achieving the purpose of skeleton extraction;
and 5: taking the foreign object target outer frame in the step 3 as constraint, and repairing the single pixel boundary in the step 4 to obtain a complete and fine polygonal foreign object target boundary; the detailed steps are as follows:
step 5.1: finding out an unclosed part in a ground object target boundary line, and dividing the unclosed part into three types, namely a boundary broken line at the edge of an image, a boundary broken line in the image and a boundary line head in the image, wherein the boundary broken line at the edge of the image is the part which is not correctly identified by a boundary detection model, and the boundary broken line in the image is the part which is not correctly identified by the boundary detection model, and the boundary line head in the image is the redundant boundary line of the target boundary caused by the skeleton extraction algorithm;
step 5.2: processing the three types of boundary broken lines by three different methods;
step 5.2.1: for the boundary broken line at the edge of the image, taking an edge strength image at a corresponding position as an indication, selecting pixel points with higher edge strength values to fill the target boundary broken line, if a gap still exists between the pixel points and the image boundary, adopting a straight line perpendicular to the image boundary to connect the pixel points with the target boundary broken line and form a closed target boundary line by connecting the pixel points with the image boundary;
step 5.2.2: resetting a threshold value for the boundary broken line in the image, mapping pixel points with intensity values higher than the new threshold value in the edge intensity graph into a binary image by taking the edge intensity graph at the corresponding position as an indication, and connecting the two broken points by the original geometric characteristics if the target boundary broken line is not closed; wherein, the step of repairing the breakpoint by the original geometric characteristic is as follows;
step 5.2.2.1: taking out the boundary line of the innermost circle connected with the two breakpoints;
step 5.2.2.2: dividing the boundary line of the innermost circle into a plurality of parts according to the change condition of the slope, and determining the approximate geometric shape of the boundary line;
step 5.2.2.3: repairing the broken point into a closed graph according to the geometric shape, so that the repaired part keeps the original geometric characteristic;
step 5.2.3: deleting the isolated boundary lines in the image, namely the isolated boundary lines which are far away from other broken lines;
step 5.3: performing skeleton extraction again, and refining the part filled in the step 5.2 into a target boundary line with a single pixel width;
step 5.4: and traversing the image again, and deleting all unclosed target boundary lines to obtain a complete and fine polygonal ground object target boundary.
CN201910638370.0A 2019-07-16 2019-07-16 Semantic edge-assisted high-resolution remote sensing target fine extraction method Active CN110443822B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910638370.0A CN110443822B (en) 2019-07-16 2019-07-16 Semantic edge-assisted high-resolution remote sensing target fine extraction method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910638370.0A CN110443822B (en) 2019-07-16 2019-07-16 Semantic edge-assisted high-resolution remote sensing target fine extraction method

Publications (2)

Publication Number Publication Date
CN110443822A CN110443822A (en) 2019-11-12
CN110443822B true CN110443822B (en) 2021-02-02

Family

ID=68430338

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910638370.0A Active CN110443822B (en) 2019-07-16 2019-07-16 Semantic edge-assisted high-resolution remote sensing target fine extraction method

Country Status (1)

Country Link
CN (1) CN110443822B (en)

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111818557B (en) * 2020-08-04 2023-02-28 中国联合网络通信集团有限公司 Network coverage problem identification method, device and system
CN112084871B (en) * 2020-08-10 2024-02-13 浙江工业大学 High-resolution remote sensing target boundary extraction method based on weak supervised learning
CN112084872A (en) * 2020-08-10 2020-12-15 浙江工业大学 High-resolution remote sensing target accurate detection method fusing semantic segmentation and edge
CN111967526B (en) * 2020-08-20 2023-09-22 东北大学秦皇岛分校 Remote sensing image change detection method and system based on edge mapping and deep learning
US11836223B2 (en) * 2020-11-13 2023-12-05 Meta Platforms, Inc. Systems and methods for automated detection of building footprints
CN113160258B (en) * 2021-03-31 2022-11-29 武汉汉达瑞科技有限公司 Method, system, server and storage medium for extracting building vector polygon
CN113128388B (en) * 2021-04-14 2022-09-02 湖南大学 Optical remote sensing image change detection method based on space-time spectrum characteristics
CN114241326B (en) * 2022-02-24 2022-05-27 自然资源部第三地理信息制图院 Progressive intelligent production method and system for ground feature elements of remote sensing images
CN115273154B (en) * 2022-09-26 2023-01-17 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) Thermal infrared pedestrian detection method and system based on edge reconstruction and storage medium
CN115797633B (en) * 2022-12-02 2023-06-27 中国科学院空间应用工程与技术中心 Remote sensing image segmentation method, remote sensing image segmentation system, storage medium and electronic equipment

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7341841B2 (en) * 2003-07-12 2008-03-11 Accelr8 Technology Corporation Rapid microbial detection and antimicrobial susceptibility testing
CN103247032B (en) * 2013-04-26 2015-12-02 中国科学院光电技术研究所 A kind of faint Extended target localization method based on pose compensation
JP6834417B2 (en) * 2016-11-30 2021-02-24 ブラザー工業株式会社 Image processing equipment and programs
JP2018147199A (en) * 2017-03-03 2018-09-20 ブラザー工業株式会社 Image processing device, and computer program
CN109712140B (en) * 2019-01-02 2021-01-26 国电内蒙古东胜热电有限公司 Method and device for training fully-connected classification network for leakage detection

Also Published As

Publication number Publication date
CN110443822A (en) 2019-11-12

Similar Documents

Publication Publication Date Title
CN110443822B (en) Semantic edge-assisted high-resolution remote sensing target fine extraction method
CN111027547B (en) Automatic detection method for multi-scale polymorphic target in two-dimensional image
CN109146948B (en) Crop growth phenotype parameter quantification and yield correlation analysis method based on vision
CN109446992B (en) Remote sensing image building extraction method and system based on deep learning, storage medium and electronic equipment
CN109580630B (en) Visual inspection method for defects of mechanical parts
Khoshelham et al. Performance evaluation of automated approaches to building detection in multi-source aerial data
CN103927526B (en) Vehicle detecting method based on Gauss difference multi-scale edge fusion
CN102704215B (en) Automatic cutting method of embroidery cloth based on combination of DST file parsing and machine vision
Wolf et al. Automatic extraction and delineation of single trees from remote sensing data
CN107133623B (en) Pointer position accurate detection method based on background difference and circle center positioning
CN108537169B (en) High-resolution remote sensing image road extraction method based on center line and road width detection algorithm
CN111414826A (en) Method, device and storage medium for identifying landmark arrow
Chen et al. Automatic building extraction via adaptive iterative segmentation with LiDAR data and high spatial resolution imagery fusion
Song et al. Automated geospatial conflation of vector road maps to high resolution imagery
CN112861654A (en) Famous tea picking point position information acquisition method based on machine vision
CN112183301A (en) Building floor intelligent identification method and device
CN111487643B (en) Building detection method based on laser radar point cloud and near-infrared image
CN115841669A (en) Pointer instrument detection and reading identification method based on deep learning technology
CN115731257A (en) Leaf form information extraction method based on image
Gavrilov et al. A method for aircraft labeling in aerial and satellite images based on continuous morphological models
CN117496401A (en) Full-automatic identification and tracking method for oval target points of video measurement image sequences
CN117541652A (en) Dynamic SLAM method based on depth LK optical flow method and D-PROSAC sampling strategy
CN115330762B (en) Fuse wire breakage detection method of X-ray image
CN116385477A (en) Tower image registration method based on image segmentation
CN110929739A (en) Automatic impervious surface range remote sensing iterative extraction method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant