CN110111346A - Remote sensing images semantic segmentation method based on parallax information - Google Patents

Remote sensing images semantic segmentation method based on parallax information Download PDF

Info

Publication number
CN110111346A
CN110111346A CN201910399019.0A CN201910399019A CN110111346A CN 110111346 A CN110111346 A CN 110111346A CN 201910399019 A CN201910399019 A CN 201910399019A CN 110111346 A CN110111346 A CN 110111346A
Authority
CN
China
Prior art keywords
parallax
result
remote sensing
sensing images
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910399019.0A
Other languages
Chinese (zh)
Other versions
CN110111346B (en
Inventor
焦李成
唐旭
***
李英萍
孙龙
屈嵘
李玲玲
郭雨薇
冯志玺
张梦璇
丁静怡
张丹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xidian University
Original Assignee
Xidian University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xidian University filed Critical Xidian University
Priority to CN201910399019.0A priority Critical patent/CN110111346B/en
Publication of CN110111346A publication Critical patent/CN110111346A/en
Application granted granted Critical
Publication of CN110111346B publication Critical patent/CN110111346B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/73Deblurring; Sharpening
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/80Geometric correction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20021Dividing image into blocks, subimages or windows
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20228Disparity calculation for image-based rendering

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

The present invention proposes a kind of remote sensing images semantic segmentation method based on parallax information, for solving the lower technical problem of segmentation precision existing in the prior art, realizes step are as follows: obtain training sample set and test sample collection;Training sample set is pre-processed;Parallax network is trained;Parallax detection is carried out to test sample collection;Obtain the parallax information of test sample collection;Semantic segmentation network is trained;Semantic detection is carried out to test sample collection;Preliminary semantic segmentation result is modified;Obtain final semantic segmentation result.The present invention carries out left and right consistency detection LRC to the parallax result of remote sensing images, and the parallax information for recycling detection to obtain is modified semantic segmentation result, and proposes a completely new semantic stripping and slicing fusion detection method, significantly improves the precision of semantic segmentation.It can be used for the practical applications such as ground quality detection, land use, urban planning, automatic Pilot, human-computer interaction, medical image identification.

Description

Remote sensing images semantic segmentation method based on parallax information
Technical field
The invention belongs to technical field of image processing, are related to a kind of remote sensing images semantic segmentation method, and in particular to a kind of Remote sensing images semantic segmentation method based on parallax information can be applied to ground quality detection, land use, urban planning, drive automatically Sail, human-computer interaction, medical image identification etc. fields.
Background technique
Semantic segmentation refers to according to color, gray scale and Texture eigenvalue, divides an image into several and mutually disjoints and have There is the pixel region of certain certain semantic meaning, and identify the classification in each region, the pixel of the same area is endowed identical Color, the final image for obtaining a width and there is pixel semantic tagger.
Remote sensing is to carry out the imaging of specific electromagnetic wave spectral coverage to the earth by the sensor on satellite, be with aeroplane photography The a special kind of skill to grow up based on technology.It, can in a short time, to regional on a large scale on the earth by remote sensing Multi-level, multi-angle of view observation is carried out, is the important means for obtaining environmental information and earth resource.Therefore, the letter of remote sensing image Breath extractive technique is particularly important.It can be said that the final goal of remote sensing be just to be able to extract from image it is useful Information gets knowledge.It, can be by image, semantic cutting techniques, by the specific objective in image in satellite remote sensing images Division mark comes out, and the specific information in remote sensing images is extracted with this.Currently, as remote sensing images semantic segmentation technology is in geology The extensive use in the fields such as detection, land use, urban planning, automatic Pilot, human-computer interaction, medical image identification, to semanteme The requirement of segmentation precision is higher and higher.
Before deep learning is applied to computer vision field, researcher generally uses texture primitive forest (TextonForest) or the conventional methods such as random forest (Random Forest) construct the classification for semantic segmentation Device.Random forest is gone out the classification of sample by multiple decision tree independent predictions, and by voting all prediction classifications, poll is most High classification is elected as final prediction result.Texture primitive forest then combines image texture characteristic building classifier, has Improve to effect the precision of classification.But conventional method is relative complex, computationally intensive, the manual feature utilized needs field Special knowledge takes time and effort, and cannot achieve real-time requirement, and too fining sometimes is so that can only be specific at some It is used under scene, for possessing the image of complicated, changeable details, is extremely difficult to high-precision.
Currently, the semantic segmentation method based on deep learning is just receiving significant attention, main thought is to utilize to refer in advance The true tag for determining training sample learns image by network, then does not occur with the model prediction training set of study New samples, and provide the classification of new samples.Relatively traditional semantic segmentation method, the method based on deep learning has powerful Learn suitable character representation ability for current problem, detection accuracy and speed, which have, to be increased substantially.For example, application is public Cloth number be CN 107610141A, the patent application of entitled " a kind of remote sensing images semantic segmentation method based on deep learning ", Disclose a kind of remote sensing images semantic segmentation method based on deep learning, this method be every kind of species Target Assignment rgb value and Gray value obtains original remote sensing images, select species target and paint, gray processing and assign gray value processing, marked Image is signed, data enhancing and edge extracting are carried out to original remote sensing images, the image after obtaining edge extracting;By original remote sensing Image training sample after image and edge extracting is trained complete convolutional neural networks, obtains best semantic segmentation network Test remote sensing images are inputted in best semantic segmentation network model, get semantic segmentation result images by model;For semanteme point It cuts result images and carries out colouring processing, obtain final semantic segmentation result images.Although the complete convolution mind that this method proposes It is realized instead of the penalty values computation layer in basic convolutional neural networks by convolutional calculation junior scholar through network using warp lamination The remote sensing features image reconstruction that acquistion is arrived is to original image size, so that the precision of remote sensing images semantic segmentation is effectively increased, There is stronger applicability for the segmentation problem of remote sensing images, still, there are still some shortcomingss for this method: due to the party Method is extracted the marginal information of image as training data, to the accuracy requirement height of image border, therefore is suitable only for scene Simple remote sensing images, for the remote sensing images of complex scene, marginal information is often fuzzy, non-closed, uses the party Rule is easy the marginal information for making network excessively learn image, over-fitting occurs, causes semantic segmentation precision low.
Summary of the invention
It is an object of the invention to overcome above-mentioned the deficiencies in the prior art, a kind of remote sensing images based on parallax information are proposed Semantic segmentation method, for solving the lower technical problem of segmentation precision existing in the prior art.
To achieve the above object, the technical solution that the present invention takes includes the following steps:
(1) training sample set and test sample collection are obtained:
Obtain the remote sensing that the satellite of two different locations is w × h to the s width different zones size that Same Scene is shot respectively Image, s > 100, w > 1000, h > 1000, the shared c semantic classes of all pixels point of 2s width remote sensing images, c=1,2, 3 ..., the image of the same area of two satellites shooting is matched, obtains s remote sensing images pair, and select s therein1 A remote sensing images are to as training sample set, remaining s2A remote sensing images to as test sample collection,s2=s- s1
(2) training sample set is pre-processed:
The every width remote sensing images concentrated to training sample are sharpened processing, and to each image after Edge contrast into The correction of row epipolar geom etry, obtains pretreated training sample set;
(3) parallax network is trained:
All remote sensing images that pretreated training sample is concentrated are to the view in input parallax network, after being trained Poor network;
(4) parallax detection is carried out to test sample collection:
(4a) using belong in all remote sensing images pair that test sample collection includes a satellite shooting remote sensing images as Test sample collection test1s, the remote sensing images of another satellite shooting are as test sample collection test2s
(4b) is with test sample collection test1sOn the basis of, using the parallax network after training to test sample collection s2In it is every A image obtains test to parallax detection is carried out1sParallax result disa,aiIt is i-th Size be w × h parallax as a result, Indicate the parallax of j-th of pixel of i as a result, simultaneously With test sample collection test2sOn the basis of, using the parallax network after training to test sample collection s2In each image to progress Parallax detection, obtains test2sParallax result disb,biIt is w × h's for i-th of size Parallax as a result, Indicate the parallax result of j-th of pixel of i;
(5) parallax information of test sample collection is obtained:
(5a) is to parallax result disaCarry out left and right consistency detection LRC:
(5a1) enables i=1, j=1;
(5a2) calculates parallax result disbIn with parallax result disa'sMatch the coordinate of pixel:
It is by coordinate value'sParallax valueCalculate parallax result disbIn withMatch pixel Coordinate valueWith
(5a3) calculates parallax result disbIn with parallax result disa'sMatch the exhausted of pixel parallax value difference To value:
Coordinates computed value isWithPoint parallax value absolute value of the differenceAnd coordinate Value isWithPoint parallax value absolute value of the difference Indicates coordinate value is Point parallax value,Indicates coordinate value isPoint parallax value;
(5a4) updates parallax result disaMiddle coordinate value isPoint parallax value
JudgementIt is whether true, δ ∈ (1,3), if so,It is not blocked, keepsIt is constant, And step (5a5) is executed, otherwise, thenIt is blocked, in parallax result aiIn fromIt is horizontal to find first to the left not It is blocked a littleFirst is found horizontally to the right not being blocked a littleIt enables Indicates coordinate value isPoint parallax value,Indicates coordinate value isPoint parallax value, and execute step (5a5);
(5a5) obtains disaLeft and right consistency detection LRC result dis 'a:
Whether true judge j≤w × h, if so, enabling j=j+1, and executes step (5a2) and otherwise judge i≤s2Whether It sets up, if so, enabling j=1, i=i+1, and executes step (5a2), otherwise, obtain parallax result disaLeft and right consistency detection LRC result dis 'a
(5b) is to parallax result disbCarry out left and right consistency detection LRC:
According to the method for step (5a) to parallax result disbLeft and right consistency detection LRC is carried out, parallax result dis is obtainedb Left and right consistency detection LRC result disb', and by dis 'aAnd disb' parallax information the dis ' as test sample collection,puFor u-th of parallax information;
(6) semantic segmentation network is trained:
According to the sequence of Row Column or Column Row, the sliding window for being m × m by size, is step-length in training using n It is slided, is obtained on every pretreated remote sensing images in sample setA figure As block, and all image blocks are input in semantic segmentation network, the semantic segmentation network after being trained, m > 128, n > 64, m > n;
(7) Semantic detection is carried out to test sample collection:
(7a) be arranged using the long w of remote sensing images as x-axis, wide h be y-axis datum level, and by test sample concentrate every width The semantic segmentation network that image rotates clockwise after the input training of each image after e degree relative to the datum level carries out stripping and slicing and melts Detection is closed, dimension probability matrix matrix_e, e=0 identical with remote sensing images classification number c, 90,180 are obtained;
Probability matrix matrix_e is rotated e degree by (7b) counterclockwise, obtains the remote sensing images size phase for being w × h with size Same, dimension is the probability matrix matrix_e ' of c, Indicate probability matrix T-th of pixel of matrix_e ', e '=0,90,180, It indicatesBelong to The probability value of semantic classes r;
(7c) is calculatedIn e '=0, the average value of 90,180 3 probability matrixs, obtain that size is w × h, dimension is the flat of c Equal probability matrix matrix_avg, Indicate t-th of pixelBelong to the probability value of semantic classes r;
(7d) calculates pixelSemantic classes r,It obtains just Semantic segmentation result seg is walked,quIndicate u-th of preliminary semantic segmentation result;
(8) preliminary semantic segmentation result is modified:
(8a) is to parallax information puIt is clustered, obtains the parallax cluster result g of test sample collection,guIndicate u-th of parallax cluster result, guWith kuA parallax classification ku=1,2, 3,…;
(8b) divides parallax cluster result guIn pixel region area included by all parallax classificationsu, Indicate areauIn pixel included by z-th of parallax classification Region;
(8c) passes through preliminary semantic segmentation result qu'sIn the most semantic classes pair of frequency of occurrenceIn The semantic classes of all pixels point is replaced, and obtains the revised semantic segmentation result of test sample collection;
(9) final semantic segmentation result is obtained:
Each revised semantic segmentation of pair to test sample collection is every class semanteme mesh as a result, according to different classifications Mark distribution rgb value, obtains the final semantic segmentation result of test sample collection.
Compared with prior art, the present invention having the advantage that
First, the present invention is due to carrying out parallax inspection to test sample collection first during obtaining semantic segmentation result It surveys, and left and right consistency detection LRC, then the parallax information obtained by detection is carried out to test sample collection to parallax testing result The primary segmentation result for carrying out Semantic detection acquisition is modified, finally to every class semanteme mesh of revised semantic segmentation result Mark distribution rgb value, combine be all pixel scale parallax Detection task, sufficiently excavate remote sensing images multi-source information, avoid The prior art detect and distribute semantic segmentation network to rgb value after image as final semantic segmentation result to segmentation essence The influence of degree effectively improves the precision of semantic segmentation.
Second, the present invention carries out left and right consistency detection LRC to the parallax result that parallax network detects, and passes through left and right The smaller value for finding first unshielding point parallax value is replaced the parallax value blocked a little, and blocking is a little only in a sub-picture Middle appearance solves in the detection of remote sensing images parallax in the pixel that another piece image does not occur due to blocking a presence, nothing Method finds match point, and then the problem for causing parallax precision low, obtains accurate parallax information, further improves semantic segmentation Precision.
Third when the present invention carries out Semantic detection to remote sensing images, carries out stripping and slicing fusion inspection to remote sensing images to be detected It surveys, first carries out remote sensing images to be detected up and down and bilateral symmetry, and sliding is carried out to the remote sensing images after symmetrical and takes block, Then the central part that each image block is sent into the probability matrix that semantic network detects is intercepted, then splices each probability matrix Central part, finally interception splicing posterior probability matrix central part, obtain the probability matrix of institute's detection image, both realized Stripping and slicing small size detection to image, increases the detection accuracy to Small object, in turn avoids the erroneous detection to image border part, Further improve the precision of semantic segmentation.
Detailed description of the invention
Fig. 1 is implementation flow chart of the invention;
Fig. 2 is the present invention to parallax result disaCarry out the implementation flow chart of left and right consistency detection LRC;
Fig. 3 is the implementation flow chart that the present invention carries out stripping and slicing fusion detection to remote sensing images.
Specific embodiment
In the following with reference to the drawings and specific embodiments, present invention is further described in detail.
Referring to Fig.1, the present invention includes the following steps:
Step 1) obtains training sample set and test sample collection:
Obtain the remote sensing that the satellite of two different locations is w × h to the s width different zones size that Same Scene is shot respectively Image, s > 100, w > 1000, h > 1000, the shared c semantic classes of all pixels point of 2s width remote sensing images, c=1,2, 3 ..., the image of the same area of two satellites shooting is matched, obtains s remote sensing images pair, and select s therein1 A remote sensing images are to as training sample set, remaining s2A remote sensing images to as test sample collection,s2=s- s1
In the present embodiment, remote sensing images are obtained from the video that worldview-3 satellite is shot, s=4292, w =1024, h=1024, c=5 randomly select 80% sample as training sample set s from s remote sensing images pair1, remaining As test sample collection s2
Step 2) pre-processes training sample set:
The every width remote sensing images concentrated to training sample are sharpened processing, and to each image after Edge contrast into The correction of row epipolar geom etry, obtains pretreated training sample set.
The present embodiment is sharpened processing to every width remote sensing images, compensates the profile of image, enhances the edge and ash of image The part for spending jump, is apparent from image, then to each image after Edge contrast to epipolar geom etry correction is carried out, so that often Corresponding pixel points are in same horizontal line in the two images of a image pair, provide the training corrected for subsequent network training Sample set.
Step 3) is trained parallax network:
All remote sensing images that pretreated training sample is concentrated are to the view in input parallax network, after being trained Poor network.
Step 4) carries out parallax detection to test sample collection:
Step 4a) remote sensing images for belonging to a satellite shooting in all remote sensing images pair that test sample collection includes are made For test sample collection test1s, the remote sensing images of another satellite shooting are as test sample collection test2s
Step 4b) with test sample collection test1sOn the basis of, using the parallax network after training to test sample collection s2In Each image obtains test to parallax detection is carried out1sParallax result disa,aiIt is i-th A size be w × h parallax as a result, The parallax of j-th of pixel of i is indicated as a result, same When with test sample collection test2sOn the basis of, using the parallax network after training to test sample collection s2In each image into The detection of row parallax, obtains test2sParallax result disb,biIt is w × h for i-th of size Parallax as a result, Indicate the parallax result of j-th of pixel of i.
The parallax information of step 5) acquisition test sample collection:
Step 5a) to parallax result disaLeft and right consistency detection LRC is carried out, implementation flow chart is as shown in Figure 2:
Step 5a1) enable i=1, j=1;
Step 5a2) calculate parallax result disbIn with parallax result disa'sMatch the coordinate of pixel:
It is by coordinate value'sParallax valueCalculate parallax result disbIn withMatch pixel Coordinate valueWith
Step 5a3) calculate parallax result disbIn with parallax result disa'sMatch pixel parallax value difference Absolute value:
Coordinates computed value isWithPoint parallax value absolute value of the differenceAnd it sits Scale value isWithPoint parallax value absolute value of the difference Indicates coordinate value isPoint parallax value,Indicates coordinate value isPoint parallax value;
Step 5a4) update parallax result disaMiddle coordinate value isPoint parallax value
JudgementIt is whether true, δ ∈ (1,3), if so,It is not blocked, keepsIt is constant, And step (5a5) is executed, otherwise, thenIt is blocked, in parallax result aiIn fromIt is horizontal to find first to the left not It is blocked a littleFirst is found horizontally to the right not being blocked a littleIt enables Table Show that coordinate value isPoint parallax value,Indicates coordinate value isPoint parallax value, and execute step 5a5);
Step 5a5) obtain disaLeft and right consistency detection LRC result dis 'a:
Whether true judge j≤w × h, if so, enabling j=j+1, and execute step 5a2), otherwise, judge i≤s2Whether at It is vertical, if so, enabling j=1, i=i+1, and execute step 5a2), otherwise, obtain parallax result disaLeft and right consistency detection LRC As a result dis 'a
Step 5b) to parallax result disbCarry out left and right consistency detection LRC:
With to parallax result disaThe step of carrying out left and right consistency detection LRC is the same, is only distinguished as a variable a and changes into B, variable b change a into, according to the method for step 5a) to parallax result disbLeft and right consistency detection LRC is carried out, parallax knot is obtained Fruit disbLeft and right consistency detection LRC result disb', and by dis 'aAnd disb' the parallax information as test sample collection Dis ',puFor u-th of parallax information.
Left and right consistency detection LRC is carried out to the parallax result that parallax network detects, finds first by left and right The smaller value of unshielding point parallax value is replaced the parallax value blocked a little, solve remote sensing images parallax detection in due to A presence is blocked, the problem that can not be found match point, and then cause parallax precision low obtains accurate parallax information.
Step 6) is trained semantic segmentation network:
According to the sequence of Row Column or Column Row, the sliding window for being m × m by size, is step-length in training using n It is slided, is obtained on every pretreated remote sensing images in sample setA figure As block, and all image blocks are input in semantic segmentation network, the semantic segmentation network after being trained, m > 128, n > 64, m > n.
In the present embodiment, m=512, n=256, semantic segmentation are the detections of pixel scale, and each in remote sensing images For the pixel size of classification target there is biggish difference, Small object is fuzzy, it is difficult to learn, by the image for acquiring small size Block, amplification Small object allow network preferably to extract feature, improve the essence to Small object semantic segmentation in the ratio of image Degree.
Step 7) carries out Semantic detection to test sample collection:
Step 7a) stripping and slicing fusion detection is carried out to remote sensing images, implementation flow chart is as shown in Figure 3:
Step 7a1) symmetrical above and below to remote sensing images to be detected progress:
Be arranged using the long w of remote sensing images as x-axis, wide h be y-axis datum level, and by test sample concentrate each image Rotate clockwise e degree relative to the datum level, then be arranged e degree is rotated using every width after image long ω as x-axis, width ξ be y-axis, a left side Inferior horn is the rectangular coordinate system of origin, by each image after rotation e degreeRegion is carried out with y =ξ is the symmetry operation of axis,Region is carried out using y=0 as the symmetrical of axis Operation, obtaining size isImage;
Step 7a2) bilateral symmetry is carried out to the remote sensing images after symmetrical above and below:
It is by sizeImageRegion is carried out with x=0 The symmetry operation of axis,Region is carried out using x=ω as the symmetrical behaviour of axis Make, obtaining size isImage;
Step 7a3) it sliding is carried out to the remote sensing images after symmetrical takes block:
It is to sizeImage, according to Row Column or antecedent Capable sequence afterwards, the sliding window for being λ × λ by size, withIt is slided for step-length, obtains image block set f, f= {f1,f2,…,fo,…,fnum, foIndicate o-th of image block,
Step 7a4) Semantic detection is carried out to image block:
Semantic segmentation network after image block set f input training is detected, obtaining num size is the general of λ × λ Rate matrix;
Step 7a5) intercept the probability matrix for detecting obtained image block:
Be arranged using num size for the probability matrix of λ × λ long λ as x-axis, width λ be y-axis, the right angle that the lower left corner is origin Coordinate system intercepts each probability matrixRegion, obtain num interception after probability matrix collection It closes;
Step 7a6) stitching image block probability matrix:
According to image block fwIt is in sizeThe opposite position of image It sets, the probability matrix after num interception is spliced into the probability matrix that a size is (ω+λ) × (ξ+λ);
Step 7a7) obtain the probability matrixs of detected remote sensing images:
Being arranged with the long ω+λ for the probability matrix that size is (ω+λ) × (ξ+λ) is x-axis, width ξ+λ is y-axis, the lower left corner is The rectangular coordinate system of origin intercepts the region of (λ < x < λ+ω, the λ < y < λ+ξ) of probability matrix, and obtaining size is ω × ξ language Justice divides dimension probability matrix matrix_e, e=0 identical with remote sensing images classification number c after network detection, 90,180.
In the present embodiment, λ=512, the stripping and slicing fusion detection utilized had both realized the stripping and slicing small size inspection to image It surveys, increases the detection accuracy to Small object, in turn avoid further improving semantic segmentation to the erroneous detection of image border part Precision.
Step 7b) probability matrix matrix_e rotated into e degree counterclockwise, it obtains with size being w × h remote sensing images size phase Same, dimension is the probability matrix matrix_e ' of c, Indicate probability matrix T-th of pixel of matrix_e ', e '=0,90,180, It indicatesBelong to The probability value of semantic classes r;
Step 7c) it calculatesIn e '=0, the average value of 90,180 3 probability matrixs, obtaining size is w × h, dimension For the average probability matrix matrix_avg of c,
Indicate t-th of pixelBelong to the probability value of semantic classes r;
Step 7d) calculate pixelSemantic classes r,? To preliminary semantic segmentation result seg,quIndicate u-th of preliminary semantic segmentation result;
The pixel distributions of various forms of pictures is different, testing result also can difference, compared to by single network It makes a policy, a possibility that mistake for the average value for taking multiple networks to make a policy is low, improves the accuracy of semantic segmentation.
Step 8) is modified preliminary semantic segmentation result:
Step 8a) to parallax information puIt is clustered, obtains the parallax cluster result g of test sample collection,guIndicate u-th of parallax cluster result, guWith kuA parallax classification ku=1,2, 3,…;
Step 8b) divide parallax cluster result guIn pixel region area included by all parallax classificationsu, Indicate areauIn pixel included by z-th of parallax classification Region;
Step 8c) pass through preliminary semantic segmentation result qu'sIn the most semantic classes pair of frequency of occurrence In the semantic classes of all pixels point be replaced, obtain the final semantic segmentation result of test sample collection.
Disparity computation result contains the atural objects such as building, trees parallax information abundant, the target category of same disparity value It is bigger in the same category a possibility that, so parallax information has very great help for segmentation task.Therefore parallax result is used Semantic results are modified.
Step 9) obtains final semantic segmentation result:
Each revised semantic segmentation of pair to test sample collection is every class semanteme mesh as a result, according to different classifications Mark distribution rgb value, obtains the final semantic segmentation result of test sample collection.

Claims (2)

1. a kind of remote sensing images semantic segmentation method based on parallax information, which comprises the following steps:
(1) training sample set and test sample collection are obtained:
Obtain the remote sensing figure that the satellite of two different locations is w × h to the s width different zones size that Same Scene is shot respectively Picture, s > 100, w > 1000, h > 1000, the shared c semantic classes of all pixels point of 2s width remote sensing images, c=1,2, 3 ..., the image of the same area of two satellites shooting is matched, obtains s remote sensing images pair, and select s therein1 A remote sensing images are to as training sample set, remaining s2A remote sensing images to as test sample collection,
(2) training sample set is pre-processed:
Processing is sharpened to every width remote sensing images that training sample is concentrated, and to each image after Edge contrast to progress pole Line geometry correction, obtains pretreated training sample set;
(3) parallax network is trained:
All remote sensing images that pretreated training sample is concentrated are to the parallax net in input parallax network, after being trained Network;
(4) parallax detection is carried out to test sample collection:
(4a) will belong to the remote sensing images of satellite shooting as test in all remote sensing images pair that test sample collection includes Sample set test1s, the remote sensing images of another satellite shooting are as test sample collection test2s
(4b) is with test sample collection test1sOn the basis of, using the parallax network after training to test sample collection s2In each figure As obtaining test to parallax detection is carried out1sParallax result disa,aiFor i-th of size For w × h parallax as a result, The parallax of j-th of pixel of i is indicated as a result, simultaneously to survey Try sample set test2sOn the basis of, using the parallax network after training to test sample collection s2In each image to carry out parallax Detection, obtains test2sParallax result disb,biThe parallax for being w × h for i-th of size As a result, Indicate the parallax result of j-th of pixel of i;
(5) parallax information of test sample collection is obtained:
(5a) is to parallax result disaCarry out left and right consistency detection LRC:
(5a1) enables i=1, j=1;
(5a2) calculates parallax result disbIn with parallax result disa'sMatch the coordinate of pixel:
It is by coordinate value'sParallax valueCalculate parallax result disbIn withMatch the coordinate of pixel ValueWith
(5a3) calculates parallax result disbIn with parallax result disa'sMatch pixel parallax value absolute value of the difference:
Coordinates computed value isWithPoint parallax value absolute value of the differenceAnd coordinate value isWithPoint parallax value absolute value of the difference Indicates coordinate value isPoint Parallax value,Indicates coordinate value isPoint parallax value;
(5a4) updates parallax result disaMiddle coordinate value isPoint parallax value
JudgementIt is whether true, δ ∈ (1,3), if so,It is not blocked, keepsIt is constant, and hold Row step (5a5), otherwise, thenIt is blocked, in parallax result aiIn fromHorizontal searching first to the left is not hidden Catch pointFirst is found horizontally to the right not being blocked a littleIt enables It indicates to sit Scale value isPoint parallax value,Indicates coordinate value isPoint parallax value, and execute step (5a5);
(5a5) obtains disaLeft and right consistency detection LRC result dis 'a:
Whether true judge j≤w × h, if so, enabling j=j+1, and executes step (5a2) and otherwise judge i≤s2It is whether true, If so, enabling j=1, i=i+1, and step (5a2) is executed, otherwise, obtains parallax result disaLeft and right consistency detection LRC knot Fruit dis 'a
(5b) is to parallax result disbCarry out left and right consistency detection LRC:
According to the method for step (5a) to parallax result disbLeft and right consistency detection LRC is carried out, parallax result dis is obtainedbA left side Right uniformity detects LRC result dis 'b, and by dis 'aWith dis 'bAs the parallax information dis ' of test sample collection,puFor u-th of parallax information;
(6) semantic segmentation network is trained:
According to the sequence of Row Column or Column Row, the sliding window for being m × m by size, is step-length in training sample using n It concentrates and is slided on every pretreated remote sensing images, obtainedA image Block, and all image blocks are input in semantic segmentation network, the semantic segmentation network after being trained, m > 128, n > 64, m > n;
(7) Semantic detection is carried out to test sample collection:
(7a) be arranged using the long w of remote sensing images as x-axis, wide h be y-axis datum level, and by test sample concentrate each image The semantic segmentation network after each image input training after rotating clockwise e degree relative to the datum level carries out stripping and slicing fusion inspection It surveys, obtains dimension probability matrix matrix_e, e=0 identical with remote sensing images classification number c, 90,180;
Probability matrix matrix_e is rotated e degree by (7b) counterclockwise, obtains, dimension identical for the remote sensing images size of w × h as size Degree is the probability matrix matrix_e ' of c, Indicate probability matrix T-th of pixel of matrix_e ', e '=0,90,180, It indicatesBelong to The probability value of semantic classes r;
(7c) is calculatedIn e '=0, the average value of 90,180 3 probability matrixs, obtain that size is w × h, dimension is the average general of c Rate matrix matrix_avg, Indicate t-th of pixelBelong to the probability value of semantic classes r;
(7d) calculates pixelSemantic classes r,Obtain preliminary semanteme Segmentation result seg,quIndicate u-th of preliminary semantic segmentation result;
(8) preliminary semantic segmentation result is modified:
(8a) is to parallax information puIt is clustered, obtains the parallax cluster result g of test sample collection,guIndicate u-th of parallax cluster result, guWith kuA parallax classification ku=1,2, 3,…;
(8b) divides parallax cluster result guIn pixel region area included by all parallax classificationsu, Indicate areauIn pixel included by z-th of parallax classification Region;
(8c) passes through preliminary semantic segmentation result qu'sIn the most semantic classes pair of frequency of occurrenceIn it is all The semantic classes of pixel is replaced, and obtains the revised semantic segmentation result of test sample collection;
(9) final semantic segmentation result is obtained:
Each revised semantic segmentation of pair to test sample collection is as a result, according to different classifications, for every class semantic objects point With rgb value, the final semantic segmentation result of test sample collection is obtained.
2. the remote sensing images semantic segmentation method according to claim 1 based on parallax information, which is characterized in that step Each image after each image that test sample is concentrated to be rotated clockwise relative to the datum level to e degree described in (7a) is defeated Semantic segmentation network after entering training carries out stripping and slicing fusion detection, realizes step are as follows:
(7a1) carries out remote sensing images to be detected symmetrical above and below:
Be arranged the long ω of image after rotating e degree using every width of test sample collection as x-axis, width ξ be y-axis, the lower left corner is the straight of origin Angular coordinate system, by each image after rotation e degreeRegion is carried out using y=ξ as the symmetrical behaviour of axis Make,Region is carried out using y=0 as the symmetry operation of axis, and obtaining size isImage;
(7a2) carries out bilateral symmetry to the remote sensing images after symmetrical above and below:
It is by sizeImageRegion is carried out using x=0 as axis Symmetry operation,Region obtain using x=ω as the symmetry operation of axis It is to sizeImage;
(7a3) carries out sliding to the remote sensing images after symmetrical and takes block:
It is to sizeImage, according to Row Column or Column Row Sequence, by size be λ × λ sliding window, withIt is slided for step-length, obtains image block set f, f={ f1, f2,…,fo,…,fnum, foIndicate o-th of image block,
(7a4) carries out Semantic detection to image block:
Semantic segmentation network after image block set f input training is detected, the probability square that num size is λ × λ is obtained Battle array;
The probability matrix for the image block that (7a5) interception detection obtains:
Be arranged using num size for the probability matrix of λ × λ long λ as x-axis, width λ be y-axis, the rectangular co-ordinate that the lower left corner is origin System, intercepts each probability matrixRegion, obtain num interception after probability matrix set;
The probability matrix of (7a6) stitching image block:
According to image block foIt is in sizeThe relative position of image, will Probability matrix after num interception is spliced into the probability matrix that a size is (ω+λ) × (ξ+λ);
(7a7) obtains the probability matrix of detected remote sensing images:
Be arranged using size be (ω+λ) × (ξ+λ) probability matrix long ω+λ be x-axis, width ξ+λ is y-axis, the lower left corner as origin Rectangular coordinate system, intercept the region of (λ < x < λ+ω, the λ < y < λ+ξ) of probability matrix, obtain size as ω × ξ semanteme point Probability matrix matrix_e after cutting network detection.
CN201910399019.0A 2019-05-14 2019-05-14 Remote sensing image semantic segmentation method based on parallax information Active CN110111346B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910399019.0A CN110111346B (en) 2019-05-14 2019-05-14 Remote sensing image semantic segmentation method based on parallax information

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910399019.0A CN110111346B (en) 2019-05-14 2019-05-14 Remote sensing image semantic segmentation method based on parallax information

Publications (2)

Publication Number Publication Date
CN110111346A true CN110111346A (en) 2019-08-09
CN110111346B CN110111346B (en) 2020-12-08

Family

ID=67489883

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910399019.0A Active CN110111346B (en) 2019-05-14 2019-05-14 Remote sensing image semantic segmentation method based on parallax information

Country Status (1)

Country Link
CN (1) CN110111346B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111292340A (en) * 2020-01-23 2020-06-16 北京市商汤科技开发有限公司 Semantic segmentation method, device, equipment and computer readable storage medium
CN112288650A (en) * 2020-10-28 2021-01-29 武汉大学 Multi-source remote sensing satellite image geometric and semantic integrated processing method and system
CN115294489A (en) * 2022-06-22 2022-11-04 太原理工大学 Semantic segmentation method and system for disaster video data
CN116129525A (en) * 2023-01-24 2023-05-16 中国人民解放军陆军防化学院 Respiratory protection training evaluation system and method

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0997334A (en) * 1995-07-24 1997-04-08 Nec Corp Method and device for three-dimensional reference image segmentation and object judging device
CN103237228A (en) * 2013-04-28 2013-08-07 清华大学 Time-space consistency segmentation method for binocular stereoscopic video
CN104469386A (en) * 2014-12-15 2015-03-25 西安电子科技大学 Stereoscopic video perception and coding method for just-noticeable error model based on DOF
CN108038866A (en) * 2017-12-22 2018-05-15 湖南源信光电科技股份有限公司 A kind of moving target detecting method based on Vibe and disparity map Background difference

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0997334A (en) * 1995-07-24 1997-04-08 Nec Corp Method and device for three-dimensional reference image segmentation and object judging device
CN103237228A (en) * 2013-04-28 2013-08-07 清华大学 Time-space consistency segmentation method for binocular stereoscopic video
CN104469386A (en) * 2014-12-15 2015-03-25 西安电子科技大学 Stereoscopic video perception and coding method for just-noticeable error model based on DOF
CN108038866A (en) * 2017-12-22 2018-05-15 湖南源信光电科技股份有限公司 A kind of moving target detecting method based on Vibe and disparity map Background difference

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
YAN LIU,ET AL.: "Efficient Patch-Wise Semantic Segmentation for Large-Scale Remote Sensing Images", 《SENSORS》 *
吴止锾 等: "类别非均衡遥感图像语义分割的全卷积网络方法", 《光学学报》 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111292340A (en) * 2020-01-23 2020-06-16 北京市商汤科技开发有限公司 Semantic segmentation method, device, equipment and computer readable storage medium
CN112288650A (en) * 2020-10-28 2021-01-29 武汉大学 Multi-source remote sensing satellite image geometric and semantic integrated processing method and system
CN112288650B (en) * 2020-10-28 2021-07-20 武汉大学 Multi-source remote sensing satellite image geometric and semantic integrated processing method and system
CN115294489A (en) * 2022-06-22 2022-11-04 太原理工大学 Semantic segmentation method and system for disaster video data
CN115294489B (en) * 2022-06-22 2023-06-09 太原理工大学 Semantic segmentation method and system for disaster video data
CN116129525A (en) * 2023-01-24 2023-05-16 中国人民解放军陆军防化学院 Respiratory protection training evaluation system and method
CN116129525B (en) * 2023-01-24 2023-11-14 中国人民解放军陆军防化学院 Respiratory protection training evaluation system and method

Also Published As

Publication number Publication date
CN110111346B (en) 2020-12-08

Similar Documents

Publication Publication Date Title
CN110111346A (en) Remote sensing images semantic segmentation method based on parallax information
Jia et al. Detection and segmentation of overlapped fruits based on optimized mask R-CNN application in apple harvesting robot
CN108596101B (en) Remote sensing image multi-target detection method based on convolutional neural network
CN106023157B (en) A kind of micro- deformation data extracting method of mountain area earth&#39;s surface based on SAR image
CN109961006A (en) A kind of low pixel multiple target Face datection and crucial independent positioning method and alignment schemes
CN101326514B (en) Background removal in a live video
CN109584213B (en) Multi-target number selection tracking method
Bi et al. Rethinking camouflaged object detection: Models and datasets
Kirthika et al. Automated road network extraction using artificial neural network
CN110378997A (en) A kind of dynamic scene based on ORB-SLAM2 builds figure and localization method
CN110827312B (en) Learning method based on cooperative visual attention neural network
CN106897681A (en) A kind of remote sensing images comparative analysis method and system
CN109635634A (en) A kind of pedestrian based on stochastic linear interpolation identifies data enhancement methods again
CN108846416A (en) The extraction process method and system of specific image
Xu et al. Recognition of weeds in wheat fields based on the fusion of RGB images and depth images
CN104182968A (en) Method for segmenting fuzzy moving targets by wide-baseline multi-array optical detection system
CN106570885A (en) Background modeling method based on brightness and texture fusion threshold value
CN110472081A (en) A kind of cross-domain search method of shoes picture based on metric learning
CN108256567A (en) A kind of target identification method and system based on deep learning
CN111191531A (en) Rapid pedestrian detection method and system
CN104933691B (en) Image interfusion method based on the detection of phase spectrum vision significance
CN109785359A (en) A kind of video object detection method based on depth characteristic pyramid and tracking loss
CN107423771B (en) Two-time-phase remote sensing image change detection method
Tarchoun et al. Adversarial attacks in a multi-view setting: an empirical study of the adversarial patches inter-view transferability
CN107292268A (en) The SAR image semantic segmentation method of quick ridge ripple deconvolution Structure learning model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant