CN115131559A - Road scene semantic segmentation method based on multi-scale feature self-adaptive fusion - Google Patents
Road scene semantic segmentation method based on multi-scale feature self-adaptive fusion Download PDFInfo
- Publication number
- CN115131559A CN115131559A CN202210661080.XA CN202210661080A CN115131559A CN 115131559 A CN115131559 A CN 115131559A CN 202210661080 A CN202210661080 A CN 202210661080A CN 115131559 A CN115131559 A CN 115131559A
- Authority
- CN
- China
- Prior art keywords
- scale
- network
- feature
- segmentation
- fusion
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/42—Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/771—Feature selection, e.g. selecting representative features from a multi-dimensional feature space
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/806—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Evolutionary Computation (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- General Health & Medical Sciences (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Image Analysis (AREA)
Abstract
The invention relates to a road scene semantic segmentation method based on multi-scale feature adaptive fusion, which fully utilizes inherent multi-level characteristics of a convolution network to carry out multi-scale reasoning on each level of feature layer output by each sub-network in the network, and utilizes an attention mechanism to carry out adaptive fusion on prediction results under each scale. The invention carries out multi-scale reasoning by digging the inherent characteristics of the convolution network, avoids reducing the calculated amount by operations such as convolution decomposition and the like, thereby realizing high-precision real-time reasoning on a light-weight backbone network; since the low-level features of the backbone network contain detailed information and have a higher spatial resolution, reasoning on the low-level features can improve the segmentation accuracy for small-sized objects.
Description
Technical Field
The invention belongs to the field of image processing and an image semantic segmentation method, and relates to a road scene semantic segmentation method based on multi-scale feature self-adaptive fusion.
Background
Image semantic segmentation is one of basic tasks in the field of computer vision, classifies each pixel point in an image according to semantic information and spatial information thereof, and is a point-to-point classification mode. Semantic segmentation is a key step for realizing scene understanding, is also an important technical means for machine perception environment, and is widely applied to multiple fields of automatic driving, indoor navigation, emotion recognition, attitude detection and the like. The patent provides a semantic segmentation technology which can be applied to road scenes, and the semantic segmentation of the road scenes is an important link in an automatic driving technology. Meanwhile, the automatic driving system can respond to the change of the external environment in time, so that the real-time performance of the algorithm is high.
The document "real-time segmentation algorithm based on attention mechanism and efficient deconvolution, computer application, 2022, 1-1" discloses a road scene segmentation algorithm based on attention mechanism and deconvolution. The convolution layer is constructed by utilizing a one-dimensional Non-bottleneck structure (Non-bottle-1D) so as to reduce the operation amount of convolution operation; the pooling operation and attention module are then combined to capture global context information to improve the segmentation effect on larger sized objects. The method described in the literature uses decomposed convolution to reduce the amount of computation of the convolutional layer, which, although effective in improving speed, significantly reduces the segmentation accuracy; the method is focused on the global context information acquisition, and the segmentation of small-scale objects in the road scene is ignored.
Disclosure of Invention
Technical problem to be solved
In order to avoid the defects of the prior art, the invention provides a road scene semantic segmentation method based on multi-scale feature self-adaptive fusion, and solves the problems of high speed, low precision and poor segmentation effect on small-size objects of the conventional real-time semantic segmentation algorithm.
Technical scheme
A road scene semantic segmentation method based on multi-scale feature adaptive fusion is characterized by comprising the following steps:
step 1: extracting bottom multi-scale features of the input image by using Resnet-18 neural network to obtain a feature map S from shallow to deep under 4 scales 1 、S 2 、S 3 、S 4 ;
Step 2: constructing a backbone network based on bidirectional branch information fusion on the Resnet-18 neural network, and constructing a characteristic diagram S 1 、S 2 、S 3 、S 4 Fusing the left branch from bottom to top and the right branch from top to bottom to obtain a characteristic diagram D of each branch 1 、D 2 、D 3 、D 4 And U 1 、U 2 、U 3 、U 4 Finally, fusing the feature maps with the same scale in the two branches to obtain an output F based on the bidirectional branch information fusion backbone network 1 、F 2 、F 3 、F 4 ;
And step 3: pairing the output F in step 2 with a progressive upsampling module based on optical flow estimation 1 、F 2 、F 3 、F 4 Performing step-by-step upsampling;
calculating the distance between adjacent feature maps in the 4 scale feature maps, namely F, by an optical flow method 1 And F 2 、F 2 And F 3 、F 3 And F 4 The optical flow grid of (1). Then F is sampled by a step-by-step up-sampling module by utilizing an optical flow grid 2 、F 3 、F 4 Up-sampling to 1 The same size is obtained to obtain an aligned characteristic diagram P 1 、P 2 、P 3 、P 4 ;
And 4, step 4: the feature map P under each scale 1 、P 2 、P 3 、P 4 The input segmentation module obtains the segmentation result Scale under each Scale 1 、Scale 2 、Scale 3 、Scale 4 Feature map P at simultaneous scales 1 、P 2 、P 3 、P 4 The input attention module obtains the Weight of the segmentation result under each scale 1 、Weight 2 、Weight 3 、Weight 4 And finally, carrying out linear fusion on the segmentation results under all scales by using the weight to obtain a final segmentation result.
Characteristic diagram S in the step 1 1 、S 2 、S 3 、S 4 Are 1/4, 1/8, 1/16 and 1/32, respectively, of the original image.
Advantageous effects
The invention provides a road scene semantic segmentation method based on multi-scale feature adaptive fusion, which fully utilizes inherent multi-level characteristics of a convolutional network to carry out multi-scale reasoning on each level of feature layer output by each sub-network in the network, and utilizes an attention mechanism to carry out adaptive fusion on prediction results under each scale. The invention carries out multi-scale reasoning by digging the inherent characteristics of the convolution network, avoids reducing the calculated amount by operations such as convolution decomposition and the like, thereby realizing high-precision real-time reasoning on a light-weight backbone network; since the low-level features of the backbone network contain detailed information and have a higher spatial resolution, reasoning on the low-level features can improve the segmentation accuracy for small-sized objects.
The invention has the beneficial effects that:
(1) the method comprises the following steps of 1-2, fully utilizing inherent multi-scale characteristics of a convolutional network, obtaining multi-scale characteristics from sub-networks of different stages of the network, fusing the multi-scale characteristics for subsequent processing, and making up for the problem of insufficient learning capacity of a shallow lightweight backbone network, so that rapid and accurate segmentation is realized;
(2) and 3-4, performing multi-scale reasoning on the obtained fusion multi-scale features. The low-level features have detail and texture information and higher spatial resolution, and are suitable for segmentation of small-scale objects. And the high-level features contain semantic information and have larger receptive field, so that the method is suitable for segmenting large-scale objects. After the predicted values of all scales are adaptively fused, the final result can have higher segmentation precision on the object of each scale.
Drawings
Fig. 1 is a diagram showing an overall configuration of a network.
Fig. 2 is a method flow diagram.
Fig. 3 is a graph of test results under the citrescaps data set.
Detailed Description
The invention will now be further described with reference to the following examples and drawings:
the technical scheme adopted by the invention for solving the technical problems is as follows: a road scene semantic segmentation method based on multi-scale feature self-adaptive fusion is characterized by comprising the following steps:
and step 1, extracting bottom-layer multi-scale features by using a Resnet-18 neural network. The Resnet-18 network rapidly reduces the size of an input image through pooling operation at an initial network layer, and builds an integral network through fewer network layers, so that the network is faster and more efficient than other networks, and is widely used for building a real-time semantic segmentation network. The invention also builds a network on the basis of Resnet-18, and simplifies the network by removing all network layers (including a pooling layer and a full-link layer) after the last residual module, wherein the simplified ResNet-18 comprises 1 convolution layer of 7x7, 1 maximum pooling layer and 8 residual modules. According to the difference of the feature diagram size output by each module, 8 residual modules are divided into 4 sub-networks (stages 1-4 in fig. 1 (a)). The feature map is reduced in size 1/2 for each pass through a sub-network. Thus, a characteristic diagram S under 4 scales from shallow to deep can be obtained 1 、S 2 、S 3 、S 4 (size 1/4, 1/8, 1/16, and 1/32 of the original image, respectively).
And 2, building a backbone network based on bidirectional branch information fusion on the basis of the simplified Resnet-18 constructed in the step 1 (figure 1 (a)). Respectively by applying a characteristic map S 1 、S 2 、S 3 、S 4 The feature map D of each branch is obtained by fusing two branches from bottom to top (left branch in figure 1(a)) and from top to bottom (right branch in figure 1(a)) 1 、D 2 、D 3 、D 4 And U 1 、U 2 、U 3 、U 4 Finally, fusing the feature maps with the same scale in the two branches to obtain an output F based on the bidirectional branch information fusion backbone network 1 、F 2 、F 3 、F 4 。
Step 3, utilizing a progressive up-sampling module based on optical flow estimation to output F in the step 2 1 、F 2 、F 3 、F 4 And performing progressive up-sampling. Computing the distance between adjacent feature maps (i.e., F) in the 4 scale feature maps by an optical flow method 1 And F 2 、F 2 And F 3 、F 3 And F 4 ) The optical flow grid of (1). Then F is sampled by a progressive upsampling module by using an optical flow grid 2 、F 3 、F 4 Up-sampling to 1 Same size, get P 2 、P 3 、P 4 (due to F) 1 Without upsampling, P is thus obtained directly 1 )。
And 4, acquiring the segmentation result under each scale by using a segmentation module, and performing self-adaptive fusion on the segmentation result under each scale by using an attention mechanism. Feature map P at each scale 1 、P 2 、P 3 、P 4 On the basis, a segmentation module is utilized to obtain a segmentation result Scale under each Scale 1 、Scale 2 、Scale 3 、Scale 4 Then, the attention module is utilized to obtain the Weight of the segmentation result under each scale 1 、Weight 2 、Weight 3 、Weight 4 . And finally, performing linear fusion on the segmentation results under all scales by using the weight to obtain a final segmentation result.
The specific embodiment is as follows:
the following describes a specific embodiment of the present invention with reference to an example of identifying a semantically segmented data set citrespaces under city streets, but the technical content of the present invention is not limited to the scope described above.
The invention provides a road scene semantic segmentation method based on multi-scale feature self-adaptive fusion, which comprises the following steps of: step 1: obtaining feature maps under different scales by using the simplified ResNet-18 neural network; step 2: constructing a backbone network based on bidirectional branch information fusion, and performing efficient fusion among multi-scale feature maps; and step 3: up-sampling feature maps of different scales by utilizing a progressive up-sampling module based on optical flow estimation; and 4, step 4: constructing an attention mechanism-based adaptive multi-scale fusion module, and performing adaptive fusion on segmentation results under different scale characteristic graphs; and 5: and training the built neural network model by using the data set.
And step one, acquiring characteristic diagrams under different scales by using the simplified ResNet-18 neural network.
The Resnet-18 network model is suitable for building a real-time semantic segmentation model because the size of an input image is rapidly reduced through the pooling operation at the initial network layer and the number of network layers is small. The Resnet-18 network is simplified by removing all network layers after the last residual module, and the simplified ResNet-18 contains 1 convolution layer of 7x7, 1 max pooling layer, and 8 residual modules. According to the difference of the feature diagram size output by each module, 8 residual modules can be divided into 4 sub-networks (stages 1-4 in fig. 1 (a)). And then inputting picture samples in the citrespaces data set into a Resnet-18 network for feature extraction. The feature map size is reduced 1/2 for each pass through a sub-network. Thus, a characteristic diagram S under 4 scales from shallow to deep can be obtained 1 、S 2 、S 3 、S 4 。
And step two, fusing the feature graphs of 4 scales by using a backbone network based on bidirectional branch information fusion.
(1) Feature fusion is performed using the bottom-up fusion leg (left leg in fig. 1 (a)). Because the lower layer feature size is large, the lower layer feature layer needs to be downsampled and then merged with the upper layer feature. At each fusion, the feature map D is fused through the previous scale n-1 Feature map S of current scale after being subjected to bilinear interpolation downsampling n Adding the two images, and fusing by convolution of 3x3 to obtain a fused feature map D n The features under the n-th and n + 1-th scales are included, and a fused feature map D under a plurality of scales is obtained in sequence through recursive operation 1 、D 2 、D 3 、D 4 :
Wherein conv _3x3 represents a convolution operation with a convolution kernel size of 3x3 for feature fusion; conv _1x1 represents a convolution operation with a convolution kernel size of 1x1, used for adjustment of the feature map depth dimension; down represents a bilinear interpolation downsampling operation.
(2) Feature fusion is performed using the top-to-bottom fusion branch (right branch in fig. 1 (a)). Because the upper layer feature size is smaller, the upper layer feature graph needs to be fused with the lower layer feature graph after upsampling, other operations are the same in fused branches from bottom to top, and the fused feature graph U is obtained in sequence through recursive operation 4 、U 3 、U 2 、U 1 :
Where up represents a bilinear upsampling operation.
(3) Fusing the characteristic graphs of the through scales in the two branches to obtain an output F based on the bidirectional branch information fusion backbone network 1 、F 2 、F 3 、F 4 :
F n =conv_3x3(D n +U n ),n=1,2,3,4
Step three, utilizing a step-by-step up-sampling module based on optical flow estimation to output F in the step two 2 、F 3 、F 4 Up-sampling to 1 The same size to align the spatial positions between the feature maps of each scale.
(1) Computing the distance between adjacent feature maps (i.e., F) in the 4 scale feature maps by using an optical flow method 1 And F 2 、F 2 And F 3 、F 3 And F 4 ) Spatial warped mesh of cells:
firstly, the optical flow field delta is calculated by utilizing the characteristic graphs of two adjacent scales n-1 :
Δ n-1 =conv_3x3(concat(F n-1 ,up(F n )))
Wherein concat represents splicing the feature graph along the depth dimension; delta n-1 Is represented by the formula n-1 The optical flow field is the same size and the depth is 2.
Then, using g n-1 Is represented by F n-1 Of the space grid omega n-1 Position point of (1), then F n Relative to F n-1 Is spatially distorted grid omega n Offset in (1) n Can be expressed as:
wherein, when n is 2, 3, 4, the sub-table represents F 1 And F 2 、F 2 And F 3 、F 3 And F 4 The space between them distorts the grid.
(2) Up-sampling step by using space distortion network to convert F into 2 、F 3 、F 4 Up-sampled to and F, respectively 1 The same size is obtained to obtain an up-sampled characteristic diagram P 2 、P 3 、P 4 (due to F) 1 Without upsampling, P is thus obtained directly 1 ). The recursive process can be expressed as:
up l indicating the use of offset in warped mesh l And performing bilinear interpolation upsampling on the feature map F. This process can be expressed as:
wherein, N (omega) l ) Representing a spatially warped mesh omega n 4 fields (upper left, lower left, upper right, lower right); w is a i To representUsing omega n Offset in (1) l The weight of the bilinear kernel calculated.
Step four, characteristic diagram P under each scale 1 、P 2 、P 3 、P 4 On the basis, a segmentation module is used for obtaining segmentation results under each scale, and an attention mechanism is used for carrying out self-adaptive fusion on the segmentation results under each scale.
(1) Obtaining the Scale of the segmentation result under each Scale by utilizing the segmentation module 1 、Scale 2 、Scale 3 、Scale 4 Then, the attention module is utilized to obtain the Weight of the segmentation result under each scale 1 、Weight 2 、Weight 3 、Weight 4 The segmentation module and attention module both use a convolution kernel of 3x3 to construct:
Scale n =conv_3x3(P n ),n=1,2,3,4
Weight n =conv_3x3(P n ),n=1,2,3,4
(2) and performing linear fusion on the segmentation results under all scales by using the weight to obtain a final segmentation result:
wherein Output represents a final division structure; an even line indicates a matrix dot product.
And finally, performing linear fusion on the segmentation results under all scales by using the weight to obtain a final segmentation result.
And fifthly, training the built neural network model by using the data set.
Claims (2)
1. A road scene semantic segmentation method based on multi-scale feature adaptive fusion is characterized by comprising the following steps:
step 1: extracting bottom multi-scale features of the input image by using Resnet-18 neural network to obtain a feature map S from shallow to deep under 4 scales 1 、S 2 、S 3 、S 4 ;
And 2, step: constructing a backbone network based on bidirectional branch information fusion on the Resnet-18 neural network, and constructing a characteristic diagram S 1 、S 2 、S 3 、S 4 Fusing the left branch from bottom to top and the right branch from top to bottom to obtain a characteristic diagram D of each branch 1 、D 2 、D 3 、D 4 And U 1 、U 2 、U 3 、U 4 Finally, fusing the feature maps with the same scale in the two branches to obtain an output F based on the bidirectional branch information fusion backbone network 1 、F 2 、F 3 、F 4 ;
And step 3: pairing the output F in step 2 with a progressive upsampling module based on optical flow estimation 1 、F 2 、F 3 、F 4 Performing progressive up-sampling;
calculating the distance between adjacent feature maps in the 4 scale feature maps, namely F, by an optical flow method 1 And F 2 、F 2 And F 3 、F 3 And F 4 The optical flow grid of (2). Then F is sampled by a step-by-step up-sampling module by utilizing an optical flow grid 2 、F 3 、F 4 Up-sampling to 1 The same size is obtained to obtain an aligned characteristic diagram P 1 、P 2 、P 3 、P 4 ;
And 4, step 4: the feature map P under each scale 1 、P 2 、P 3 、P 4 The input segmentation module obtains a segmentation result Scale under each Scale 1 、Scale 2 、Scale 3 、Scale 4 Feature map P at simultaneous scales 1 、P 2 、P 3 、P 4 The input attention module obtains the Weight of the segmentation result under each scale 1 、Weight 2 、Weight 3 、Weight 4 And finally, carrying out linear fusion on the segmentation results under all scales by using the weight to obtain a final segmentation result.
2. The road scene semantic segmentation method based on the multi-scale feature adaptive fusion according to claim 1,the method is characterized in that: characteristic diagram S in the step 1 1 、S 2 、S 3 、S 4 Are 1/4, 1/8, 1/16 and 1/32, respectively, of the original image.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210661080.XA CN115131559A (en) | 2022-06-12 | 2022-06-12 | Road scene semantic segmentation method based on multi-scale feature self-adaptive fusion |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210661080.XA CN115131559A (en) | 2022-06-12 | 2022-06-12 | Road scene semantic segmentation method based on multi-scale feature self-adaptive fusion |
Publications (1)
Publication Number | Publication Date |
---|---|
CN115131559A true CN115131559A (en) | 2022-09-30 |
Family
ID=83377278
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210661080.XA Pending CN115131559A (en) | 2022-06-12 | 2022-06-12 | Road scene semantic segmentation method based on multi-scale feature self-adaptive fusion |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115131559A (en) |
-
2022
- 2022-06-12 CN CN202210661080.XA patent/CN115131559A/en active Pending
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110287849B (en) | Lightweight depth network image target detection method suitable for raspberry pi | |
CN110232394B (en) | Multi-scale image semantic segmentation method | |
CN110728200B (en) | Real-time pedestrian detection method and system based on deep learning | |
CN113033570B (en) | Image semantic segmentation method for improving void convolution and multilevel characteristic information fusion | |
CN114255238A (en) | Three-dimensional point cloud scene segmentation method and system fusing image features | |
CN113657388A (en) | Image semantic segmentation method fusing image super-resolution reconstruction | |
CN114048822A (en) | Attention mechanism feature fusion segmentation method for image | |
CN115082675B (en) | Transparent object image segmentation method and system | |
CN113888547A (en) | Non-supervision domain self-adaptive remote sensing road semantic segmentation method based on GAN network | |
CN113554032B (en) | Remote sensing image segmentation method based on multi-path parallel network of high perception | |
CN114724155A (en) | Scene text detection method, system and equipment based on deep convolutional neural network | |
CN116797787B (en) | Remote sensing image semantic segmentation method based on cross-modal fusion and graph neural network | |
CN115631344B (en) | Target detection method based on feature self-adaptive aggregation | |
CN116612288B (en) | Multi-scale lightweight real-time semantic segmentation method and system | |
CN114019467A (en) | Radar signal identification and positioning method based on MobileNet model transfer learning | |
CN115424017B (en) | Building inner and outer contour segmentation method, device and storage medium | |
CN115775316A (en) | Image semantic segmentation method based on multi-scale attention mechanism | |
CN113313176A (en) | Point cloud analysis method based on dynamic graph convolution neural network | |
CN115482518A (en) | Extensible multitask visual perception method for traffic scene | |
CN116863194A (en) | Foot ulcer image classification method, system, equipment and medium | |
CN115049945A (en) | Method and device for extracting lodging area of wheat based on unmanned aerial vehicle image | |
CN114445620A (en) | Target segmentation method for improving Mask R-CNN | |
CN113436198A (en) | Remote sensing image semantic segmentation method for collaborative image super-resolution reconstruction | |
CN113160291A (en) | Change detection method based on image registration | |
CN115131559A (en) | Road scene semantic segmentation method based on multi-scale feature self-adaptive fusion |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
CB03 | Change of inventor or designer information |
Inventor after: Zhang Ke Inventor after: Zhang Yanhua Inventor after: Su Yu Inventor after: Wang Jingyu Inventor before: Zhang Ke Inventor before: Yan Hua Inventor before: Su Yu Inventor before: Wang Jingyu |
|
CB03 | Change of inventor or designer information |