CN115131559A - Road scene semantic segmentation method based on multi-scale feature self-adaptive fusion - Google Patents

Road scene semantic segmentation method based on multi-scale feature self-adaptive fusion Download PDF

Info

Publication number
CN115131559A
CN115131559A CN202210661080.XA CN202210661080A CN115131559A CN 115131559 A CN115131559 A CN 115131559A CN 202210661080 A CN202210661080 A CN 202210661080A CN 115131559 A CN115131559 A CN 115131559A
Authority
CN
China
Prior art keywords
scale
network
feature
segmentation
fusion
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210661080.XA
Other languages
Chinese (zh)
Inventor
张科
彦华
苏雨
王靖宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Northwestern Polytechnical University
Original Assignee
Northwestern Polytechnical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Northwestern Polytechnical University filed Critical Northwestern Polytechnical University
Priority to CN202210661080.XA priority Critical patent/CN115131559A/en
Publication of CN115131559A publication Critical patent/CN115131559A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/42Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/771Feature selection, e.g. selecting representative features from a multi-dimensional feature space
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a road scene semantic segmentation method based on multi-scale feature adaptive fusion, which fully utilizes inherent multi-level characteristics of a convolution network to carry out multi-scale reasoning on each level of feature layer output by each sub-network in the network, and utilizes an attention mechanism to carry out adaptive fusion on prediction results under each scale. The invention carries out multi-scale reasoning by digging the inherent characteristics of the convolution network, avoids reducing the calculated amount by operations such as convolution decomposition and the like, thereby realizing high-precision real-time reasoning on a light-weight backbone network; since the low-level features of the backbone network contain detailed information and have a higher spatial resolution, reasoning on the low-level features can improve the segmentation accuracy for small-sized objects.

Description

Road scene semantic segmentation method based on multi-scale feature self-adaptive fusion
Technical Field
The invention belongs to the field of image processing and an image semantic segmentation method, and relates to a road scene semantic segmentation method based on multi-scale feature self-adaptive fusion.
Background
Image semantic segmentation is one of basic tasks in the field of computer vision, classifies each pixel point in an image according to semantic information and spatial information thereof, and is a point-to-point classification mode. Semantic segmentation is a key step for realizing scene understanding, is also an important technical means for machine perception environment, and is widely applied to multiple fields of automatic driving, indoor navigation, emotion recognition, attitude detection and the like. The patent provides a semantic segmentation technology which can be applied to road scenes, and the semantic segmentation of the road scenes is an important link in an automatic driving technology. Meanwhile, the automatic driving system can respond to the change of the external environment in time, so that the real-time performance of the algorithm is high.
The document "real-time segmentation algorithm based on attention mechanism and efficient deconvolution, computer application, 2022, 1-1" discloses a road scene segmentation algorithm based on attention mechanism and deconvolution. The convolution layer is constructed by utilizing a one-dimensional Non-bottleneck structure (Non-bottle-1D) so as to reduce the operation amount of convolution operation; the pooling operation and attention module are then combined to capture global context information to improve the segmentation effect on larger sized objects. The method described in the literature uses decomposed convolution to reduce the amount of computation of the convolutional layer, which, although effective in improving speed, significantly reduces the segmentation accuracy; the method is focused on the global context information acquisition, and the segmentation of small-scale objects in the road scene is ignored.
Disclosure of Invention
Technical problem to be solved
In order to avoid the defects of the prior art, the invention provides a road scene semantic segmentation method based on multi-scale feature self-adaptive fusion, and solves the problems of high speed, low precision and poor segmentation effect on small-size objects of the conventional real-time semantic segmentation algorithm.
Technical scheme
A road scene semantic segmentation method based on multi-scale feature adaptive fusion is characterized by comprising the following steps:
step 1: extracting bottom multi-scale features of the input image by using Resnet-18 neural network to obtain a feature map S from shallow to deep under 4 scales 1 、S 2 、S 3 、S 4
Step 2: constructing a backbone network based on bidirectional branch information fusion on the Resnet-18 neural network, and constructing a characteristic diagram S 1 、S 2 、S 3 、S 4 Fusing the left branch from bottom to top and the right branch from top to bottom to obtain a characteristic diagram D of each branch 1 、D 2 、D 3 、D 4 And U 1 、U 2 、U 3 、U 4 Finally, fusing the feature maps with the same scale in the two branches to obtain an output F based on the bidirectional branch information fusion backbone network 1 、F 2 、F 3 、F 4
And step 3: pairing the output F in step 2 with a progressive upsampling module based on optical flow estimation 1 、F 2 、F 3 、F 4 Performing step-by-step upsampling;
calculating the distance between adjacent feature maps in the 4 scale feature maps, namely F, by an optical flow method 1 And F 2 、F 2 And F 3 、F 3 And F 4 The optical flow grid of (1). Then F is sampled by a step-by-step up-sampling module by utilizing an optical flow grid 2 、F 3 、F 4 Up-sampling to 1 The same size is obtained to obtain an aligned characteristic diagram P 1 、P 2 、P 3 、P 4
And 4, step 4: the feature map P under each scale 1 、P 2 、P 3 、P 4 The input segmentation module obtains the segmentation result Scale under each Scale 1 、Scale 2 、Scale 3 、Scale 4 Feature map P at simultaneous scales 1 、P 2 、P 3 、P 4 The input attention module obtains the Weight of the segmentation result under each scale 1 、Weight 2 、Weight 3 、Weight 4 And finally, carrying out linear fusion on the segmentation results under all scales by using the weight to obtain a final segmentation result.
Characteristic diagram S in the step 1 1 、S 2 、S 3 、S 4 Are 1/4, 1/8, 1/16 and 1/32, respectively, of the original image.
Advantageous effects
The invention provides a road scene semantic segmentation method based on multi-scale feature adaptive fusion, which fully utilizes inherent multi-level characteristics of a convolutional network to carry out multi-scale reasoning on each level of feature layer output by each sub-network in the network, and utilizes an attention mechanism to carry out adaptive fusion on prediction results under each scale. The invention carries out multi-scale reasoning by digging the inherent characteristics of the convolution network, avoids reducing the calculated amount by operations such as convolution decomposition and the like, thereby realizing high-precision real-time reasoning on a light-weight backbone network; since the low-level features of the backbone network contain detailed information and have a higher spatial resolution, reasoning on the low-level features can improve the segmentation accuracy for small-sized objects.
The invention has the beneficial effects that:
(1) the method comprises the following steps of 1-2, fully utilizing inherent multi-scale characteristics of a convolutional network, obtaining multi-scale characteristics from sub-networks of different stages of the network, fusing the multi-scale characteristics for subsequent processing, and making up for the problem of insufficient learning capacity of a shallow lightweight backbone network, so that rapid and accurate segmentation is realized;
(2) and 3-4, performing multi-scale reasoning on the obtained fusion multi-scale features. The low-level features have detail and texture information and higher spatial resolution, and are suitable for segmentation of small-scale objects. And the high-level features contain semantic information and have larger receptive field, so that the method is suitable for segmenting large-scale objects. After the predicted values of all scales are adaptively fused, the final result can have higher segmentation precision on the object of each scale.
Drawings
Fig. 1 is a diagram showing an overall configuration of a network.
Fig. 2 is a method flow diagram.
Fig. 3 is a graph of test results under the citrescaps data set.
Detailed Description
The invention will now be further described with reference to the following examples and drawings:
the technical scheme adopted by the invention for solving the technical problems is as follows: a road scene semantic segmentation method based on multi-scale feature self-adaptive fusion is characterized by comprising the following steps:
and step 1, extracting bottom-layer multi-scale features by using a Resnet-18 neural network. The Resnet-18 network rapidly reduces the size of an input image through pooling operation at an initial network layer, and builds an integral network through fewer network layers, so that the network is faster and more efficient than other networks, and is widely used for building a real-time semantic segmentation network. The invention also builds a network on the basis of Resnet-18, and simplifies the network by removing all network layers (including a pooling layer and a full-link layer) after the last residual module, wherein the simplified ResNet-18 comprises 1 convolution layer of 7x7, 1 maximum pooling layer and 8 residual modules. According to the difference of the feature diagram size output by each module, 8 residual modules are divided into 4 sub-networks (stages 1-4 in fig. 1 (a)). The feature map is reduced in size 1/2 for each pass through a sub-network. Thus, a characteristic diagram S under 4 scales from shallow to deep can be obtained 1 、S 2 、S 3 、S 4 (size 1/4, 1/8, 1/16, and 1/32 of the original image, respectively).
And 2, building a backbone network based on bidirectional branch information fusion on the basis of the simplified Resnet-18 constructed in the step 1 (figure 1 (a)). Respectively by applying a characteristic map S 1 、S 2 、S 3 、S 4 The feature map D of each branch is obtained by fusing two branches from bottom to top (left branch in figure 1(a)) and from top to bottom (right branch in figure 1(a)) 1 、D 2 、D 3 、D 4 And U 1 、U 2 、U 3 、U 4 Finally, fusing the feature maps with the same scale in the two branches to obtain an output F based on the bidirectional branch information fusion backbone network 1 、F 2 、F 3 、F 4
Step 3, utilizing a progressive up-sampling module based on optical flow estimation to output F in the step 2 1 、F 2 、F 3 、F 4 And performing progressive up-sampling. Computing the distance between adjacent feature maps (i.e., F) in the 4 scale feature maps by an optical flow method 1 And F 2 、F 2 And F 3 、F 3 And F 4 ) The optical flow grid of (1). Then F is sampled by a progressive upsampling module by using an optical flow grid 2 、F 3 、F 4 Up-sampling to 1 Same size, get P 2 、P 3 、P 4 (due to F) 1 Without upsampling, P is thus obtained directly 1 )。
And 4, acquiring the segmentation result under each scale by using a segmentation module, and performing self-adaptive fusion on the segmentation result under each scale by using an attention mechanism. Feature map P at each scale 1 、P 2 、P 3 、P 4 On the basis, a segmentation module is utilized to obtain a segmentation result Scale under each Scale 1 、Scale 2 、Scale 3 、Scale 4 Then, the attention module is utilized to obtain the Weight of the segmentation result under each scale 1 、Weight 2 、Weight 3 、Weight 4 . And finally, performing linear fusion on the segmentation results under all scales by using the weight to obtain a final segmentation result.
The specific embodiment is as follows:
the following describes a specific embodiment of the present invention with reference to an example of identifying a semantically segmented data set citrespaces under city streets, but the technical content of the present invention is not limited to the scope described above.
The invention provides a road scene semantic segmentation method based on multi-scale feature self-adaptive fusion, which comprises the following steps of: step 1: obtaining feature maps under different scales by using the simplified ResNet-18 neural network; step 2: constructing a backbone network based on bidirectional branch information fusion, and performing efficient fusion among multi-scale feature maps; and step 3: up-sampling feature maps of different scales by utilizing a progressive up-sampling module based on optical flow estimation; and 4, step 4: constructing an attention mechanism-based adaptive multi-scale fusion module, and performing adaptive fusion on segmentation results under different scale characteristic graphs; and 5: and training the built neural network model by using the data set.
And step one, acquiring characteristic diagrams under different scales by using the simplified ResNet-18 neural network.
The Resnet-18 network model is suitable for building a real-time semantic segmentation model because the size of an input image is rapidly reduced through the pooling operation at the initial network layer and the number of network layers is small. The Resnet-18 network is simplified by removing all network layers after the last residual module, and the simplified ResNet-18 contains 1 convolution layer of 7x7, 1 max pooling layer, and 8 residual modules. According to the difference of the feature diagram size output by each module, 8 residual modules can be divided into 4 sub-networks (stages 1-4 in fig. 1 (a)). And then inputting picture samples in the citrespaces data set into a Resnet-18 network for feature extraction. The feature map size is reduced 1/2 for each pass through a sub-network. Thus, a characteristic diagram S under 4 scales from shallow to deep can be obtained 1 、S 2 、S 3 、S 4
And step two, fusing the feature graphs of 4 scales by using a backbone network based on bidirectional branch information fusion.
(1) Feature fusion is performed using the bottom-up fusion leg (left leg in fig. 1 (a)). Because the lower layer feature size is large, the lower layer feature layer needs to be downsampled and then merged with the upper layer feature. At each fusion, the feature map D is fused through the previous scale n-1 Feature map S of current scale after being subjected to bilinear interpolation downsampling n Adding the two images, and fusing by convolution of 3x3 to obtain a fused feature map D n The features under the n-th and n + 1-th scales are included, and a fused feature map D under a plurality of scales is obtained in sequence through recursive operation 1 、D 2 、D 3 、D 4
Figure BDA0003689760050000051
Wherein conv _3x3 represents a convolution operation with a convolution kernel size of 3x3 for feature fusion; conv _1x1 represents a convolution operation with a convolution kernel size of 1x1, used for adjustment of the feature map depth dimension; down represents a bilinear interpolation downsampling operation.
(2) Feature fusion is performed using the top-to-bottom fusion branch (right branch in fig. 1 (a)). Because the upper layer feature size is smaller, the upper layer feature graph needs to be fused with the lower layer feature graph after upsampling, other operations are the same in fused branches from bottom to top, and the fused feature graph U is obtained in sequence through recursive operation 4 、U 3 、U 2 、U 1
Figure BDA0003689760050000061
Where up represents a bilinear upsampling operation.
(3) Fusing the characteristic graphs of the through scales in the two branches to obtain an output F based on the bidirectional branch information fusion backbone network 1 、F 2 、F 3 、F 4
F n =conv_3x3(D n +U n ),n=1,2,3,4
Step three, utilizing a step-by-step up-sampling module based on optical flow estimation to output F in the step two 2 、F 3 、F 4 Up-sampling to 1 The same size to align the spatial positions between the feature maps of each scale.
(1) Computing the distance between adjacent feature maps (i.e., F) in the 4 scale feature maps by using an optical flow method 1 And F 2 、F 2 And F 3 、F 3 And F 4 ) Spatial warped mesh of cells:
firstly, the optical flow field delta is calculated by utilizing the characteristic graphs of two adjacent scales n-1
Δ n-1 =conv_3x3(concat(F n-1 ,up(F n )))
Wherein concat represents splicing the feature graph along the depth dimension; delta n-1 Is represented by the formula n-1 The optical flow field is the same size and the depth is 2.
Then, using g n-1 Is represented by F n-1 Of the space grid omega n-1 Position point of (1), then F n Relative to F n-1 Is spatially distorted grid omega n Offset in (1) n Can be expressed as:
Figure BDA0003689760050000062
wherein, when n is 2, 3, 4, the sub-table represents F 1 And F 2 、F 2 And F 3 、F 3 And F 4 The space between them distorts the grid.
(2) Up-sampling step by using space distortion network to convert F into 2 、F 3 、F 4 Up-sampled to and F, respectively 1 The same size is obtained to obtain an up-sampled characteristic diagram P 2 、P 3 、P 4 (due to F) 1 Without upsampling, P is thus obtained directly 1 ). The recursive process can be expressed as:
Figure BDA0003689760050000071
up l indicating the use of offset in warped mesh l And performing bilinear interpolation upsampling on the feature map F. This process can be expressed as:
Figure BDA0003689760050000072
wherein, N (omega) l ) Representing a spatially warped mesh omega n 4 fields (upper left, lower left, upper right, lower right); w is a i To representUsing omega n Offset in (1) l The weight of the bilinear kernel calculated.
Step four, characteristic diagram P under each scale 1 、P 2 、P 3 、P 4 On the basis, a segmentation module is used for obtaining segmentation results under each scale, and an attention mechanism is used for carrying out self-adaptive fusion on the segmentation results under each scale.
(1) Obtaining the Scale of the segmentation result under each Scale by utilizing the segmentation module 1 、Scale 2 、Scale 3 、Scale 4 Then, the attention module is utilized to obtain the Weight of the segmentation result under each scale 1 、Weight 2 、Weight 3 、Weight 4 The segmentation module and attention module both use a convolution kernel of 3x3 to construct:
Scale n =conv_3x3(P n ),n=1,2,3,4
Weight n =conv_3x3(P n ),n=1,2,3,4
(2) and performing linear fusion on the segmentation results under all scales by using the weight to obtain a final segmentation result:
Figure BDA0003689760050000073
wherein Output represents a final division structure; an even line indicates a matrix dot product.
And finally, performing linear fusion on the segmentation results under all scales by using the weight to obtain a final segmentation result.
And fifthly, training the built neural network model by using the data set.

Claims (2)

1. A road scene semantic segmentation method based on multi-scale feature adaptive fusion is characterized by comprising the following steps:
step 1: extracting bottom multi-scale features of the input image by using Resnet-18 neural network to obtain a feature map S from shallow to deep under 4 scales 1 、S 2 、S 3 、S 4
And 2, step: constructing a backbone network based on bidirectional branch information fusion on the Resnet-18 neural network, and constructing a characteristic diagram S 1 、S 2 、S 3 、S 4 Fusing the left branch from bottom to top and the right branch from top to bottom to obtain a characteristic diagram D of each branch 1 、D 2 、D 3 、D 4 And U 1 、U 2 、U 3 、U 4 Finally, fusing the feature maps with the same scale in the two branches to obtain an output F based on the bidirectional branch information fusion backbone network 1 、F 2 、F 3 、F 4
And step 3: pairing the output F in step 2 with a progressive upsampling module based on optical flow estimation 1 、F 2 、F 3 、F 4 Performing progressive up-sampling;
calculating the distance between adjacent feature maps in the 4 scale feature maps, namely F, by an optical flow method 1 And F 2 、F 2 And F 3 、F 3 And F 4 The optical flow grid of (2). Then F is sampled by a step-by-step up-sampling module by utilizing an optical flow grid 2 、F 3 、F 4 Up-sampling to 1 The same size is obtained to obtain an aligned characteristic diagram P 1 、P 2 、P 3 、P 4
And 4, step 4: the feature map P under each scale 1 、P 2 、P 3 、P 4 The input segmentation module obtains a segmentation result Scale under each Scale 1 、Scale 2 、Scale 3 、Scale 4 Feature map P at simultaneous scales 1 、P 2 、P 3 、P 4 The input attention module obtains the Weight of the segmentation result under each scale 1 、Weight 2 、Weight 3 、Weight 4 And finally, carrying out linear fusion on the segmentation results under all scales by using the weight to obtain a final segmentation result.
2. The road scene semantic segmentation method based on the multi-scale feature adaptive fusion according to claim 1,the method is characterized in that: characteristic diagram S in the step 1 1 、S 2 、S 3 、S 4 Are 1/4, 1/8, 1/16 and 1/32, respectively, of the original image.
CN202210661080.XA 2022-06-12 2022-06-12 Road scene semantic segmentation method based on multi-scale feature self-adaptive fusion Pending CN115131559A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210661080.XA CN115131559A (en) 2022-06-12 2022-06-12 Road scene semantic segmentation method based on multi-scale feature self-adaptive fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210661080.XA CN115131559A (en) 2022-06-12 2022-06-12 Road scene semantic segmentation method based on multi-scale feature self-adaptive fusion

Publications (1)

Publication Number Publication Date
CN115131559A true CN115131559A (en) 2022-09-30

Family

ID=83377278

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210661080.XA Pending CN115131559A (en) 2022-06-12 2022-06-12 Road scene semantic segmentation method based on multi-scale feature self-adaptive fusion

Country Status (1)

Country Link
CN (1) CN115131559A (en)

Similar Documents

Publication Publication Date Title
CN110287849B (en) Lightweight depth network image target detection method suitable for raspberry pi
CN110232394B (en) Multi-scale image semantic segmentation method
CN110728200B (en) Real-time pedestrian detection method and system based on deep learning
CN113033570B (en) Image semantic segmentation method for improving void convolution and multilevel characteristic information fusion
CN114255238A (en) Three-dimensional point cloud scene segmentation method and system fusing image features
CN113657388A (en) Image semantic segmentation method fusing image super-resolution reconstruction
CN114048822A (en) Attention mechanism feature fusion segmentation method for image
CN115082675B (en) Transparent object image segmentation method and system
CN113888547A (en) Non-supervision domain self-adaptive remote sensing road semantic segmentation method based on GAN network
CN113554032B (en) Remote sensing image segmentation method based on multi-path parallel network of high perception
CN114724155A (en) Scene text detection method, system and equipment based on deep convolutional neural network
CN116797787B (en) Remote sensing image semantic segmentation method based on cross-modal fusion and graph neural network
CN115631344B (en) Target detection method based on feature self-adaptive aggregation
CN116612288B (en) Multi-scale lightweight real-time semantic segmentation method and system
CN114019467A (en) Radar signal identification and positioning method based on MobileNet model transfer learning
CN115424017B (en) Building inner and outer contour segmentation method, device and storage medium
CN115775316A (en) Image semantic segmentation method based on multi-scale attention mechanism
CN113313176A (en) Point cloud analysis method based on dynamic graph convolution neural network
CN115482518A (en) Extensible multitask visual perception method for traffic scene
CN116863194A (en) Foot ulcer image classification method, system, equipment and medium
CN115049945A (en) Method and device for extracting lodging area of wheat based on unmanned aerial vehicle image
CN114445620A (en) Target segmentation method for improving Mask R-CNN
CN113436198A (en) Remote sensing image semantic segmentation method for collaborative image super-resolution reconstruction
CN113160291A (en) Change detection method based on image registration
CN115131559A (en) Road scene semantic segmentation method based on multi-scale feature self-adaptive fusion

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB03 Change of inventor or designer information

Inventor after: Zhang Ke

Inventor after: Zhang Yanhua

Inventor after: Su Yu

Inventor after: Wang Jingyu

Inventor before: Zhang Ke

Inventor before: Yan Hua

Inventor before: Su Yu

Inventor before: Wang Jingyu

CB03 Change of inventor or designer information