CN114529878B - Cross-domain road scene semantic segmentation method based on semantic perception - Google Patents

Cross-domain road scene semantic segmentation method based on semantic perception Download PDF

Info

Publication number
CN114529878B
CN114529878B CN202210083793.2A CN202210083793A CN114529878B CN 114529878 B CN114529878 B CN 114529878B CN 202210083793 A CN202210083793 A CN 202210083793A CN 114529878 B CN114529878 B CN 114529878B
Authority
CN
China
Prior art keywords
semantic
feature
image
features
perception
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210083793.2A
Other languages
Chinese (zh)
Other versions
CN114529878A (en
Inventor
雷印杰
彭铎
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan University
Original Assignee
Sichuan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan University filed Critical Sichuan University
Priority to CN202210083793.2A priority Critical patent/CN114529878B/en
Publication of CN114529878A publication Critical patent/CN114529878A/en
Application granted granted Critical
Publication of CN114529878B publication Critical patent/CN114529878B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computational Linguistics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to the technical field of semantic segmentation, and discloses a semantic-perception-based cross-domain road scene semantic segmentation method, which comprises the following steps: s1, sending source domain images into a feature encoder according to batches to obtain features of a plurality of images in the batches; s2, sending the image features in the batch to a central alignment module based on semantic perception, and realizing global central alignment of source domain images with different styles in a feature space; and S3, sending the characteristics with the aligned centers into a distribution alignment module based on semantic perception, and further realizing local distribution alignment on the image characteristics of various styles in the batch. According to the semantic segmentation method for the cross-domain road scene based on semantic perception, the labeling cost is reduced, the source domain image which is the most basic of the training network can be directly obtained from the game, the corresponding label is directly generated through the engine, manual labeling is not needed, and a large amount of manpower and material resources are saved.

Description

Cross-domain road scene semantic segmentation method based on semantic perception
Technical Field
The invention relates to the technical field of semantic segmentation, in particular to a semantic-perception-based cross-domain road scene semantic segmentation method.
Background
The image semantic segmentation refers to that a computer realizes deep understanding of an image according to semantic content of the image and then performs pixel-level visual segmentation. In recent years, with the continuous development of artificial intelligence, semantic segmentation technology based on a deep neural network is increasingly applied to various aspects such as industrial production, social security, transportation and the like, wherein the realization of unmanned semantic segmentation is a popular research field, and has a good development prospect. The semantic segmentation is a core algorithm of unmanned vehicle driving, when the vehicle-mounted camera probes the image, the image is input into the neural network, and the background operation equipment can automatically segment and classify the image so as to realize obstacle avoidance of pedestrians and vehicles.
In recent years, the performance of deep learning semantic segmentation has tended to be perfect, but researchers have found that models that perform well on training datasets are not ideal when applied in other scenarios because the training image (source domain) and the applied image (target domain) exist in two different domains, with inconsistent data distributions. In this regard, researchers have proposed a large number of artificial intelligence methods based on deep learning to cope with the problem of efficiency attenuation of semantic segmentation after cross-domain, and generally these methods need to acquire images of target domains in advance to further adapt the source domain distribution to the target domain distribution, so as to improve the segmentation robustness in the target domain, but there are two problems: first, most of the current methods need to acquire a part of image data of a new target area in advance to adapt to the new target area, which is clearly very labor-and material-consuming, and for example, an operator cannot acquire road images of all target areas in advance. Secondly, these methods aim at performing cross-domain segmentation on a certain known target domain, and a fixed model can only be applied to the specific domain, cannot be generalized to other domains, and cannot meet the needs of practical applications, so that a method which does not depend on target domain data and has considerable generalization performance is very required to be provided. In order to solve the problems, the invention processes data in a feature space, aligns the data from a center and a distribution layer, and respectively aligns the features of different categories in consideration of the distribution difference of the features of different categories, thereby realizing domain-invariant feature transformation with high discrimination and enhancing the multi-domain generalization performance of the model.
Disclosure of Invention
(one) solving the technical problems
Aiming at the defects of the prior art, the invention provides a semantic perception-based cross-domain road scene semantic segmentation method, which has the advantages of realizing high discrimination and invariable feature conversion by training a model only through source domain data without depending on any target domain data and data enhancement method, ensuring the model to have considerable multi-domain generalization performance and the like, and solving the problems in the background technology.
(II) technical scheme
The invention provides the following technical scheme: a semantic-aware-based cross-domain road scene semantic segmentation method, comprising the steps of:
s1, sending the source domain images into a feature encoder according to batches to obtain features of a plurality of images in the batches.
S2, sending the image features in the batch to a central alignment module based on semantic perception, and realizing global central alignment of source domain images with different styles in a feature space.
And S3, sending the characteristics with the aligned centers into a distribution alignment module based on semantic perception, and further realizing local distribution alignment on the image characteristics of various styles in the batch.
And S4, sending the processed features to a feature decoder to obtain a semantic segmentation prediction result consistent with the original image in size.
S5, calculating a loss value of the segmentation prediction result of the current batch of images and training the network according to the loss value.
And S6, saving the training model to obtain the model which is applied to any scene for semantic segmentation.
Preferably, in the step S1, the number of images in the batch process is at least 2 images.
Preferably, in the steps S2 and S3, the segmentation classes are ordered according to the proportion occupied in the data set, and the first 16 classes are selected as semantic perception objects.
Preferably, in the step S4, the decoded image is consistent with the original image, where the original image refers to an original image after clipping and scaling, the clipping mode is clipping at the long side of the image with equal step length according to the width of the image, and then scaling uniformly to an image with a resolution of 640 x 640, and the network model is one of three of deep convolutional neural networks VGG-16, resnet-50 and Resnet-101.
Preferably, in the step S1, feature extraction is performed on the source domain batch image X by a feature encoder F (), so as to obtain an image feature F, which is shown in the following formula:
original image feature f=f (X)
Preferably, in the step S2, the source domain original features are first aligned in the center, and in order to improve the feature discrimination, the category features are first roughly separated:
category characteristics
Figure GDA0004141135760000032
Wherein M is c Class c mask, generated by semantic segmentation classifier, F' c The obtained class c features are coarsely extracted.
Preferably, the coarse features are refined to obtain accurate class features:
optimized class features
Figure GDA0004141135760000033
Wherein Sigm () is a Sigmoid activation function, compressing the feature values to [0,1 ]]Inside, f 3×3 (.) a convolution of 3X 3, F' c,max And F' c,avg The characteristic value obtained after maximum value pooling and average value pooling for the category coarse characteristic,
Figure GDA0004141135760000034
is element-by-element dot product.
Preferably, each obtained fine category feature is independently normalized to realize semantic perception center alignment, and finally fused into a single feature map:
post center alignment feature
Figure GDA0004141135760000031
Where c=16 is the number of categories, IN (F c ,M c ) To M c In-range for category characteristic F c And carrying out instance normalization to realize semantic perception center alignment.
Compared with the prior art, the invention provides a semantic perception-based cross-domain road scene semantic segmentation method, which has the following beneficial effects:
1. according to the semantic segmentation method for the cross-domain road scene based on semantic perception, the labeling cost is reduced, the source domain image which is the most basic of the training network can be directly obtained from the game, the corresponding label is directly generated through the engine, manual labeling is not needed, and a large amount of manpower and material resources are saved.
2. The cross-domain road scene semantic segmentation method based on semantic perception is convenient to train, the model does not need to additionally capture a target domain image, and meanwhile, data enhancement is not required to be carried out on source domain data, and the influence of any image style is avoided, so that the method can be universally applied to various real application scenes.
3. The semantic-perception-based cross-domain road scene semantic segmentation method has high precision and good universality, and on the basis of not contacting any target domain data, four cross-domain semantic segmentation settings of GTA5 to Cityscapes, GTA to BDDS, GTA5 to Mapilary and GTA5 to SYNTHIA respectively reach 38.21, 36.30, 36.87 and 28.45 mIoU on a backbone network of VGG-16; on the backbone network of Resnet-50, mIoU of 39.75, 37.34, 41.86 and 30.79 are respectively reached on four cross-domain semantic segmentation settings; on the backbone network of Resnet-101, mIoU of 45.33, 41.18, 40.77 and 31.84 are achieved on four cross-domain semantic segmentation settings, respectively.
4. The cross-domain road scene semantic segmentation method based on semantic perception has good development prospect, data does not depend on any target domain data, does not depend on any data enhancement means, can be multiplexed with other feature normalization and data enhancement methods, and can meet the requirements of deeper networks and higher computational power conditions in the future.
Drawings
FIG. 1 is a schematic flow chart of the present invention;
FIG. 2 is a schematic diagram of a network architecture according to the present invention;
fig. 3 is a cross-domain semantic segmentation effect contrast diagram under four real scenes.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
Referring to fig. 1-3, a semantic segmentation method for a cross-domain road scene based on semantic perception, the semantic segmentation method comprises the following steps:
s1, sending the source domain images into a feature encoder according to batches to obtain features of a plurality of images in the batches.
S2, sending the image features in the batch to a central alignment module based on semantic perception, and realizing global central alignment of source domain images with different styles in a feature space.
And S3, sending the characteristics with the aligned centers into a distribution alignment module based on semantic perception, and further realizing local distribution alignment on the image characteristics of various styles in the batch.
And S4, sending the processed features to a feature decoder to obtain a semantic segmentation prediction result consistent with the original image in size.
S5, calculating a loss value of the segmentation prediction result of the current batch of images and training the network according to the loss value.
And S6, saving the training model to obtain the model which is applied to any scene for semantic segmentation.
In this embodiment, in step S1, no less than 2 images are required to participate in the network training as a composition batch, and feature extraction is performed on the source domain batch image X by a feature encoder F (), so as to obtain an image feature F, as shown in the following formula:
original image feature f=f (X)
In this embodiment, in steps S2 and S3, the segmentation classes are ordered according to the proportion occupied in the dataset, and the first 16 classes are selected as semantic perception objects. In step S2, the original features of the source domain are first aligned in the center, and in order to improve the feature discrimination, the category features are first roughly separated:
category characteristics
Figure GDA0004141135760000051
Wherein M is c Class c mask, generated by semantic segmentation classifier, F' c The obtained class c features are coarsely extracted. And then refining the coarse features to obtain accurate class features:
optimized class features
Figure GDA0004141135760000061
Wherein Sigm () is a Sigmoid activation function, compressing the feature values to [0,1 ]]Inside, f 3×3 (.) a convolution of 3X 3, F' c,max And F' c,avg The characteristic value obtained after maximum value pooling and average value pooling for the category coarse characteristic,
Figure GDA0004141135760000062
is element-by-element dot product. Finally, each obtained fine category feature is independently normalized to realize semantic perception center alignment, and finally, the fine category features are fused into a single feature map:
post center alignment feature
Figure GDA0004141135760000063
Where c=16 is the number of categories, IN (F c ,M c ) To M c In-range for category characteristic F c And carrying out instance normalization to realize semantic perception center alignment.
In this embodiment, step S3 further aligns the center-aligned features to distribution, and groups the features according to channels first to obtain an mth group of nth image features in the batch:
feature channel grouping
Figure GDA0004141135760000064
Then, the intra-group features are whitened by the example according to the channel grouping, and the specific formula is as follows:
semantic level distribution alignment loss
Figure GDA0004141135760000065
Wherein, ψ (), I is a unit matrix, and the decorrelation of the channels is realized by constraining the channel covariance matrix in the packet as a unit matrix, thereby achieving the aim of aligning semantic perception distribution.
The semantic perception-based cross-domain road scene semantic segmentation method realizes center alignment and distribution alignment in a feature space by batched image data, and simultaneously maintains the distance between class features to realize domain-invariant feature conversion with high discrimination; the method fully considers that the data of the target domain is difficult to acquire in advance in the application scene, and proposes that the network is trained only in the source domain, thereby realizing the reliable cross-domain segmentation effect and having strong model universality.
In the encoding process, conventional VGG-16, resnet-50 and Resnet-101 are used for feature extraction; the network structure is an end-to-end 'encoding-decoding' structure, in the decoding process, each module receives the output of the previous module as input, then nearest neighbor interpolation is carried out, so that the size of the feature map becomes 2 times of the input, and in the training process, the current segmentation effect of the network is measured by adopting a cross entropy loss function form, and the network weight is penalized.
Although embodiments of the present invention have been shown and described, it will be understood by those skilled in the art that various changes, modifications, substitutions and alterations can be made therein without departing from the principles and spirit of the invention, the scope of which is defined in the appended claims and their equivalents.

Claims (5)

1. A semantic perception-based cross-domain road scene semantic segmentation method is characterized by comprising the following steps of: the semantic segmentation method comprises the following steps:
s1, sending source domain images into a feature encoder according to batches to obtain features of a plurality of images in the batches;
s2, sending the image features in the batch to a central alignment module based on semantic perception, and realizing global central alignment of source domain images with different styles in a feature space;
firstly, aligning the centers of original features of a source domain, and in order to improve the feature discrimination, roughly separating category features:
category characteristics
Figure QLYQS_1
Wherein M is c Class c mask, generated by semantic segmentation classifier, F c ' is a class c feature obtained by rough extraction;
refining the coarse features to obtain accurate class features:
optimized class features
Figure QLYQS_2
Wherein Sigm () is a Sigmoid activation function, compressing the feature values to [0,1 ]]Inside, f 3×3 (.) a convolution of 3X 3, F' c,max And F' c,avg The characteristic value obtained after maximum value pooling and average value pooling for the category coarse characteristic,
Figure QLYQS_3
is element-by-element dot multiplication;
normalizing each obtained fine category feature independently to realize semantic perception center alignment, and finally fusing into a single feature map:
post center alignment feature
Figure QLYQS_4
Where c=16 is the number of categories, IN (F c ,M c ) To M c In-range for category characteristic F c Performing instance normalization to realize semantic perception center alignment;
s3, sending the characteristics with the aligned centers into a distribution alignment module based on semantic perception, and further realizing local distribution alignment on the image characteristics of various styles in the batch;
the center-aligned features are further distributed and aligned, and the features are firstly grouped according to channels to obtain an mth group of nth image features in the batch:
feature channel grouping
Figure QLYQS_5
Then, the intra-group features are whitened by the example according to the channel grouping, and the specific formula is as follows:
semantic level distribution alignment loss
Figure QLYQS_6
Wherein, ψ () is a channel covariance matrix, I is a unit matrix, and the decorrelation of the channels is realized by restricting the channel covariance matrix in the group as the unit matrix, thereby achieving the aim of aligning semantic perception distribution;
s4, sending the processed features to a feature decoder to obtain a semantic segmentation prediction result consistent with the original image in size;
s5, calculating a loss value according to the segmentation prediction result of the current batch of images and training a network according to the loss value;
and S6, saving the training model to obtain the model which is applied to any scene for semantic segmentation.
2. The semantic perception-based cross-domain road scene semantic segmentation method according to claim 1, wherein the method is characterized by comprising the following steps of: in the step S1, the number of images in batch processing is at least 2 images.
3. The semantic perception-based cross-domain road scene semantic segmentation method according to claim 1, wherein the method is characterized by comprising the following steps of: in the steps S2 and S3, the segmentation classes are ordered according to the proportion occupied in the data set, and the first 16 classes are selected as semantic perception objects.
4. The semantic perception-based cross-domain road scene semantic segmentation method according to claim 1, wherein the method is characterized by comprising the following steps of: in the step S4, the decoded image and the original image are kept consistent in size, where the original image refers to an original image after clipping and scaling, the clipping mode is clipping at the long side of the image with equal step length according to the width of the image, and then scaling is performed uniformly to an image with 640 x 640 resolution, and the network model is one of three kinds of deep convolutional neural networks VGG-16, resnet-50 and Resnet-101.
5. The semantic perception-based cross-domain road scene semantic segmentation method according to claim 1, wherein the method is characterized by comprising the following steps of: in the step S1, feature extraction is performed on the source domain batch image X by a feature encoder F (), so as to obtain an image feature F, which is shown in the following formula:
original image feature f=f (X).
CN202210083793.2A 2022-01-21 2022-01-21 Cross-domain road scene semantic segmentation method based on semantic perception Active CN114529878B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210083793.2A CN114529878B (en) 2022-01-21 2022-01-21 Cross-domain road scene semantic segmentation method based on semantic perception

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210083793.2A CN114529878B (en) 2022-01-21 2022-01-21 Cross-domain road scene semantic segmentation method based on semantic perception

Publications (2)

Publication Number Publication Date
CN114529878A CN114529878A (en) 2022-05-24
CN114529878B true CN114529878B (en) 2023-04-25

Family

ID=81621541

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210083793.2A Active CN114529878B (en) 2022-01-21 2022-01-21 Cross-domain road scene semantic segmentation method based on semantic perception

Country Status (1)

Country Link
CN (1) CN114529878B (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110322446A (en) * 2019-07-01 2019-10-11 华中科技大学 A kind of domain adaptive semantic dividing method based on similarity space alignment
CN112488229A (en) * 2020-12-10 2021-03-12 西安交通大学 Domain self-adaptive unsupervised target detection method based on feature separation and alignment
CN112766089A (en) * 2021-01-04 2021-05-07 武汉大学 Cross-domain road extraction method based on global-local countermeasure learning framework
CN112991353A (en) * 2021-03-12 2021-06-18 北京航空航天大学 Unsupervised semantic segmentation method for cross-domain remote sensing image
WO2021159742A1 (en) * 2020-02-10 2021-08-19 腾讯科技(深圳)有限公司 Image segmentation method and apparatus, and storage medium
CN113902913A (en) * 2021-08-31 2022-01-07 际络科技(上海)有限公司 Image semantic segmentation method and device

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110322446A (en) * 2019-07-01 2019-10-11 华中科技大学 A kind of domain adaptive semantic dividing method based on similarity space alignment
WO2021159742A1 (en) * 2020-02-10 2021-08-19 腾讯科技(深圳)有限公司 Image segmentation method and apparatus, and storage medium
CN112488229A (en) * 2020-12-10 2021-03-12 西安交通大学 Domain self-adaptive unsupervised target detection method based on feature separation and alignment
CN112766089A (en) * 2021-01-04 2021-05-07 武汉大学 Cross-domain road extraction method based on global-local countermeasure learning framework
CN112991353A (en) * 2021-03-12 2021-06-18 北京航空航天大学 Unsupervised semantic segmentation method for cross-domain remote sensing image
CN113902913A (en) * 2021-08-31 2022-01-07 际络科技(上海)有限公司 Image semantic segmentation method and device

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
Duo Peng 等.Semantic-Aware Domain Generalized Segmentation.《arXiv》.2022,1-16. *
Li Gao 等.DSP: Dual Soft-Paste for Unsupervised Domain Adaptive Semantic Segmentation.《arXiv》.2021,1-9. *
Wonwoong Cho 等.Image-to-Image Translation via Group-wise Deep Whitening-and-Coloring Transformation.《arXiv》.2019,1-15. *
Yijun Li 等.Universal Style Transfer via Feature Transforms.《arXiv》.2017,1-11. *
周思航.用于低质量医疗图像分割的表示学习.《中国博士学位论文全文数据库 医药卫生科技辑》.2022,(第第01期期),E080-6. *
张晓伟 等.基于局部语义特征不变性的跨域行人重识别.《北京航空航天大学学报》.2020,1-10. *

Also Published As

Publication number Publication date
CN114529878A (en) 2022-05-24

Similar Documents

Publication Publication Date Title
CN111639564B (en) Video pedestrian re-identification method based on multi-attention heterogeneous network
CN109285162A (en) A kind of image, semantic dividing method based on regional area conditional random field models
CN112287941B (en) License plate recognition method based on automatic character region perception
CN114092917B (en) MR-SSD-based shielded traffic sign detection method and system
CN113034506B (en) Remote sensing image semantic segmentation method and device, computer equipment and storage medium
CN114913493A (en) Lane line detection method based on deep learning
CN111008979A (en) Robust night image semantic segmentation method
CN114782949B (en) Traffic scene semantic segmentation method for boundary guide context aggregation
CN112419163B (en) Single image weak supervision defogging method based on priori knowledge and deep learning
CN115187945A (en) Lane line recognition method, lane line recognition device, electronic device, and storage medium
CN114973199A (en) Rail transit train obstacle detection method based on convolutional neural network
CN113869396A (en) PC screen semantic segmentation method based on efficient attention mechanism
CN111160282B (en) Traffic light detection method based on binary Yolov3 network
CN114529878B (en) Cross-domain road scene semantic segmentation method based on semantic perception
CN116994164A (en) Multi-mode aerial image fusion and target detection combined learning method
CN114821510B (en) Lane line detection method and device based on improved U-Net network
CN112767275B (en) Single image defogging method based on artificial sparse annotation information guidance
CN113379711B (en) Image-based urban road pavement adhesion coefficient acquisition method
CN115496764A (en) Dense feature fusion-based foggy image semantic segmentation method
CN114972752A (en) Real-time semantic segmentation method based on lightweight cross attention network
CN111008986B (en) Remote sensing image segmentation method based on multitasking semi-convolution
Wang et al. Fusion attention network for autonomous cars semantic segmentation
CN118155012A (en) YOLOv5 s-based lightweight vehicle model training method
CN117649635B (en) Method, system and storage medium for detecting shadow eliminating point of narrow water channel scene
CN116895029B (en) Aerial image target detection method and aerial image target detection system based on improved YOLO V7

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant