CN112016556B - Multi-type license plate recognition method - Google Patents
Multi-type license plate recognition method Download PDFInfo
- Publication number
- CN112016556B CN112016556B CN202010851716.8A CN202010851716A CN112016556B CN 112016556 B CN112016556 B CN 112016556B CN 202010851716 A CN202010851716 A CN 202010851716A CN 112016556 B CN112016556 B CN 112016556B
- Authority
- CN
- China
- Prior art keywords
- license plate
- character
- semantic
- position information
- information
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 39
- 230000011218 segmentation Effects 0.000 claims description 64
- 108091006146 Channels Proteins 0.000 claims description 27
- 238000000605 extraction Methods 0.000 claims description 23
- 238000012549 training Methods 0.000 claims description 16
- 238000010586 diagram Methods 0.000 claims description 5
- 238000005070 sampling Methods 0.000 claims description 4
- 239000011159 matrix material Substances 0.000 claims description 3
- 230000004913 activation Effects 0.000 claims 1
- 238000012545 processing Methods 0.000 abstract description 6
- 238000012805 post-processing Methods 0.000 abstract description 5
- 230000002349 favourable effect Effects 0.000 abstract description 2
- 230000006870 function Effects 0.000 description 14
- 230000004927 fusion Effects 0.000 description 6
- 238000004458 analytical method Methods 0.000 description 5
- 238000013135 deep learning Methods 0.000 description 5
- 230000000694 effects Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 4
- 238000013527 convolutional neural network Methods 0.000 description 2
- 238000005286 illumination Methods 0.000 description 2
- 238000002372 labelling Methods 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 125000004122 cyclic group Chemical group 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000000306 recurrent effect Effects 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/14—Image acquisition
- G06V30/148—Segmentation of character regions
- G06V30/153—Segmentation of character regions using recognition of characters or words
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2413—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
- G06F18/24133—Distances to prototypes
- G06F18/24137—Distances to cluster centroïds
- G06F18/2414—Smoothing the distance, e.g. radial basis function networks [RBFN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/62—Text, e.g. of license plates, overlay texts or captions on TV images
- G06V20/625—License plates
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Multimedia (AREA)
- Character Discrimination (AREA)
- Image Analysis (AREA)
- Traffic Control Systems (AREA)
Abstract
The invention discloses a multi-type license plate recognition method, which comprises the following steps: extracting the features of the license plate image to be recognized to obtain a corresponding feature map; extracting semantic information and position information of the feature map; and (3) fusing semantic information and position information by adopting an attention mechanism, and realizing the prediction of the license plate character strings through a shared classifier. The method comprises the following steps: on one hand, by explicitly modeling the position information of the characters of the license plate, the position information of the characters can be efficiently acquired, and then the characters are identified by using the position information without respectively processing according to the types of the license plate; on the other hand, the attention mechanism is adopted to fuse semantic information and position information, the semantic information does not need to be manually processed, any complicated post-processing step is avoided, the algorithm efficiency is high, and the robustness is strong; finally, through sharing the classifier, improve data utilization and only need carry out once the identification process, be favorable to alleviating the unbalanced problem of data.
Description
Technical Field
The invention relates to the field of intelligent transportation, in particular to a multi-type license plate recognition method.
Background
The intelligent transportation system improves the traffic safety and mobility by using advanced technology, thereby improving productivity and having a wide impact on people's lives. The license plate number of the motor vehicle is used as the unique 'identity' mark of the vehicle, the automatic identification technology of the license plate number is one of the basic core contents of the intelligent traffic system, the license plate number has been widely researched, and the license plate number can be applied to the traffic related fields of road charging, parking management, traffic law enforcement and the like.
Before outbreak of deep learning application, license plate recognition is mainly divided into two steps: character segmentation and character recognition. The character segmentation divides the license plate image into images containing single characters, so that the subsequent character recognition is facilitated, and the license plate recognition effect is directly influenced by the character segmentation quality. The character segmentation technology mainly comprises a connected domain analysis method and a projection analysis method, wherein the connected domain analysis method is simple to operate, and the projection analysis method is not influenced by the inclination of a license plate and can be applied to simple scenes with clear license plates and uniform illumination, such as parking lots, toll stations and the like. For complex scenes such as vehicle-mounted equipment and traffic monitoring, the segmentation effect is easily influenced by natural environment conditions, such as uneven illumination, low resolution, motion blur and the like. In contrast, the character recognition effect is greatly improved after the CNN is introduced. Therefore, the traditional method still faces a great obstacle to extracting the position information of the characters in the license plate.
The deep convolutional neural network is widely introduced in license plate recognition and achieves good effect due to strong feature extraction and learning capacity. In a patent of a license plate detection and integral recognition method based on deep learning, features extracted from a network are directly classified through seven classifiers, and each classifier is forced to pay attention to characters at a corresponding position. The patent 'efficient and accurate license plate recognition method' deconstructs license plate recognition into semantic segmentation tasks, and license plate character strings are obtained by performing post-processing analysis on a semantic segmentation map of a license plate. Both methods avoid explicitly extracting character position information, but the classifier for recognition in the former is difficult to locate characters, and a series of hyper-parameters in the latter reduces the generalization performance of the method.
In addition, the above method mainly considers the recognition of a single type of license plate (such as a single-row car license plate), but in practical application, the license plate recognition system is required to be capable of processing various license plates, that is, the recognition of multiple types of license plates needs to be considered. A few approaches have been tried to address this problem. The patents 'license plate identification method, license plate identification device, license plate identification equipment and medium' and 'a license plate identification method based on license plate classification and LSTM' divide and then splice double-row license plates into single-row license plate identification, and the operation steps are complex. In the patent license plate recognition method and device, a recurrent neural network is adopted to recognize single and double-row license plates, but the cyclic decoding brings low recognition efficiency.
In summary, the license plate recognition technology based on deep learning has been vigorously developed, but how to adopt a unified model to perform efficient and accurate license plate recognition for multiple types of license plates has not been solved yet.
Disclosure of Invention
The invention aims to provide a multi-type license plate recognition method, which does not need to carry out classification processing according to the types of license plates, has wide application prospect, is light and efficient, has strong robustness and can be used for unified recognition of different types of license plates in a complex traffic environment.
The purpose of the invention is realized by the following technical scheme:
a multi-type license plate recognition method comprises the following steps:
extracting the features of the license plate image to be recognized to obtain a corresponding feature map;
extracting semantic information and position information of the feature map;
and (3) fusing semantic information and position information by adopting an attention mechanism, and realizing the prediction of the license plate character strings through a shared classifier.
According to the technical scheme provided by the invention, on one hand, the position information of the characters can be efficiently acquired by explicitly modeling the position information of the characters of the license plate, and then the characters are identified by utilizing the position information without respectively processing according to the types of the license plate; on the other hand, the attention mechanism is adopted to fuse semantic information and position information, the semantic information does not need to be manually processed, any complicated post-processing step is avoided, the algorithm efficiency is high, and the robustness is strong; finally, through sharing the classifier, improve data utilization and only need carry out once the identification process, be favorable to alleviating the unbalanced problem of data.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a network structure diagram of a multi-type license plate recognition method according to an embodiment of the present invention;
FIG. 2 is a flowchart of a multi-type license plate recognition network including training and testing according to an embodiment of the present invention;
FIG. 3 is an example of a map of the location segmentation of different types of license plates provided by an embodiment of the present invention;
fig. 4 is a schematic diagram of fusion of semantic information and location information provided by an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention are clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present invention without making any creative effort, shall fall within the protection scope of the present invention.
The embodiment of the invention provides a multi-type license plate recognition method, which adopts an attention mechanism to process multi-type problems, constructs an end-to-end multi-type license plate recognition network and can directly output license plate character strings for input different license plate images. Fig. 1 provides a main composition structure and a working process of a multi-type license plate recognition network, and the main process of the recognition method is as follows:
1. and extracting the features of the license plate image to be recognized to obtain a corresponding feature map.
As shown in fig. 1, this step is implemented by a feature extraction module, which mainly includes a convolution layer, an active layer, and a down-sampling layer connected in sequence. Illustratively, a spatial Path (SpatialPath) and a Context Path (Context Path) in the BiSeNet may be used as the feature extraction module, and the downsampling multiple n may be set to 8.
The license plate image to be recognized is converted into a feature map with smaller space size through a feature extraction module, and the spatial resolution and the processing calculated amount can be balanced by adjusting the down-sampling multiple n.
Assuming that the size of the license plate image to be recognized is h multiplied by w multiplied by c ', the size of the extracted feature image is h multiplied by w multiplied by c' through the feature extraction moduleWhere h represents the image height, w represents the image width, c' represents the number of channels of the license plate image to be recognized (e.g., for an RGB image, the number of channels is 3), n represents the downsampling multiple, and c represents the number of channels of the feature map.
2. And extracting semantic information and position information of the feature map.
As shown in fig. 1, the extraction of the semantic information and the position information is realized by a semantic information extraction module and a position information extraction module, respectively. Illustratively, a Feature Fusion Module (Feature Fusion Module) in the BiSeNet may be used as the semantic information extraction Module and the position information extraction Module.
1) Semantic information extraction, namely predicting semantic information of all characters of license plate, with the size ofThe semantic information output by the semantic information extraction module is the sizeThe semantic segmentation graph of the license plate characters is expressed, wherein CsThe number of the character categories is the number of the character categories, the characters of all types of license plates are covered, 1 represents the background category, and each element of the license plate character semantic segmentation graph represents the character semantic category score of the corresponding pixel point.
2) Extracting position information, namely predicting the position information of all characters of the license plate, wherein the size of the position information isThe position information output by the position information extraction module is the same as the position information output by the position information extraction moduleLicense plate wordIs represented by a symbol position segmentation map, wherein CpThe maximum length of the license plate character string is represented, 1 represents a background category, and each element of the license plate character position segmentation graph represents the position category score of the character to which the corresponding pixel point belongs; the ith channel of the license plate character position segmentation graph represents the position segmentation graph of the ith character in the license plate character string, namely the channel sequence expresses the character sequence and is not influenced by the type of the license plate, i is 1p。
3. And (3) fusing semantic information and position information by adopting an attention mechanism, and realizing the prediction of the license plate character strings through a shared classifier.
As shown in fig. 1, this step is implemented by a license plate string prediction module with attention fused, and the main steps are as follows:
1) according to the license plate character position segmentation map, regionality selection is carried out on the license plate character semantic segmentation map in a spatial attention (spatial attention) mode, namely each channel of the license plate character position segmentation map except a background channel is multiplied by the license plate character semantic segmentation map to obtain CpAnd each new license plate character semantic segmentation map only keeps the semantic information of a single character, so that a proprietary semantic feature is constructed for each character and is irrelevant to the type of the license plate.
Each channel of the license plate character position segmentation graph corresponds to the position of one character, the pixel value of the character position represented by one channel is close to 1, and the pixel values of the rest positions are close to 0; the semantic segmentation map comprises semantic information of all characters, and in the information fusion operation, the license plate character position segmentation map is split according to channels, copied and expanded to the same dimension of the license plate semantic segmentation map, and then multiplied by the license plate character semantic segmentation map; and after multiplication, each new license plate character semantic segmentation graph only keeps the semantic information of one character. For example, an 8-channel license plate character position segmentation map is multiplied by a 24-channel license plate character semantic segmentation map channel by channel to obtain 8 new license plate character semantic segmentation maps with 24 channels, and each new license plate character semantic segmentation map only retains semantic information of one character.
Illustratively, a location attention map can be generated from a license plate character location segmentation map by using softmax, each location category score is distributed between 0 and 1, and the location is more prominent, and then the information fusion operation is carried out. Particularly, channels belonging to the background in the license plate character position segmentation map and the license plate character semantic segmentation map are discarded during information fusion.
As shown in FIG. 1, the symbolsThe representation adopts an attention mechanism to fuse semantic information and position information.
2) After the semantic features of all characters are spliced in the dimension of batch size (batch size), all characters are processed once through a shared classifier to obtain Cp×ClWherein C ispFor the longest character string length in all types of license plates, corresponds to CpA position, ClAs the number of character classes CsPlus a no character category.
3) To Cp×ClAnd (4) performing post-processing on the probability matrix, and obtaining a final license plate recognition result by taking the maximum probability character of each position and discarding non-characters.
Illustratively, the shared classifier may be implemented by a convolutional layer, a downsample layer, and a fully-connected layer, which are connected in sequence.
Compared with the traditional license plate recognition algorithm, the scheme of the invention does not need character segmentation, and improves the recognition efficiency and the algorithm robustness; compared with other deep learning algorithms, the method explicitly learns the positions of characters, constructs an end-to-end model, avoids complex post-processing steps and improves the robustness of the algorithm; the system is light in weight by sharing the character classifier; a single model can process multiple types of license plates, and has better scene adaptability.
The main process for realizing the multi-type license plate recognition based on the multi-type license plate recognition network is described above. The multi-type license plate recognition network needs to be trained in advance and used for multi-type license plate recognition after training. As shown in fig. 2, in the training stage, a plurality of types of license plate images are collected first, and a license plate recognition data set used for training is constructed. And training a multi-type license plate recognition network on the constructed license plate recognition data set to obtain a trained multi-type license plate recognition network model. In the using stage, firstly, the license plate image to be recognized is zoomed to a proper size, and then forward reasoning (namely the process introduced above) is carried out through the multi-type license plate recognition network to obtain a final recognition result. The main description is as follows:
firstly, training.
1. And constructing a license plate recognition data set used for training.
Collecting license plate images of different types such as yellow color of a large automobile, yellow-green color matching of a large new energy automobile, blue color of a small automobile, green color of a small new energy automobile and the like, zooming the license plate images to a fixed size (such as 50 x 160), wherein the number of the license plate images of different types is as close as possible, and then making semantic segmentation labels (namely license plate character semantic segmentation label images), position segmentation labels (namely license plate character position segmentation label images) and license plate character string labels for the license plate images, wherein the semantic segmentation labels and the position segmentation labels are optional.
1) The semantic segmentation labels are used for labeling the types of characters in the license plate image and are in the form of matrixes with the same size as the license plate image. Specifically, each character is respectively marked with a compact rectangular frame column, pixels in the frame are marked as the category of the corresponding character, and other pixels are marked as the background category. The license plate characters mainly comprise Chinese characters, numeric characters and alphabetical characters, and the characters are totally 65 types according to the license plate specification of the Chinese motor vehicle, wherein the Chinese characters mainly comprise 22 provinces, 5 autonomous regions and 4 direct municipalities for short, and totally 31 types; the number characters are 0-9, and 10 types are total; the alphabetic characters are of the 24 classes A-Z (I and O excluded).
2) The position segmentation labels are used for labeling the positions of characters in the license plate image, the form of the position segmentation labels is the same as that of the semantic segmentation labels, and the labels in the rectangular frame are character position categories, such as: for a 7-character frame of a car license plate, representing the character frame by discrete values {1, 2.., 7} from left to right; for a 7-character large yellow double-row automobile license plate, the character position sequence is from top to bottom and from left to right, and is represented by discrete values {1, 2., 7 }; for a new energy license plate with 8 characters, the license plate is represented by {1, 2.., 8} from left to right.
3) The license plate character string label is used for marking a character sequence of a license plate, the maximum length of the character sequence is the maximum length of a license plate number (such as a new energy license plate, total 8 bits), and when the number of the license plate characters is insufficient (such as a common 7-bit car license plate), the characters without categories are used for complementing behind the license plate characters, so that all the license plate character sequences have the same length.
2. And (5) training.
During training, the input of the network is a license plate recognition data set license plate image, and the output is a predicted license plate character string; the internal processing procedures are also 1 to 3 described above, and are not described herein again.
After the network outputs the result, updating the network parameters through a loss function; the loss function for the training phase is:
L=α1Ls+α2Lp+LC
wherein L iss、Lp、LCRespectively extracting a loss function of semantic information, a loss function of position information extraction and a loss function of license plate character string prediction; alpha is alpha1And alpha2To balance the three-term loss function.
Illustratively, a PyTorch deep learning framework can be used for establishing a multi-type license plate recognition network, and a back propagation algorithm and a gradient descent strategy are used for reducing a loss function value and converging a model, so that a well-trained license plate recognition model is finally obtained.
In the embodiment of the invention, the loss function L for semantic information extractionsLoss function L for extracting position informationpA cross entropy loss function may be employed; the formula is as follows:
where M is the number of classes, yiFor indicating variables, if the class i is the same as the labeled class of the training sample, it is 1, otherwise it is 0, piIs the predicted probability that the training sample belongs to class i.
As will be appreciated by those skilled in the art, Ls、LpIs prepared from Cs+1、Cp+1 losses resulting from bringing, respectively, C into the above formulas+1 for M to give Ls。
Predicted loss function L of license plate character stringCIs mainly used for training a shared classifier, and can complete all license plate numbers to the maximum license plate length C during trainingp(completion at the end of the string with the missing character category for the deficiency), the penalty function LCThe formula is as follows:
wherein L islIs the cross entropy loss of a single character.
And secondly, a testing stage (namely, a subsequent using stage).
As described above, after the network training is completed, the recognition result can be obtained through forward reasoning, and the specific process is 1 to 3 as described above; specific numerical values will be described below as examples.
As shown in fig. 1, a license plate image of the left side as an example; the image is adjusted to a fixed size close to the aspect ratio of the license plate as an input of the license plate recognition model, and for example, the size of the image may be adjusted to 50 × 160.
Through the semantic information extraction module, the obtained license plate semantic segmentation graph has the size ofThrough the position information extraction module, the obtained position segmentation graph has the size ofIllustratively, since the downsampling multiple n of the network is 8, h is 50, w is 160, there are 65 character classes, and there are 8 characters at most, the semantic segmentation map size excluding the background channel is 65 × 7 × 20, the position segmentation map size is 8 × 7 × 20, and there are manyThe position segmentation map of the type license plate samples is shown in fig. 3, and is arranged from left to right in the channel order, i.e., the character order. Therefore it isThe 7 is obtained because the calculation process is rounded up, and the downsampling is 8 times, and is usually 2 times for 3 times, and padding may occur after each downsampling, and thus the final result is 7.
Then, fusing the semantics and the position segmentation graph, as shown in fig. 4, multiplying each channel of the position segmentation graph by the semantic segmentation graph, and only retaining the semantic information belonging to the channel, thereby constructing a special semantic feature for each character, wherein the number in fig. 4 represents the semantic information corresponding to the character before and after the operation, the fused features are spliced in sequence, the character sequence is retained, and then the character sequence is sent to a shared classifier to obtain Cp×ClAnd obtaining a final license plate recognition result by taking the maximum probability character of each position and discarding non-characters.
Illustratively, since the longest license plate number is 8 bits and the number of character classes is 65, CpIs 8, ClIs 66. For the four types of license plates in fig. 2, the network output result is shown in the right part of fig. 1, and "-" represents a no-character category, and the sequence of each character is consistent with the channel sequence of the position segmentation map.
Through the description of the above embodiments, it is clear to those skilled in the art that the above embodiments may be implemented by software, or by software plus a necessary general hardware platform. With this understanding, the technical solutions of the embodiments can be embodied in the form of a software product, which can be stored in a non-volatile storage medium (which can be a CD-ROM, a usb disk, a removable hard disk, etc.), and includes several instructions for enabling a computer device (which can be a personal computer, a server, or a network device, etc.) to execute the methods according to the embodiments of the present invention.
It can be clearly understood by those skilled in the art that, for convenience and simplicity of description, the foregoing division of each functional module is merely used as an example, and in practical applications, the foregoing function distribution may be completed by different functional modules as needed, that is, the internal structure of the network is divided into different functional modules to complete all or part of the above-described functions.
The above description is only a preferred embodiment of the present invention, but the scope of the present invention is not limited thereto, and any changes or substitutions that can be easily conceived by those skilled in the art within the technical scope of the present invention are also within the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the claims.
Claims (5)
1. A multi-type license plate recognition method is characterized by comprising the following steps:
extracting the features of the license plate image to be recognized to obtain a corresponding feature map;
extracting semantic information and position information of the feature map;
an attention mechanism is adopted to fuse semantic information and position information, and a shared classifier is used for predicting license plate character strings;
the method comprises the following steps of adopting an attention mechanism to fuse semantic information and position information, and realizing the prediction of the license plate character strings through a shared classifier, wherein the steps comprise:
the semantic information is a license plate character semantic segmentation map, and the position information is a license plate character position segmentation map;
according to the license plate character position segmentation map, regionality selection is carried out on the license plate character semantic segmentation map in a space attention mode, namely each channel of the license plate character position segmentation map except a background channel is multiplied by the license plate character semantic segmentation map to obtain CpEach new license plate character semantic segmentation map only keeps semantic information of a single character, so that a special semantic feature is constructed for each character; cpRepresenting a maximum license plate string length;
performing semantic features of all characters in a batch size dimensionAfter splicing, all characters are processed once through a shared classifier to obtain Cp×ClA probability matrix of (2), wherein CpFor the longest character string length in all types of license plates, corresponds to CpA position, ClAs the number of character classes CsAdding a no character category;
to Cp×ClThe probability matrix is post-processed, and a final license plate recognition result is obtained by taking the maximum probability character of each position and discarding non-characters.
2. The method of claim 1, wherein the extracting features of the license plate image to be recognized to obtain the corresponding feature map comprises:
for the license plate image to be recognized with the size of h multiplied by w multiplied by c', the size of the license plate image to be recognized is extracted through a characteristic extraction moduleA characteristic diagram of (2);
the characteristic extraction module comprises a convolution layer, an activation layer and a down-sampling layer which are sequentially connected; wherein h represents the image height, w represents the image width, c' represents the number of channels of the license plate image to be recognized, n represents the down-sampling multiple, and c represents the number of channels of the feature map.
3. The method of claim 1, wherein the extracting semantic information from the feature map comprises:
for a size ofThe semantic information is extracted through a semantic information extraction module; wherein,c represents the height, width and channel number of the characteristic diagram respectively;
semantic information is defined by a size ofThe license plate character semantic segmentation graph is expressed, wherein CsThe number of the character categories is the number of the character categories, the characters of all types of license plates are covered, 1 represents the background category, and each element of the license plate character semantic segmentation graph represents the character semantic category score of the corresponding pixel point.
4. The method of claim 1, wherein the extracting the location information of the feature map comprises:
for a size ofThe position information is extracted through a position information extraction module; wherein,c represents the height, width and channel number of the characteristic diagram respectively;
the position information is of sizeThe character position segmentation graph of the license plate is represented, wherein CpThe maximum length of the license plate character string is represented, 1 represents a background category, and each element of the license plate character position segmentation graph represents the position category score of the character to which the corresponding pixel point belongs; the ith channel represents a position segmentation graph of the ith character in the license plate character string, namely the channel sequence expresses the character sequence, and i is 1p。
5. The method according to claim 1, wherein the method is implemented by constructing an end-to-end multi-type license plate recognition network, and the loss function of the multi-type license plate recognition network in the training stage is:
L=α1Ls+α2Lp+LC
wherein L iss、Lp、LCRespectively extracting a loss function of semantic information, a loss function of position information extraction and a loss function of license plate character string prediction; alpha is alpha1And alpha2To balance the three-term loss function.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010851716.8A CN112016556B (en) | 2020-08-21 | 2020-08-21 | Multi-type license plate recognition method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010851716.8A CN112016556B (en) | 2020-08-21 | 2020-08-21 | Multi-type license plate recognition method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112016556A CN112016556A (en) | 2020-12-01 |
CN112016556B true CN112016556B (en) | 2022-07-15 |
Family
ID=73505498
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010851716.8A Active CN112016556B (en) | 2020-08-21 | 2020-08-21 | Multi-type license plate recognition method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112016556B (en) |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112329787B (en) * | 2020-12-07 | 2023-06-30 | 北京同方软件有限公司 | Water level character positioning method based on information attention degree |
CN112488108A (en) * | 2020-12-11 | 2021-03-12 | 广州小鹏自动驾驶科技有限公司 | Parking space number identification method and device, electronic equipment and storage medium |
CN112580628B (en) * | 2020-12-22 | 2023-08-01 | 浙江智慧视频安防创新中心有限公司 | Attention mechanism-based license plate character recognition method and system |
KR20220127555A (en) * | 2021-03-11 | 2022-09-20 | 한국전자통신연구원 | Apparatus and Method for Recognizing Number Plate based on Weakly-Supervised Localization |
CN113313115B (en) * | 2021-06-11 | 2023-08-04 | 浙江商汤科技开发有限公司 | License plate attribute identification method and device, electronic equipment and storage medium |
CN113554030B (en) * | 2021-07-27 | 2022-08-16 | 上海大学 | Multi-type license plate recognition method and system based on single character attention |
CN115131776B (en) * | 2022-04-12 | 2023-05-12 | 公安部交通管理科学研究所 | License plate image character classification recognition method integrating new energy automobile features |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107239778A (en) * | 2017-06-09 | 2017-10-10 | 中国科学技术大学 | The licence plate recognition method of efficiently and accurately |
WO2019169532A1 (en) * | 2018-03-05 | 2019-09-12 | 深圳前海达闼云端智能科技有限公司 | License plate recognition method and cloud system |
CN111008632A (en) * | 2019-10-17 | 2020-04-14 | 安徽清新互联信息科技有限公司 | License plate character segmentation method based on deep learning |
CN111008639A (en) * | 2019-10-17 | 2020-04-14 | 安徽清新互联信息科技有限公司 | Attention mechanism-based license plate character recognition method |
CN111027539A (en) * | 2019-10-17 | 2020-04-17 | 安徽清新互联信息科技有限公司 | License plate character segmentation method based on spatial position information |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10719743B2 (en) * | 2018-01-19 | 2020-07-21 | Arcus Holding A/S | License plate reader using optical character recognition on plural detected regions |
-
2020
- 2020-08-21 CN CN202010851716.8A patent/CN112016556B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107239778A (en) * | 2017-06-09 | 2017-10-10 | 中国科学技术大学 | The licence plate recognition method of efficiently and accurately |
WO2019169532A1 (en) * | 2018-03-05 | 2019-09-12 | 深圳前海达闼云端智能科技有限公司 | License plate recognition method and cloud system |
CN111008632A (en) * | 2019-10-17 | 2020-04-14 | 安徽清新互联信息科技有限公司 | License plate character segmentation method based on deep learning |
CN111008639A (en) * | 2019-10-17 | 2020-04-14 | 安徽清新互联信息科技有限公司 | Attention mechanism-based license plate character recognition method |
CN111027539A (en) * | 2019-10-17 | 2020-04-17 | 安徽清新互联信息科技有限公司 | License plate character segmentation method based on spatial position information |
Non-Patent Citations (2)
Title |
---|
A CNN-Based Approach for Automatic License Plate Recognition in the Wild;Meng Dong 等;《SEMANTIC SCHOLAR》;20171205;第1-12页 * |
基于深度学习的车牌相似字符识别;潘翔等;《计算机科学》;20170615;第239-241,257页 * |
Also Published As
Publication number | Publication date |
---|---|
CN112016556A (en) | 2020-12-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112016556B (en) | Multi-type license plate recognition method | |
CN109902806B (en) | Method for determining target bounding box of noise image based on convolutional neural network | |
CN109949317B (en) | Semi-supervised image example segmentation method based on gradual confrontation learning | |
de Geus et al. | Single network panoptic segmentation for street scene understanding | |
Huo et al. | Vehicle type classification and attribute prediction using multi-task RCNN | |
CN112508977A (en) | Deep learning-based semantic segmentation method for automatic driving scene | |
CN109145900A (en) | A kind of licence plate recognition method based on deep learning | |
CN111104903A (en) | Depth perception traffic scene multi-target detection method and system | |
CN111008639B (en) | License plate character recognition method based on attention mechanism | |
CN108830254B (en) | Fine-grained vehicle type detection and identification method based on data balance strategy and intensive attention network | |
CN108596240B (en) | Image semantic segmentation method based on discriminant feature network | |
CN111461129B (en) | Context prior-based scene segmentation method and system | |
CN112101117A (en) | Expressway congestion identification model construction method and device and identification method | |
CN114299286A (en) | Road scene semantic segmentation method based on category grouping in abnormal weather | |
CN112990065A (en) | Optimized YOLOv5 model-based vehicle classification detection method | |
Chen et al. | Enhancing the robustness of object detection via 6G vehicular edge computing | |
CN115115915A (en) | Zebra crossing detection method and system based on intelligent intersection | |
CN112785610B (en) | Lane line semantic segmentation method integrating low-level features | |
CN113869396A (en) | PC screen semantic segmentation method based on efficient attention mechanism | |
CN113505640A (en) | Small-scale pedestrian detection method based on multi-scale feature fusion | |
Pan et al. | A hybrid deep learning algorithm for the license plate detection and recognition in vehicle-to-vehicle communications | |
CN116363072A (en) | Light aerial image detection method and system | |
CN116206201A (en) | Monitoring target detection and identification method, device, equipment and storage medium | |
CN115240271A (en) | Video behavior identification method and system based on space-time modeling | |
Wang et al. | Fusion attention network for autonomous cars semantic segmentation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |