CN108921162A - Licence plate recognition method and Related product based on deep learning - Google Patents

Licence plate recognition method and Related product based on deep learning Download PDF

Info

Publication number
CN108921162A
CN108921162A CN201810593633.6A CN201810593633A CN108921162A CN 108921162 A CN108921162 A CN 108921162A CN 201810593633 A CN201810593633 A CN 201810593633A CN 108921162 A CN108921162 A CN 108921162A
Authority
CN
China
Prior art keywords
license plate
network structure
anchor
context module
convolutional layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810593633.6A
Other languages
Chinese (zh)
Inventor
陈书楷
童飞扬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xiamen Central Intelligent Information Technology Co Ltd
Original Assignee
Xiamen Central Intelligent Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xiamen Central Intelligent Information Technology Co Ltd filed Critical Xiamen Central Intelligent Information Technology Co Ltd
Priority to CN201810593633.6A priority Critical patent/CN108921162A/en
Publication of CN108921162A publication Critical patent/CN108921162A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/14Image acquisition
    • G06V30/148Segmentation of character regions
    • G06V30/153Segmentation of character regions using recognition of characters or words
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Multimedia (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The embodiment of the present application discloses a kind of licence plate recognition method based on deep learning, and this method is applied to electronic device, and this method includes:Original input picture is obtained, the picture element matrix of the original input picture is extracted;Trained network model execution multilayer forward operation is preset into picture element matrix input and obtains forward operation as a result, determining license plate set of coordinates according to the forward operation result;The license plate range in the original input picture is determined according to the license plate set of coordinates, identifies the license plate content in the license plate range.

Description

Licence plate recognition method and Related product based on deep learning
Technical field
This application involves computer vision recognition technology fields, and in particular to a kind of Car license recognition side based on deep learning Method and Related product.
Background technique
Car plate detection and identification are sport technique segments important in intelligent transportation system, to other portions of intelligent transportation system Point, such as vehicle-logo recognition, body color analysis, model analysis, bayonet test, electronic police and other traffic events correlation modules The quality of processing have an important influence.
It is several to be roughly divided into car plate detection, License Plate, Character segmentation and character recognition etc. for current Vehicle License Plate Recognition System Step, but network model memory is big, arithmetic speed is slow, is all made of the same network model, effect to all types of license plates It is single, it can not precisely identify the license plate of all sizes.
Summary of the invention
The embodiment of the present application provides a kind of licence plate recognition method and Related product based on deep learning, to reduce net The memory of network model improves the speed of Car license recognition.
In a first aspect, the embodiment of the present application provides a kind of licence plate recognition method based on deep learning, the method application In electronic device, the method includes:
Original input picture is obtained, the picture element matrix of the original input picture is extracted;
Trained network model execution multilayer forward operation is preset into picture element matrix input and obtains forward operation knot Fruit determines license plate set of coordinates according to the forward operation result;
The license plate range in the original input picture is determined according to the license plate set of coordinates, is identified in the license plate range License plate content.
Second aspect, the embodiment of the present application provide a kind of electronic device of Car license recognition, and the electronic device includes:
Acquiring unit extracts the picture element matrix of the original input picture for obtaining original input picture;
Arithmetic element is obtained for trained network model execution multilayer forward operation to be preset in picture element matrix input To forward operation as a result, determining license plate set of coordinates according to the forward operation result;
Determination unit is identified for determining the license plate range in the original input picture according to the license plate set of coordinates License plate content in the license plate range.
The third aspect, the embodiment of the present application provide a kind of terminal, including one or more processors, one or more storages Device, one or more transceivers, and one or more programs, one or more of programs are stored in the memory In, and be configured to be executed by one or more of processors, described program includes for executing as described in relation to the first aspect The instruction of step in method.
Fourth aspect, the embodiment of the present application provide a kind of computer readable storage medium, and storage is handed over for electronic data The computer program changed, wherein the computer program makes the method for computer execution as described in relation to the first aspect.
5th aspect, the embodiment of the present application provide a kind of computer program product, and the computer program product includes depositing The non-transient computer readable storage medium of computer program is stored up, the computer is operable to make computer to execute such as the Method described in one side.
Implement the embodiment of the present application, has the advantages that:
As can be seen that establishing multiple convolutional layers in the network model of the embodiment of the present application, sufficiently extract in license plate image Vehicle license plate characteristic, and context module network structure of having connected, for license plate actual size by context module net The anchor of network structure is designed and sized to 1:2, to accurately determine license plate area, improve network model positioning licence plate coordinate Accuracy.
Detailed description of the invention
In order to more clearly explain the technical solutions in the embodiments of the present application, make required in being described below to embodiment Attached drawing is briefly described, it should be apparent that, the accompanying drawings in the following description is some embodiments of the present application, for ability For the those of ordinary skill of domain, without creative efforts, it can also be obtained according to these attached drawings other attached Figure.
Fig. 1 is a kind of flow diagram of licence plate recognition method based on deep learning provided by the embodiments of the present application;
Fig. 2 is a kind of structural schematic diagram of deep learning network model provided by the embodiments of the present application;
Fig. 3 is the knot of the inception network structure in a kind of deep learning network model provided by the embodiments of the present application Structure schematic diagram;
Fig. 4 is the context module network structure in a kind of deep learning network model provided by the embodiments of the present application Structural schematic diagram;
Fig. 5 is a kind of flow diagram of convolution algorithm disclosed in the embodiment of the present application;
Fig. 6 is the flow diagram of another kind convolution algorithm disclosed in the embodiment of the present application;
Fig. 7 is a kind of functional structure of the Car license recognition electronic device based on deep learning disclosed in the embodiment of the present application Figure.
Specific embodiment
Below in conjunction with the attached drawing in the embodiment of the present application, technical solutions in the embodiments of the present application carries out clear, complete Site preparation description, it is clear that described embodiment is some embodiments of the present application, instead of all the embodiments.Based on this Shen Please in embodiment, every other implementation obtained by those of ordinary skill in the art without making creative efforts Example, shall fall in the protection scope of this application.
The description and claims of this application and term " first ", " second ", " third " and " in the attached drawing Four " etc. are not use to describe a particular order for distinguishing different objects.In addition, term " includes " and " having " and it Any deformation, it is intended that cover and non-exclusive include.Such as it contains the process, method of a series of steps or units, be System, product or equipment are not limited to listed step or unit, but optionally further comprising the step of not listing or list Member, or optionally further comprising other step or units intrinsic for these process, methods, product or equipment.
Referenced herein " embodiment " is it is meant that the special characteristic, result or the characteristic that describe can wrap in conjunction with the embodiments It is contained at least one embodiment of the application.Each position in the description occur the phrase might not each mean it is identical Embodiment, nor the independent or alternative embodiment with other embodiments mutual exclusion.Those skilled in the art explicitly and Implicitly understand, embodiment described herein can be combined with other embodiments.
Electronic device involved by the embodiment of the present application may include the various handheld devices with wireless communication function, Mobile unit, wearable device calculate equipment or are connected to other processing equipments and various forms of radio modem User equipment (User Equipment, UE), mobile station (Mobile Station, MS), terminal device (terminal Device) etc..For convenience of description, apparatus mentioned above is referred to as terminal.Operation system involved by the embodiment of the present invention System is the software systems for being managed collectively to hardware resource, and providing a user business interface.
Refering to fig. 1, Fig. 1 is that a kind of process of licence plate recognition method based on deep learning provided by the embodiments of the present application is shown It is intended to, the method is applied to electronic device, the method includes:
Step S101, original input picture is obtained, the picture element matrix of the original input picture is extracted.
Optionally, after getting the original input picture, this is extracted to image progress gray proces and is originally inputted figure The tonal gradation of each pixel as in, group become the picture element matrix of the original input picture.Certainly, as original input picture is Color image can also extract the original input picture rgb pixel value, i.e. R picture element matrix, G picture element matrix and B picture element matrix, R picture element matrix, G picture element matrix and B picture element matrix group are become into triple channel input model.
Step S102, trained network model execution multilayer forward operation is preset in picture element matrix input to obtain just To operation result, license plate set of coordinates is determined according to the forward operation result.
Optionally, before trained network model is preset in picture element matrix input, initial depth is trained first Learning network model construction this preset trained network model.
Optionally, as shown in Fig. 2, the initial deep learning network model includes:First convolutional layer Conv1, the first pond Change layer Pool1, the second convolutional layer Conv2, the second pond layer Pool2, the first inception network structure, the 2nd inception Network structure, the 3rd inception network structure, third convolutional layer Conv3, Volume Four lamination Conv4, the first context Module network structure and the 2nd context module network structure, first convolutional layer Conv1, the first pond layer Pool1, Second convolutional layer Conv2, the second pond layer Pool2, the first inception network structure, the 2nd inception network structure, 3rd inception network structure, third convolutional layer Conv3 and Volume Four lamination Conv4 are sequentially connected in series, this first Context module network structure and the third convolutional layer Conv3 are connected in series, the 2nd context module network knot Structure and the Volume Four lamination Conv4 are connected in series.As shown in Fig. 2, the step-length stride=4 of the first convolutional layer Conv1;It is specific next The convolution kernel of the first convolutional layer Conv1 is said having a size of 7*7, the number of convolution kernel is 24;The stride=of first pond layer Pool1 2, convolution kernel size can be 3*3;The stride=2 of second convolutional layer Conv2, convolution kernel are having a size of 5*5, convolution kernel number 64;The stride=2 of second pond layer Pool2, convolution kernel is having a size of 3*3;The stride=1 of third convolutional layer Conv3, convolution For core having a size of 1*1, convolution kernel number is 128;The stride=2 of Volume Four lamination Conv4, convolution kernel is having a size of 3*3, convolution kernel Number is 256, it can be seen that total step-length of first convolutional layer, the first pond layer, the second convolutional layer and the second pond layer Stride=4*2*2*2=32, therefore in convolution algorithm, 32 times of picture element matrix rapid drop of original input picture are compared Current car plate detection network improves the speed for extracting vehicle license plate characteristic.
Optionally, the convolution kernel size in each convolutional layer in the initial deep learning network model can for 7*7, 5*5,3*3 or other values, above-mentioned convolution kernel size are to do for example, but not limiting the tool of convolution kernel in each convolutional layer Body size.
Optionally, the first inception network structure, the 2nd inception network structure and the third Inception network structure is all made of inception-v1Network structure, due to the first inception network structure, second Inception network structure and the 3rd inception network structure are all made of inception-v1Network structure, therefore with first The composition of inception network structure is illustrated for inception network structure.As shown in figure 3, this first Inception network structure includes four branches, wherein the first branch only includes a convolutional layer, and the convolution of the convolutional layer For core having a size of 1*1, convolution kernel number is 32;Second branch includes a pond layer and a convolutional layer, and the volume of the pond layer Product core is having a size of 3*3, and for the convolution kernel of the convolutional layer having a size of 1*1, convolution kernel number is 32;Third branch includes two convolution Layer, the convolution kernel of first convolutional layer is having a size of 1*1, and convolution nuclear volume is 24, and the convolution kernel of second convolutional layer is having a size of 3* 3, convolution kernel number is 32;4th branch includes three convolutional layers, and the convolution kernel of first convolutional layer is having a size of 1*1, convolution Core number is 24, and the convolution kernel of second convolutional layer is having a size of 3*3, and the number of convolution kernel is 32, the convolution of third convolutional layer For core having a size of 3*3, convolution kernel number is 32.It is understood that total step-length of this four branches is identical, i.e., each branch's output Characteristic pattern dimension it is identical.For example, as the first branch convolutional layer stride=4, then the pond layer of the second branch with The stride of convolutional layer is respectively 2 and 2, i.e., total step-length is 4, and the step-length of two convolutional layers of third branch may be alternatively provided as 2 Hes 2, three convolutional layers of the 4th branch can be set to 1,2 and 2, that is, guarantee that total step-length of each branch in four branches is 4.
Spy of the concat function for exporting four branches in further, in the first inception network structure Sign figure carries out dimension splicing, in order to increase the width width of network model.
Optionally, the first context module network structure and the 2nd context module network structure are main For multiple characteristic patterns of the convolutional layer output of front to be carried out full attended operation, a various dimensions feature vector is obtained, that is, is given birth to At a characteristic pattern.
Optionally, the first context module network structure and the 2nd context module network structure are adopted With the building region candidate network RPN of the convolutional network FAST R-CNN based on region (Region Proposal Network, Referred to as:RPN strategy) constructs respective RPN to obtain respective anchor coordinate, due to the first context module Network structure and the 2nd context module network structure, which are all made of in FAST R-CNN, constructs the strategy of RPN to construct The RPN of first context module network structure and the 2nd context module network structure, therefore the first context Module network structure and the 2nd context module network structure structure having the same, below with the first context The structure of context module network structure is illustrated for module network structure.As shown in figure 4, this first Context module network structure includes Liang Ge branch, wherein and first branch only includes the convolutional layer of a 3*3, and second A branch includes the convolutional layer of a 3*3, and the convolutional layer of the 3*3 includes Kuo Liangge sub-branch again, and first sub-branch includes The convolutional layer of one 3*3, second sub-branch include the convolutional layer of two concatenated 3*3.Such as above-mentioned inception network knot Total step-length of structure, the first context module Liang Ge branch is identical, no longer describes herein.
Further, the anchor dimension scale of the first context module network structure is respectively 16*32,32* 64 and 64*128, wherein the anchor number of dimension scale 16*32,32*64 and 64*128 are respectively 16,4 and 1, this second The anchor ratio of context module network structure is respectively 128*256 and 256*512, wherein dimension scale 128* The anchor number of 256 and 256*512 is respectively 1 and 1, further, dimension scale 16*32,32*64,64*128,128* The anchor separation of 256 and 256*512 is respectively 8,16,32,64 and 64.
Optionally, the first context module network structure and the 2nd context module network structure is complete Articulamentum loss function is:
Wherein, LclsIt is the objective function of classification task, using Cross Entropy function, is mainly used for estimation range It is divided into contexts region;LregIt is the loss function of Detection task, uses smooth L1 function;pi, tiIt is predicted value respectively With prediction coordinate, pi *, ti *It is ground-truth label and actual coordinate respectively.
Optionally, the initial deep learning network model specific method of training includes:Training sample is obtained, by the training sample Originally it is input to initial network model execution multilayer forward operation and obtains the prediction license plate set of coordinates of the training sample, according to preset The anchor ratio-dependent prediction license plate set of coordinates corresponding multiple prediction license plate frame regions in the training sample, it is more to obtain this Degree of overlapping IOU (Intersection over Union, the abbreviation of a prediction license plate frame region and actual license plate frame region:IOU) Sample greater than preset threshold is positive sample, and the sample less than the preset threshold is negative sample, according to the positive sample and the negative sample This determination exports result gradient, and the output result gradient execution reversed operation of multilayer is updated the initial deep learning network mould The weight gradient of every layer of convolutional layer in type obtains this and presets trained network model.
Wherein, which is specifically as follows 0.35,0.5,0.7 or other values, is usually set to 0.35 or 0.5.
Optionally, if the picture element matrix of original input picture is single channel input data (the corresponding pixel of such as gray level image Matrix or the corresponding picture element matrix of bianry image), for convolution operation, not due to convolution kernel size and the picture element matrix Picture element matrix need to be cut into multiple basic granularities when executing convolution operation and carry out convolution with convolution kernel respectively by matching.Such as Fig. 5 Shown, Fig. 5 shows the picture element matrix that input data is 9*9, and convolution kernel is having a size of 3*3, when stride=2, generates 4*4 output The convolution process of matrix.
Optionally, as picture element matrix be multichannel input data (such as rgb pixel matrix of color image), in every secondary volume When product operation, the characteristic pattern that each channel obtains is overlapped and generates the final characteristic pattern of the secondary convolution operation, therefore characteristic pattern Quantity it is related with the quantity of convolution kernel, it is unrelated with input channel number.It is rolled up as shown in fig. 6, Fig. 6 shows RGB triple channel Product operation is to generate the detailed process of characteristic pattern, wherein three picture element matrixs input size of RGB is 9*9, convolution kernel size 3*3*3,3*3*3 refers to the convolution kernel for having a 3*3 on each channel here, without referring to three-dimensional data, therefore the volume in each channel A 4*4 output matrix is generated after product core and respective picture element matrix convolution, three 4*4 output matrixes is superimposed to obtain final 4*4 characteristic pattern.
In the application, illustrated so that picture element matrix is single channel input data as an example.
Optionally, it which is inputted this presets trained network model and executes multilayer forward operation and obtain forward direction Operation result determines license plate set of coordinates according to the forward operation result, specifically includes:The picture element matrix is inputted into the first volume Lamination carries out convolution algorithm and exports 24 fisrt feature figures, which is input to the second pond layer and carries out pond Change processing 24 the first ponds of output as a result, this 24 the first pond results, which are input to second convolutional layer, carries out convolution algorithm 64 second feature figures are exported, which is input to the second pond layer carries out pondization processing and export 64 the Two ponds as a result, this 64 the second pond results are input to the first inception network structure, by this first The convolution algorithm of inception network structure, the 2nd inception network structure and the 3rd inception network structure 128 third feature figures are exported afterwards, which is input to the third convolutional layer and carries out convolution algorithm output 128 fourth feature figures are inputted the first context module network structure respectively and rolled up by 128 fourth feature figures Product operation obtains fifth feature figure, obtains N number of first anchor coordinate of N number of characteristic point in the fifth feature figure, according to this Anchor dimension scale in one context module network structure determines N number of first anchor coordinate corresponding N number of One license plate range extracts M the first license plate ranges that IOU in N number of first license plate range is greater than the preset threshold, determines the M Corresponding M the first anchor coordinate of a first license plate range is the license plate set of coordinates, wherein N, the M are positive integer, M≤N. Alternatively, 128 fourth feature figures are inputted Volume Four product 256 sixth feature figures of output, by 256 sixth feature figures The 2nd context module network structure progress convolution algorithm is inputted respectively and obtains seventh feature figure, obtains the seventh feature P the 2nd anchor coordinates of P characteristic point in figure, according to the anchor ruler in the 2nd context module network structure It is big to extract IOU in the P the second license plate ranges for corresponding P the second license plate range of very little the 2nd anchor coordinate of ratio-dependent In Q the second license plate ranges of the preset threshold, determine that corresponding Q the 2nd anchor coordinate of Q the second license plate range is to be somebody's turn to do License plate set of coordinates, P, the Q are positive integer, P≤Q.
The first context network structure and the 2nd context network structure is arranged in size based on license plate, it can be seen that The anchor ratio of 2nd context network structure is 2 times of the first context network structure, therefore the first context network structure Mainly for small size license plate, the 2nd context network structure is mainly for large-sized license plate, therefore when identifying small license plate, It obtains and predicts license plate range identification license plate content in the first context network structure, when identifying big license plate, obtain second Predict that license plate range identifies license plate content in context network structure,
Step S103, the license plate range in the original input picture is determined according to the license plate set of coordinates, described in identification License plate content in license plate range.
Based on the license plate set of coordinates in step S102, license plate range is determined according to anchor dimension scale, to license plate range Interior license plate image carries out Character segmentation, identifies the corresponding license plate content of the license plate image.
As can be seen that three inception structures of having connected in the network model of the embodiment of the present application, pass through concat letter Number increases the width of network model, and the network model is made to adapt to the license plate of various sizes, improves network model identification license plate Precision, in addition, the actual ratio based on license plate constructs 1:2 anchor, improve identification license plate range efficiency and Precision, and network model extraction vehicle license plate characteristic speed is fast, and memory is small.
It is consistent with above-mentioned Fig. 1 embodiment, referring to Fig. 7, Fig. 7 is provided by the embodiments of the present application a kind of based on depth The possible functional unit of the electronic device 700 of habit forms block diagram, which includes:Acquiring unit 710, operation list First 720, determination unit 730, wherein;
Acquiring unit 710 extracts the picture element matrix of the original input picture for obtaining original input picture;
Arithmetic element 720 executes multilayer forward direction fortune for trained network model to be preset in picture element matrix input Calculation obtains forward operation as a result, determining license plate set of coordinates according to the forward operation result;
Determination unit 730 is known for determining the license plate range in the original input picture according to the license plate set of coordinates License plate content in the not described license plate range.
The embodiment of the present application also provides a kind of computer storage medium, wherein computer storage medium storage is for electricity The computer program of subdata exchange, it is as any in recorded in above method embodiment which execute computer A kind of some or all of the licence plate recognition method based on deep learning step.
The embodiment of the present application also provides a kind of computer program product, and the computer program product includes storing calculating The non-transient computer readable storage medium of machine program, the computer program are operable to that computer is made to execute such as above-mentioned side Some or all of any licence plate recognition method based on deep learning recorded in method embodiment step.
It should be noted that for the various method embodiments described above, for simple description, therefore, it is stated as a series of Combination of actions, but those skilled in the art should understand that, the application is not limited by the described action sequence because According to the application, some steps may be performed in other sequences or simultaneously.Secondly, those skilled in the art should also know It knows, embodiment described in this description belongs to alternative embodiment, related actions and modules not necessarily the application It is necessary.
In the above-described embodiments, it all emphasizes particularly on different fields to the description of each embodiment, there is no the portion being described in detail in some embodiment Point, reference can be made to the related descriptions of other embodiments.
In several embodiments provided herein, it should be understood that disclosed device, it can be by another way It realizes.For example, the apparatus embodiments described above are merely exemplary, such as the division of the unit, it is only a kind of Logical function partition, there may be another division manner in actual implementation, such as multiple units or components can combine or can To be integrated into another system, or some features can be ignored or not executed.Another point, shown or discussed is mutual Coupling, direct-coupling or communication connection can be through some interfaces, the indirect coupling or communication connection of device or unit, It can be electrical or other forms.
The unit as illustrated by the separation member may or may not be physically separated, aobvious as unit The component shown may or may not be physical unit, it can and it is in one place, or may be distributed over multiple In network unit.It can select some or all of unit therein according to the actual needs to realize the mesh of this embodiment scheme 's.
It, can also be in addition, each functional unit in each embodiment of the application can integrate in one processing unit It is that each unit physically exists alone, can also be integrated in one unit with two or more units.Above-mentioned integrated list Member both can take the form of hardware realization, can also be realized in the form of software program module.
If the integrated unit is realized in the form of software program module and sells or use as independent product When, it can store in a computer-readable access to memory.Based on this understanding, the technical solution of the application substantially or Person says that all or part of the part that contributes to existing technology or the technical solution can body in the form of software products Reveal and, which is stored in a memory, including some instructions are used so that a computer equipment (can be personal computer, server or network equipment etc.) executes all or part of each embodiment the method for the application Step.And memory above-mentioned includes:USB flash disk, read-only memory (ROM, Read-Only Memory), random access memory The various media that can store program code such as (RAM, Random Access Memory), mobile hard disk, magnetic or disk.
Those of ordinary skill in the art will appreciate that all or part of the steps in the various methods of above-described embodiment is can It is completed with instructing relevant hardware by program, which can store in a computer-readable memory, memory May include:Flash disk, read-only memory (English:Read-Only Memory, referred to as:ROM), random access device (English: Random Access Memory, referred to as:RAM), disk or CD etc..
The embodiment of the present application is described in detail above, specific case used herein to the principle of the application and Embodiment is expounded, the description of the example is only used to help understand the method for the present application and its core ideas; At the same time, for those skilled in the art can in specific embodiments and applications according to the thought of the application There is change place, in conclusion the contents of this specification should not be construed as limiting the present application.

Claims (10)

1. a kind of licence plate recognition method based on deep learning, the method is applied to electronic device, which is characterized in that the side Method includes:
Original input picture is obtained, the picture element matrix of the original input picture is extracted;
Trained network model execution multilayer forward operation is preset into picture element matrix input and obtains forward operation as a result, root License plate set of coordinates is determined according to the forward operation result;
The license plate range in the original input picture is determined according to the license plate set of coordinates, identifies the vehicle in the license plate range Board content.
2. the method is also wrapped the method according to claim 1, wherein described before obtaining original image It includes:
Training sample is obtained, the training sample is input to initial network model execution multilayer forward operation and obtains the training The prediction license plate set of coordinates of sample, prediction license plate set of coordinates is in the training sample according to preset anchor ratio-dependent In corresponding multiple prediction license plate frame regions, obtain it is the multiple prediction license plate frame region and actual license plate frame region degree of overlapping The sample that IOU is greater than preset threshold is positive sample, and the sample less than preset threshold is negative sample, according to the positive sample and described Negative sample determines output result gradient, and the output result gradient is executed the reversed operation of multilayer and updates the initial network model In the weight gradient of every layer of convolutional layer obtain described presetting trained network model.
3. according to the method described in claim 2, it is characterized in that, the initial network model includes:First convolutional layer, first Pond layer, the second convolutional layer, the second pond layer, the first inception network structure, the 2nd inception network structure, third Inception network structure, third convolutional layer, Volume Four lamination, the first context module network structure and second Context module network structure, first convolutional layer, first pond layer, second convolutional layer, described second Pond layer, the first inception network structure, the 2nd inception network structure, the 3rd inception Network structure, the third convolutional layer and the Volume Four lamination are sequentially connected in series, the first context module net Network structure and the third convolutional layer are connected in series, the 2nd context module network structure and the Volume Four lamination It is connected in series.
4. according to the method described in claim 3, it is characterized in that, the first inception network structure, described second Inception network structure and the 3rd inception network structure are all made of inception-v1Network structure.
5. according to the method described in claim 2, it is characterized in that, first convolutional layer, first pond layer, described Two convolutional layers, second pond layer, the third convolutional layer and the Volume Four lamination step-length be respectively 4,2,2,2,1 and 2, the convolution kernel number of first convolutional layer is 24, and the convolution kernel number of second convolutional layer is 64, the third convolution The convolution kernel number of layer is 128, and the convolution kernel number of the Volume Four lamination is 256.
6. according to the method described in claim 5, it is characterized in that, the first context module network structure and described 2nd context module network structure is using building region candidate network in the convolutional network FAST R-CNN based on region The respective RPN of the construction of strategy of RPN is to obtain respective anchor coordinate, and the first context module network knot The anchor dimension scale of structure and the 2nd context module network structure is set as 1:2, and described first Context module network structure is different with the anchor size of the 2nd context module network structure.
7. according to the method described in claim 6, it is characterized in that, the first context module network structure Anchor dimension scale is respectively 16*32,32*64 and 64*128, wherein dimension scale 16*32,32*64 and 64*128's Anchor number is respectively 16,4 and 1, and the anchor dimension scale of the 2nd context module network structure is respectively 128*256 and 256*512, wherein the anchor number that dimension scale is 128*256 and 256*512 is respectively 1 and 1, the ruler Very little ratio is that the anchor separation of 16*32,32*64,64*128,128*256 and 256*512 are respectively 8,16,32,64 and 64。
8. method as claimed in claim 2, which is characterized in that described that trained network mould is preset in picture element matrix input Type executes multilayer forward operation and obtains forward operation as a result, determining license plate set of coordinates according to the forward operation result, including:
The picture element matrix is inputted into first convolutional layer and exports 24 fisrt feature figures, 24 fisrt feature figures is defeated Enter to second pond layer and exports 24 the first ponds as a result, 24 first pond results are input to the volume Two Lamination exports 64 second feature figures, and 64 second feature figures are input to second pond layer and export 64 the second ponds Change as a result, 64 second pond results are input to the first inception network structure, by described first The convolution of inception network structure, the 2nd inception network structure and the 3rd inception network structure 128 third feature figures are exported after operation, and 128 third feature figures are input to the third convolutional layer and export 128 128 fourth feature figures are inputted the first context module network structure respectively and obtain by fourth feature figure Five characteristic patterns obtain N number of first anchor coordinate of N number of characteristic point in the fifth feature figure, according to the first context Anchor dimension scale in module network structure determines the corresponding N number of first license plate model of N number of first anchor coordinate It encloses, extracts M the first license plate ranges that IOU in N number of first license plate range is greater than the preset threshold, determine the M Corresponding M the first anchor coordinate of first license plate range is the license plate set of coordinates, and described N, M are positive integer, M≤N.
9. according to the method described in claim 8, it is characterized in that, the method also includes:
128 fourth feature figures are inputted into Volume Four product 256 sixth feature figures of output, by described 256 the 6th Characteristic pattern inputs the 2nd context module network structure respectively and obtains seventh feature figure, obtains the seventh feature figure P the 2nd anchor coordinates of middle P characteristic point, according to the anchor ratio in the 2nd context module network structure Example determines corresponding P the second license plate range of P the 2nd anchor coordinate, extracts IOU in the P the second license plate ranges Greater than Q the second license plate ranges of the preset threshold, determine that corresponding Q the 2nd anchor of the second license plate of Q range is sat It is designated as the license plate set of coordinates, described P, Q are positive integer, P≤Q.
10. a kind of Car license recognition electronic device based on deep learning, which is characterized in that the electronic device includes:
Acquiring unit extracts the picture element matrix of the original input picture for obtaining original input picture;
Arithmetic element obtains just for trained network model execution multilayer forward operation to be preset in picture element matrix input To operation result, license plate set of coordinates is determined according to the forward operation result;
Determination unit, for determining the license plate range in the original input picture according to the license plate set of coordinates, described in identification License plate content in license plate range.
CN201810593633.6A 2018-06-11 2018-06-11 Licence plate recognition method and Related product based on deep learning Pending CN108921162A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810593633.6A CN108921162A (en) 2018-06-11 2018-06-11 Licence plate recognition method and Related product based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810593633.6A CN108921162A (en) 2018-06-11 2018-06-11 Licence plate recognition method and Related product based on deep learning

Publications (1)

Publication Number Publication Date
CN108921162A true CN108921162A (en) 2018-11-30

Family

ID=64418043

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810593633.6A Pending CN108921162A (en) 2018-06-11 2018-06-11 Licence plate recognition method and Related product based on deep learning

Country Status (1)

Country Link
CN (1) CN108921162A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI670610B (en) * 2018-12-06 2019-09-01 中華電信股份有限公司 Customizable method for detecting specific style objects
CN110378245A (en) * 2019-06-26 2019-10-25 平安科技(深圳)有限公司 Football match Activity recognition method, apparatus and terminal device based on deep learning
CN111028160A (en) * 2019-11-21 2020-04-17 西北工业大学 Remote sensing image noise suppression method based on convolutional neural network
CN111460996A (en) * 2020-03-31 2020-07-28 东南大学 Truck number plate identification method based on extreme learning deep network fusion model
CN113139579A (en) * 2021-03-23 2021-07-20 广东省科学院智能制造研究所 Image classification method and system based on image feature adaptive convolution network

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107679531A (en) * 2017-06-23 2018-02-09 平安科技(深圳)有限公司 Licence plate recognition method, device, equipment and storage medium based on deep learning

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107679531A (en) * 2017-06-23 2018-02-09 平安科技(深圳)有限公司 Licence plate recognition method, device, equipment and storage medium based on deep learning

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
CHRISTIAN SZEGEDY ET AL.: "Rethinking the Inception Architecture for Computer Vision", 《ARXIV》 *
CHRISTIAN SZEGEDY: "Going deeper with convolutions", 《2015 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 *
SHAOQING REN: "Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks", 《IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE》 *
赵振兴: "基于深度学习的车牌识别技术研究", 《中国优秀硕士学位论文全文数据库 工程科技Ⅱ辑》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI670610B (en) * 2018-12-06 2019-09-01 中華電信股份有限公司 Customizable method for detecting specific style objects
CN110378245A (en) * 2019-06-26 2019-10-25 平安科技(深圳)有限公司 Football match Activity recognition method, apparatus and terminal device based on deep learning
CN110378245B (en) * 2019-06-26 2023-07-21 平安科技(深圳)有限公司 Football match behavior recognition method and device based on deep learning and terminal equipment
CN111028160A (en) * 2019-11-21 2020-04-17 西北工业大学 Remote sensing image noise suppression method based on convolutional neural network
CN111460996A (en) * 2020-03-31 2020-07-28 东南大学 Truck number plate identification method based on extreme learning deep network fusion model
CN111460996B (en) * 2020-03-31 2022-02-15 东南大学 Truck number plate identification method based on extreme learning deep network fusion model
CN113139579A (en) * 2021-03-23 2021-07-20 广东省科学院智能制造研究所 Image classification method and system based on image feature adaptive convolution network
CN113139579B (en) * 2021-03-23 2024-02-02 广东省科学院智能制造研究所 Image classification method and system based on image feature self-adaptive convolution network

Similar Documents

Publication Publication Date Title
CN108921162A (en) Licence plate recognition method and Related product based on deep learning
CN109522874B (en) Human body action recognition method and device, terminal equipment and storage medium
CN108537292A (en) Semantic segmentation network training method, image, semantic dividing method and device
CN108830211A (en) Face identification method and Related product based on deep learning
CN111860398B (en) Remote sensing image target detection method and system and terminal equipment
CN105930402A (en) Convolutional neural network based video retrieval method and system
CN113822209B (en) Hyperspectral image recognition method and device, electronic equipment and readable storage medium
CN112163601B (en) Image classification method, system, computer device and storage medium
CN108846404B (en) Image significance detection method and device based on related constraint graph sorting
CN112712546A (en) Target tracking method based on twin neural network
CN109146891B (en) Hippocampus segmentation method and device applied to MRI and electronic equipment
CN111626184A (en) Crowd density estimation method and system
CN104699781B (en) SAR image search method based on double-deck anchor figure hash
CN107506350A (en) A kind of method and apparatus of identification information
CN114783021A (en) Intelligent detection method, device, equipment and medium for wearing of mask
CN109740752A (en) Depth model training method and device, electronic equipment and storage medium
CN108875826B (en) Multi-branch object detection method based on coarse and fine granularity composite convolution
CN111709415B (en) Target detection method, device, computer equipment and storage medium
CN108197567A (en) For the method, apparatus and computer-readable medium of image procossing
CN109901835B (en) Method, device and equipment for laying out elements and storage medium
CN111145196A (en) Image segmentation method and device and server
CN106503386A (en) The good and bad method and device of assessment luminous power prediction algorithm performance
CN111027670B (en) Feature map processing method and device, electronic equipment and storage medium
CN116227573B (en) Segmentation model training method, image segmentation device and related media
CN114494782B (en) Image processing method, model training method, related device and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information

Address after: Room 1301, No.132 Fengqi Road, phase III, software park, Xiamen City, Fujian Province

Applicant after: Xiamen Entropy Technology Co., Ltd

Address before: 361000, Xiamen three software park, Fujian Province, 8 North Street, room 2001

Applicant before: XIAMEN ZKTECO BIOMETRIC IDENTIFICATION TECHNOLOGY Co.,Ltd.

CB02 Change of applicant information
RJ01 Rejection of invention patent application after publication

Application publication date: 20181130

RJ01 Rejection of invention patent application after publication