CN110298791B - Super-resolution reconstruction method and device for license plate image - Google Patents

Super-resolution reconstruction method and device for license plate image Download PDF

Info

Publication number
CN110298791B
CN110298791B CN201910610927.XA CN201910610927A CN110298791B CN 110298791 B CN110298791 B CN 110298791B CN 201910610927 A CN201910610927 A CN 201910610927A CN 110298791 B CN110298791 B CN 110298791B
Authority
CN
China
Prior art keywords
feature maps
local
branch
deep
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910610927.XA
Other languages
Chinese (zh)
Other versions
CN110298791A (en
Inventor
王殿伟
郝元杰
宋鸽
李娜
李大湘
刘颖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian University of Posts and Telecommunications
Original Assignee
Xian University of Posts and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian University of Posts and Telecommunications filed Critical Xian University of Posts and Telecommunications
Priority to CN201910610927.XA priority Critical patent/CN110298791B/en
Publication of CN110298791A publication Critical patent/CN110298791A/en
Application granted granted Critical
Publication of CN110298791B publication Critical patent/CN110298791B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/62Text, e.g. of license plates, overlay texts or captions on TV images
    • G06V20/625License plates

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Multimedia (AREA)
  • Image Processing (AREA)

Abstract

The disclosure relates to the field of image processing, in particular to a super-resolution reconstruction method and device for a license plate image, and is applied to a terminal. The method comprises the following steps: firstly, shallow feature extraction is carried out on a low-resolution license plate image to be processed through a first convolution layer; secondly, depth feature extraction is carried out through N identical residual error sampling blocks and a second convolution layer which are sequentially connected in series, and different branches built in the residual error sampling blocks fully utilize the scale and depth information of features; and finally, amplifying the plurality of deep characteristic maps obtained by the characteristic extraction module by multiple times step by step through a reconstruction module, and obtaining a reconstructed target high-resolution license plate image according to the plurality of high-resolution characteristic maps. According to the method, the problem that the extracted characteristic information is lost is solved by performing super-resolution reconstruction processing on the low-resolution license plate image, so that the reconstructed license plate avoids the conditions of character blurring, character adhesion, insufficient information and the like, and the recognition degree of the license plate image is improved.

Description

Super-resolution reconstruction method and device for license plate image
Technical Field
The disclosure relates to the field of image processing, in particular to a super-resolution reconstruction method and device for a license plate image.
Background
The super-resolution reconstruction of the license plate image means that the corresponding most possible high-resolution license plate image is obtained from the degraded low-resolution license plate image by a software method. The super-resolution reconstruction technology is widely applied and mainly applied to the fields of public safety, medical images, remote sensing images, high-definition televisions and the like, so the super-resolution reconstruction of the images is very important.
At present, a super-resolution reconstruction method of a license plate image is an interpolation-based method, and the method obtains a value of an estimation point by using a neighborhood sampling point of a low-resolution image through a certain functional relationship, and inserts the value between the neighborhood sampling points. However, the number plate image reconstructed by the interpolation-based method has less detail information, and characters in the number plate image are adhered to each other, so that the characters in the number plate image are difficult to recognize, that is, the recognition degree of the number plate image is low.
Disclosure of Invention
In order to solve the problems in the prior art, the embodiment of the disclosure provides a super-resolution reconstruction method and device for a license plate image. The technical scheme is as follows:
on one hand, the super-resolution reconstruction method of the license plate image is provided and applied to a terminal, an image processing network is deployed on the terminal and comprises a feature extraction module and a reconstruction module, the feature extraction module comprises an initialization feature extraction module and a depth feature extraction module which are connected, the initialization feature extraction module comprises a first convolution layer, the depth feature extraction module comprises N identical residual sampling blocks and a second convolution layer which are sequentially connected in series, N is an integer greater than or equal to 1, each residual sampling block comprises a first branch, a second branch and a third branch, the first branch comprises an upper sampling layer, a third convolution layer and a lower sampling layer which are sequentially connected, the second branch comprises a fourth convolution layer, the third branch is an identity mapping branch, an output feature map of the identity mapping branch is identical to an input feature map, and the method comprises the following steps:
performing shallow feature extraction on the low-resolution license plate image to be processed through the first convolution layer to obtain a plurality of shallow feature maps;
carrying out first depth feature extraction on the shallow feature maps through the N identical residual error sampling blocks which are sequentially connected in series to obtain a plurality of local fourth deep feature maps output by the Nth residual error sampling block;
performing second depth feature extraction on a plurality of local fourth deep feature maps output by the Nth residual sampling block through the second convolution layer to obtain a plurality of fifth deep feature maps, and performing feature fusion on the plurality of fifth deep feature maps and the plurality of shallow feature maps in a one-to-one correspondence manner to obtain a plurality of sixth deep feature maps;
the reconstruction module is used for amplifying the sixth deep feature maps obtained by the feature extraction module by multiple times step by step to obtain a plurality of high-resolution feature maps, and a target high-resolution license plate image is obtained according to the high-resolution feature maps and is obtained by performing super-resolution reconstruction on the low-resolution license plate image through the image processing network;
the first depth feature extraction is performed on the multiple shallow feature maps through the N identical residual sampling blocks connected in series in sequence to obtain multiple local fourth deep feature maps output by the nth residual sampling block, and the method includes:
carrying out scale transformation on a plurality of target feature maps through a first branch in an ith residual sampling block in the N same residual sampling blocks which are sequentially connected in series to obtain a plurality of local first deep feature maps, carrying out feature extraction on the plurality of target feature maps through a second branch in the ith residual sampling block to obtain a plurality of local second deep feature maps, and carrying out identity mapping on the plurality of target feature maps through a third branch in the ith residual sampling block to obtain a plurality of local third deep feature maps; performing feature fusion on the plurality of local first deep feature maps, the plurality of local second deep feature maps and the plurality of local third deep feature maps in a one-to-one correspondence manner through the ith residual sampling block to obtain a plurality of local fourth deep feature maps; acquiring a plurality of local fourth deep feature maps output by the Nth residual error sampling block; when i is not equal to 1, the target feature maps are local fourth deep feature maps output by the i-1 th residual sample block.
Optionally, the performing scale transformation on the multiple target feature maps through a first branch in an ith residual sample block of the N identical residual sample blocks connected in series in sequence to obtain multiple local first deep feature maps includes:
the target feature maps are up-sampled through an up-sampling layer in the first branch to obtain a plurality of up-sampling feature maps;
performing feature extraction on the plurality of upsampled feature maps through a third convolution layer in the first branch to obtain a plurality of third convolution feature maps;
and downsampling the plurality of third convolution feature maps through a downsampling layer in the first branch to obtain a plurality of downsampling feature maps, and determining the plurality of downsampling feature maps as the plurality of local first deep feature maps, wherein the upsampling adopts transposition convolution and the downsampling adopts convolution.
Optionally, the upsampling the multiple target feature maps by the upsampling layer in the first branch to obtain multiple upsampled feature maps includes:
through an upsampling layer in the first branch, upsampling the multiple target feature maps according to a first formula as follows to obtain multiple upsampled feature maps:
the first formula:
Figure BDA0002122343600000031
wherein the content of the first and second substances,
Figure BDA0002122343600000032
is the ith residueA plurality of upsampled feature maps of the difference sample blocks phi being the non-linear mapping function or activation function of the ith residual sample block R i-1 For the plurality of target feature maps, W i And b i Weights and offsets, respectively, # e @ f, for the ith residual sample block s Represents the upsampling by a factor of s;
performing feature extraction on the plurality of upsampled feature maps through a third convolutional layer in the first branch to obtain a plurality of third convolutional feature maps, including:
performing feature extraction on the plurality of upsampled feature maps through a third convolution layer in the first branch according to a second formula as follows to obtain a plurality of third convolution feature maps:
the second formula:
Figure BDA0002122343600000033
wherein the content of the first and second substances,
Figure BDA0002122343600000034
a plurality of third convolution maps for an ith residual sample block;
the downsampling the plurality of third convolution feature maps by the downsampling layer in the first branch to obtain a plurality of downsampled feature maps includes:
downsampling the plurality of third convolution feature maps according to a third formula to obtain a plurality of downsampled feature maps by using a downsampling layer in the first branch:
the third formula:
Figure BDA0002122343600000035
wherein the content of the first and second substances,
Figure BDA0002122343600000036
for a plurality of downsampled feature maps of the ith residual sample block, ↓ s is downsampled by s times.
The depth feature extraction is performed on the multiple target feature maps through a second branch in the ith residual sampling block to obtain multiple local second deep feature maps, and the method comprises the following steps:
performing feature extraction on the plurality of target feature maps through a second branch in the ith residual sampling block according to a fourth formula to obtain a plurality of local second deep feature maps:
the fourth formula:
Figure BDA0002122343600000037
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0002122343600000038
a plurality of local second depth profiles are represented.
Optionally, the performing feature fusion on the plurality of local first deep feature maps, the plurality of local second deep feature maps and the plurality of local third deep feature maps through the i-th residual sampling block to obtain a plurality of local fourth deep feature maps includes:
performing feature fusion on the plurality of local first deep feature maps, the plurality of local second deep feature maps and the plurality of local third deep feature maps according to a fifth formula to obtain a plurality of local fourth deep feature maps by using the ith residual sampling block:
the fifth formula:
Figure BDA0002122343600000041
wherein R is i A plurality of local fourth deep feature maps output for the ith residual sample block,
Figure BDA0002122343600000042
a plurality of local first deep feature maps from a first branch of an ith residual sample block,
Figure BDA0002122343600000043
multiple local second deep feature maps, R, obtained for a second branch in an ith residual sample block i-1 Multiple bureaus derived for the third branch of the ith residual sample blockAnd (c) a third deep level feature map.
On the other hand, a super-resolution reconstruction device for license plate images is provided, which is applied to a terminal, an image processing network is deployed on the terminal, the image processing network includes a feature extraction module and a reconstruction module, the feature extraction module includes an initialization feature extraction module and a depth feature extraction module which are connected, the initialization feature extraction module includes a first convolution layer, the depth feature extraction module includes N same residual sampling blocks and a second convolution layer which are connected in series in sequence, N is an integer greater than or equal to 1, each residual sampling block includes a first branch, a second branch and a third branch, the first branch includes an upper sampling layer, a third convolution layer and a lower sampling layer which are connected in sequence, the second branch includes a fourth convolution layer, the third branch is an identity mapping branch, an output feature map of the identity mapping branch is the same as an input feature map, and the device includes:
the initialization feature extraction module is used for performing shallow feature extraction on the low-resolution license plate image to be processed through the first convolution layer to obtain a plurality of shallow feature maps;
the depth feature extraction module is used for performing first depth feature extraction on the plurality of shallow feature maps through the N same residual error sampling blocks which are sequentially connected in series to obtain a plurality of local fourth deep feature maps output by the Nth residual error sampling block;
the feature fusion module is configured to perform second depth feature extraction on a plurality of local fourth deep feature maps output by the nth residual sampling block through the second convolution layer to obtain a plurality of fifth deep feature maps, and perform feature fusion on the plurality of fifth deep feature maps and the plurality of shallow feature maps in a one-to-one correspondence manner to obtain a plurality of sixth deep feature maps;
the super-resolution reconstruction module is used for amplifying the sixth deep feature maps obtained by the feature extraction module by multiple times step by step through the reconstruction module to obtain a plurality of high-resolution feature maps, and obtaining a target high-resolution license plate image according to the high-resolution feature maps, wherein the target high-resolution license plate image is a license plate image obtained by performing super-resolution reconstruction on the low-resolution license plate image through the image processing network;
the depth feature extraction module is further configured to perform scale transformation on multiple target feature maps through a first branch in an ith residual sampling block of the N identical residual sampling blocks connected in series in sequence to obtain multiple local first deep feature maps, perform feature extraction on the multiple target feature maps through a second branch in the ith residual sampling block to obtain multiple local second deep feature maps, and perform identity mapping on the multiple target feature maps through a third branch in the ith residual sampling block to obtain multiple local third deep feature maps; performing feature fusion on the plurality of local first deep feature maps, the plurality of local second deep feature maps and the plurality of local third deep feature maps in a one-to-one correspondence manner through the ith residual sampling block to obtain a plurality of local fourth deep feature maps; acquiring a plurality of local fourth deep feature maps output by the Nth residual error sampling block; when i is not equal to 1, the target feature maps are local fourth deep feature maps output by the i-1 th residual sample block.
Optionally, the depth feature extraction module includes:
the first sampling submodule is used for performing upsampling on the plurality of target characteristic graphs through an upsampling layer in the first branch to obtain a plurality of upsampled characteristic graphs;
the feature extraction sub-module is used for performing feature extraction on the plurality of upsampled feature maps through a third convolution layer in the first branch to obtain a plurality of third convolution feature maps;
and the second sampling sub-module is configured to perform downsampling on the plurality of third convolution feature maps through a downsampling layer in the first branch to obtain a plurality of downsampling feature maps, and determine the plurality of downsampling feature maps as the plurality of local first deep feature maps, where the upsampling is performed by using a transposed convolution and the downsampling is performed by using a convolution.
Optionally, the first sampling sub-module is configured to, through an upsampling layer in the first branch, upsample the multiple target feature maps according to a first formula as follows to obtain multiple upsampled feature maps:
the first formula:
Figure BDA0002122343600000051
wherein the content of the first and second substances,
Figure BDA0002122343600000052
a plurality of up-sampling feature maps for the ith residual sample block, phi is a non-linear mapping function or activation function, R, of the ith residual sample block i-1 For the target feature map, W i And b i Weights and offsets, respectively, # e @ f, for the ith residual sample block s Represents the upsampling by a factor of s;
the feature extraction submodule is configured to perform feature extraction on the plurality of upsampled feature maps through a plurality of third convolution layers in the first branch according to a second formula as follows to obtain a plurality of third convolution feature maps:
the second formula:
Figure BDA0002122343600000061
wherein the content of the first and second substances,
Figure BDA0002122343600000062
a plurality of third convolution maps for an ith residual sample block;
the second sampling sub-module is configured to perform downsampling on the plurality of third convolution feature maps according to a following third formula by using a downsampling layer in the first branch to obtain a plurality of downsampled feature maps:
the third formula:
Figure BDA0002122343600000063
wherein the content of the first and second substances,
Figure BDA0002122343600000064
a plurality of down-sampled feature maps, ↓, for the i-th residual sample block s Is downsampled by a factor of s.
Optionally, the depth feature extraction module is further configured to perform feature extraction on the multiple target feature maps through a second branch in the ith residual sample block according to a fourth formula as follows to obtain multiple local second deep feature maps:
the fourth formula:
Figure BDA0002122343600000065
wherein the content of the first and second substances,
Figure BDA0002122343600000066
a plurality of local second depth profiles are represented.
Optionally, the depth feature extraction module is configured to perform feature fusion on the multiple local first deep feature maps, the multiple local second deep feature maps, and the multiple local third deep feature maps according to a fifth formula below by using the ith residual sampling block to obtain multiple local fourth deep feature maps:
the fifth formula:
Figure BDA0002122343600000067
wherein R is i A plurality of local fourth deep feature maps output for the ith residual sample block,
Figure BDA0002122343600000068
a plurality of local first deep feature maps obtained from a first branch of an ith residual sample block,
Figure BDA0002122343600000069
multiple local second deep feature maps, R, for a second branch of the ith residual sample block i-1 A plurality of local third deep feature maps for a third branch of the ith residual sample block.
The technical scheme provided by the embodiment of the disclosure at least has the following beneficial effects:
in the embodiment of the disclosure, shallow feature extraction is performed on a low-resolution license plate image to be processed through a first convolution layer to obtain a plurality of shallow feature maps, first depth feature extraction is performed on the plurality of shallow feature maps through N residual sampling blocks which are sequentially connected in series to obtain a plurality of local fourth deep feature maps output by the Nth residual sampling block, and then, based on the plurality of local fourth deep feature maps, a target high-resolution license plate image is finally obtained through a second convolution layer and a reconstruction module. Each residual sampling block comprises three branches, so that when depth feature extraction is carried out for the first time, a plurality of target feature maps can be processed through the three branches of each residual sampling block, feature fusion is carried out on a plurality of local first deep feature maps, a plurality of local second deep feature maps and a plurality of local third deep feature maps which are obtained through processing in a one-to-one correspondence mode, a plurality of local fourth deep feature maps are obtained, and then a plurality of local fourth deep feature maps output by the Nth residual sampling blocks are obtained. Because the first branch includes the upsampling layer, the third convolution layer and the downsampling layer that connect gradually, can increase the detailed information of characteristic map, the second branch fully draws the characteristic of low resolution characteristic map through the fourth convolution layer, the third branch can avoid appearing the gradient disappearance and lead to rebuilding the poor problem of effect among the reconstruction process, consequently, carry out the feature fusion with a plurality of local deep characteristic maps that these three branches obtained and can provide more information for the reconstruction module, the problem of the characteristic information loss of having solved the extraction makes the license plate of rebuilding avoid the characters fuzzy, the character adhesion, the condition such as information is not abundant, the degree of discerning of license plate image has been improved.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the disclosure.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present disclosure, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present disclosure, and it is obvious for those skilled in the art to obtain other drawings based on the drawings without creative efforts.
Fig. 1 is a schematic diagram of an image processing network according to an embodiment of the present disclosure.
FIG. 2 is a flowchart of a super-resolution reconstruction method for license plate images provided by an embodiment of the present disclosure
Fig. 3 is a flowchart of a super-resolution reconstruction method for a license plate image according to an embodiment of the present disclosure.
Fig. 4 is a schematic diagram of a super-resolution reconstruction apparatus for a license plate image according to an embodiment of the present disclosure.
Detailed Description
To make the objects, technical solutions and advantages of the present disclosure more apparent, embodiments of the present disclosure will be described in detail with reference to the accompanying drawings.
Fig. 1 is a schematic diagram of an image processing network provided by an embodiment of the present disclosure, and referring to fig. 1, the image processing network may be deployed on a terminal. The image processing network comprises a feature extraction module 1 and a reconstruction module 2, the feature extraction module 1 comprises a shallow feature extraction module 11 and a depth feature extraction module 12 which are connected, the initialization feature extraction module 11 comprises a first convolution layer 111, the depth feature extraction module 12 comprises N identical residual error sampling blocks 121 and a second convolution layer 122 which are sequentially connected in series, each residual error sampling block 121 comprises a first branch A, a second branch B and a third branch C, the first branch A comprises an upper sampling layer A1, a third convolution layer A2 and a lower sampling layer A3 which are sequentially connected, the second branch B comprises a fourth convolution layer B1, and the third branch C is an identity mapping branch. In the image processing network, the solid line "a" may indicate that feature fusion is performed between a plurality of fifth deep-layer feature maps output from the second convolution layer 122 and a plurality of shallow-layer feature maps in one-to-one correspondence.
The embodiment of the disclosure provides a super-resolution reconstruction method of a license plate image, and referring to fig. 2, the method comprises the following steps:
step 201: and performing shallow feature extraction on the low-resolution license plate image to be processed through the first convolution layer to obtain a plurality of shallow feature maps.
Step 202: and performing first depth feature extraction on the plurality of shallow feature maps through the N same residual error sampling blocks which are sequentially connected in series to obtain a plurality of local fourth deep feature maps output by the Nth residual error sampling block.
Carrying out scale transformation on a plurality of target characteristic maps through first branches in an ith residual sampling block in N identical residual sampling blocks which are sequentially connected in series to obtain a plurality of local first deep characteristic maps, carrying out characteristic extraction on the plurality of target characteristic maps through second branches in the ith residual sampling block to obtain a plurality of local second deep characteristic maps, and carrying out identity mapping on the plurality of target characteristic maps through third branches in the ith residual sampling block to obtain a plurality of local third deep characteristic maps; performing feature fusion on the plurality of local first deep feature maps, the plurality of local second deep feature maps and the plurality of local third deep feature maps in a one-to-one correspondence manner through the ith residual sampling block to obtain a plurality of local fourth deep feature maps; obtaining a plurality of local fourth deep characteristic maps output by the Nth residual error sampling block; when i is not equal to 1, the target feature maps are local fourth deep feature maps output by the i-1 th residual sampling block.
Step 203: and performing second depth feature extraction on a plurality of local fourth deep feature maps output by the Nth residual sampling block through the second convolution layer to obtain a plurality of fifth deep feature maps, and performing feature fusion on the plurality of fifth deep feature maps and the plurality of shallow feature maps in a one-to-one correspondence manner to obtain a plurality of sixth deep feature maps.
Step 204: and the reconstruction module is used for amplifying the plurality of sixth deep feature maps obtained by the feature extraction module by multiple times step by step to obtain a plurality of high-resolution feature maps, and obtaining a target high-resolution license plate image according to the plurality of high-resolution feature maps, wherein the target high-resolution license plate image is a license plate image obtained by performing super-resolution reconstruction on a low-resolution license plate image through an image processing network.
Optionally, performing scale transformation on the multiple target feature maps through a first branch in an ith residual sample block of the N identical residual sample blocks connected in series in sequence to obtain multiple local first deep feature maps, where the scale transformation includes:
the plurality of target characteristic graphs are subjected to upsampling through an upsampling layer in the first branch to obtain a plurality of upsampling characteristic graphs;
performing feature extraction on the plurality of upsampling feature maps through a third convolution layer in the first branch to obtain a plurality of third convolution feature maps;
and downsampling the plurality of third convolution feature maps through a downsampling layer in the first branch to obtain a plurality of downsampling feature maps, and determining the plurality of downsampling feature maps into a plurality of local first deep feature maps, wherein the transposed convolution is adopted for upsampling, and the convolution is adopted for downsampling.
Optionally, the upsampling the multiple target feature maps by the upsampling layer in the first branch to obtain multiple upsampled feature maps includes:
through an upsampling layer in the first branch, upsampling the multiple target feature maps according to a first formula as follows to obtain multiple upsampled feature maps:
the first formula:
Figure BDA0002122343600000091
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0002122343600000092
a plurality of up-sampling feature maps for the ith residual sample block, phi is a non-linear mapping function or activation function, R, for the ith residual sample block i-1 For a plurality of target feature maps, W i And b i Weights and offsets, respectively, # e @ f, for the ith residual sample block s Represents the upsampling by a factor of s;
performing feature extraction on the plurality of upsampled feature maps through a third convolutional layer in the first branch to obtain a plurality of third convolutional feature maps, including:
and performing feature extraction on the plurality of upsampled feature maps through a third convolution layer in the first branch according to a second formula to obtain a plurality of third convolution feature maps:
the second formula:
Figure BDA0002122343600000093
wherein the content of the first and second substances,
Figure BDA0002122343600000094
a plurality of third convolution maps for an ith residual sample block;
the method for downsampling the third convolution feature map through the downsampling layer in the first branch to obtain a plurality of downsampled feature maps comprises the following steps:
and downsampling the plurality of third convolution feature maps according to a following third formula by a downsampling layer in the first branch to obtain a plurality of downsampled feature maps:
the third formula:
Figure BDA0002122343600000095
wherein the content of the first and second substances,
Figure BDA0002122343600000096
for a plurality of down-sampled feature maps of the i-th residual sample block, ↓ s is the down-sample s times.
Optionally, performing depth feature extraction on the multiple target feature maps through a second branch in the ith residual sample block to obtain multiple local second deep feature maps, including:
performing feature extraction on the plurality of target feature maps through a second branch in the ith residual error sampling block according to a fourth formula to obtain a plurality of local second deep feature maps:
the fourth formula:
Figure BDA0002122343600000097
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0002122343600000098
representing a plurality of local second depth profiles。
Optionally, feature fusing the multiple local first deep feature maps, the multiple local second deep feature maps, and the multiple local third deep feature maps through an ith residual sampling block to obtain multiple local fourth deep feature maps, including:
performing feature fusion on the multiple local first deep feature maps, the multiple local second deep feature maps and the multiple local third deep feature maps through an ith residual sampling block according to a fifth formula to obtain multiple local fourth deep feature maps:
the fifth formula:
Figure BDA0002122343600000101
wherein R is i A plurality of local fourth deep feature maps output for the ith residual sample block,
Figure BDA0002122343600000102
a plurality of local first deep feature maps from a first branch of an ith residual sample block,
Figure BDA0002122343600000103
multiple local second deep feature maps, R, for a second branch of the ith residual sample block i-1 A plurality of local third deep feature maps for a third branch of the ith residual sample block.
In summary, in the embodiment of the present disclosure, shallow feature extraction is performed on a low-resolution license plate image to be processed through a first convolution layer to obtain a plurality of shallow feature maps, first depth feature extraction is performed on the plurality of shallow feature maps through N residual sampling blocks connected in series in sequence to obtain a plurality of local fourth deep feature maps output by the nth residual sampling block, and then, based on the plurality of local fourth deep feature maps, a target high-resolution license plate image is finally obtained through a second convolution layer and a reconstruction module. Each residual sampling block comprises three branches, so that when depth feature extraction is carried out for the first time, a plurality of target feature maps can be processed through the three branches of each residual sampling block, feature fusion is carried out on a plurality of local first deep feature maps, a plurality of local second deep feature maps and a plurality of local third deep feature maps which are obtained through processing in a one-to-one correspondence mode, a plurality of local fourth deep feature maps are obtained, and then a plurality of local fourth deep feature maps output by the Nth residual sampling blocks are obtained. Because the first branch comprises an upper sampling layer, a third convolution layer and a lower sampling layer which are sequentially connected, the detailed information of the feature map can be increased, the second branch fully extracts the features of the low-resolution feature map through the fourth convolution layer, and the third branch can avoid the problem of poor reconstruction effect caused by gradient disappearance in the reconstruction process.
The embodiment of the present disclosure provides a super-resolution reconstruction method for a license plate image, referring to fig. 3, the implementation environment includes deep learning frames such as tensrflow, tensrlayer, and the like, that is, the operation environment of the embodiment of the present disclosure is applied to a terminal, and the method includes:
step 301: and performing shallow feature extraction on the low-resolution license plate image to be processed through the first convolution layer to obtain a plurality of shallow feature maps.
Optionally, before the low-resolution image is input into the image processing network, certain processing is required to be performed and a license plate image training library is constructed. The license plate image is acquired through a monitoring video or a picture shot by a mobile phone camera, a license plate area is cut out, the cut image containing the license plate area forms an original high-resolution license plate image, and the low-resolution license plate image to be processed can be obtained by down-sampling the original high-resolution license plate image through an interpolation method.
It should be noted that the parameter setting of the first convolution layer includes: the step size is 1, the convolution kernel size is 3 x 3, and the convolution operation adopts a zero filling mode to keep the consistency of the image resolution. The number of input channels is RGB three-channel images, the number of output shallow feature maps is 64, the feature extraction aims to map the low-resolution license plate image to a feature space from an image space, the feature space is a high-dimensional vector formed by a plurality of vectors, the number of the vectors is equal to the number of the feature maps, and features such as shape edges of the license plate image are mainly extracted.
Step 302: and performing first depth feature extraction on the shallow feature maps through the N identical residual error sampling blocks which are sequentially connected in series to obtain a plurality of local fourth deep feature maps output by the Nth residual error sampling block.
Carrying out scale transformation on a plurality of target characteristic maps through first branches in an ith residual sampling block in N identical residual sampling blocks which are sequentially connected in series to obtain a plurality of local first deep characteristic maps, carrying out characteristic extraction on the plurality of target characteristic maps through second branches in the ith residual sampling block to obtain a plurality of local second deep characteristic maps, and carrying out identity mapping on the plurality of target characteristic maps through third branches in the ith residual sampling block to obtain a plurality of local third deep characteristic maps; performing feature fusion on the plurality of local first deep feature maps, the plurality of local second deep feature maps and the plurality of local third deep feature maps in a one-to-one correspondence manner through the ith residual sampling block to obtain a plurality of local fourth deep feature maps; acquiring a plurality of local fourth deep characteristic maps output by the Nth residual error sampling block; when i is not equal to 1, the target feature maps are local fourth deep feature maps output by the i-1 th residual sampling block.
Optionally, performing scale transformation on the multiple target feature maps through a first branch in an ith residual sample block of N identical residual sample blocks connected in series in sequence to obtain multiple local first deep feature maps, including:
the method comprises the steps that a plurality of target characteristic graphs are up-sampled through an up-sampling layer in a first branch to obtain a plurality of up-sampling characteristic graphs;
performing feature extraction on the plurality of upsampling feature maps through a third convolution layer in the first branch to obtain a plurality of third convolution feature maps;
and downsampling the plurality of third convolution feature maps through a downsampling layer in the first branch to obtain a plurality of downsampling feature maps, and determining the plurality of downsampling feature maps into a plurality of local first deep feature maps, wherein the transposed convolution is adopted for upsampling, and the convolution is adopted for downsampling.
Upsampling the plurality of target feature maps by an upsampling layer in the first branch to obtain a plurality of upsampled feature maps, comprising:
through an upsampling layer in the first branch, upsampling the multiple target feature maps according to a first formula as follows to obtain multiple upsampled feature maps:
the first formula:
Figure BDA0002122343600000121
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0002122343600000122
a plurality of up-sampling feature maps for the ith residual sample block, phi is a non-linear mapping function or activation function, R, for the ith residual sample block i-1 For a plurality of target feature maps, W i And b i Weights and offsets, respectively, # e @ f, for the ith residual sample block s Represents an upsampling by a factor of s;
performing feature extraction on the plurality of upsampled feature maps through a third convolutional layer in the first branch to obtain a plurality of third convolutional feature maps, including:
and performing feature extraction on the plurality of upsampled feature maps through a third convolutional layer in the first branch according to a second formula to obtain a plurality of third convolutional feature maps:
the second formula:
Figure BDA0002122343600000123
wherein the content of the first and second substances,
Figure BDA0002122343600000124
a plurality of third convolution maps for an ith residual sample block;
the downsampling of the plurality of third convolution feature maps by the downsampling layer in the first branch to obtain a plurality of downsampled feature maps includes:
and downsampling the plurality of third convolution feature maps according to a following third formula through a downsampling layer in the first branch to obtain a plurality of downsampled feature maps:
the third formula:
Figure BDA0002122343600000125
wherein the content of the first and second substances,
Figure BDA0002122343600000126
for a plurality of downsampled feature maps of the ith residual sample block, ↓ s is downsampled by s times.
Performing feature extraction on the plurality of target feature maps through a second branch in the ith residual sampling block to obtain a plurality of local second deep feature maps, wherein the method comprises the following steps:
and performing depth feature extraction on the plurality of target feature maps through a second branch in the ith residual error sampling block according to a fourth formula to obtain a plurality of local second deep feature maps:
the fourth formula:
Figure BDA0002122343600000127
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0002122343600000128
a plurality of local second depth profiles are represented.
Performing feature fusion on the plurality of local first deep feature maps, the plurality of local second deep feature maps and the plurality of local third deep feature maps through the ith residual sampling block to obtain a plurality of local fourth deep feature maps, including:
performing feature fusion on the multiple local first deep feature maps, the multiple local second deep feature maps and the multiple local third deep feature maps through an ith residual sampling block according to a fifth formula to obtain multiple local fourth deep feature maps:
the fifth formula:
Figure BDA0002122343600000131
wherein R is i A plurality of local fourth deep feature maps output for the ith residual sample block,
Figure BDA0002122343600000132
a plurality of local first deep feature maps from a first branch of an ith residual sample block,
Figure BDA0002122343600000133
multiple local second deep feature maps, R, for a second branch of the ith residual sample block i-1 A plurality of local third deep feature maps for a third branch of the ith residual sample block.
It should be noted that 1, the first branch does not include a batch normalization layer, which is applied in the super-resolution field to cause loss of image pixel value scale information and affect the change of network parameters in a free range, the present disclosure removes the batch normalization layer, and applies computing resources to feature extraction, such as a third convolution layer in the first branch and a fourth convolution layer in the second branch; 2. the first branch mainly comprises an upper sampling layer, a third convolution layer and a lower sampling layer which are connected in sequence, and the scale transformation is performed to be beneficial to enhancing the expression capability of the image processing network and improve the information extracted by the network; the upper sampling layer adopts transposition convolution, and parameters of the transposition convolution are set as follows: the step size is 4, the input and output characteristic graphs are the same and are 64, the convolution kernel size is 6 multiplied by 6, the third convolution layer parameter setting is that the step size is 1, the input and output characteristic graphs are the same and are 64, the convolution kernel size is 3 multiplied by 3, the down-sampling layer adopts convolution, and the parameter setting of the convolution is as follows: the step length is 4, the input characteristic diagram and the output characteristic diagram are the same and are both 64, the size of a convolution kernel is 6 multiplied by 6, and zero padding operation is adopted in the operations; 3. the method has the advantages that the multiple local first deep characteristic maps, the multiple local second deep characteristic maps and the multiple local third deep characteristic maps are subjected to characteristic fusion in a one-to-one correspondence mode through the ith residual sampling block, the multiple local fourth deep characteristic maps are obtained, fusion is carried out, more information can be provided for the reconstruction module, the problem that extracted characteristic information is lost is solved, meanwhile, information flow is increased in the local part, and the situation that training errors caused by gradient disappearance are larger and larger is avoided.
Step 303: and performing second depth feature extraction on a plurality of local fourth deep feature maps output by the Nth residual sampling block through the second convolution layer to obtain a plurality of fifth deep feature maps, and performing feature fusion on the plurality of fifth deep feature maps and the plurality of shallow feature maps in a one-to-one correspondence manner to obtain a plurality of sixth deep feature maps.
Optionally, the image processing network is gradually deepened, which may weaken the backward propagation capability of the network, because the gradient of each layer in the backward propagation is calculated on the basis of the previous layer, the larger the number of layers is, which may cause the gradient to become smaller and smaller during propagation until the gradient disappears, and the larger the training error is, so that feature fusion is performed by one-to-one correspondence between the fifth deep layer feature maps and the shallow layer feature maps, which may increase information flow, further improve the expression capability of the network, and avoid the training error caused by the disappearance of the gradient from becoming larger.
Step 304: and the reconstruction module is used for amplifying the sixth deep feature maps obtained by the feature extraction module by multiple times step by step to obtain a plurality of high-resolution feature maps, and a target high-resolution license plate image is obtained according to the high-resolution feature maps and is obtained by performing super-resolution reconstruction on a low-resolution license plate image through an image processing network.
Optionally, the reconstruction module is configured to amplify the resolutions of the sixth deep feature maps obtained by the feature extraction module by a certain factor, where the step-by-step magnification is performed step by step when the magnification factor is higher, and if the magnification factor is 4, the step may be divided into two steps, where each step amplifies the resolution of the feature map by 2 times. The total number of pixels is unchanged, the pixels in the low-resolution feature map space are recombined and arranged in the high-resolution feature map space according to a certain multiple, the number of the corresponding high-resolution feature maps is reduced according to a certain multiple compared with the number of the low-resolution feature maps before arrangement due to the length and width amplification of the high-resolution feature maps, for example, for an amplification factor of 2, the specific operation of the sub-pixel layer is to convert the spatial positions of 256 feature maps with the resolution of H multiplied by W to obtain 64 feature maps with the resolution of 2H multiplied by 2W; and finally, inputting the multiple high-resolution feature maps amplified by multiple times into a3 x 3 convolutional layer to obtain a three-channel target high-resolution license plate image. Generally, since the number of feature maps is reduced after each pass through the sub-pixel layer, and when step-by-step amplification is performed, the number of feature maps is increased by a convolution layer before the sub-pixel layer to ensure that there is enough feature maps, each step of step-by-step multiple amplification includes two modules: 1. the number of the characteristic graphs is increased according to a certain multiple; 2. a sub-pixel layer.
Step 305: a difference between the original high-resolution license plate image and the target high-resolution license plate image is minimized through a loss function to generate a first high-resolution license plate image.
The original high-resolution license plate image is a license plate image which is not processed through a network, the first high-resolution license plate image is a license plate image which is obtained through the steps that network parameters are continuously updated through a loss function, and when the network parameters approach convergence, the finally obtained super-resolution reconstruction quality is the optimal license plate image.
Optionally, the loss function measures the difference between the predicted value and the true value, and the disclosure employs a minimum mean square error criterion:
the sixth formula:
Figure BDA0002122343600000141
wherein l SR Represents the loss value, r represents the magnification factor, W and H represent the width and height of the low-resolution license plate image respectively,
Figure BDA0002122343600000142
representing the corresponding pixel points of the original high-resolution license plate image,
Figure BDA0002122343600000143
and representing corresponding pixel points of the first high-resolution license plate image.
In summary, in the embodiment of the present disclosure, shallow feature extraction is performed on a low-resolution license plate image to be processed through a first convolution layer to obtain a plurality of shallow feature maps, first depth feature extraction is performed on the plurality of shallow feature maps through N residual sampling blocks connected in series in sequence to obtain a plurality of local fourth deep feature maps output by the nth residual sampling block, and then, based on the plurality of local fourth deep feature maps, a target high-resolution license plate image is finally obtained through a second convolution layer and a reconstruction module. Each residual sampling block comprises three branches, so that when depth feature extraction is carried out for the first time, a plurality of target feature maps can be processed through the three branches of each residual sampling block, feature fusion is carried out on a plurality of local first deep feature maps, a plurality of local second deep feature maps and a plurality of local third deep feature maps which are obtained through processing in a one-to-one correspondence mode, a plurality of local fourth deep feature maps are obtained, and then a plurality of local fourth deep feature maps output by the Nth residual sampling blocks are obtained. Because the first branch comprises an upper sampling layer, a third convolution layer and a lower sampling layer which are sequentially connected, the detailed information of the feature map can be increased, the second branch fully extracts the features of the low-resolution feature map through the fourth convolution layer, and the third branch can avoid the problem of poor reconstruction effect caused by gradient disappearance in the reconstruction process.
The embodiment of the present disclosure provides a super-resolution reconstruction device for license plate images, referring to fig. 4, the device includes:
the initialization feature extraction module 401 is configured to perform shallow feature extraction on the to-be-processed low-resolution license plate image through the first convolution layer to obtain a plurality of shallow feature maps.
A depth feature extraction module 402, configured to perform first depth feature extraction on the multiple shallow feature maps through N identical residual sampling blocks connected in series in sequence, so as to obtain multiple local fourth deep feature maps output by an nth residual sampling block.
The depth feature extraction module is further configured to perform scale transformation on the multiple target feature maps through a first branch in an ith residual sampling block of the N identical residual sampling blocks connected in series in sequence to obtain multiple local first deep feature maps, perform feature extraction on the multiple target feature maps through a second branch in the ith residual sampling block to obtain multiple local second deep feature maps, and perform identity mapping on the multiple target feature maps through a third branch in the ith residual sampling block to obtain multiple local third deep feature maps; performing feature fusion on the plurality of local first deep feature maps, the plurality of local second deep feature maps and the plurality of local third deep feature maps in a one-to-one correspondence manner through the ith residual sampling block to obtain a plurality of local fourth deep feature maps; acquiring a plurality of local fourth deep characteristic maps output by the Nth residual error sampling block; when i is not equal to 1, the target feature maps are local fourth deep feature maps output by the i-1 th residual sampling block.
The feature fusion module 403 is configured to perform second depth feature extraction on the multiple local fourth deep feature maps output by the nth residual sampling block through the second convolution layer to obtain multiple fifth deep feature maps, and perform feature fusion on the multiple fifth deep feature maps and the multiple shallow feature maps in a one-to-one correspondence manner to obtain multiple sixth deep feature maps.
The super-resolution reconstruction module 404 is configured to perform progressive multiple amplification on the sixth deep feature maps obtained by the feature extraction module through the reconstruction module to obtain a plurality of high-resolution feature maps, and obtain a target high-resolution license plate image according to the plurality of high-resolution feature maps, where the target high-resolution license plate image is a license plate image obtained by performing super-resolution reconstruction on a low-resolution license plate image through an image processing network.
Optionally, the depth feature extraction module 402 includes:
the first sampling submodule is used for performing up-sampling on the plurality of target characteristic graphs through an up-sampling layer in the first branch to obtain a plurality of up-sampling characteristic graphs;
the feature extraction submodule is used for performing feature extraction on the plurality of up-sampling feature maps through a third convolution layer in the first branch to obtain a plurality of third convolution feature maps;
and the second sampling sub-module is used for carrying out downsampling on the plurality of third convolution feature maps through a downsampling layer in the first branch to obtain a plurality of downsampling feature maps, and determining the plurality of downsampling feature maps into a plurality of local first deep feature maps, wherein the up-sampling adopts transposed convolution and the downsampling adopts convolution.
Optionally, the first sampling sub-module performs upsampling on the multiple target feature maps through an upsampling layer in the first branch according to a first formula as follows to obtain multiple upsampled feature maps:
the first formula:
Figure BDA0002122343600000161
wherein the content of the first and second substances,
Figure BDA0002122343600000162
a plurality of up-sampling feature maps for the ith residual sample block, phi is a non-linear mapping function or activation function, R, of the ith residual sample block i-1 For a plurality of target feature maps, W i And b i Weights and offsets, respectively, # e @ f, for the ith residual sample block s Represents the upsampling by a factor of s;
the feature extraction sub-module is configured to perform feature extraction on the multiple upsampled feature maps through the third convolution layer in the first branch according to a second formula as follows to obtain multiple third convolution feature maps:
the second formula:
Figure BDA0002122343600000163
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0002122343600000171
a plurality of third convolutions for the ith residual sample blockA feature map;
the second sampling submodule is used for downsampling the plurality of third convolution feature maps according to a following third formula through a downsampling layer in the first branch to obtain a plurality of downsampling feature maps:
the third formula:
Figure BDA0002122343600000172
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0002122343600000173
a plurality of down-sampled feature maps, ↓, for the i-th residual sample block s Is s times the down sampling.
Optionally, the depth feature extraction module is further configured to perform feature extraction on the multiple target feature maps through a second branch in the ith residual sampling block according to a fourth formula as follows to obtain multiple local second deep feature maps:
the fourth formula:
Figure BDA0002122343600000174
wherein the content of the first and second substances,
Figure BDA0002122343600000175
a plurality of local second depth profiles are represented.
Optionally, the depth feature extraction module is configured to perform feature fusion on the multiple local first deep feature maps, the multiple local second deep feature maps, and the multiple local third deep feature maps through the ith residual sampling block according to a fifth formula as follows to obtain multiple local fourth deep feature maps:
the fifth formula:
Figure BDA0002122343600000176
wherein R is i A plurality of local fourth deep feature maps output for the ith residual sample block,
Figure BDA0002122343600000177
a plurality of local first deep feature maps from a first branch of an ith residual sample block,
Figure BDA0002122343600000178
multiple local second deep feature maps, R, obtained for a second branch in an ith residual sample block i-1 A plurality of local third deep feature maps for a third branch of the ith residual sample block.
In summary, in the embodiment of the present disclosure, shallow feature extraction is performed on a to-be-processed low-resolution license plate image through a first convolution layer to obtain a plurality of shallow feature maps, first depth feature extraction is performed on the plurality of shallow feature maps through N residual sampling blocks connected in series in sequence to obtain a plurality of local fourth deep feature maps output by an nth residual sampling block, and then, based on the plurality of local fourth deep feature maps, a target high-resolution license plate image is finally obtained through a second convolution layer and a reconstruction module. Each residual sampling block comprises three branches, so that when depth feature extraction is carried out for the first time, a plurality of target feature maps can be processed through the three branches of each residual sampling block, feature fusion is carried out on a plurality of local first deep feature maps, a plurality of local second deep feature maps and a plurality of local third deep feature maps which are obtained through processing in a one-to-one correspondence mode, a plurality of local fourth deep feature maps are obtained, and then a plurality of local fourth deep feature maps output by the Nth residual sampling blocks are obtained. Because the first branch includes the upsampling layer, the third convolution layer and the downsampling layer that connect gradually, can increase the detailed information of characteristic map, the second branch fully draws the characteristic of low resolution characteristic map through the fourth convolution layer, the third branch can avoid appearing the gradient disappearance and lead to rebuilding the poor problem of effect among the reconstruction process, consequently, carry out the feature fusion with a plurality of local deep characteristic maps that these three branches obtained and can provide more information for the reconstruction module, the problem of the characteristic information loss of having solved the extraction makes the license plate of rebuilding avoid the characters fuzzy, the character adhesion, the condition such as information is not abundant, the degree of discerning of license plate image has been improved.
The disclosed embodiment provides a super-resolution reconstruction method and device for a license plate image, and the method and device are applied to a terminal, wherein at least one instruction, at least one section of program, a code set or an instruction set is stored in a computer-readable storage medium, and the instruction, the program, the code set or the instruction set is loaded and executed by a processor to implement the operations executed by the terminal in the above-mentioned license plate image super-resolution reconstruction method.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosed invention. This application is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains.
It should be understood that the present disclosure is not limited to the precise construction and arrangements shown and described above, and that various modifications, equivalent substitutions, improvements, etc. can be made without departing from the spirit and scope of the present disclosure.

Claims (10)

1. The super-resolution reconstruction method of the license plate image is characterized by being applied to a terminal, wherein an image processing network is deployed on the terminal and comprises a feature extraction module and a reconstruction module, the feature extraction module comprises an initialization feature extraction module and a depth feature extraction module which are connected, the initialization feature extraction module comprises a first convolution layer, the depth feature extraction module comprises N identical residual sampling blocks and a second convolution layer which are sequentially connected in series, N is an integer greater than or equal to 1, each residual sampling block comprises a first branch, a second branch and a third branch, the first branch comprises an upper sampling layer, a third convolution layer and a lower sampling layer which are sequentially connected, the second branch comprises a fourth convolution layer, the third branch is an identity mapping branch, an output feature map of the identity mapping branch is identical to an input feature map, and the method comprises the following steps:
performing shallow feature extraction on the low-resolution license plate image to be processed through the first convolution layer to obtain a plurality of shallow feature maps;
carrying out first depth feature extraction on the shallow feature maps through the N identical residual error sampling blocks which are sequentially connected in series to obtain a plurality of local fourth deep feature maps output by the Nth residual error sampling block;
performing second depth feature extraction on a plurality of local fourth deep feature maps output by the Nth residual sampling block through the second convolution layer to obtain a plurality of fifth deep feature maps, and performing feature fusion on the plurality of fifth deep feature maps and the plurality of shallow feature maps in a one-to-one correspondence manner to obtain a plurality of sixth deep feature maps;
amplifying the sixth deep feature maps obtained by the feature extraction module by multiple times step by step through the reconstruction module to obtain a plurality of high-resolution feature maps, and obtaining a target high-resolution license plate image according to the high-resolution feature maps, wherein the target high-resolution license plate image is a license plate image obtained by performing super-resolution reconstruction on the low-resolution license plate image through the image processing network;
the first depth feature extraction is performed on the multiple shallow feature maps through the N identical residual sampling blocks connected in series in sequence to obtain multiple local fourth deep feature maps output by the nth residual sampling block, and the method includes:
carrying out scale transformation on a plurality of target feature maps through a first branch in an ith residual sampling block in the N identical residual sampling blocks which are sequentially connected in series to obtain a plurality of local first deep feature maps, carrying out feature extraction on the plurality of target feature maps through a second branch in the ith residual sampling block to obtain a plurality of local second deep feature maps, and carrying out identity mapping on the plurality of target feature maps through a third branch in the ith residual sampling block to obtain a plurality of local third deep feature maps; performing feature fusion on the plurality of local first deep feature maps, the plurality of local second deep feature maps and the plurality of local third deep feature maps in a one-to-one correspondence manner through the ith residual sampling block to obtain a plurality of local fourth deep feature maps; acquiring a plurality of local fourth deep feature maps output by the Nth residual error sampling block; when i is not equal to 1, the target feature maps are local fourth deep feature maps output by the i-1 th residual sample block.
2. The method of claim 1, wherein scaling the plurality of target feature maps through the first branch in an ith residual sample block of the N identical residual sample blocks connected in series in sequence to obtain a plurality of local first deep feature maps comprises:
the plurality of target characteristic graphs are subjected to upsampling through an upsampling layer in the first branch to obtain a plurality of upsampling characteristic graphs;
performing feature extraction on the plurality of upsampled feature maps through a third convolution layer in the first branch to obtain a plurality of third convolution feature maps;
and downsampling the plurality of third convolution feature maps through a downsampling layer in the first branch to obtain a plurality of downsampling feature maps, and determining the plurality of downsampling feature maps as the plurality of local first deep feature maps, wherein the upsampling adopts transposition convolution, and the downsampling adopts convolution.
3. The method of claim 2, wherein the upsampling the plurality of target feature maps by the upsampling layer in the first branch to obtain a plurality of upsampled feature maps, comprises:
through an upsampling layer in the first branch, upsampling the multiple target feature maps according to a first formula as follows to obtain multiple upsampled feature maps:
the first formula:
Figure FDA0002122343590000021
wherein the content of the first and second substances,
Figure FDA0002122343590000022
a plurality of up-sampling feature maps for the ith residual sample block, phi is a non-linear mapping function or activation function, R, for the ith residual sample block i-1 For said plurality of target feature maps, W i And b i Weight and bias of the ith residual sample block, # r ≠ f s Represents an upsampling by a factor of s;
performing feature extraction on the plurality of upsampled feature maps through a third convolutional layer in the first branch to obtain a plurality of third convolutional feature maps, including:
performing feature extraction on the plurality of upsampled feature maps through a third convolutional layer in the first branch according to a second formula as follows to obtain a plurality of third convolutional feature maps:
the second formula:
Figure FDA0002122343590000023
wherein the content of the first and second substances,
Figure FDA0002122343590000024
a plurality of third convolution maps for an ith residual sample block;
the downsampling the plurality of third convolution feature maps by the downsampling layer in the first branch to obtain a plurality of downsampled feature maps includes:
downsampling the plurality of third convolution feature maps according to a third formula to obtain a plurality of downsampled feature maps by using a downsampling layer in the first branch:
the third formula:
Figure FDA0002122343590000031
wherein, the first and the second end of the pipe are connected with each other,
Figure FDA0002122343590000032
a plurality of down-sampling feature maps for the ith residual sample block, ↓ s is down-sampling sAnd (4) multiplying.
4. The method of claim 1, wherein said feature extracting the plurality of target feature maps through a second branch in the i-th residual sample block to obtain a plurality of local second deep feature maps comprises:
performing depth feature extraction on the plurality of target feature maps through a second branch in the ith residual sampling block according to a fourth formula to obtain a plurality of local second deep feature maps:
the fourth formula:
Figure FDA0002122343590000033
wherein, the first and the second end of the pipe are connected with each other,
Figure FDA0002122343590000034
a plurality of local second depth profiles are represented.
5. The method of claim 1, wherein feature fusing the plurality of local first, second, and third deep feature maps by the ith residual sample block to obtain a plurality of local fourth deep feature maps comprises:
performing feature fusion on the plurality of local first deep feature maps, the plurality of local second deep feature maps and the plurality of local third deep feature maps according to a fifth formula to obtain a plurality of local fourth deep feature maps by using the ith residual sampling block:
the fifth formula:
Figure FDA0002122343590000035
wherein R is i A plurality of local fourth deep feature maps output for the ith residual sample block,
Figure FDA0002122343590000036
a plurality of local first deep feature maps from a first branch of an ith residual sample block,
Figure FDA0002122343590000037
multiple local second deep feature maps, R, for a second branch of the ith residual sample block i-1 A plurality of local third deep feature maps for a third branch of the ith residual sample block.
6. The utility model provides a super-resolution of license plate image rebuilds device which characterized in that is applied to the terminal, the terminal is last to be deployed and has been had the image processing network, the image processing network includes feature extraction module and rebuild the module, the feature extraction module is including the initialized feature extraction module and the depth feature extraction module of connecting, the initialized feature extraction module includes first volume layer, the depth feature extraction module is including N the same residual error sampling block and a second volume layer that connect gradually in series, N is the integer that is more than or equal to 1, and each residual error sampling block includes first branch, second branch and third branch, first branch is including the upsampling layer, third volume layer and the downsampling layer that connect gradually, the second branch includes the fourth volume layer, the third branch is the mapping branch of identity, the output characteristic map of mapping branch is the same with the input characteristic map, the device includes:
the initialization feature extraction module is used for performing shallow feature extraction on the low-resolution license plate image to be processed through the first convolution layer to obtain a plurality of shallow feature maps;
the depth feature extraction module is used for performing first depth feature extraction on the shallow feature maps through the N identical residual error sampling blocks which are sequentially connected in series to obtain a plurality of local fourth deep feature maps output by the Nth residual error sampling block;
the feature fusion module is configured to perform second depth feature extraction on a plurality of local fourth deep feature maps output by the nth residual sampling block through the second convolution layer to obtain a plurality of fifth deep feature maps, and perform feature fusion on the plurality of fifth deep feature maps and the plurality of shallow feature maps in a one-to-one correspondence manner to obtain a plurality of sixth deep feature maps;
the super-resolution reconstruction module is used for amplifying the sixth deep feature maps obtained by the feature extraction module by multiple times step by step through the reconstruction module to obtain a plurality of high-resolution feature maps, and obtaining a target high-resolution license plate image according to the high-resolution feature maps, wherein the target high-resolution license plate image is a license plate image obtained by performing super-resolution reconstruction on the low-resolution license plate image through the image processing network;
the depth feature extraction module is further configured to perform scale transformation on multiple target feature maps through a first branch in an ith residual sampling block of the N identical residual sampling blocks connected in series in sequence to obtain multiple local first deep feature maps, perform feature extraction on the multiple target feature maps through a second branch in the ith residual sampling block to obtain multiple local second deep feature maps, and perform identity mapping on the multiple target feature maps through a third branch in the ith residual sampling block to obtain multiple local third deep feature maps; performing feature fusion on the plurality of local first deep feature maps, the plurality of local second deep feature maps and the plurality of local third deep feature maps in a one-to-one correspondence manner through the ith residual sampling block to obtain a plurality of local fourth deep feature maps; acquiring a plurality of local fourth deep feature maps output by the Nth residual error sampling block; when i is equal to 1, the plurality of target feature maps are the plurality of shallow feature maps, and when i is not equal to 1, the plurality of target feature maps are a plurality of local fourth deep feature maps output by the i-1 th residual sampling block.
7. The apparatus of claim 6, wherein the depth feature extraction module comprises:
the first sampling submodule is used for performing upsampling on the plurality of target characteristic graphs through an upsampling layer in the first branch to obtain a plurality of upsampling characteristic graphs;
the feature extraction submodule is used for performing feature extraction on the plurality of up-sampling feature maps through a third convolution layer in the first branch to obtain a plurality of third convolution feature maps;
and the second sampling sub-module is configured to perform downsampling on the plurality of third convolution feature maps through a downsampling layer in the first branch to obtain a plurality of downsampling feature maps, and determine the plurality of downsampling feature maps as the plurality of local first deep feature maps, where the upsampling is performed by using a transposed convolution and the downsampling is performed by using a convolution.
8. The apparatus of claim 7,
the first sampling sub-module is configured to, through an upsampling layer in the first branch, upsample the multiple target feature maps according to a first formula as follows to obtain multiple upsampled feature maps:
the first formula:
Figure FDA0002122343590000051
wherein, the first and the second end of the pipe are connected with each other,
Figure FDA0002122343590000052
a plurality of up-sampling feature maps for the ith residual sample block, phi is a non-linear mapping function or activation function, R, of the ith residual sample block i-1 For said plurality of target feature maps, W i And b i Weight and bias of the ith residual sample block, # r ≠ f s Represents the upsampling by a factor of s;
the feature extraction sub-module is configured to perform feature extraction on the plurality of upsampled feature maps through a third convolution layer in the first branch according to a second formula as follows to obtain a plurality of third convolution feature maps:
the second formula:
Figure FDA0002122343590000053
wherein the content of the first and second substances,
Figure FDA0002122343590000054
a plurality of third convolution maps for an ith residual sample block;
the second sampling sub-module is configured to perform downsampling on the plurality of third convolution feature maps according to a following third formula by using a downsampling layer in the first branch to obtain a plurality of downsampled feature maps:
the third formula:
Figure FDA0002122343590000055
wherein, the first and the second end of the pipe are connected with each other,
Figure FDA0002122343590000056
a plurality of down-sampled feature maps, ↓, for the i-th residual sample block s Is downsampled by a factor of s.
9. The apparatus of claim 6,
the depth feature extraction module is further configured to perform feature extraction on the multiple target feature maps through a second branch in the ith residual sampling block according to a fourth formula as follows to obtain multiple local second deep feature maps:
the fourth formula:
Figure FDA0002122343590000057
wherein the content of the first and second substances,
Figure FDA0002122343590000058
a plurality of local second depth profiles are represented.
10. The apparatus of claim 6,
the depth feature extraction module is configured to perform feature fusion on the multiple local first deep feature maps, the multiple local second deep feature maps, and the multiple local third deep feature maps according to a fifth formula below by using the ith residual sampling block to obtain multiple local fourth deep feature maps:
the fifth formula:
Figure FDA0002122343590000061
wherein R is i A plurality of local fourth deep feature maps output for the ith residual sample block,
Figure FDA0002122343590000062
a plurality of local first deep feature maps from a first branch of an ith residual sample block,
Figure FDA0002122343590000063
multiple local second deep feature maps, R, obtained for a second branch in an ith residual sample block i-1 A plurality of local third deep feature maps for a third branch of the ith residual sample block.
CN201910610927.XA 2019-07-08 2019-07-08 Super-resolution reconstruction method and device for license plate image Active CN110298791B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910610927.XA CN110298791B (en) 2019-07-08 2019-07-08 Super-resolution reconstruction method and device for license plate image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910610927.XA CN110298791B (en) 2019-07-08 2019-07-08 Super-resolution reconstruction method and device for license plate image

Publications (2)

Publication Number Publication Date
CN110298791A CN110298791A (en) 2019-10-01
CN110298791B true CN110298791B (en) 2022-10-28

Family

ID=68030646

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910610927.XA Active CN110298791B (en) 2019-07-08 2019-07-08 Super-resolution reconstruction method and device for license plate image

Country Status (1)

Country Link
CN (1) CN110298791B (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112950463A (en) * 2019-12-11 2021-06-11 香港理工大学深圳研究院 Image super-resolution method, image super-resolution device and terminal equipment
CN111429355A (en) * 2020-03-30 2020-07-17 新疆大学 Image super-resolution reconstruction method based on generation countermeasure network
US11900563B2 (en) 2020-04-01 2024-02-13 Boe Technology Group Co., Ltd. Computer-implemented method, apparatus, and computer-program product
CN112053287B (en) * 2020-09-11 2023-04-18 北京邮电大学 Image super-resolution method, device and equipment
CN112329603B (en) * 2020-11-03 2022-09-13 西南科技大学 Dam face crack defect positioning method based on image cascade
CN112767427A (en) * 2021-01-19 2021-05-07 西安邮电大学 Low-resolution image recognition algorithm for compensating edge information
CN114037624B (en) * 2021-10-27 2023-05-23 成都市第二人民医院 Image enhancement method and device for classifying diabetic nephropathy
CN114550137B (en) * 2022-02-22 2024-04-09 智道网联科技(北京)有限公司 Method and device for identifying traffic sign board and electronic equipment

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101780057B1 (en) * 2016-08-02 2017-09-19 한양대학교 에리카산학협력단 Apparatus and method for restroing high resolution image
CN109064398A (en) * 2018-07-14 2018-12-21 深圳市唯特视科技有限公司 A kind of image super-resolution implementation method based on residual error dense network
CN109064407A (en) * 2018-09-13 2018-12-21 武汉大学 Intensive connection network image super-resolution method based on multi-layer perception (MLP) layer
CN109377530A (en) * 2018-11-30 2019-02-22 天津大学 A kind of binocular depth estimation method based on deep neural network
CN109829421A (en) * 2019-01-29 2019-05-31 西安邮电大学 The method, apparatus and computer readable storage medium of vehicle detection
CN109903226A (en) * 2019-01-30 2019-06-18 天津城建大学 Image super-resolution rebuilding method based on symmetrical residual error convolutional neural networks

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20130001213A (en) * 2010-01-28 2013-01-03 이섬 리서치 디벨러프먼트 컴파니 오브 더 히브루 유니버시티 오브 예루살렘 엘티디. Method and system for generating an output image of increased pixel resolution from an input image
US11024009B2 (en) * 2016-09-15 2021-06-01 Twitter, Inc. Super resolution using a generative adversarial network

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101780057B1 (en) * 2016-08-02 2017-09-19 한양대학교 에리카산학협력단 Apparatus and method for restroing high resolution image
CN109064398A (en) * 2018-07-14 2018-12-21 深圳市唯特视科技有限公司 A kind of image super-resolution implementation method based on residual error dense network
CN109064407A (en) * 2018-09-13 2018-12-21 武汉大学 Intensive connection network image super-resolution method based on multi-layer perception (MLP) layer
CN109377530A (en) * 2018-11-30 2019-02-22 天津大学 A kind of binocular depth estimation method based on deep neural network
CN109829421A (en) * 2019-01-29 2019-05-31 西安邮电大学 The method, apparatus and computer readable storage medium of vehicle detection
CN109903226A (en) * 2019-01-30 2019-06-18 天津城建大学 Image super-resolution rebuilding method based on symmetrical residual error convolutional neural networks

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
Deep Differential Convolutional Network for Single Image Super-Resolution;PENG LIU 等;《IEEE Access》;20190307;第7卷;第37555-37564页 *
DENSE BYNET: RESIDUAL DENSE NETWORK FOR IMAGE SUPER RESOLUTION;Jiu Xu 等;《2018 25th IEEE International Conference on Image Processing (ICIP)》;20181010;第71-75页 *
Residual Dense Network for Image Super-Resolution;Yulun Zhang 等;《2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition》;20181231;第2472-2481页 *
基于改进卷积神经网络的单幅图像超分辨率重建方法;刘月峰 等;《计算机应用》;20190510;第39卷(第5期);第1441-1447页 *

Also Published As

Publication number Publication date
CN110298791A (en) 2019-10-01

Similar Documents

Publication Publication Date Title
CN110298791B (en) Super-resolution reconstruction method and device for license plate image
US10853916B2 (en) Convolution deconvolution neural network method and system
CN112750082B (en) Human face super-resolution method and system based on fusion attention mechanism
CN110322400B (en) Image processing method and device, image processing system and training method thereof
CN112037129B (en) Image super-resolution reconstruction method, device, equipment and storage medium
CN111598778A (en) Insulator image super-resolution reconstruction method
CN111696038A (en) Image super-resolution method, device, equipment and computer-readable storage medium
CN111784582A (en) DEC-SE-based low-illumination image super-resolution reconstruction method
CN111797834B (en) Text recognition method and device, computer equipment and storage medium
CN114170167B (en) Polyp segmentation method and computer device based on attention-guided context correction
CN112889084B (en) Method, system and computer readable medium for improving color quality of image
CN115578255B (en) Super-resolution reconstruction method based on inter-frame sub-pixel block matching
CN107220934B (en) Image reconstruction method and device
CN115345866A (en) Method for extracting buildings from remote sensing images, electronic equipment and storage medium
CN113298716A (en) Image super-resolution reconstruction method based on convolutional neural network
CN111861888A (en) Image processing method, image processing device, electronic equipment and storage medium
CN112801879B (en) Image super-resolution reconstruction method and device, electronic equipment and storage medium
CN113674154B (en) Single image super-resolution reconstruction method and system based on generation countermeasure network
CN111223046A (en) Image super-resolution reconstruction method and device
CN115311145A (en) Image processing method and device, electronic device and storage medium
CN113947521A (en) Image resolution conversion method and device based on deep neural network and terminal equipment
CN114445277A (en) Depth image pixel enhancement method and device and computer readable storage medium
CN111382845B (en) Template reconstruction method based on self-attention mechanism
CN112419146B (en) Image processing method and device and terminal equipment
JP5181345B2 (en) Image processing apparatus and image processing method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant