CN112085106A - Image identification method and device applied to multi-image fusion, electronic equipment and storage medium - Google Patents
Image identification method and device applied to multi-image fusion, electronic equipment and storage medium Download PDFInfo
- Publication number
- CN112085106A CN112085106A CN202010947804.8A CN202010947804A CN112085106A CN 112085106 A CN112085106 A CN 112085106A CN 202010947804 A CN202010947804 A CN 202010947804A CN 112085106 A CN112085106 A CN 112085106A
- Authority
- CN
- China
- Prior art keywords
- image
- target
- feature
- template image
- detection frame
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 41
- 230000004927 fusion Effects 0.000 title claims abstract description 21
- 238000001514 detection method Methods 0.000 claims abstract description 90
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 37
- 238000012216 screening Methods 0.000 claims abstract description 15
- 238000000605 extraction Methods 0.000 claims description 21
- 238000013459 approach Methods 0.000 claims description 16
- 239000011159 matrix material Substances 0.000 claims description 14
- 238000004590 computer program Methods 0.000 claims description 12
- 238000012549 training Methods 0.000 abstract description 6
- 239000000284 extract Substances 0.000 abstract description 5
- 238000010801 machine learning Methods 0.000 abstract description 3
- 230000008569 process Effects 0.000 abstract description 3
- 230000000717 retained effect Effects 0.000 abstract description 2
- 238000010586 diagram Methods 0.000 description 5
- 230000000694 effects Effects 0.000 description 4
- 230000003044 adaptive effect Effects 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 230000008030 elimination Effects 0.000 description 2
- 238000003379 elimination reaction Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 238000000926 separation method Methods 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 238000007792 addition Methods 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000000354 decomposition reaction Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000007499 fusion processing Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 238000007500 overflow downdraw method Methods 0.000 description 1
- 238000005215 recombination Methods 0.000 description 1
- 230000006798 recombination Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4038—Image mosaicing, e.g. composing plane images from plane sub-images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
- G06V10/443—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/75—Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
- G06V10/751—Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Data Mining & Analysis (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Multimedia (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Medical Informatics (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Health & Medical Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Databases & Information Systems (AREA)
- General Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses an image identification method and device applied to multi-image fusion, electronic equipment and a storage medium, and belongs to the technical field of image identification. The method avoids the process of fusing images, and firstly extracts and matches the image characteristics; then, respectively using the original input of the image to perform target detection, comparing the position of a target boundary frame obtained after the detection with the characteristic matching position, and screening to leave a target object; therefore, the original feature information of the image is retained to the maximum extent, the original feature of the image is convenient to be input into a machine learning model for extracting, the accuracy of image recognition is improved, and the problem that the algorithm for extracting the image target is not accurate due to difference between the training image and the actually used image sample is solved.
Description
Technical Field
The invention belongs to the technical field of image recognition, and particularly relates to an image recognition method and device applied to multi-image fusion, electronic equipment and a storage medium.
Background
In recent years, deep learning methods are mature in target detection and identification applications of images, and algorithm accuracy is greatly improved compared with traditional feature extraction. For remote sensing images and aerial photography applications, a camera often cannot capture all objects at one time, and splicing and fusion processing is performed after a large number of photos of repeated superposition of target objects are shot. And then the spliced image is used as a network model for inputting deep learning. The method has the advantages that all target objects can be completed by one image, but the image effect after image splicing and fusion is not good in many cases along with the influence of the quality of the shot image and the lens angle, so that the validity of the detection result is difficult to ensure when the image with poor fusion effect is input into a network for feature extraction. Secondly, since the deep learning algorithm is trained based on samples, the images used in the training are all original images taken individually. And the images after splicing and fusion all adopt a weighted superposition mode. Therefore, the sample acquisition difference between the training image and the actually used image is caused, and the detection result is influenced by the splicing and fusion effect, so that the fused algorithm for extracting the target has great uncertainty.
Disclosure of Invention
Problems to be solved
Aiming at the problem that in the existing image recognition, spliced and fused images are adopted as training images, and the algorithm of image target extraction is not accurate due to the difference between the training images and the samples of the actually used images, the method avoids the process of fusing the images, and extracts and matches the image characteristics firstly; then, respectively using the original input of the image to perform target detection, comparing the position of a target boundary frame obtained after the detection with the characteristic matching position, and screening to leave a target object; therefore, the original feature information of the image is retained to the maximum extent, the original feature of the image is conveniently input into the machine learning model to be extracted, and the accuracy of image identification is improved.
Technical scheme
In order to solve the above problems, the present invention adopts the following technical solutions.
The invention provides an image identification method applied to multi-image fusion, which comprises the following steps:
s102: acquiring an image to be identified, calibrating the image to be identified into a template image and a target image, and respectively extracting feature points of the template image and the target image through a feature extraction algorithm;
s104: traversing the feature points of the template image, screening the feature point information through distance, and drawing a feature matching area according to the screened feature point information;
s106: detecting the template image and the target image through a target detection algorithm, and acquiring a detection frame of a target object in the template image and the target image;
s108: and calculating the relative position of the detection frame and the feature matching area, and cutting the image to be recognized according to the relative position.
In some embodiments, before the step of acquiring the image to be recognized and calibrating the image to be recognized as the template image and the target image, the method further includes:
acquiring two images to be identified, and calculating the width and the height of the two images to be identified;
in the horizontal direction, taking the maximum value of the heights in the two images to be identified as the splicing height; adding the widths of the two images to be identified in the vertical direction to obtain a spliced width;
and constructing an image drawing board according to the spliced height and width, and randomly calibrating one image as a template image and the other image as a target image.
In some embodiments, the step of extracting the feature points of the template image and the target image respectively by a feature extraction algorithm includes:
and extracting the characteristics of the template image and the target image through an AKAZE algorithm, and respectively storing the extracted characteristic point information into matrix data structures.
In some embodiments, the step of traversing feature points of the template image, and the step of filtering the feature point information by distance includes:
traversing the characteristic points of the template image through a proximity algorithm to obtain a proximity distance and a secondary proximity distance;
calculating the ratio of the approach distance to the secondary approach distance, and keeping the characteristic points of which the ratio accords with a preset numerical range as matching values in the matrix data structure; feature points for which the ratio does not meet a predetermined range of values are subtracted from the matrix data structure.
In some embodiments, the step of drawing the feature matching region according to the filtered feature point information includes:
and acquiring pixel coordinates corresponding to the feature points, and drawing a feature matching area according to the splicing direction of the target image and the template image and the pixel coordinates.
In some embodiments, calculating the relative positional relationship of the detection box and the feature matching region includes:
according to the coordinate information of the detection frame and the feature matching area;
judging whether the detection frame is in a feature matching area or not; if not, cutting the template image and the target image according to the size of the detection frame;
when the detection frame is overlapped with the feature matching area, detecting whether the target object is in the overlapped area; cutting the template image and the target image according to the judgment result; wherein the detection frame comprises a template image and a detection frame of a target object in a target image.
In some embodiments, when the detection frame overlaps the feature matching region, detecting whether the target object is within the overlap region; the step of cutting the template image and the target image according to the judgment result comprises the following steps:
if the corresponding characteristic line in the overlapping area does not detect the target object, cutting the template image and the target image according to the size of the detection frame;
if there is a detection target object in the overlapping detection corresponding to the characteristic line, determining whether the target object and the target object in the step S106 are of one type;
if not, correspondingly cutting the template image and the target image according to the size of the detection frame;
if so, cutting the detection frame with the largest area and outputting.
The second aspect of the present invention provides an image recognition method and apparatus applied to multi-image fusion, including:
the system comprises a characteristic extraction module, a feature extraction module and a feature extraction module, wherein the characteristic extraction module is used for acquiring an image to be identified, calibrating the image to be identified into a template image and a target image, and respectively extracting characteristic points of the template image and the target image through a characteristic extraction algorithm;
the screening module is used for traversing the feature points of the template image, screening the feature point information through distance, and drawing a feature matching area according to the screened feature point information;
the target object detection module is used for detecting the template image through a target detection algorithm and acquiring a detection frame of a target object in the template image; and
and the identification module is used for calculating the relative position of the detection frame and the feature matching area and cutting the template image and the target image according to the result of the relative position.
A third aspect of the present invention provides an electronic device, including a processor, an input device, an output device, and a memory, where the processor, the input device, the output device, and the memory are connected in sequence, the memory is used to store a computer program, the computer program includes program instructions, and the processor is configured to call the program instructions to execute the above method.
A fourth aspect of the invention provides a readable storage medium, the storage medium storing a computer program comprising program instructions which, when executed by a processor, cause the processor to perform the method described above.
Has the beneficial effects of
Compared with the prior art, the invention has the beneficial effects that:
(1) the method avoids the process of fusing images, and firstly extracts and matches the image characteristics; then, respectively using the original input of the image to perform target detection, comparing the position of a target boundary frame obtained after the detection with the characteristic matching position, and screening to leave a target object; original feature information of the image is reserved to the greatest extent, the original feature of the image is conveniently input into the machine learning model to be extracted, and the accuracy of image recognition is improved;
(2) the error matching elimination algorithm is used, the ratio of the approach distance to the secondary approach distance is obtained through the KNN algorithm for screening, and the accuracy of the image feature points can be guaranteed to the maximum extent;
(3) the invention abandons the fusion algorithm with longer detection time, so that the image detection speed is higher.
Drawings
The above and other objects, features and advantages of the present application will become more apparent by describing in more detail embodiments of the present application with reference to the attached drawings. The accompanying drawings are included to provide a further understanding of the embodiments of the application and are incorporated in and constitute a part of this specification, illustrate embodiments of the application and together with the description serve to explain the principles of the application. In the drawings, like reference numbers generally represent like parts or steps. In the drawings:
FIG. 1 is a flowchart of an image recognition method applied to multi-image fusion according to an embodiment of the present invention;
FIG. 2 is a block diagram of an adaptive image stitching and fusing apparatus according to an embodiment of the present invention;
FIG. 3 is a flowchart of an image adaptive boundary fusion method according to an embodiment of the present invention;
fig. 4 is a block diagram of an electronic device provided by an embodiment of the invention.
Detailed Description
Hereinafter, embodiments according to the present application will be described in detail with reference to the accompanying drawings. It should be understood that the described embodiments are only some of the embodiments of the present application, and not all of the embodiments of the present application, and it should be understood that the present application is not limited by the embodiments described herein.
Exemplary method
As shown in fig. 1 and 3, an image recognition method applied to multi-image fusion includes the following steps:
s102: the method comprises the steps of obtaining an image to be identified, calibrating the image to be identified into a template image and a target image, and respectively extracting feature points of the template image and the target image through a feature extraction algorithm.
Specifically, the image to be recognized in this example may obtain an original image of the pointer code table to be recognized in a photo or a video, for example, the image is read by a camera, and a frame of the current image stream is directly read by using a camera reading Api; randomly selecting one image as a template image, and taking the rest images as target images; and extracting the characteristics of the template image and the target image through an AKAZE algorithm, and respectively storing the extracted characteristic point information into a matrix data structure, wherein the characteristic point information comprises the characteristic point information of the template image and the characteristic point information of the target image. It should be understood by those skilled in the art that the feature extraction algorithm herein may be a SIFT (scale invariant feature transform) algorithm, etc., and is not limited herein.
As a variation column, before the steps of acquiring an image to be recognized and calibrating the image to be recognized as a template image and a target image, the method further comprises the following steps:
acquiring two images to be identified, and calculating the width and the height of the two images to be identified;
in the horizontal direction, taking the maximum value of the heights in the two images to be identified as the splicing height; adding the widths of the two images to be identified in the vertical direction to obtain a spliced width;
and constructing an image drawing board according to the spliced height and width, and randomly calibrating one image as a template image and the other image as a target image.
It should be noted that, in the embodiment of the present invention, generally, a template image may be used as a template, and features (referred to as "template features" herein) characterizing the template image or an object of interest therein are obtained from the template image, where the template features may be points and/or edges that may characterize the template image or the object image; extracting features from the target image or template image, which may also be points and/or edges; the present example extracts feature point information, but not limited thereto.
S104: traversing the feature points of the template image, screening the feature point information through distance, and drawing a feature matching area according to the screened feature point information.
Specifically, traversing the feature points of the template image through a proximity algorithm to obtain a proximity distance and a secondary proximity distance; the feature points here refer to a large number of feature points that can be obtained by using the AKAZE detection algorithm, and these feature points are described by some numerical values (e.g., illuminance, angle, gray value, etc.). The approach distance and the next approach distance are used for explaining the similarity of the KNN screening feature points, the absolute value of the subtraction of the values of the feature points of the two images is used as the distance (generally called the feature distance because the calculation mode is very similar to the distance formula), the KNN algorithm provides a plurality of similarities, the similarities are sorted according to the distance calculation, the value with the minimum feature distance of each feature is selected as the approach distance, and the next minimum value is used as the next approach distance.
Calculating a ratio of the approach distance to a next approach distance, and keeping a close-range point of which the ratio meets a preset numerical range as a matching value in the matrix data structure; deleting from the matrix data structure the near distance points for which the ratio does not conform to a predetermined range of values.
In this example, the threshold is defined as thread, the near distance value is defined as, and the next near distance value is defined as, calculated according to the method; if yes, the feature point is used as a matching point and is kept in the matrix data structure determined in the step S102; if the feature point is deleted from the matrix data structure determined in step S102, in this example, the selection of the threshold value thread may be determined according to the actual situation, and is not limited herein. The error matching elimination algorithm used in the example can obtain the ratio of the approach distance to the secondary approach distance through the KNN algorithm for screening, and the accuracy of the image feature points can be guaranteed to the greatest extent.
And acquiring pixel coordinates corresponding to the feature points, and drawing a feature matching area according to the splicing direction of the target image and the template image and the pixel coordinates.
It should be noted that the feature points detected in this example are pixel coordinates, and when horizontal stitching is required, the series of coordinates are sorted according to the x-axis (image horizontal direction); ordering according to the y axis (image vertical reversal) needing vertical splicing; the sorting way here is to bubble sort to find the maximum value of the y-axis or x-axis coordinate. If the two images are spliced in the left-right horizontal direction, the x axis (image horizontal pixels) is used as a separation area; similarly, how to vertically stitch the top and bottom, the y-axis (vertical pixels of the image) is used as the separation region.
In one possible embodiment, the target image maximum coordinate position (x 1, y 1) and the source image minimum coordinate position (x 2, y 2) are determined, where x1 is the abscissa of the target image maximum coordinate position, x2 is the abscissa of the target image minimum coordinate position, and y1 is the ordinate of the target image maximum coordinate position; y2 is the ordinate of the maximum coordinate position of the target image; and taking the coordinates of the two vertexes as the coordinates of the feature matching area, and taking the coordinates of (x 1, y 2) and (x 2, y 1) as the other two vertexes to obtain a rectangular feature matching area.
S106: and detecting the template image and the target image through a target detection algorithm, and acquiring a detection frame of a target object in the template image and the target image.
Specifically, the target detection algorithm may be that a target image training set is used to train a target detection YOLO model to obtain a detection model of the target object, and then the target detection model is used to detect a test sample to be identified to obtain a coordinate position of a rectangular bounding box with the target object in the test set.
S108: and calculating the relative position of the detection frame and the feature matching area, and cutting the image to be recognized according to the result of the relative position.
Specifically, according to the coordinates of the detection frame and the feature matching area; judging whether the detection frame is in a feature matching area or not; if not, cutting the target image and the template image according to the size of the detection frame; wherein the detection frame comprises a template image and a detection frame of a target object in a target image.
When the detection frame is overlapped with the feature matching area, detecting whether the target object is in the overlapped area; cutting the target image and the template image according to the judgment result;
if the corresponding characteristic line in the overlapping area does not detect the target object, the target image and the template image are detected according to the size of the detection frame;
if there is a detection target object in the overlapping detection corresponding to the characteristic line, determining whether the target object and the target object in the step S106 are of one type; the class is the meaning of class category in target detection, for example, if the detected categories are all people, the logic judgment is carried out to determine whether the areas are repeated; if the detection categories are one person and one vase, both the person and the vase need to be kept.
If not, cutting the target image and the template image according to the size of the detection frame; if so, the detection frame with the largest area is cut out and output as the original image. It should be noted that the maximum area here refers to the detection frame with the largest area when the detection frame obtained in the target image is compared with the detection frame in the template image. The class is the meaning of class category in target detection, for example, if the detected categories are all people, the logic judgment is carried out to determine whether the areas are repeated; if the detection categories are one person and one vase, both the person and the vase need to be kept.
Exemplary devices
As shown in fig. 2, an adaptive image stitching and fusing apparatus includes:
the feature extraction module 20 is configured to acquire an image to be identified, calibrate the image to be identified into a template image and a target image, and extract feature points of the template image and the target image respectively through a feature extraction algorithm.
The module is further used for acquiring two images to be identified and calculating the width and the height of the two images to be identified; in the horizontal direction, taking the maximum value of the heights in the two images to be identified as the splicing height; adding the widths of the two images to be identified in the vertical direction to obtain a spliced width; constructing an image drawing board according to the spliced height and width, and randomly calibrating one image as a template image and the other image as a target image; and extracting the characteristics of the template image and the target image through an AKAZE algorithm, and respectively storing the extracted characteristic point information into matrix data structures.
And the screening module 30 is configured to traverse the feature points of the template image, screen the feature point information by distance, and draw a feature matching area according to the screened feature point information. Specifically, traversing the feature points of the template image through a proximity algorithm to obtain a proximity distance and a secondary proximity distance; calculating a ratio of the approach distance to a next approach distance, and keeping a close-range point of which the ratio meets a preset numerical range as a matching value in the matrix data structure; deleting the close range points of which the ratio does not conform to a predetermined numerical range from the matrix data structure; and acquiring pixel coordinates corresponding to the feature points, and drawing a feature matching area according to the splicing direction of the target image and the template image and the pixel coordinates.
The target object detection module 40 is configured to detect the template image through a target detection algorithm, and acquire a detection frame of a target object in the template image; and
and the identification module 50 is used for calculating the relative position of the detection frame and the feature matching area, and cutting the image to be identified according to the result of the relative position. Specifically, according to the coordinates of the detection frame and the feature matching area; judging whether the detection frame is in a feature matching area or not; if not, cutting the template image and the target image according to the size of the detection frame; when the detection frame is overlapped with the feature matching area, detecting whether the target object is in the overlapped area; and cutting the template image and the target image according to the judgment result. If the corresponding characteristic line in the overlapping area does not detect the target object, cutting the template image and the target image according to the size of the detection frame; if there is a detection target object in the overlapping detection corresponding to the characteristic line, determining whether the target object and the target object in the step S106 are of one type; if not, cutting the template image and the target image according to the size of the detection frame; if so, the detection frame with the largest area is cut out and output as the original image.
Exemplary electronic device
Next, an electronic apparatus according to an embodiment of the present application is described with reference to fig. 4. The electronic device may be the removable device itself or a stand-alone device separate therefrom that may communicate with the removable device to receive the captured input signals therefrom and to transmit the combined image information thereto.
FIG. 4 illustrates a block diagram of an electronic device in accordance with an embodiment of the present application.
As shown in fig. 4, the electronic device 10 includes one or more processors 11 and memory 12.
The processor 11 may be a Central Processing Unit (CPU) or other form of processing unit having data processing capabilities and/or instruction execution capabilities, and may control other components in the electronic device 10 to perform desired functions.
In one example, the electronic device 10 may further include: an input device 13 and an output device 14, which are interconnected by a bus system and/or other form of connection mechanism (not shown). For example, the input device 13 may include various devices such as a camera, a video player, and the like. The input device 13 may also include, for example, a keyboard, a mouse, and the like. The output devices 14 may include, for example, a display, speakers, a printer, and a communication network and its connected remote output devices, among others.
Of course, for simplicity, only some of the components of the electronic device 10 relevant to the present application are shown in fig. 4, omitting components such as buses, input/output interfaces, and the like. In addition, the electronic device 10 may include any other suitable components depending on the particular application.
Exemplary computer program product and computer-readable storage Medium
In addition to the above-described methods and apparatus, embodiments of the present application may also be a computer program product comprising computer program instructions that, when executed by a processor, cause the processor to perform the steps in the decision-making method according to various embodiments of the present application described in the "exemplary methods" section of this specification above.
The computer program product may be written with program code for performing the operations of embodiments of the present application in any combination of one or more programming languages, including an object oriented programming language such as Java, C + + or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computing device, partly on the user's device, as a stand-alone software package, partly on the user's computing device and partly on a remote computing device, or entirely on the remote computing device or server.
Furthermore, embodiments of the present application may also be a computer-readable storage medium having stored thereon computer program instructions that, when executed by a processor, cause the processor to perform steps in a decision method according to various embodiments of the present application described in the "exemplary methods" section above of this specification.
The computer-readable storage medium may take any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. A readable storage medium may include, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the foregoing. More specific examples (a non-exhaustive list) of the readable storage medium include: an electrical connection having one or more wires, a portable disk, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
The foregoing describes the general principles of the present application in conjunction with specific embodiments, however, it is noted that the advantages, effects, etc. mentioned in the present application are merely examples and are not limiting, and they should not be considered essential to the various embodiments of the present application. Furthermore, the foregoing disclosure of specific details is for the purpose of illustration and description and is not intended to be limiting, since the foregoing disclosure is not intended to be exhaustive or to limit the disclosure to the precise details disclosed.
The block diagrams of devices, apparatuses, systems referred to in this application are only given as illustrative examples and are not intended to require or imply that the connections, arrangements, configurations, etc. must be made in the manner shown in the block diagrams. These devices, apparatuses, devices, systems may be connected, arranged, configured in any manner, as will be appreciated by those skilled in the art. Words such as "including," "comprising," "having," and the like are open-ended words that mean "including, but not limited to," and are used interchangeably therewith. The words "or" and "as used herein mean, and are used interchangeably with, the word" and/or, "unless the context clearly dictates otherwise. The word "such as" is used herein to mean, and is used interchangeably with, the phrase "such as but not limited to".
It should also be noted that in the devices, apparatuses, and methods of the present application, the components or steps may be decomposed and/or recombined. These decompositions and/or recombinations are to be considered as equivalents of the present application.
The previous description of the disclosed aspects is provided to enable any person skilled in the art to make or use the present application. Various modifications to these aspects will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other aspects without departing from the scope of the application. Thus, the present application is not intended to be limited to the aspects shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.
The foregoing description has been presented for purposes of illustration and description. Furthermore, the description is not intended to limit embodiments of the application to the form disclosed herein. While a number of example aspects and embodiments have been discussed above, those of skill in the art will recognize certain variations, modifications, alterations, additions and sub-combinations thereof.
Claims (10)
1. An image recognition method applied to multi-image fusion is characterized by comprising the following steps:
s102: acquiring an image to be identified, calibrating the image to be identified into a template image and a target image, and respectively extracting feature points of the template image and the target image through a feature extraction algorithm;
s104: traversing the feature points of the template image, screening the feature point information through distance, and drawing a feature matching area according to the screened feature point information;
s106: detecting the template image and the target image through a target detection algorithm, and acquiring a detection frame of a target object in the template image and the target image;
s108: and calculating the relative position of the detection frame and the feature matching area, and cutting the image to be recognized according to the relative position.
2. The image recognition method applied to multi-image fusion according to claim 1, wherein before the step of obtaining the image to be recognized and calibrating the image to be recognized as the template image and the target image, the method further comprises:
acquiring two images to be identified, and calculating the width and the height of the two images to be identified;
in the horizontal direction, taking the maximum value of the heights in the two images to be identified as the splicing height; adding the widths of the two images to be identified in the vertical direction to obtain a spliced width;
and constructing an image drawing board according to the spliced height and width, and randomly calibrating one image as a template image and the other image as a target image.
3. The image recognition method applied to multi-image fusion according to claim 2, wherein the step of extracting the feature points of the template image and the target image respectively through a feature extraction algorithm comprises:
and extracting the characteristics of the template image and the target image through an AKAZE algorithm, and respectively storing the extracted characteristic point information into matrix data structures.
4. The image recognition method applied to multi-image fusion according to claim 3, wherein the step of traversing the feature points of the template image and screening the feature point information by distance comprises:
traversing the characteristic points of the template image through a proximity algorithm to obtain a proximity distance and a secondary proximity distance;
calculating the ratio of the approach distance to the secondary approach distance, and keeping the characteristic points of which the ratio accords with a preset numerical range as matching values in the matrix data structure; feature points for which the ratio does not meet a predetermined range of values are subtracted from the matrix data structure.
5. The image recognition method applied to multi-image fusion according to claim 4, wherein the step of drawing the feature matching region according to the filtered feature point information comprises:
and acquiring pixel coordinates corresponding to the feature points, and drawing a feature matching area according to the splicing direction of the target image and the template image and the pixel coordinates.
6. The image recognition method applied to multi-image fusion according to claim 1, wherein calculating the relative positional relationship between the detection frame and the feature matching region comprises:
according to the coordinate information of the detection frame and the feature matching area;
judging whether the detection frame is in a feature matching area or not; if not, cutting the template image and the target image according to the size of the detection frame;
when the detection frame is overlapped with the feature matching area, detecting whether the target object is in the overlapped area; cutting the template image and the target image according to the judgment result; wherein the detection frame comprises a template image and a detection frame of a target object in a target image.
7. The image recognition method applied to multi-image fusion according to claim 6, wherein when the detection frame overlaps the feature matching region, whether the target object is within the overlapping region is detected; the step of cutting the template image and the target image according to the judgment result comprises the following steps:
if the corresponding characteristic line in the overlapping area does not detect the target object, cutting the template image and the target image according to the size of the detection frame;
if there is a detection target object in the overlapping detection corresponding to the characteristic line, determining whether the target object and the target object in the step S106 are of one type;
if not, correspondingly cutting the template image and the target image according to the size of the detection frame;
if so, cutting the detection frame with the largest area and outputting.
8. An image recognition method and device applied to multi-image fusion is characterized by comprising the following steps:
the system comprises a characteristic extraction module, a feature extraction module and a feature extraction module, wherein the characteristic extraction module is used for acquiring an image to be identified, calibrating the image to be identified into a template image and a target image, and respectively extracting characteristic points of the template image and the target image through a characteristic extraction algorithm;
the screening module is used for traversing the feature points of the template image, screening the feature point information through distance, and drawing a feature matching area according to the screened feature point information;
the target object detection module is used for detecting the template image through a target detection algorithm and acquiring a detection frame of a target object in the template image; and
and the identification module is used for calculating the relative position of the detection frame and the feature matching area and cutting the template image and the target image according to the result of the relative position.
9. An electronic device comprising a processor, an input device, an output device, and a memory, the processor, the input device, the output device, and the memory being connected in series, the memory being configured to store a computer program comprising program instructions, the processor being configured to invoke the program instructions to perform the method of any of claims 1-7.
10. A readable storage medium, characterized in that the storage medium stores a computer program comprising program instructions which, when executed by a processor, cause the processor to carry out the method according to any one of claims 1-7.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010947804.8A CN112085106A (en) | 2020-09-10 | 2020-09-10 | Image identification method and device applied to multi-image fusion, electronic equipment and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010947804.8A CN112085106A (en) | 2020-09-10 | 2020-09-10 | Image identification method and device applied to multi-image fusion, electronic equipment and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112085106A true CN112085106A (en) | 2020-12-15 |
Family
ID=73736349
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010947804.8A Pending CN112085106A (en) | 2020-09-10 | 2020-09-10 | Image identification method and device applied to multi-image fusion, electronic equipment and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112085106A (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112926439A (en) * | 2021-02-22 | 2021-06-08 | 深圳中科飞测科技股份有限公司 | Detection method and device, detection equipment and storage medium |
CN113220927A (en) * | 2021-05-08 | 2021-08-06 | 百度在线网络技术(北京)有限公司 | Image detection method, device, equipment and storage medium |
CN113837270A (en) * | 2021-09-18 | 2021-12-24 | 广东人工智能与先进计算研究院 | Target identification method, device, equipment and storage medium |
WO2023066147A1 (en) * | 2021-10-19 | 2023-04-27 | 中国第一汽车股份有限公司 | Image processing method and apparatus, and electronic device and medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102821238A (en) * | 2012-03-19 | 2012-12-12 | 北京泰邦天地科技有限公司 | Wide-field ultra-high-resolution imaging system |
US9990753B1 (en) * | 2017-01-11 | 2018-06-05 | Macau University Of Science And Technology | Image stitching |
CN110991283A (en) * | 2019-11-21 | 2020-04-10 | 北京格灵深瞳信息技术有限公司 | Re-recognition and training data acquisition method and device, electronic equipment and storage medium |
CN111091091A (en) * | 2019-12-16 | 2020-05-01 | 北京迈格威科技有限公司 | Method, device and equipment for extracting target object re-identification features and storage medium |
CN111104538A (en) * | 2019-12-06 | 2020-05-05 | 深圳久凌软件技术有限公司 | Fine-grained vehicle image retrieval method and device based on multi-scale constraint |
-
2020
- 2020-09-10 CN CN202010947804.8A patent/CN112085106A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102821238A (en) * | 2012-03-19 | 2012-12-12 | 北京泰邦天地科技有限公司 | Wide-field ultra-high-resolution imaging system |
US9990753B1 (en) * | 2017-01-11 | 2018-06-05 | Macau University Of Science And Technology | Image stitching |
CN110991283A (en) * | 2019-11-21 | 2020-04-10 | 北京格灵深瞳信息技术有限公司 | Re-recognition and training data acquisition method and device, electronic equipment and storage medium |
CN111104538A (en) * | 2019-12-06 | 2020-05-05 | 深圳久凌软件技术有限公司 | Fine-grained vehicle image retrieval method and device based on multi-scale constraint |
CN111091091A (en) * | 2019-12-16 | 2020-05-01 | 北京迈格威科技有限公司 | Method, device and equipment for extracting target object re-identification features and storage medium |
Non-Patent Citations (2)
Title |
---|
ZHAO YI-LI: "Automatic panorama recognition and stitching method based on graph structure", COMPUTER ENGINEERING AND DESIGN, vol. 34, no. 6, 3 April 2014 (2014-04-03), pages 2067 - 2070 * |
么鸿原;王海鹏;焦莉;林雪原;: "基于改进SURF算法的无人机遥感图像拼接方法", 海军航空工程学院学报, no. 02, 30 April 2018 (2018-04-30) * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112926439A (en) * | 2021-02-22 | 2021-06-08 | 深圳中科飞测科技股份有限公司 | Detection method and device, detection equipment and storage medium |
CN113220927A (en) * | 2021-05-08 | 2021-08-06 | 百度在线网络技术(北京)有限公司 | Image detection method, device, equipment and storage medium |
CN113837270A (en) * | 2021-09-18 | 2021-12-24 | 广东人工智能与先进计算研究院 | Target identification method, device, equipment and storage medium |
CN113837270B (en) * | 2021-09-18 | 2022-08-30 | 广东人工智能与先进计算研究院 | Target identification method, device, equipment and storage medium |
WO2023066147A1 (en) * | 2021-10-19 | 2023-04-27 | 中国第一汽车股份有限公司 | Image processing method and apparatus, and electronic device and medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112085106A (en) | Image identification method and device applied to multi-image fusion, electronic equipment and storage medium | |
CN112381075B (en) | Method and system for carrying out face recognition under specific scene of machine room | |
CN109426835B (en) | Information processing apparatus, control method of information processing apparatus, and storage medium | |
US8019164B2 (en) | Apparatus, method and program product for matching with a template | |
CN110390229B (en) | Face picture screening method and device, electronic equipment and storage medium | |
CN102365645A (en) | Organizing digital images by correlating faces | |
CN109389105B (en) | Multitask-based iris detection and visual angle classification method | |
CN107272899B (en) | VR (virtual reality) interaction method and device based on dynamic gestures and electronic equipment | |
CN114549993B (en) | Method, system and device for grading line segment image in experiment and readable storage medium | |
US11200694B2 (en) | Apparatus and method for extracting object information | |
CN111986214B (en) | Construction method of pedestrian crossing in map and electronic equipment | |
CN110737785A (en) | picture labeling method and device | |
CN114627482A (en) | Method and system for realizing table digital processing based on image processing and character recognition | |
CN113281780B (en) | Method and device for marking image data and electronic equipment | |
JP2022009474A (en) | System and method for detecting lines in vision system | |
CN109978829B (en) | Detection method and system for object to be detected | |
EP2887261B1 (en) | Information processing device, information processing method, and program | |
CN111488776A (en) | Object detection method, object detection device and electronic equipment | |
CN113269153B (en) | Form identification method and device | |
CN114140782A (en) | Text recognition method and device, electronic equipment and storage medium | |
KR20220083347A (en) | Method, apparatus, and computer program for measuring volume of objects by using image | |
CN110956130A (en) | Method and system for four-level face detection and key point regression | |
KR102243884B1 (en) | Method for inspecting product based on vector modeling and Apparatus thereof | |
CN113537199B (en) | Image boundary box screening method, system, electronic device and medium | |
CN113255632B (en) | Camera parameter adjusting method, device, equipment and medium based on license plate recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |