CN117496532A - Intelligent recognition tool based on 0CR - Google Patents

Intelligent recognition tool based on 0CR Download PDF

Info

Publication number
CN117496532A
CN117496532A CN202311487549.3A CN202311487549A CN117496532A CN 117496532 A CN117496532 A CN 117496532A CN 202311487549 A CN202311487549 A CN 202311487549A CN 117496532 A CN117496532 A CN 117496532A
Authority
CN
China
Prior art keywords
image
module
features
feature
pixel value
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311487549.3A
Other languages
Chinese (zh)
Inventor
刘晓龙
孙强
刘新语
陈洋洋
陆昕雨
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhongtong Hexin Technology Co ltd
Original Assignee
Zhongtong Hexin Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhongtong Hexin Technology Co ltd filed Critical Zhongtong Hexin Technology Co ltd
Priority to CN202311487549.3A priority Critical patent/CN117496532A/en
Publication of CN117496532A publication Critical patent/CN117496532A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/19Recognition using electronic means
    • G06V30/191Design or setup of recognition systems or techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
    • G06V30/19173Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/14Image acquisition
    • G06V30/1431Illumination control
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/16Image preprocessing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/18Extraction of features or characteristics of the image
    • G06V30/1801Detecting partial patterns, e.g. edges or contours, or configurations, e.g. loops, corners, strokes or intersections
    • G06V30/18019Detecting partial patterns, e.g. edges or contours, or configurations, e.g. loops, corners, strokes or intersections by matching or filtering
    • G06V30/18038Biologically-inspired filters, e.g. difference of Gaussians [DoG], Gabor filters
    • G06V30/18048Biologically-inspired filters, e.g. difference of Gaussians [DoG], Gabor filters with interaction between the responses of different filters, e.g. cortical complex cells
    • G06V30/18057Integrating the filters into a hierarchical structure, e.g. convolutional neural networks [CNN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/19Recognition using electronic means
    • G06V30/19007Matching; Proximity measures
    • G06V30/19013Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/19Recognition using electronic means
    • G06V30/191Design or setup of recognition systems or techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
    • G06V30/1912Selecting the most significant subset of features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/24Character recognition characterised by the processing or recognition method
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/26Techniques for post-processing, e.g. correcting the recognition result
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Biodiversity & Conservation Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a 0 CR-based intelligent recognition tool, which belongs to the field of image recognition tools and comprises a data acquisition module, a preprocessing module, a feature extraction and selection module, a classification decision module and an information recognition module; the data acquisition module is used for acquiring the image file to be processed transmitted by the terminal equipment; the preprocessing module is used for preprocessing the image file to be processed, wherein the preprocessing comprises the steps of increasing the definition of the image to be processed, adjusting the angle of the image and normalizing; the feature extraction and selection module is used for extracting and selecting the same feature information of the preprocessed image file; the classification decision module is used for acquiring the information of the image to be processed and classifying the image according to the scene type; the information identification module is used for matching the classification and result of the image acquisition with the original template in the image identification information base and outputting a matching identification result in the image to be processed. The intelligent recognition tool can effectively improve recognition effect and recognition speed.

Description

Intelligent recognition tool based on 0CR
Technical Field
The invention relates to an image recognition tool, in particular to a smart recognition tool based on 0 CR.
Background
0CR (image recognition technology) refers to a process in which an electronic device (e.g., a scanner or a digital camera) checks characters printed on paper, determines the shape thereof by detecting dark and light patterns, and translates the shape into computer text by a character recognition method. OCR image recognition technology belongs to artificial intelligence field, and the main index of measuring the performance of an OCR system is: rejection rate, false recognition rate, recognition speed, user interface friendliness, product stability, usability, feasibility and the like.
At present, the image recognition technology on the market basically performs training and learning based on a large amount of data with high quality of original images, and cannot meet the problems of memory storage, recognition failure and the like. The following problems are mainly categorized: affected by light: under the influence of light, the light sensing capability of the camera is disturbed, and particularly, the model required at different stages in the daytime is different from the model required at night, so that the influence of random events cannot be completely solved. Limited to image effects: if the interference elements increase during image recognition due to the problems of external interference, poor image quality (such as image blurring) and the like, the accuracy of image recognition may be affected. A large amount of annotation data is needed for training: many deep learning algorithms require a large amount of annotation data to train to achieve higher recognition accuracy. However, acquiring these annotation data requires a significant amount of manpower, material resources, and time. The calculated amount is large: image recognition algorithms typically require extensive computation and high performance computer hardware support, which limits their application to large-scale data sets. Therefore, how to improve the recognition effect and recognition speed of the image recognition tool is a current urgent problem to be solved.
Accordingly, one skilled in the art would provide a 0 CR-based intelligent recognition tool to solve the problems set forth in the background.
Disclosure of Invention
The invention aims to provide a 0 CR-based intelligent recognition tool which can effectively improve recognition effect and recognition speed so as to solve the problems in the prior art.
In order to achieve the above purpose, the present invention provides the following technical solutions:
the intelligent recognition tool based on 0CR comprises a data acquisition module, a preprocessing module, a feature extraction and selection module, a classification decision module and an information recognition module;
the data acquisition module is used for acquiring an image file to be processed transmitted by the terminal equipment; the preprocessing module is used for preprocessing the image file to be processed, wherein the preprocessing comprises the steps of increasing the definition of the image to be processed, adjusting the angle of the image and normalizing; the feature extraction and selection module is used for extracting and selecting the same feature information of the preprocessed image file; the classification decision module is used for acquiring the information of the image to be processed and classifying the image according to the scene type; the information identification module is used for matching the classification and result of the image acquisition with the original template in the image identification information base and outputting a matching identification result in the image to be processed.
As a further scheme of the invention: the specific process of preprocessing the image file to be processed by the preprocessing module is as follows:
normalizing the size of the acquired image, and acquiring illumination intensity parameters related to the acquired image;
judging whether the collected image has abnormal illumination or not based on the illumination intensity parameter, and simultaneously analyzing the abnormal type of the abnormal illumination image;
carrying out illumination correction on the illumination abnormal image;
performing dotting and positioning on the image file with the increased definition of the image to be processed to obtain image features with background colors;
image segmentation is performed on the image based on the set of image background colors.
As still further aspects of the invention: the feature extraction and selection module is used for carrying out reprocessing on the image file before extracting and selecting the same feature information of the preprocessed image file, and specifically comprises the following steps:
based on an image segmentation recognition result of the illumination correction image, invoking a corresponding shielding elimination algorithm to clear shielding information of image features of the object to be recognized;
and calling a corresponding background splashing algorithm to perform background removing and subtracting operation on the image features of the objects to be identified, and forming feature vectors by utilizing the image features of the objects to be identified.
As still further aspects of the invention: the specific process of extracting and selecting the same characteristic information of the preprocessed image file by the characteristic extracting and selecting module is as follows:
(1) Feature classification: physical features, structural features, mathematical features;
(2) Feature formation generating a set of basic features from the identified image, such generated features being called raw features;
(3) Extracting features, namely representing samples by a low-dimensional space through a mapping method, wherein the mapped secondary features are linear combinations of original features;
(4) Feature selection a number of most efficient features are selected from a set of features to achieve a reduction in feature space dimension.
As still further aspects of the invention: the specific process of classifying the images according to the scene types by the classification decision module is as follows:
preprocessing an illumination correction image, specifically: determining a difference operator, and carrying out gray level difference in the neighborhood to obtain a difference gray level map; determining a color difference valve value, and separating an image area to be identified in the differential gray scale image from background colors by using the color difference valve value;
and determining a pixel value interval to which a pixel value of each background color area pixel point in the image belongs, and obtaining a first pixel identification information sequence based on a corresponding relation between the pixel value interval to which the pixel value of each pixel point belongs and preset pixel value interval identification information.
As still further aspects of the invention: the specific obtaining process of the first pixel identification information sequence comprises the following steps:
determining a plurality of continuous pixel value intervals divided in advance, and determining a pixel value interval to which each pixel value of each pixel point in the first area belongs;
dividing the first region pixel points based on the pixel value intervals to obtain at least one first pixel point pixel value string, wherein the pixel value intervals of the pixel points in any two adjacent first pixel point pixel value strings are different;
and ordering the pixel identification information of each first pixel point pixel value string to obtain a corresponding first pixel identification information sequence.
As still further aspects of the invention: the specific process of the information identification module for matching the classification and result of the image acquisition with the original template in the image identification information base is as follows:
acquiring image features of an object to be identified, and carrying out feature extraction on the image features based on a convolution algorithm to obtain an intermediate feature matrix;
sorting and reorganizing the intermediate feature matrix to obtain a post feature matrix;
performing feature extraction on the post feature matrix based on a convolution algorithm to obtain an image feature set;
and matching the image feature set with the original template in the image identification information base.
As still further aspects of the invention: and the output matching recognition result in the image to be processed is displayed through the mobile terminal and the computer terminal, and meanwhile, the manual adjustment of the recognition result is supported, so that the accuracy of the recognition result is ensured.
Compared with the prior art, the invention has the beneficial effects that:
1. the automatic and cost-saving image recognition technology can automatically process a large number of images, automatically increase the image recognizability, reduce the cost of manual intervention and the error rate, and improve the working efficiency and the recognition speed.
2. Real-time performance and accuracy, image recognition can acquire real-time data through real-time processing, compares with manual processing, promotes the rate of accuracy and the recognition effect of discernment greatly.
3. The expansibility and the intellectualization, the image recognition technology can be continuously upgraded and expanded through various algorithms and models, and the requirements of different fields can be better met.
Drawings
FIG. 1 is a block diagram of a smart identification tool based on 0 CR.
Detailed Description
In order to better understand the above technical solutions, the following detailed description will refer to the accompanying drawings and specific embodiments.
Referring to fig. 1, in the embodiment of the invention, a smart recognition tool based on 0CR includes a data acquisition module, a preprocessing module, a feature extraction and selection module, a classification decision module, and an information recognition module; the data acquisition module is used for acquiring the image file to be processed transmitted by the terminal equipment; the preprocessing module is used for preprocessing the image file to be processed, wherein the preprocessing comprises the steps of increasing the definition of the image to be processed, adjusting the angle of the image and normalizing; the feature extraction and selection module is used for extracting and selecting the same feature information of the preprocessed image file; the classification decision module is used for acquiring the information of the image to be processed and classifying the image according to the scene type; the information identification module is used for matching the classification and result of the image acquisition with the original template in the image identification information base and outputting a matching identification result in the image to be processed. The intelligent recognition tool can effectively improve recognition effect and recognition speed, and high-accuracy information recognition and classification of images are achieved through an image recognition technology.
In this embodiment: the specific process of preprocessing the image file to be processed by the preprocessing module is as follows: normalizing the size of the acquired image, and acquiring illumination intensity parameters related to the acquired image; judging whether the collected image has abnormal illumination or not based on the illumination intensity parameter, and simultaneously analyzing the abnormal type of the abnormal illumination image; carrying out illumination correction on the illumination abnormal image; performing dotting and positioning on the image file with the increased definition of the image to be processed to obtain image features with background colors; image segmentation is performed on the image based on the set of image background colors. The image file is kept available to the maximum extent by preprocessing (such as image enhancement and normalization) the input image file. The image normalization processing is a method for preprocessing image data, and aims to adjust pixel values in an image to be within a reasonable range. There are two common methods of image normalization. One is to subtract the minimum pixel value of the image from all the pixel values in the image, then divide the resulting difference by the difference between the maximum pixel value and the minimum pixel value of the image, and the result is the normalized pixel value. The other is to subtract the average pixel value of the image from all the pixel values in the image, and then divide the obtained difference by the standard deviation of the image, so that the normalized pixel value is obtained. Illumination correction of an image refers to adjusting the brightness, contrast, and hue of the image to improve the quality and readability of the image. In the process of illumination correction of an image, common algorithms include histogram equalization, adaptive histogram equalization, retinex algorithm, and the like. Pointing an image file is finding one or more points on the image and determining its location. In image processing, description point positioning is often used for image stitching, image registration, object tracking, and other tasks. By finding the feature points between the two images and determining the transformation relationship between them, the two images can be aligned and stitched into a larger image. In addition, tracking and identification of the target can be achieved by locating the position of the target in the image.
In this embodiment: the feature extraction and selection module is used for carrying out reprocessing on the image file before extracting and selecting the same feature information of the preprocessed image file, and specifically comprises the following steps: based on an image segmentation recognition result of the illumination correction image, invoking a corresponding shielding elimination algorithm to clear shielding information of image features of the object to be recognized; and calling a corresponding background splashing algorithm to perform background removing and subtracting operation on the image features of the objects to be identified, and forming feature vectors by utilizing the image features of the objects to be identified. The setting can facilitate the subsequent extraction and selection of the same characteristic information. The occlusion elimination algorithm is a commonly used algorithm in graphic rendering, and is mainly used for eliminating models occluded by other objects, so that the rendering efficiency is improved. The optimal shielding and eliminating algorithm can select visible objects to render, namely, before graphic rendering, judging which models are shielded, eliminating the shielded models, and only rendering the model which is not shielded, so that the workload of rendering is reduced. The background removing operation of the background splashing algorithm is realized mainly by calculating the projection of liquid drops on the background, and the specific steps are as follows:
calculating a projection of the droplet on the background, first, calculating a projection of the droplet on the background using hydrodynamic simulation software, which projection may represent the position, size and shape of the droplet on the background;
a mask is created, based on the projection of the drop onto the background, a mask is created for separating the drop from the background. The mask is a gray scale image in which the shape of the drop is represented as white (or high brightness) and the background is represented as black (or low brightness).
A mask is applied. A mask is applied to the input image to effect the background subtraction operation. Specifically, the mask is bit-operated from the input image, separating the drop from the background.
The color and brightness are adjusted. After the background subtraction operation, the color and brightness of the droplets and the background may need to be adjusted to make them more natural and realistic.
In this embodiment: the specific process of extracting and selecting the same characteristic information of the preprocessed image file by the characteristic extracting and selecting module is as follows: (1) Feature classification: physical features, structural features, mathematical features; (2) Feature formation generating a set of basic features from the identified image, the features so generated being called original features by calculation (when the identified object is a digital image); (3) The number of original features is probably large, a sample is represented by a low-dimensional space through a mapping (or transformation) method, the process is called feature extraction, and the mapped secondary features are linear combinations of the original features; (4) Feature selection a number of most efficient features are selected from a set of features to achieve a reduction in feature space dimension. The setting can accelerate the extraction and selection of the same characteristic information of the preprocessed image file.
In this embodiment: the specific process of classifying the images according to the scene types by the classifying decision module is as follows: preprocessing an illumination correction image, specifically: determining a difference operator, and carrying out gray level difference in the neighborhood to obtain a difference gray level map; determining a color difference valve value, and separating an image area to be identified in the differential gray scale image from background colors by using the color difference valve value; and determining a pixel value interval to which a pixel value of each background color area pixel point in the image belongs, and obtaining a first pixel identification information sequence based on a corresponding relation between the pixel value interval to which the pixel value of each pixel point belongs and preset pixel value interval identification information. The setting automatically identifies the scene type in the image and various elements in the scene by classifying and annotating different positions of the image file. The differential operator is an operator, and if Δf (x) =f (x+1) -f (x) is recorded for any real function f (x), Δ is referred to as a forward differential operator, and is simply referred to as a differential operator. The difference is one of the basic concepts of computational mathematics, and refers to the change amount of a discrete function on discrete nodes.
In this embodiment: the specific obtaining process of the first pixel identification information sequence is as follows: determining a plurality of continuous pixel value intervals divided in advance, and determining a pixel value interval to which each pixel value of each pixel point in the first area belongs; dividing the first region pixel points based on the pixel value intervals to obtain at least one first pixel point pixel value string, wherein the pixel value intervals of the pixel points in any two adjacent first pixel point pixel value strings are different; and ordering the pixel identification information of each first pixel point pixel value string to obtain a corresponding first pixel identification information sequence. The first pixel identification information sequence refers to information carried by each pixel point in the image, and is used for describing various properties of the image, such as color, brightness, contrast and the like. In image processing, a sequence of pixel identification information is typically used to perform various operations on an image, such as scaling, cropping, rotation, and the like. By processing the pixel identification information sequence, various analyses and processes of the image, such as image segmentation, feature extraction, object detection, and the like, can be realized. Therefore, the first pixel identification information sequence has an important meaning in image processing.
In this embodiment: the specific process of the information identification module for matching the classification and result of the image acquisition with the original template in the image identification information base is as follows: acquiring image features of an object to be identified, and carrying out feature extraction on the image features based on a convolution algorithm to obtain an intermediate feature matrix; sorting and reorganizing the intermediate feature matrix to obtain a post feature matrix; performing feature extraction on the post feature matrix based on a convolution algorithm to obtain an image feature set; and matching the image feature set with the original template in the image identification information base. The image recognition technique employed in the present application can customize the optimization algorithm, improve the model structure to ensure that its effect reaches a desired level, wherein the convolution algorithm is an algorithm for image processing that can apply a function (convolution kernel) to each pixel of the image and store the calculation result in a new image. This process is typically done in the frequency domain, and the computation can be accelerated by fourier transformation. In convolution algorithms, the convolution kernel is typically a small matrix whose center corresponds to each pixel in the image. Each element in the convolution kernel has a weight for calculation with the pixel value at the corresponding position in the image. This calculation is to multiply each element of the convolution kernel with the image pixel value it covers and then add the results of all the products to get the new pixel value for that location. Convolution algorithms may be used for a variety of different image processing tasks such as filtering, edge detection, feature extraction, etc. It is a very flexible and powerful tool that can be used to process a variety of different types of images and the weights of the convolution kernels can be adjusted to achieve different processing effects. The middle feature matrix refers to the feature matrix output by the convolution layer in the convolution neural network. The post feature matrix refers to the feature matrix processed by the convolution layer and the activation function in the convolution neural network.
In this embodiment: the output matching recognition result in the image to be processed is displayed through the mobile terminal and the computer terminal, and meanwhile, the manual adjustment of the recognition result is supported, so that the accuracy of the recognition result is ensured.
The invention has the advantages of automation and cost saving, and the image recognition technology can automatically process a large number of images, automatically increase the image recognizability, reduce the cost of manual intervention and error rate, and improve the working efficiency and recognition speed. In terms of instantaneity and accuracy, the image recognition of the method can acquire real-time data through real-time processing, and compared with manual processing, the method greatly improves recognition accuracy and recognition effect. In addition, the invention has expansibility and intellectualization, the image recognition technology can be continuously upgraded and expanded through various algorithms and models, and the requirements of different fields can be better met.
It will be apparent to those skilled in the art that various modifications and variations can be made to the present invention without departing from the spirit or scope of the invention. Thus, it is intended that the present invention also include such modifications and alterations insofar as they come within the scope of the appended claims or the equivalents thereof.
The foregoing description is only a preferred embodiment of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art, who is within the scope of the present invention, should make equivalent substitutions or modifications according to the technical solution of the present invention and the inventive concept thereof, and should be covered by the scope of the present invention.

Claims (8)

1. The intelligent recognition tool based on 0CR is characterized by comprising a data acquisition module, a preprocessing module, a feature extraction and selection module, a classification decision module and an information recognition module;
the data acquisition module is used for acquiring an image file to be processed transmitted by the terminal equipment; the preprocessing module is used for preprocessing the image file to be processed, wherein the preprocessing comprises the steps of increasing the definition of the image to be processed, adjusting the angle of the image and normalizing; the feature extraction and selection module is used for extracting and selecting the same feature information of the preprocessed image file; the classification decision module is used for acquiring the information of the image to be processed and classifying the image according to the scene type; the information identification module is used for matching the classification and result of the image acquisition with the original template in the image identification information base and outputting a matching identification result in the image to be processed.
2. The intelligent recognition tool based on 0CR as set forth in claim 1, wherein the preprocessing module performs preprocessing of the image file to be processed by:
normalizing the size of the acquired image, and acquiring illumination intensity parameters related to the acquired image;
judging whether the collected image has abnormal illumination or not based on the illumination intensity parameter, and simultaneously analyzing the abnormal type of the abnormal illumination image;
carrying out illumination correction on the illumination abnormal image;
performing dotting and positioning on the image file with the increased definition of the image to be processed to obtain image features with background colors;
image segmentation is performed on the image based on the set of image background colors.
3. The intelligent recognition tool based on 0CR as set forth in claim 2, wherein the feature extraction and selection module re-processes the image file before extracting and selecting the same feature information of the preprocessed image file, specifically as follows:
based on an image segmentation recognition result of the illumination correction image, invoking a corresponding shielding elimination algorithm to clear shielding information of image features of the object to be recognized;
and calling a corresponding background splashing algorithm to perform background removing and subtracting operation on the image features of the objects to be identified, and forming feature vectors by utilizing the image features of the objects to be identified.
4. A smart recognition tool based on 0CR as recited in claim 3, wherein the feature extraction and selection module extracts and selects the same feature information of the preprocessed image file by:
(1) Feature classification: physical features, structural features, mathematical features;
(2) Feature formation generating a set of basic features from the identified image, such generated features being called raw features;
(3) Extracting features, namely representing samples by a low-dimensional space through a mapping method, wherein the mapped secondary features are linear combinations of original features;
(4) Feature selection a number of most efficient features are selected from a set of features to achieve a reduction in feature space dimension.
5. The intelligent recognition tool based on 0CR as set forth in claim 4, wherein the classification decision module classifies the image according to scene type by:
preprocessing an illumination correction image, specifically: determining a difference operator, and carrying out gray level difference in the neighborhood to obtain a difference gray level map; determining a color difference valve value, and separating an image area to be identified in the differential gray scale image from background colors by using the color difference valve value;
and determining a pixel value interval to which a pixel value of each background color area pixel point in the image belongs, and obtaining a first pixel identification information sequence based on a corresponding relation between the pixel value interval to which the pixel value of each pixel point belongs and preset pixel value interval identification information.
6. The 0 CR-based intelligent recognition tool according to claim 5, wherein the specific obtaining process of the first pixel identification information sequence is:
determining a plurality of continuous pixel value intervals divided in advance, and determining a pixel value interval to which each pixel value of each pixel point in the first area belongs;
dividing the first region pixel points based on the pixel value intervals to obtain at least one first pixel point pixel value string, wherein the pixel value intervals of the pixel points in any two adjacent first pixel point pixel value strings are different;
and ordering the pixel identification information of each first pixel point pixel value string to obtain a corresponding first pixel identification information sequence.
7. The intelligent recognition tool based on 0CR as set forth in claim 6, wherein the information recognition module performs the steps of matching the classification and result of the image acquisition with the original template in the image recognition information base:
acquiring image features of an object to be identified, and carrying out feature extraction on the image features based on a convolution algorithm to obtain an intermediate feature matrix;
sorting and reorganizing the intermediate feature matrix to obtain a post feature matrix;
performing feature extraction on the post feature matrix based on a convolution algorithm to obtain an image feature set;
and matching the image feature set with the original template in the image identification information base.
8. The intelligent recognition tool based on 0CR of claim 1, wherein the output matching recognition result in the image to be processed is displayed through a mobile terminal and a computer terminal, and meanwhile, manual adjustment of the recognition result is supported to ensure the accuracy of the recognition result.
CN202311487549.3A 2023-11-09 2023-11-09 Intelligent recognition tool based on 0CR Pending CN117496532A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311487549.3A CN117496532A (en) 2023-11-09 2023-11-09 Intelligent recognition tool based on 0CR

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311487549.3A CN117496532A (en) 2023-11-09 2023-11-09 Intelligent recognition tool based on 0CR

Publications (1)

Publication Number Publication Date
CN117496532A true CN117496532A (en) 2024-02-02

Family

ID=89684432

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311487549.3A Pending CN117496532A (en) 2023-11-09 2023-11-09 Intelligent recognition tool based on 0CR

Country Status (1)

Country Link
CN (1) CN117496532A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117911796A (en) * 2024-03-18 2024-04-19 深圳市玛塔创想科技有限公司 Intelligent data processing system and method based on image recognition

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117911796A (en) * 2024-03-18 2024-04-19 深圳市玛塔创想科技有限公司 Intelligent data processing system and method based on image recognition
CN117911796B (en) * 2024-03-18 2024-06-21 深圳市玛塔创想科技有限公司 Intelligent data processing system and method based on image recognition

Similar Documents

Publication Publication Date Title
CN109154978B (en) System and method for detecting plant diseases
CN111553837B (en) Artistic text image generation method based on neural style migration
CN107909081B (en) Method for quickly acquiring and quickly calibrating image data set in deep learning
CN110119741A (en) A kind of card card image information recognition methods having powerful connections
Alkoffash et al. A survey of digital image processing techniques in character recognition
CN109711268B (en) Face image screening method and device
CN111680690A (en) Character recognition method and device
CN113435407B (en) Small target identification method and device for power transmission system
CN117496532A (en) Intelligent recognition tool based on 0CR
CN113537037A (en) Pavement disease identification method, system, electronic device and storage medium
CN112417955A (en) Patrol video stream processing method and device
CN113052215A (en) Sonar image automatic target identification method based on neural network visualization
CN117557784B (en) Target detection method, target detection device, electronic equipment and storage medium
CN112348018A (en) Digital display type instrument reading identification method based on inspection robot
CN116843581B (en) Image enhancement method, system, device and storage medium for multi-scene graph
CN113989814A (en) Image generation method and device, computer equipment and storage medium
CN110188693B (en) Improved complex environment vehicle feature extraction and parking discrimination method
CN113435358B (en) Sample generation method, device, equipment and program product for training model
Krishna et al. Digital image processing techniques in character recognition-a survey
CN115187788A (en) Crop seed automatic counting method based on machine vision
CN115063679A (en) Pavement quality assessment method based on deep learning
CN110276260B (en) Commodity detection method based on depth camera
Rani et al. Object Detection in Natural Scene Images Using Thresholding Techniques
CN113012167A (en) Combined segmentation method for cell nucleus and cytoplasm
CN111242047A (en) Image processing method and apparatus, electronic device, and computer-readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination