CN106709494B - Scene character recognition method based on coupling space learning - Google Patents

Scene character recognition method based on coupling space learning Download PDF

Info

Publication number
CN106709494B
CN106709494B CN201710014236.4A CN201710014236A CN106709494B CN 106709494 B CN106709494 B CN 106709494B CN 201710014236 A CN201710014236 A CN 201710014236A CN 106709494 B CN106709494 B CN 106709494B
Authority
CN
China
Prior art keywords
dictionary
feature
scene
recognition
dist
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201710014236.4A
Other languages
Chinese (zh)
Other versions
CN106709494A (en
Inventor
张重
王红
刘爽
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhongfang Information Technology Tianjin Co ltd
Original Assignee
Tianjin Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin Normal University filed Critical Tianjin Normal University
Priority to CN201710014236.4A priority Critical patent/CN106709494B/en
Publication of CN106709494A publication Critical patent/CN106709494A/en
Application granted granted Critical
Publication of CN106709494B publication Critical patent/CN106709494B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/62Text, e.g. of license plates, overlay texts or captions on TV images
    • G06V20/63Scene text, e.g. street names
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
    • G06F18/23213Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Probability & Statistics with Applications (AREA)
  • Multimedia (AREA)
  • Character Discrimination (AREA)
  • Image Analysis (AREA)

Abstract

The embodiment of the invention discloses a scene character recognition method based on coupling space learning, which comprises the following steps: preprocessing an input scene character image to obtain a training scene character image; performing recognition feature extraction on the character image of the training scene to obtain a space dictionary; carrying out spatial coding on the recognition features of the corresponding images by using a spatial dictionary to obtain corresponding spatial coding vectors; performing maximum extraction on the space coding vector to obtain a characteristic vector; training by utilizing a linear support vector machine based on the feature vectors to obtain a scene character recognition classification model; and acquiring a feature vector of the text and image of the test scene, and inputting the text and image recognition classification model of the test scene to obtain a text and image recognition result of the test scene. The invention can effectively combine the spatial context information into the feature vector by creating the spatial dictionary and utilizing the spatial dictionary to carry out spatial coding, thereby achieving the purpose of effectively mining the spatial information and further improving the accuracy of scene character recognition.

Description

Scene character recognition method based on coupling space learning
Technical Field
The invention belongs to the technical field of pattern recognition, and particularly relates to a scene character recognition method based on coupling space learning.
Background
Scene character recognition plays an important role in the field of pattern recognition, and can be directly applied to the fields of image retrieval, intelligent transportation, man-machine interaction and the like. In practical applications, scene text recognition is a very challenging research direction because the scene text is affected by external factors such as uneven illumination, distortion, complex background, and the like.
Scene character recognition has been extensively studied in recent decades, and some early methods utilize optical character recognition technology for scene character recognition. However, the optical character recognition technology has great limitations, such as a scene text image binarization operation. In recent years, a large number of methods for scene character recognition have been proposed and made great progress. Among them, the most representative work is a scene character recognition method based on object recognition. The method based on the target recognition skips the binarization process of the scene character image and considers each scene character as a special target, and achieves certain success in the field of pattern recognition. Such as: newell et al use multi-scale HOG (Histogram of Gradients) for characterization. Zhang et al extracts and characterizes sparse code Histograms (HSCs). Shi et al consider both local feature information and global structure information. While these approaches have met with some success, they have largely ignored spatial context information. To solve this problem, Gao et al propose a stroke library in the feature representation stage to consider spatial context information, since different characters may contain the same feature information at different positions, which may cause reconstruction errors. The method proposed by Shi et al is an extension of the Gao et al method, which uses discriminative multi-scale stroke libraries to represent features. Tian et al propose to add spatial context information considering symbiotic relationships between HOG features. In addition, Gao et al also propose to consider spatial context information based on a location-embedded dictionary. Although the above method has been successful, the spatial context information, i.e., the dictionary learning stage or the encoding stage, is considered in a single aspect, and thus effective spatial context information cannot be sufficiently retained.
Disclosure of Invention
The invention aims to solve the technical problem that spatial context information has a large influence on a scene character recognition result, and therefore, the invention provides a scene character recognition method based on coupling space learning.
In order to achieve the purpose, the scene character recognition method based on the coupling space learning comprises the following steps:
step S1, respectively carrying out preprocessing operation on the N input scene character images to obtain N training scene character images;
step S2, respectively extracting the recognition features of the N training scene character images to obtain N space dictionaries;
step S3, carrying out space coding on the recognition characteristics of each training scene character image by using the space dictionary of the image to obtain a corresponding space coding vector;
step S4, performing maximum extraction on the space coding vector of each training scene character image to obtain a characteristic vector corresponding to the training scene character image;
step S5, training by using a linear support vector machine based on the feature vector of the training scene character image to obtain a scene character recognition classification model;
and S6, acquiring the feature vectors of the text and image of the test scene according to the steps S1-S4, and inputting the feature vectors into the scene character recognition classification model to obtain a scene character recognition result.
Optionally, the step S1 includes the following steps:
step S11, converting the input scene text image into a grayscale scene text image;
step S12, the size of the grayscale scene character image is normalized to H × W, and the normalized grayscale scene character image is used as the training scene character image, where H and W represent the height and width of the grayscale scene character image, respectively.
Optionally, the step S2 includes the following steps:
step S21, P of character image in each training sceneiExtracting identification features at positions (i-1, 2, …, m), wherein m is the number of the identification feature extraction positions of each training scene character image;
step S22, for N training scene character images, for the slave PiClustering all the recognition features extracted from the positions to obtain a sub-dictionary Ci(i ═ 1,2, …, m), and combines the sub-dictionaries CiIs denoted as Pi
In step S23, m sub-dictionaries carrying position information are concatenated to obtain a spatial dictionary.
Optionally, the identifying feature is a HOG feature.
Optionally, in the step S22, the identifying features are clustered by using a k-means clustering algorithm.
Optionally, the spatial dictionary is represented as:
D={C,P}={(C1,P1),(C2,P2),...,(Cm,Pm)},
wherein D represents a space dictionary, and C ═ C1,C2,…,Cm) For a set of m sub-dictionaries, P ═ P1,P2,…,Pm) A set of position information representing the set of sub-dictionaries C.
Optionally, in step S3, the recognition features of the text image of the training scene are spatially encoded by an objective function shown in the following formula:
wherein | · | purple sweet2Is represented by2A norm,. indicates a dot product operation of corresponding elements in two matrices, fjRepresenting a recognition feature, ajDenotes fjCorresponding spatial coding vector, a ═ a1,a2,…,aj,…]Represents the set of all spatial encoding vectors, | | fj-Caj||2Representing an error generated by reconstructing the recognition feature by using a spatial dictionary; | d |jF⊙aj||2The local regular term represents the distance constraint relation between the recognition features in the feature space and the code words in the sub-dictionary; | d |jE⊙aj||2The space regular term represents the position relation between the constraint characteristics in the Euclidean space and the code words in the sub-dictionary; alpha and beta are the regularization parameters,representing a spatially encoded vector ajThe sum of all elements in (a) is equal to 1; djFRepresenting the distance between the recognition feature in the feature space and the code word in the sub-dictionary, djERepresenting a recognition feature f in Euclidean spacejThe distance between the corresponding position and the position corresponding to the codeword in the sub-dictionary.
Optionally, the distance d between the recognition feature in the feature space and the codeword in the sub-dictionaryjFExpressed as:
wherein σFIs a device for adjusting djFParameter of weight descent speed, dist (f)jAnd C) is defined as:
dist(fj,C)=[dist(fj,C1),dist(fj,C2),...,dist(fj,Cm)]T
wherein, dist (f)j,Ci) (i-1, 2, …, m) represents the feature fjAnd sub-dictionary CiThe euclidean distance between all codewords in (a).
Optionally, identifying the feature f in Euclidean spacejDistance d between the corresponding position and the position corresponding to the code word in the sub-dictionaryjEExpressed as:
wherein σEIs a device for adjusting djEParameter of the weight descent speed, dist (l)jAnd, P) is defined as:
dist(lj,P)=[dist(lj,P1),…,dist(lj,P1),dist(lj,P2),…,dist(lj,P2),…,dist(lj,Pm),…,dist(lj,Pm)]Twherein, dist (l)j,Pi) (i-1, 2, …, m) represents the identification feature fjPosition l ofjAnd sub-dictionary CiPosition PiThe euclidean distance between them.
Optionally, in step S4, the spatial coding vector of each text image of the training scene is maximally extracted by using the following formula:
a=max{a1,a2,...,aj,...,am},
wherein, a represents the characteristic vector of the character image of the training scene, aj(j ═ 1, 2.. times, m) denotes a spatial coding vector.
The invention has the beneficial effects that: the invention can effectively combine the spatial context information into the feature vector by creating the spatial dictionary and performing spatial coding by using the created spatial dictionary, thereby achieving the purpose of effectively mining the spatial information and improving the accuracy of scene character recognition.
It should be noted that the invention obtains the subsidies of national science fund projects No.61401309 and No.61501327, Tianjin City applied foundation and leading edge technology research plan youth fund project No.15JCQNJC01700, Tianjin Master university doctor fund project No.5RL134 and No. 52XB1405.
Drawings
Fig. 1 is a flowchart of a scene character recognition method based on coupled space learning according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be described in further detail with reference to the accompanying drawings in conjunction with the following detailed description. It should be understood that the description is intended to be exemplary only, and is not intended to limit the scope of the present invention. Moreover, in the following description, descriptions of well-known structures and techniques are omitted so as to not unnecessarily obscure the concepts of the present invention.
Fig. 1 is a flowchart of a scene character recognition method based on coupled space learning according to an embodiment of the present invention, and some implementation flows of the present invention are described below with reference to fig. 1 as an example. The invention relates to a scene character recognition method based on coupling space learning, which comprises the following specific steps:
step S1, respectively carrying out preprocessing operation on the N input scene character images to obtain N training scene character images;
wherein the pre-processing operation comprises the steps of:
step S11, converting the input scene text image into a grayscale scene text image;
step S12, the size of the grayscale scene character image is normalized to H × W, and the normalized grayscale scene character image is used as the training scene character image, where H and W represent the height and width of the grayscale scene character image, respectively.
Step S2, respectively extracting the recognition features of the N training scene character images to obtain N space dictionaries;
further, the step S2 includes the following steps:
step S21, P of character image in each training sceneiExtracting an identification feature at the position (i-1, 2, …, m), wherein m is the number of feature extraction positions of each training scene character image, so that each training scene character image can obtain m identification features;
the identification feature may be an HOG feature or another identification feature.
Step S22, for N training scene character images, for the slave PiClustering all the recognition features extracted from the positions to obtain a sub-dictionary Ci(i ═ 1,2, …, m), and combines the sub-dictionaries CiIs denoted as PiThus, m sub-dictionaries are obtained for m feature extraction positions;
wherein, clustering operation can be carried out by using a clustering algorithm such as k-means and the like.
In step S23, m sub-dictionaries carrying position information are concatenated to obtain a spatial dictionary.
Wherein the spatial dictionary may be expressed as:
D={C,P}={(C1,P1),(C2,P2),...,(Cm,Pm)},
wherein D represents a space dictionary, and C ═ C1,C2,…,Cm) For a set of m sub-dictionaries, the corresponding P ═ P1,P2,…,Pm) A set of position information representing the set of sub-dictionaries C.
Step S3, using the space dictionary of each training scene character image to perform space coding on m recognition characteristics of the image to obtain corresponding m space coding vectors;
in step S3, the m recognition features of each training scene character image are spatially encoded by the spatial dictionary through the following objective function:
wherein | · | purple sweet2Is represented by2A norm,. indicates a dot product operation of corresponding elements in two matrices, fjRepresenting a recognition feature, ajDenotes fjCorresponding spatial encoding vector, corresponding a ═ a1,a2,…,aj,…]Represents the set of all spatial encoding vectors, | | fj-Caj||2Representing an error generated by reconstructing the recognition feature by using a spatial dictionary; | d |jF⊙aj||2The local regular term represents the distance constraint relation between the recognition features in the feature space and the code words in the sub-dictionary; | d |jE⊙aj||2The space regular term represents the position relation between the constraint characteristics in the Euclidean space and the code words in the sub-dictionary; alpha and beta are the regularization parameters,representing a spatially encoded vector ajThe sum of all elements in (a) is equal to 1; djFAnd representing the distance between the recognition features in the feature space and the code words in the sub-dictionary, wherein the specific expression form is as follows:
wherein σFIs a device for adjusting djFParameter of weight descent speed, dist (f)jAnd C) is defined as follows:
dist(fj,C)=[dist(fj,C1),dist(fj,C2),...,dist(fj,Cm)]T
wherein, dist (f)j,Ci) (i-1, 2, …, m) represents the feature fjAnd sub-dictionary CiThe euclidean distance between all codewords in (a).
djERepresenting a recognition feature f in Euclidean spacejCorresponding position ljAnd P, the specific expression form is as follows:
wherein σEIs a device for adjusting djEThe weight drop speed parameter. dist (l)jP) is defined as follows: dist (l)j,P)=[dist(lj,P1),…,dist(lj,P1),dist(lj,P2),…,dist(lj,P2),…,dist(lj,Pm),…,dist(lj,Pm)]TWherein, dist (l)j,Pi) (i-1, 2, …, m) represents the identification feature fjPosition l ofjAnd sub-dictionary CiPosition PiThe euclidean distance between them.
The target function selects a group of code words to reconstruct the identification characteristics by using a local regular term in a characteristic space, and simultaneously restricts the position relation between the identification characteristics and the code words in the sub-dictionary by using a space regular term in an Euclidean space.
An analytical solution can be obtained by taking the derivative of the objective function as follows:
wherein A isj=(CT-1fj T)(CT-1fj T)TRepresenting a covariance matrix using a formulaCan solveAnd carrying out normalization operation.
Through the analysis solution, the situation that a complex optimization process directly solves the space coding vector corresponding to the identification feature can be avoided.
Step S4, performing maximum extraction on the space coding vector of each training scene character image to obtain a characteristic vector corresponding to the training scene character image;
in step S4, the spatial coding vector of each text image of the training scene is maximally extracted by using the following formula:
a=max{a1,a2,...,aj,...,am},
wherein, ajAnd (j ═ 1, 2.. times, m) represents a spatial coding vector, and a represents a feature vector of a character image of a training scene.
And taking the maximum value of each dimension of the m spatial coding vectors of the training scene character image through the formula to obtain the feature vector a of the training scene character image.
Step S5, training by using a linear support vector machine based on the feature vector of the training scene character image to obtain a scene character recognition classification model;
and S6, acquiring the feature vectors of the text and image of the test scene according to the steps S1-S4, and inputting the feature vectors into the scene character recognition classification model to obtain a scene character recognition result.
The validity of the method of the present invention can be seen by using a scene text image database published on the internet as a test object, for example, on an ICDAR2003 database, when H × W is 64 × 32 and position m is 128, the accuracy of scene text recognition is 83.2%.
It is to be understood that the above-described embodiments of the present invention are merely illustrative of or explaining the principles of the invention and are not to be construed as limiting the invention. Therefore, any modification, equivalent replacement, improvement and the like made without departing from the spirit and scope of the present invention should be included in the protection scope of the present invention. Further, it is intended that the appended claims cover all such variations and modifications as fall within the scope and boundaries of the appended claims or the equivalents of such scope and boundaries.

Claims (9)

1. A scene character recognition method based on coupled space learning is characterized by comprising the following steps:
step S1, respectively carrying out preprocessing operation on the N input scene character images to obtain N training scene character images;
step S2, respectively extracting the recognition features of the N training scene character images to obtain N space dictionaries;
step S3, carrying out space coding on the recognition characteristics of each training scene character image by using the space dictionary of the image to obtain a corresponding space coding vector;
step S4, performing maximum extraction on the space coding vector of each training scene character image to obtain a characteristic vector corresponding to the training scene character image;
step S5, training by using a linear support vector machine based on the feature vector of the training scene character image to obtain a scene character recognition classification model;
step S6, acquiring the feature vector of the text and image of the test scene according to the steps S1-S4, and inputting the feature vector into the scene character recognition classification model to obtain a scene character recognition result;
in step S3, the recognition features of the text images of the training scene are spatially encoded by an objective function shown in the following formula:
wherein | · | purple sweet2Is represented by2A norm,. indicates a dot product operation of corresponding elements in two matrices, fjRepresenting a recognition feature, ajDenotes fjCorresponding spatial coding vector, a ═ a1,a2,…,aj,…]Represents the set of all spatial encoding vectors, | | fj-Caj||2Representing an error generated by reconstructing the recognition feature by using a spatial dictionary; | d |jF⊙aj||2The local regular term represents the distance constraint relation between the recognition features in the feature space and the code words in the sub-dictionary; | d |jE⊙aj||2The space regular term represents the position relation between the constraint characteristics in the Euclidean space and the code words in the sub-dictionary; alpha and beta are regularization parameters, 1Taj=1,Representing a spatially encoded vector ajThe sum of all elements in (a) is equal to 1; djFRepresenting the distance between the recognition feature in the feature space and the code word in the sub-dictionary, djERepresenting a recognition feature f in Euclidean spacejThe distance between the corresponding position and the position corresponding to the codeword in the sub-dictionary.
2. The method according to claim 1, wherein the step S1 comprises the steps of:
step S11, converting the input scene text image into a grayscale scene text image;
step S12, the size of the grayscale scene character image is normalized to H × W, and the normalized grayscale scene character image is used as the training scene character image, where H and W represent the height and width of the grayscale scene character image, respectively.
3. The method according to claim 1, wherein the step S2 comprises the steps of:
step S21, P of character image in each training sceneiExtracting identification features at positions (i-1, 2, …, m), wherein m is the number of the identification feature extraction positions of each training scene character image;
step S22, for N training scene character images, for the slave PiClustering all the recognition features extracted from the positions to obtain a sub-dictionary Ci(i ═ 1,2, …, m), and combines the sub-dictionaries CiIs denoted as Pi
In step S23, m sub-dictionaries carrying position information are concatenated to obtain a spatial dictionary.
4. The method of claim 3, wherein the identified feature is a HOG feature.
5. The method according to claim 3, wherein in step S22, the identification features are clustered using a k-means clustering algorithm.
6. The method of claim 3, wherein the spatial dictionary is represented as:
D={C,P}={(C1,P1),(C2,P2),...,(Cm,Pm)},
wherein D represents a space dictionary, and C ═ C1,C2,…,Cm) For a set of m sub-dictionaries, P ═ P1,P2,…,Pm) A set of position information representing the set of sub-dictionaries C.
7. Method according to claim 1, characterized in that the distance d between the recognition features in the feature space and the code words in the sub-dictionaryjFExpressed as:
wherein σFIs a device for adjusting djFParameter of weight descent speed, dist (f)jAnd C) is defined as:
dist(fj,C)=[dist(fj,C1),dist(fj,C2),...,dist(fj,Cm)]T
wherein, dist (f)j,Ci) (i-1, 2, …, m) represents the feature fjAnd sub-dictionary CiThe euclidean distance between all codewords in (a).
8. The method of claim 1, wherein the feature f is identified in Euclidean spacejDistance d between the corresponding position and the position corresponding to the code word in the sub-dictionaryjEExpressed as:
wherein σEIs a device for adjusting djEParameter of the weight descent speed, dist (l)jAnd, P) is defined as:
dist(lj,P)=[dist(lj,P1),…,dist(lj,P1),dist(lj,P2),…,dist(lj,P2),…,dist(lj,Pm),…,dist(lj,Pm)]Twherein, dist (l)j,Pi) (i-1, 2, …, m) represents the identification feature fjPosition l ofjAnd sub-dictionary CiPosition PiThe euclidean distance between them.
9. The method according to claim 1, wherein in step S4, the spatial coding vector of each text image of the training scene is maximally extracted by using the following formula:
a=max{a1,a2,...,aj,...,am},
wherein, a represents the characteristic vector of the character image of the training scene, aj(j ═ 1, 2.. times, m) denotes a spatial coding vector.
CN201710014236.4A 2017-01-10 2017-01-10 Scene character recognition method based on coupling space learning Active CN106709494B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710014236.4A CN106709494B (en) 2017-01-10 2017-01-10 Scene character recognition method based on coupling space learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710014236.4A CN106709494B (en) 2017-01-10 2017-01-10 Scene character recognition method based on coupling space learning

Publications (2)

Publication Number Publication Date
CN106709494A CN106709494A (en) 2017-05-24
CN106709494B true CN106709494B (en) 2019-12-24

Family

ID=58908090

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710014236.4A Active CN106709494B (en) 2017-01-10 2017-01-10 Scene character recognition method based on coupling space learning

Country Status (1)

Country Link
CN (1) CN106709494B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107679505B (en) * 2017-10-13 2020-04-21 林辉 Method for realizing rejection of handwritten character
CN108764233B (en) * 2018-05-08 2021-10-15 天津师范大学 Scene character recognition method based on continuous convolution activation

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104537362A (en) * 2015-01-16 2015-04-22 中国科学院自动化研究所 Domain-based self-adaptive English scene character recognition method
CN105760821A (en) * 2016-01-31 2016-07-13 中国石油大学(华东) Classification and aggregation sparse representation face identification method based on nuclear space

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104537362A (en) * 2015-01-16 2015-04-22 中国科学院自动化研究所 Domain-based self-adaptive English scene character recognition method
CN105760821A (en) * 2016-01-31 2016-07-13 中国石油大学(华东) Classification and aggregation sparse representation face identification method based on nuclear space

Also Published As

Publication number Publication date
CN106709494A (en) 2017-05-24

Similar Documents

Publication Publication Date Title
Ahmad et al. Object detection through modified YOLO neural network
CN110532900B (en) Facial expression recognition method based on U-Net and LS-CNN
Christlein et al. Unsupervised feature learning for writer identification and writer retrieval
Kadhim et al. A Real-Time American Sign Language Recognition System using Convolutional Neural Network for Real Datasets.
CN105913053B (en) A kind of facial expression recognizing method for singly drilling multiple features based on sparse fusion
Ali et al. Urdu natural scene character recognition using convolutional neural networks
Patel et al. Handwritten character recognition in english: a survey
CN107622267B (en) Scene character recognition method based on embedded bilateral convolution activation
Alaasam et al. Experiment study on utilizing convolutional neural networks to recognize historical Arabic handwritten text
Wang et al. Concept mask: Large-scale segmentation from semantic concepts
CN106709494B (en) Scene character recognition method based on coupling space learning
Sarhan et al. HLR-net: a hybrid lip-reading model based on deep convolutional neural networks
Garg et al. Optical character recognition using artificial intelligence
Ulaganathan et al. Isolated handwritten Tamil character recognition using convolutional neural networks
Lin et al. LLC encoded BoW features and softmax regression for microscopic image classification
Srinivas et al. Learning deep and sparse feature representation for fine-grained object recognition
CN108052867B (en) Single-sample face recognition method based on bag-of-words model
CN111950592B (en) Multi-modal emotion feature fusion method based on supervised least square multi-class kernel canonical correlation analysis
Li et al. Facial expression recognition using facial-component-based bag of words and PHOG descriptors
CN109190645B (en) High-order high-dimensional image data representation and classification method
Alaasam et al. Synthesizing versus augmentation for arabic word recognition with convolutional neural networks
CN109784356B (en) Matrix variable limited Boltzmann machine image classification method based on Fisher discriminant analysis
Dinajpur Handwritten Bengali character recognition through geometry based feature extraction
Zanona et al. Arabic Hand Written Character Recognition Based on Contour Matching and Neural Network.
Venkatadri et al. Image Processing based ML Framework for Leaf Classification and Disease Detection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20221130

Address after: 300392 Room 603, Building 1, No. 1, Huixue Road, Xuefu Industrial Zone, Xiqing District, Tianjin

Patentee after: Zhongfang Information Technology (Tianjin) Co.,Ltd.

Address before: 300387 Tianjin city Xiqing District West Binshui Road No. 393

Patentee before: TIANJIN NORMAL University

TR01 Transfer of patent right