CN117523642A - Face recognition method based on optimal-spacing Bayesian classification model - Google Patents

Face recognition method based on optimal-spacing Bayesian classification model Download PDF

Info

Publication number
CN117523642A
CN117523642A CN202311633669.XA CN202311633669A CN117523642A CN 117523642 A CN117523642 A CN 117523642A CN 202311633669 A CN202311633669 A CN 202311633669A CN 117523642 A CN117523642 A CN 117523642A
Authority
CN
China
Prior art keywords
face
classification model
optimal
face recognition
bayesian classification
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311633669.XA
Other languages
Chinese (zh)
Inventor
黄玮
刘志东
徐志磊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Institute of Technology BIT
Original Assignee
Beijing Institute of Technology BIT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Institute of Technology BIT filed Critical Beijing Institute of Technology BIT
Priority to CN202311633669.XA priority Critical patent/CN117523642A/en
Publication of CN117523642A publication Critical patent/CN117523642A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/004Artificial life, i.e. computing arrangements simulating life
    • G06N3/006Artificial life, i.e. computing arrangements simulating life based on simulated virtual individual or collective life forms, e.g. social simulations or particle swarm optimisation [PSO]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/467Encoded features or binary features, e.g. local binary patterns [LBP]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/50Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/7715Feature extraction, e.g. by transforming the feature space, e.g. multi-dimensional scaling [MDS]; Mappings, e.g. subspace methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Human Computer Interaction (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a novel method for an optimal-spacing Bayesian classification model, which is applied to the field of face recognition. The face recognition method is different from the traditional naive Bayesian classification method, the proposed optimal interval Bayesian core concept is to utilize probability inference capability of naive Bayesian classification, and consider optimal interval measurement between samples at the same time so as to solve the problem of possible category overlapping, so that a classifier can better distinguish categories. The method comprises the following steps: firstly, capturing a target face in real time by using a camera, and preprocessing an image by adopting histogram equalization; then, extracting the face features by adopting a feature extraction method of a local binary pattern, and performing dimension reduction processing by using PCA; and finally, classifying the extracted features by using the proposed optimal interval Bayesian classification model, thereby outputting a final face recognition result. Experimental results show that the method provided by the invention has more excellent performance in the aspect of face recognition.

Description

Face recognition method based on optimal-spacing Bayesian classification model
Technical Field
The invention relates to the technical field of biological feature recognition, in particular to a face recognition method based on an optimal-spacing Bayesian classification model.
Background
In the current digital age, the biological recognition technology is a revolutionary technology, and makes a great contribution to the safety and convenience of human society. Biometric methods are distinguished by the unique dependence of individual biometric features, such as authentication and identification of faces, fingerprints, irises, sounds, etc. Face recognition is popular as a biometric solution with its excellent performance and wide range of applications. However, the conventional face recognition method still has some challenges in terms of processing image complexity, illumination variation, posture variation, expression variation, and the like. Naive bayes are widely used in face recognition, but there are some limitations to the traditional naive bayes classification model. One of the main limitations is its strong assumption of feature independence, which means that the model assumes that all features are conditional independent, and in practical applications, there may be a correlation between features. And also, category overlap problems may be encountered when processing complex data sets, resulting in reduced classifier accuracy. Although algorithms exist to solve the problem of the spacing between categories, there is still room for some potential drawbacks and improvements in practical applications.
Disclosure of Invention
The invention aims to overcome the technical defects and shortcomings, and provides a face recognition method based on an optimal-spacing Bayesian classification model.
The technical scheme adopted by the invention is as follows:
a face recognition method based on an optimal interval Bayesian classification model comprises the following steps:
step 1, capturing face images in real time through a camera or image acquisition equipment;
step 2, preprocessing the captured face image, and eliminating uneven illumination or other image quality problems by using a histogram equalization technology;
step 3, extracting the most representative and robust features in the face image by adopting a Local Binary Pattern (LBP) feature extraction method;
step 4, performing feature dimension reduction processing by using a Principal Component Analysis (PCA) technology to reduce feature dimensions, improve calculation efficiency and reserve main feature information;
step 5, training and identifying classification by using an optimal interval Bayesian classification model;
and step 6, outputting a face recognition result.
Step 1: capturing a face image, and detecting a face area in real time by using a Haar classifier to capture the face image in real time, wherein the method comprises the following specific steps of:
step 1.1, calculating Haar characteristics, and using pixels of a difference sum of adjacent rectangles at a specified position of a detection range;
step 1.2, accelerating to obtain Haar characteristics, and accelerating to calculate the Haar characteristics by using an integral graph:
0 sum =I(A)+I(B)+I(C)+I(D)
where O represents the detection window, A, B, C, D represents four vertices of the detection window, and I (x, y) represents the pixel value of the coordinate point;
step 1.3, normalizing the Haar characteristic values, and normalizing the characteristic values to ensure that the characteristic values have uniform scale or range.
The method comprises the following steps:
step 1.3.1, calculating the gray value and the square sum of the gray values of the images in the detection window:
sum=ΣI(x,y)
sq sum =ΣI 2 (x,y)
step 1.3.2, calculating the average value:
step 1.3.3, normalization factor:
step 1.3.4, normalized eigenvalues:
finally, comparing the normalized characteristic value with a threshold value;
step 1.4, the training of the AdaBoost algorithm is used for distinguishing the human face from the non-human face, and the training of the AdaBoost algorithm comprises the following steps:
step 1.4.1, initializing sample weights. Assigning an equal weight value to each sample;
step 1.4.2, iteratively training a plurality of weak classifiers. Training a weak classifier, and classifying the data according to the current sample weight learning. And then, the error rate of the classifier on the weighted data set is calculated, the weight of the sample is adjusted, the weight of the error classification sample is improved, and the weight of the correct classification sample is reduced. Finally, calculating the weight of the weak classifier according to the classification accuracy of the weak classifier under the current weight, and generating a plurality of weak classifiers by continuously iterating the process;
and 1.4.3, constructing an integrated strong classifier. And (3) weighting and combining all the weak classifiers according to the weights of the weak classifiers to construct an integrated strong classifier.
Step 2: preprocessing the captured face image by using a histogram equalization technology, wherein the steps are as follows:
step 2.1, converting the captured face image into a gray image, and calculating a gray histogram:
wherein N is γ The number of pixels with gamma is the gray scale, and N is the imageTotal number of elements, gray value of gamma i P for frequency of (2) γi ) Representing, and the gray level number of the image is represented by k;
step 2.2, calculating a cumulative distribution function based on the gray level histogram:
where n is the total number of pixels in the frame, k is the number of gray levels corresponding to the image, k=0, 1,2, …, L-1.L is the total number of gray levels, n i The number of pixels of the current gray level;
step 2.3, carrying out histogram equalization. For each pixel value, mapping the pixel value of the original image by using a cumulative distribution function to obtain an equalized pixel value.
Step 3: the method adopts a Local Binary Pattern (LBP) to extract the face characteristics, and comprises the following specific steps:
step 3.1, calculate the LBP value for each pixel. Comparing each pixel point in the image with surrounding pixels, comparing the gray values of the surrounding pixels with the gray values of the central pixels, and encoding the result into binary numbers, wherein finally, a binary number, namely LBP code, can be obtained for each pixel;
step 3.2, calculating LBP characteristic vector. And counting the LBP codes of all pixels to obtain an LBP characteristic vector.
Step 4: the feature dimension reduction processing is carried out by using a Principal Component Analysis (PCA) technology, and the specific steps are as follows:
step 4.1, calculating the mean value of the feature matrix, and carrying out centering treatment:
where n is the number of samples, x j For each feature;
step 4.2, calculating a covariance matrix. Solving a covariance matrix C for the standardized data matrix:
wherein n is the number of samples, and X' is the standardized data matrix;
step 4.3, calculating the eigenvalues and eigenvectors and selecting the principal component quantity. Performing eigenvalue decomposition on the covariance matrix to obtain eigenvalues and corresponding eigenvectors, and selecting the number k of principal components to be reserved according to the magnitude of the eigenvalues;
step 4.4, constructing a projection matrix and projecting data. The first k eigenvectors with the largest eigenvalues are selected to form a projection matrix W, and the projection matrix W is used for projecting the standardized data matrix X' into a new k-dimensional space:
X n =X′·W
wherein X is n Is a data matrix in the new k-dimensional space.
Step 5: the optimal pitch bayesian classification model consists of four steps. The method comprises the following specific steps:
step 5.1, calculation of prior probability and conditional probability of Gaussian naive Bayes:
wherein y is i Representing the category of the ith data in the training data set, y representing the category label, n representing the number of samples of the training data, s representing the total category number, x j Represents the j-th attribute, x ij Is the j-th attribute, μ of the i-th sample j Is the mean value of the j-th attribute,is the variance of the jth attribute;
step 5.2, defining the fitness evaluation function. And in the optimal interval Bayesian classification model, the proposed optimal interval measurement and the accuracy of the training set are used as fitness evaluation functions. The optimal spacing is defined as the difference between the joint probabilities of all samples of one class and all samples of another class:
for the multi-classification problem, a one-to-many strategy is adopted to construct a multi-classification model;
and 5.3, training and optimizing the Bayesian classification model with the optimal distance. And (3) maximizing the optimal distance and the accuracy of the training set by adopting a multi-target particle swarm algorithm, and outputting a Pareto solution set through iterative updating to obtain a plurality of groups of attribute weight vectors. Evaluating the obtained multiple groups of attribute weights by using cross verification, and finally obtaining an optimal global attribute weight vector W;
step 5.4, calculating the posterior probability of each instance with weight under each category by using the attribute weight vector and the class conditional probability, and predicting the category to which the data belongs by maximizing the posterior probability:
wherein Y represents class labels to which the data belongs, Y represents a set of all class labels in the training data set, n represents the number of all attributes in the data set, and w i Indicating the weight given to the ith attribute.
Step 6: and outputting a face recognition result.
And classifying the target face by adopting an optimal interval Bayesian classification model, and taking the face with the highest probability as a final recognition result.
Advantages and beneficial effects of the invention
The invention provides a face recognition method based on an optimal spacing Bayesian classification model. The boundaries between different categories can be effectively increased, and the accurate recognition capability of the classifier on different face features is improved. Meanwhile, the invention can complete the capture of the face image under various conditions of changing expression, posture and illumination, and adopts a histogram equalization technology to improve the image quality. In the aspect of feature extraction, a local binary pattern is adopted to extract more robust and distinguishable facial features, and dimension disasters are avoided through PCA. And finally, classifying by adopting an optimal interval Bayesian classification method, so that the performance of face recognition classification is obviously improved, and obvious improvement is brought to the aspects of safety, user experience and the like.
Drawings
FIG. 1 is a general flow chart of an implementation of the face recognition method of the present invention;
FIG. 2 is an overall flow diagram of an optimal pitch Bayesian classification model;
FIG. 3 is a flow chart of optimizing attribute weights for multiple target particle swarms;
Detailed Description
The invention is described in detail below with reference to the accompanying drawings and examples.
A face recognition method based on an optimal interval Bayesian classification model is implemented with the general flow shown in figure 1; according to the invention, face image data of 8 persons are collected, each person has 26 images, and training and testing are carried out by using an optimal interval Bayesian classification model. The specific implementation method comprises the following steps:
step 1, capturing face images in real time by using a camera, wherein the acquired images are stored in a designated folder and are used in the subsequent face feature extraction and test model training process;
step 2, carrying out gray processing on the obtained face image, and carrying out histogram equalization processing on the gray face image;
step 3, extracting face features of the grey face image by adopting a local binary pattern algorithm;
step 4, performing characteristic dimension reduction processing by using a principal component analysis technology;
step 5, training and identifying face data by using an optimal interval Bayesian classification model, wherein a specific flow chart is shown in fig. 2, and a multi-target particle swarm optimization attribute weight flow chart is shown in fig. 3;
and step 6, outputting a face recognition result. Taking the recognition result of the 8 # face image as an example, the comparison result is that the probability of being similar to the 1 # face is 0, the probability of being similar to the 2 # face is 0, the probability of being similar to the 3 # face is 0, the probability of being similar to the 4 # face is 0, the probability of being similar to the 5 # face is 0, the probability of being similar to the 6 # face is 0, the probability of being similar to the 7 # face is 0, and the probability of being similar to the 8 # face is a percentage, so the final recognition result is the 8 # face.
The optimal interval Bayesian classification model provided by the invention is repeatedly tested for a plurality of times by using the UCI standard data set, and is compared with the traditional algorithm. The 6 data sets Inonosphere, glass, spambase, kr-vs-kp, blood, rice were selected. The experimental results are shown in graph 1.
Table 1 comparison of experimental results (Classification accuracy)
KNN Traditional Bayes algorithm Optimum spacing Bayes
Ionosphere 86.79% 83.81% 90.48%
Glass 66.155 51.16% 76.92%
Spambase 79.65% 80.94% 91.23%
kr-vs-kp 63.98% 60.17% 73.93%
Blood 72.00% 74.11% 80.00%
Rice 88.45% 91.69% 92.40%
The optimal interval Bayesian classification model provided by the invention is repeatedly tested for a plurality of times by using the ORL face database, and is compared with other common face recognition methods. The ORL face database consists of face images of 40 persons, researchers acquire face images of each person under different environments such as facial expressions, illumination conditions and the like, and each person acquires 10 photos under different environments. We performed experiments on the ORL face database and compared with algorithms such as EDLPP, LDENP, 2DJLNDA, OR-NMF, pca+svm, etc. The experimental results are shown in graph 2.
Table 2 ORL face database experimental results
The above description is only a preferred example of the present invention and is not intended to limit the present invention, and various modifications and variations of the present invention will be apparent to those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention should be included in the scope of the claims of the present invention.

Claims (7)

1. The face recognition method based on the optimal interval Bayesian classification model is characterized by comprising the following steps of:
step 1, capturing face images in real time through a camera or image acquisition equipment;
step 2, preprocessing the captured face image, and eliminating uneven illumination or other image quality problems by using a histogram equalization technology;
step 3, extracting the most representative and robust features in the face image by adopting a Local Binary Pattern (LBP) feature extraction method;
step 4, performing feature dimension reduction processing by using a Principal Component Analysis (PCA) technology to reduce feature dimensions, improve calculation efficiency and reserve main feature information;
step 5, training and identifying classification by using an optimal interval Bayesian classification model;
and step 6, outputting a face recognition result.
2. The face recognition method based on the optimal pitch bayesian classification model according to claim 1, wherein: the step 1 capturing a face image in real time through a camera, and using a Haar classifier, wherein the specific steps are as follows:
step 1.1, calculating Haar characteristics, and using pixels of a difference sum of adjacent rectangles at a specified position of a detection range;
step 1.2, accelerating to obtain Haar characteristics, and accelerating to calculate the Haar characteristics by using an integral graph:
O sum =I(A)+I(B)+I(C)+I(D)
where O represents the detection window, A, B, C, D represents four vertices of the detection window, and I (x, y) represents the pixel value of the coordinate point;
step 1.3, normalizing the Haar characteristic values, and normalizing the characteristic values to ensure that the characteristic values have uniform scale or range. The method comprises the following steps:
step 1.3.1, calculating the gray value and the square sum of the gray values of the images in the detection window:
sum=∑I(x,y)
sq sum =∑I 2 (x,y)
step 1.3.2, calculating the average value:
step 1.3.3, normalization factor:
step 1.3.4, normalized eigenvalues:
finally, comparing the normalized characteristic value with a threshold value;
step 1.4, the training of the AdaBoost algorithm is used for distinguishing the human face from the non-human face, and the training of the AdaBoost algorithm comprises the following steps:
step 1.4.1, initializing sample weights. Assigning an equal weight value to each sample;
step 1.4.2, iteratively training a plurality of weak classifiers. Training a weak classifier, and classifying the data according to the current sample weight learning. And then, the error rate of the classifier on the weighted data set is calculated, the weight of the sample is adjusted, the weight of the error classification sample is improved, and the weight of the correct classification sample is reduced. Finally, calculating the weight of the weak classifier according to the classification accuracy of the weak classifier under the current weight, and generating a plurality of weak classifiers by continuously iterating the process;
and 1.4.3, constructing an integrated strong classifier. And (3) weighting and combining all the weak classifiers according to the weights of the weak classifiers to construct an integrated strong classifier.
3. The face recognition method based on the optimal pitch bayesian classification model according to claim 1, wherein: the step 2 of preprocessing the captured face image by using a histogram equalization technology, specifically comprises the following steps:
step 2.1, converting the captured face image into a gray image, and calculating a gray histogram:
wherein N is γ The number of pixels with the gray scale of gamma is N, the total number of pixels of the image, and the gray scale value of gamma i P for frequency of (2) γi ) Representing, and the gray level number of the image is represented by k;
step 2.2, calculating a cumulative distribution function based on the gray level histogram:
where n is the total number of pixels in the frame, k is the number of gray levels corresponding to the image, k=0, 1,2, …, L-1.L is the total number of gray levels, n i The number of pixels of the current gray level;
step 2.3, carrying out histogram equalization. For each pixel value, mapping the pixel value of the original image by using a cumulative distribution function to obtain an equalized pixel value.
4. The face recognition method based on the optimal pitch bayesian classification model according to claim 1, wherein: the step 3 of extracting the face features by adopting a Local Binary Pattern (LBP) specifically comprises the following steps:
step 3.1, calculate the LBP value for each pixel. Comparing each pixel point in the image with surrounding pixels, comparing the gray values of the surrounding pixels with the gray values of the central pixels, and encoding the result into binary numbers, wherein finally, a binary number, namely LBP code, can be obtained for each pixel;
step 3.2, calculating LBP characteristic vector. And counting the LBP codes of all pixels to obtain an LBP characteristic vector.
5. The face recognition method based on the optimal pitch bayesian classification model according to claim 1, wherein: the feature dimension reduction processing is performed by using Principal Component Analysis (PCA) technology, and the specific steps are as follows:
step 4.1, calculating the mean value of the feature matrix, and carrying out centering treatment:
where n is the number of samples, x j For each feature;
step 4.2, calculating a covariance matrix. Solving a covariance matrix C for the standardized data matrix:
wherein n is the number of samples, and X' is the standardized data matrix;
step 4.3, calculating the eigenvalues and eigenvectors and selecting the principal component quantity. Performing eigenvalue decomposition on the covariance matrix to obtain eigenvalues and corresponding eigenvectors, and selecting the number k of principal components to be reserved according to the magnitude of the eigenvalues;
step 4.4, constructing a projection matrix and projecting data. The first k eigenvectors with the largest eigenvalues are selected to form a projection matrix W, and the projection matrix W is used for projecting the standardized data matrix X' into a new k-dimensional space:
X n =X′·W
wherein X is n Is a data matrix in the new k-dimensional space.
6. The face recognition method based on the optimal pitch bayesian classification model according to claim 1, wherein: the optimal interval Bayesian classification model in the step 5 is trained and identified for classification, and the specific steps are as follows:
step 5.1, calculation of prior probability and conditional probability of Gaussian naive Bayes:
wherein y is i Representing the category of the ith data in the training data set, y representing the category label, n representing the number of samples of the training data, s representing the total category number, x j Represents the j-th attribute, x ij Is the j-th attribute, μ of the i-th sample j Is the mean value of the j-th attribute,is the variance of the jth attribute;
step 5.2, defining the fitness evaluation function. And in the optimal interval Bayesian classification model, the proposed optimal interval measurement and the accuracy of the training set are used as fitness evaluation functions. The optimal spacing is defined as the difference between the joint probabilities of all samples of one class and all samples of another class:
for the multi-classification problem, a one-to-many strategy is adopted to construct a multi-classification model;
and 5.3, training and optimizing the Bayesian classification model with the optimal distance. And (3) maximizing the optimal distance and the accuracy of the training set by adopting a multi-target particle swarm algorithm, and outputting a Pareto solution set through iterative updating to obtain a plurality of groups of attribute weight vectors. Evaluating the obtained multiple groups of attribute weights by using cross verification, and finally obtaining an optimal global attribute weight vector W;
step 5.4, calculating the posterior probability of each instance with weight under each category by using the attribute weight vector and the class conditional probability, and predicting the category to which the data belongs by maximizing the posterior probability:
wherein Y represents class labels to which the data belongs, Y represents a set of all class labels in the training data set, n represents the number of all attributes in the data set, and w i Indicating the weight given to the ith attribute.
7. The face recognition method based on the optimal pitch bayesian classification model according to claim 1, wherein: the method for outputting the face recognition result in the step 6 is as follows: and classifying the target face by adopting an optimal interval Bayesian classification model, and taking the face with the highest probability as a final recognition result.
CN202311633669.XA 2023-12-01 2023-12-01 Face recognition method based on optimal-spacing Bayesian classification model Pending CN117523642A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311633669.XA CN117523642A (en) 2023-12-01 2023-12-01 Face recognition method based on optimal-spacing Bayesian classification model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311633669.XA CN117523642A (en) 2023-12-01 2023-12-01 Face recognition method based on optimal-spacing Bayesian classification model

Publications (1)

Publication Number Publication Date
CN117523642A true CN117523642A (en) 2024-02-06

Family

ID=89756617

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311633669.XA Pending CN117523642A (en) 2023-12-01 2023-12-01 Face recognition method based on optimal-spacing Bayesian classification model

Country Status (1)

Country Link
CN (1) CN117523642A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117932474A (en) * 2024-03-22 2024-04-26 山东核电有限公司 Training method, device, equipment and storage medium of communication missing data determination model

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117932474A (en) * 2024-03-22 2024-04-26 山东核电有限公司 Training method, device, equipment and storage medium of communication missing data determination model

Similar Documents

Publication Publication Date Title
JP4543423B2 (en) Method and apparatus for automatic object recognition and collation
CN106295124B (en) The method of a variety of image detecting technique comprehensive analysis gene subgraph likelihood probability amounts
Stuhlsatz et al. Feature extraction with deep neural networks by a generalized discriminant analysis
CN110659665B (en) Model construction method of different-dimension characteristics and image recognition method and device
CN106529504B (en) A kind of bimodal video feeling recognition methods of compound space-time characteristic
CN113239839B (en) Expression recognition method based on DCA face feature fusion
CN108108760A (en) A kind of fast human face recognition
CN117523642A (en) Face recognition method based on optimal-spacing Bayesian classification model
Mu et al. Hyperspectral image classification based on active learning and spectral-spatial feature fusion using spatial coordinates
Zuobin et al. Feature regrouping for cca-based feature fusion and extraction through normalized cut
CN112132117A (en) Fusion identity authentication system assisting coercion detection
CN112560710A (en) Method for constructing finger vein recognition system and finger vein recognition system
CN109948662B (en) Face image depth clustering method based on K-means and MMD
JP6166981B2 (en) Facial expression analyzer and facial expression analysis program
CN109002770B (en) Face recognition method under low-resolution condition
CN112395901A (en) Improved face detection, positioning and recognition method in complex environment
CN106709442B (en) Face recognition method
WO2009047561A1 (en) Value determination
CN115310491A (en) Class-imbalance magnetic resonance whole brain data classification method based on deep learning
Türkmen et al. Global feature based female facial beauty decision system
CN112241680A (en) Multi-mode identity authentication method based on vein similar image knowledge migration network
CN107341485B (en) Face recognition method and device
CN111860529A (en) Image preprocessing method, system, device and medium
CN110751023B (en) Series pedestrian detection method for video stream
CN114693964B (en) MRI data feature extraction and classification identification method based on artificial neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination