CN113936312A - Face recognition base screening method based on deep learning graph convolution network - Google Patents

Face recognition base screening method based on deep learning graph convolution network Download PDF

Info

Publication number
CN113936312A
CN113936312A CN202111185859.0A CN202111185859A CN113936312A CN 113936312 A CN113936312 A CN 113936312A CN 202111185859 A CN202111185859 A CN 202111185859A CN 113936312 A CN113936312 A CN 113936312A
Authority
CN
China
Prior art keywords
image
network
node
model
face recognition
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111185859.0A
Other languages
Chinese (zh)
Other versions
CN113936312B (en
Inventor
王乾宇
周金明
张世坤
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Inspector Intelligent Technology Co Ltd
Original Assignee
Nanjing Inspector Intelligent Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Inspector Intelligent Technology Co Ltd filed Critical Nanjing Inspector Intelligent Technology Co Ltd
Priority to CN202111185859.0A priority Critical patent/CN113936312B/en
Publication of CN113936312A publication Critical patent/CN113936312A/en
Application granted granted Critical
Publication of CN113936312B publication Critical patent/CN113936312B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/277Analysis of motion involving stochastic approaches, e.g. using Kalman filters
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20076Probabilistic image processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a face recognition base screening method based on a deep learning graph convolution network, which comprises the following steps: the method comprises the steps of firstly, acquiring a video in a use scene; secondly, selecting a face recognition network as a main network according to the requirements of a use scene, combining a graph convolution network as a branch, constructing a face image quality evaluation model, and then training the model; and thirdly, detecting a human face object from the captured video. Fourthly, acquiring an image to be selected by using a mask detection algorithm; fifthly, adding part of interference images and the images to be selected obtained in the previous step, inputting the interference images and the images to be selected into a trained face recognition base image screening model for screening, and outputting the images with the highest confidence scores in the images to be detected by the model; and sixthly, storing the high-quality face image output by the model as a base library. The method for screening the human face bottom library not only ensures the screening efficiency, but also ensures the definition of the screened human face image and the integrity of the characteristics.

Description

Face recognition base screening method based on deep learning graph convolution network
Technical Field
The invention relates to the fields of computer vision, deep learning face recognition and intelligent monitoring, in particular to a face recognition base screening method based on a deep learning graph convolution network.
Background
With the rapid development of the artificial intelligence direction in China, the face recognition technology is widely applied to various fields, not only brings great convenience to the life of people, but also improves the safety of public places to a great extent; the face recognition is based on the premise that face base information is input, and if the input base image quality is poor, the situations of fuzziness, facial occlusion and the like exist, the face recognition information is likely to be identified wrongly when compared; therefore, the method selects a proper method to screen out the high-definition high-quality face image as the base information to play a crucial role in face recognition, and is beneficial to improving the accuracy and stability of face recognition.
At present, most of face recognition base image construction work still depends on manual screening, and high-quality clear faces are screened out from a large number of face images to serve as recognition base images; however, there are significant drawbacks to manual screening: depending on the experience of workers, the subjectivity of the workers in the screening process is overlarge, and the screening result has no unified standard; the detection and screening takes long time and the working efficiency is low.
Disclosure of Invention
In order to overcome the defects of the prior art, the invention provides a face recognition base screening method based on a deep learning graph convolution network, which screens a face base through a model, thereby ensuring the screening efficiency and ensuring the definition and the feature integrity of the screened face image; the highest-quality face image of the detected object is obtained through neural network learning and is used as a bottom library, a screening threshold value does not need to be set manually, and the influence caused by insufficient manual experience is reduced. The technical scheme is as follows:
a face recognition base screening method based on a deep learning graph convolution network comprises the following steps:
firstly, a video shot by a monitoring camera in a using scene is obtained.
And secondly, selecting a face recognition network as a main network according to the use scene, combining a graph convolution network as a branch, constructing a face image quality evaluation model, and then training the model.
And thirdly, detecting a face object from the captured video, and adopting a lightweight face detection algorithm MTCNN for face detection. And selecting a detected face object as a target, marking the target to prevent repeated selection of the next detection, tracking the selected target in each frame later by using a KCF target tracking algorithm, and intercepting and storing a frame image of the target face.
And fourthly, detecting the image stored in the third step by using a paddlehub mask detection algorithm, and storing the image without wearing the mask according to a detection result to be used as an image to be selected.
And fifthly, adding part of interference images and the images to be selected obtained in the previous step, inputting the interference images and the images to be selected into a trained human face recognition base image screening model for screening, and outputting the image with the highest confidence score in the images to be detected by the model.
And sixthly, storing the high-quality face image output by the model as a base library.
Preferably, the face recognition network in the second step is trained using the Resnet network.
Further, when the face recognition network is trained by using the Resnet network, the model is trained in two stages:
the first stage is as follows: and training a backbone network Resnet by taking a data set with a label as input for extracting the characteristics of the image to be detected.
And a second stage: training the GCN-V network:
(1) constructing a data set:
firstly, a data set with classification labels is given, and the features of each image in the given data set are extracted by utilizing a trained backbone network Resnet to form a feature set
Figure BDA0003299176750000021
Wherein f isi∈RDD represents a characteristic dimension, i ∈ {1,2,3 … … N }, N being the number of images in a given dataset; defining each image feature fiFor node i, the similarity between node i and node j is denoted as ai,j,ai,jIs fiAnd fjCosine similarity between them, j ∈ {1,2,3 … … N }.
Acquiring k adjacent nodes of all image nodes according to the similarity between the image nodes, and constructing a similar graph G (V, E) by using the k adjacent nodes, wherein the method specifically comprises the following steps: each image feature is a node belonging to V, the similarity value of each node is sorted from large to small, the first K nodes are taken as neighbor nodes of the node, namely, a neighborhood of the node is formed, the vertex is connected with the neighbor nodes to form K edges belonging to E, a similarity graph can be represented as a vertex feature matrix F 'and an adjacency matrix A, the size of the vertex feature matrix F' is N multiplied by D, the size of the adjacency matrix A is N multiplied by N, and in the adjacency matrix A, if vi and vj are not connected, the similarity a between an image node i and a node j isi,jIs updated to 0.
Solving a ground-truth confidence coefficient label; since data sets typically vary greatly within classes, there may be different confidence values even if each image belongs to the same class. The image with high confidence coefficient has more obvious characteristics in the class, the probability of belonging to the class is high, the image with low confidence coefficient is marginalized, and the characteristics in the class are weaker. Based on this feature, the confidence c of each node is defined according to its neighborhoodi: and for all adjacent nodes of the current node, if the adjacent nodes are consistent with the adjacent node types, accumulating the similarity between the adjacent nodes, if the adjacent nodes are not of the same type, subtracting the similarity between the current node and the adjacent node, and finally dividing the similarity by the number of the adjacent nodes of the current node to obtain the node confidence.
(2) Training of the model:
the inputs to the model are the vertex feature matrix F' and the adjacency matrix A in the dataset constructed in the previous step.
Secondly, performing polymerization operation on the feature matrix F 'and the adjacent matrix A, obtaining new features through L layers of convolution layers, performing regression on the new features in the last layer of linear layer of the network to obtain a predicted confidence value C',
C'=FLW+b
wherein W is a trainable regression variable, b is a trainable deviation, and L is the number of layers of graph convolution and can be adjusted according to the requirement; node viMay be selected from
Figure BDA0003299176750000031
Extracting corresponding elements from the Chinese medicinal materials with
Figure BDA0003299176750000032
And (4) showing.
(iii) training the model by minimizing the Mean Square Error (MSE) between the true confidence and the confidence score predicted by the model.
Preferably, the backbone network in the second step is not fixed, and can be replaced by any required face recognition network at will according to the use requirement.
Preferably, the fifth step specifically includes:
(1) firstly, an image characteristic set is obtained through a backbone network
Figure BDA0003299176750000033
(2) Processing F output by the backbone network: and acquiring k adjacent nodes of all image nodes according to the similarity between the image nodes, constructing a similar graph G (V, E), and expressing the similar graph G as a vertex characteristic matrix F' and an adjacent matrix A.
(3) And taking the vertex characteristic matrix F and the adjacency matrix A as the input of a GCN-V network, and calculating the confidence score of each image to be detected through GCN-V: c ═ FLW+b。
(4) And screening the image with the highest confidence score for output.
Compared with the prior art, one of the technical schemes has the following beneficial effects: constructing a face recognition base image screening model based on a deep learning face recognition network and a graph convolution network, and screening the face base through the model, so that the screening efficiency is ensured, and the definition and the feature integrity of the screened face image are also ensured; meanwhile, only one high-quality image of one object is kept as a bottom library. Meanwhile, the highest-quality face image of the detected object is obtained through neural network learning and is used as a bottom library, a screening threshold value does not need to be set manually, and the influence caused by insufficient manual experience is reduced; meanwhile, the real-time performance is high, the requirement on hardware is lower, and the hardware cost is reduced; high-quality human faces can be effectively detected and screened out to be used as bottom library images.
Detailed Description
In order to clarify the technical solution and the working principle of the present invention, the embodiments of the present disclosure will be described in further detail below. All the above optional technical solutions may be combined arbitrarily to form the optional embodiments of the present disclosure, and are not described herein again.
The terms "first step," "second step," "third step," or the like, in the description and in the claims of this application, or the like, are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It should be understood that the data so used may be interchanged under appropriate circumstances such that the embodiments of the application described herein may be practiced in sequences other than those described herein.
The embodiment of the disclosure provides a face recognition base screening method based on a deep learning graph convolution network, which comprises the following steps: the method comprises the following steps:
the method comprises the steps of firstly, acquiring a video shot by a monitoring camera in a using scene; the method can be suitable for the construction of the face recognition base in any scene, such as airports, railway stations, banks and the like.
And secondly, selecting a face recognition network as a main network according to the use scene, combining a graph convolution network as a branch, constructing a face image quality evaluation model, and then training the model.
Preferably, the backbone network is not fixed in the second step, and can be replaced by any required face recognition network at will according to the use requirement; the flexibility is stronger, and the performance can not receive the influence that uses the scene change. All face recognition networks can be used, and different networks have different influences on the accuracy of model screening and the detection speed;
preferably, the face recognition network is trained using the Resnet network.
Further, when the face recognition network is trained by using the Resnet network, the model is trained in two stages.
The first stage is as follows: and training a backbone network Resnet by taking a data set with a label as input for extracting the characteristics of the image to be detected.
And a second stage: training the GCN-V network:
(1) constructing a data set:
firstly, a data set with classification labels is given, and the features of each image in the given data set are extracted by utilizing a trained backbone network Resnet to form a feature set
Figure BDA0003299176750000041
Wherein f isi∈RDD represents a characteristic dimension, i ∈ {1,2,3 … … N }, N being the number of images in a given dataset; defining each image feature fiFor node i, the similarity between node i and node j is denoted as ai,j,ai,jIs fiAnd fjCosine similarity between them, j ∈ {1,2,3 … … N }.
Acquiring k adjacent nodes of all image nodes according to the similarity between the image nodes, and constructing a similar graph G (V, E) by using the k adjacent nodes, wherein the method specifically comprises the following steps: each image feature is a node belonging to V, the similarity value of each node is sorted from large to small, the first K nodes are taken as neighbor nodes of the node, namely, a neighborhood of the node is formed, the top point is connected with the neighbor nodes to form K edges belonging to E, and the similarity graph can be represented as a top point feature matrix F' and the neighbor nodesThe size of the adjacency matrix A is NxD, the size of the adjacency matrix A is NxN, and in the adjacency matrix A, if vi and vj are not connected, the similarity a between the image node i and the node ji,jIs updated to 0.
Solving a ground-truth confidence coefficient label; since data sets typically vary greatly within classes, there may be different confidence values even if each image belongs to the same class. The image with high confidence coefficient has more obvious characteristics in the class, the probability of belonging to the class is high, the image with low confidence coefficient is marginalized, and the characteristics in the class are weaker. Based on this feature, the confidence c of each node is defined according to its neighborhoodi: and for all adjacent nodes of the current node, if the adjacent nodes are consistent with the adjacent node types, accumulating the similarity between the adjacent nodes, if the adjacent nodes are not of the same type, subtracting the similarity between the current node and the adjacent node, and finally dividing the similarity by the number of the adjacent nodes of the current node to obtain the node confidence.
(2) Training of the model:
inputting a vertex characteristic matrix F' and an adjacent matrix A in the data set constructed in the previous step into the model;
secondly, performing polymerization operation on the feature matrix F 'and the adjacent matrix A, obtaining new features through L layers of convolution layers, performing regression on the new features in the last layer of linear layer of the network to obtain a predicted confidence value C',
C'=FLW+b
where W is a trainable regression variable and b is a trainable bias, L, i.e., the number of layers of graph convolution, may be adjusted as desired. Node viMay be selected from
Figure BDA0003299176750000051
Extracting corresponding elements from the Chinese medicinal materials with
Figure BDA0003299176750000052
And (4) showing.
(iii) training the model by minimizing the Mean Square Error (MSE) between the true confidence and the confidence score predicted by the model.
And thirdly, detecting a face object from the captured video, and adopting a lightweight face detection algorithm MTCNN for face detection. And selecting a detected face object as a target, marking the target to prevent repeated selection of the next detection, tracking the selected target in each frame later by using a KCF target tracking algorithm, and intercepting and storing a frame image of the target face.
The face recognition base screening model designed by the invention has no requirement on input images and can accept input of any size, so that the images obtained by utilizing a tracking algorithm do not need to be subjected to size constraint, and the applicability of the model is also improved. If the unit of use has a requirement on the size of the bottom library image, after all the frame images of the target object are acquired at this step, the images can be further preprocessed: and (5) adjusting the size. And after the size is uniformly adjusted, the subsequent operation is carried out.
And fourthly, detecting the image stored in the third step by using a paddlehub mask detection algorithm, and storing the image without wearing the mask according to a detection result to be used as an image to be selected. The images are preprocessed through mask detection, so that facial features of the finally screened high-quality face recognition base image are not shielded, and the influence of incomplete face images on the next extraction of similarity images is reduced.
And fifthly, adding part of interference images and the images to be selected obtained in the previous step, inputting the interference images and the images to be selected into a trained human face recognition base image screening model for screening, and outputting the image with the highest confidence score in the images to be detected by the model.
Preferably, the fifth step specifically includes:
(1) firstly, an image characteristic set is obtained through a backbone network
Figure BDA0003299176750000061
(2) Processing F output by the backbone network: and acquiring k adjacent nodes of all image nodes according to the similarity between the image nodes, constructing a similar graph G (V, E), and expressing the similar graph G as a vertex characteristic matrix F' and an adjacent matrix A.
(3) The vertex feature matrix F and adjacencyThe matrix A is used as the input of a GCN-V network, and the confidence score of each image to be detected is calculated through GCN-V: c ═ FLW+b。
(4) And screening the image with the highest confidence score for output.
The confidence degree represents the quality score of the image, and the higher the quality score is, the higher the quality of the image is, and the more the requirement on the bottom library image is met. For the image with high confidence level, the included features in the class are more comprehensive and can represent the belonging class more completely. Images with low confidence usually contain less obvious and incomplete features in the class and cannot represent the class. Therefore, the image with the highest confidence score in the same class is the image with the highest quality, which can most comprehensively represent the features in the class, in all the images in the class. Based on the face image feature information, the face image with the highest confidence level score in the same class is screened as a base library.
And sixthly, storing the high-quality face image output by the model as a base library.
The invention has been described above by way of example, it is obvious that the specific implementation of the invention is not limited by the above-described manner, and that various insubstantial modifications are possible using the method concepts and technical solutions of the invention; or the technical scheme of the invention can be directly applied to other occasions without improvement and equivalent replacement, and the invention is within the protection scope of the invention.

Claims (5)

1. A face recognition base screening method based on a deep learning graph convolution network is characterized by comprising the following steps:
the method comprises the steps of firstly, acquiring a video shot by a monitoring camera in a using scene;
secondly, selecting a face recognition network as a main network according to the requirements of a use scene, combining a graph convolution network as a branch, constructing a face image quality evaluation model, and then training the model;
thirdly, detecting a face object from the captured video, and adopting a lightweight face detection algorithm MTCNN for face detection; selecting a detected face object as a target, marking the target to prevent repeated selection of next detection, tracking the selected target in each frame later by using a KCF target tracking algorithm, and intercepting and storing a frame image of the face of the target;
fourthly, detecting the image stored in the third step by using a paddlehub mask detection algorithm, and storing the image without wearing the mask according to a detection result to be used as an image to be selected;
fifthly, adding part of interference images and the images to be selected obtained in the previous step, inputting the interference images and the images to be selected into a trained face recognition base image screening model for screening, and outputting the images with the highest confidence scores in the images to be detected by the model;
and sixthly, storing the high-quality face image output by the model as a base library.
2. The method as claimed in claim 1, wherein the face recognition base screening method based on the deep learning graph convolution network is characterized in that the face recognition network in the second step is trained by using a Resnet network.
3. The method for screening the face recognition base based on the deep learning graph convolution network as claimed in claim 2, wherein when the face recognition network is trained by using the Resnet network, the model is trained in two stages:
the first stage is as follows: training a backbone network Resnet by taking a data set with a label as input, and extracting the characteristics of an image to be detected;
and a second stage: training the GCN-V network:
(1) constructing a data set:
firstly, a data set with classification labels is given, and the features of each image in the given data set are extracted by utilizing a trained backbone network Resnet to form a feature set
Figure FDA0003299176740000011
Wherein f isi∈RDD denotes the feature dimension, i ∈ {1,2,3 … … N }, N being the image in a given datasetThe number of (2); defining each image feature fiFor node i, the similarity between node i and node j is denoted as ai,,j,ai,jIs fiAnd fjCosine similarity between them, j ∈ {1,2,3 … … N };
acquiring k adjacent nodes of all image nodes according to the similarity between the image nodes, and constructing a similar graph G (V, E) according to the k adjacent nodes, wherein the method specifically comprises the following steps: each image feature is a node belonging to V, the similarity value of each node is sorted from large to small, the first K nodes are taken as neighbor nodes of the node, namely, a neighborhood of the node is formed, the vertex is connected with the neighbor nodes to form K edges belonging to E, a similarity graph can be represented as a vertex feature matrix F 'and an adjacency matrix A, the size of the vertex feature matrix F' is NxD, the size of the adjacency matrix A is NxN, in the adjacency matrix A, if vi and vj are not connected, the similarity a between an image node i and a node j is obtainedi,jUpdating to 0;
solving a ground-truth confidence coefficient label; since data sets typically vary greatly within classes, there may be different confidence values even if each image belongs to the same class; the image with high confidence coefficient has more obvious characteristics in the class, the probability of belonging to the class is high, the image with low confidence coefficient is marginalized, and the characteristics in the class are weaker; based on this feature, the confidence c of each node is defined according to its neighborhoodi: for all adjacent nodes of the current node, if the type of the adjacent nodes is consistent with that of the adjacent nodes, the similarity between the current node and the adjacent nodes is accumulated, if the type of the adjacent nodes is not the same, the similarity between the current node and the adjacent nodes is subtracted, and finally, the number of the adjacent nodes of the current node is divided to obtain the node confidence;
(2) training of the model:
inputting a vertex characteristic matrix F' and an adjacent matrix A in the data set constructed in the previous step into the model;
secondly, the feature matrix F 'and the adjacent matrix A are subjected to polymerization operation, new features are obtained through L layers of convolution layers, the new features are regressed in the last layer of linear layer of the network to obtain a predicted confidence value C',
C′=FLW+b
wherein W is a trainable regression variableB is a trainable bias, L is the number of layers of graph convolution, which can be adjusted as required; node viMay be selected from
Figure FDA0003299176740000022
Extracting corresponding elements from the Chinese medicinal materials with
Figure FDA0003299176740000023
Represents;
(iii) training the model by minimizing the Mean Square Error (MSE) between the true confidence and the confidence score predicted by the model.
4. The method for screening the face recognition base based on the deep learning graph convolution network as claimed in claim 1, wherein the backbone network is not fixed in the second step, and any required face recognition network can be replaced as required at will according to the requirements of use.
5. The method for screening the face recognition base based on the deep learning graph convolution network according to any one of claims 1 to 4, wherein the fifth step specifically comprises:
(1) firstly, an image characteristic set is obtained through a backbone network
Figure FDA0003299176740000021
(2) Processing F output by the backbone network: acquiring k adjacent nodes of all image nodes according to the similarity between the image nodes, constructing a similar graph G (V, E), and expressing the similar graph G as a vertex characteristic matrix F' and an adjacent matrix A;
(3) and taking the vertex characteristic matrix F and the adjacency matrix A as the input of a GCN-V network, and calculating the confidence score of each image to be detected through GCN-V: c ═ FLW+b;
(4) And screening the image with the highest confidence score for output.
CN202111185859.0A 2021-10-12 2021-10-12 Face recognition base screening method based on deep learning graph convolution network Active CN113936312B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111185859.0A CN113936312B (en) 2021-10-12 2021-10-12 Face recognition base screening method based on deep learning graph convolution network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111185859.0A CN113936312B (en) 2021-10-12 2021-10-12 Face recognition base screening method based on deep learning graph convolution network

Publications (2)

Publication Number Publication Date
CN113936312A true CN113936312A (en) 2022-01-14
CN113936312B CN113936312B (en) 2024-06-07

Family

ID=79278328

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111185859.0A Active CN113936312B (en) 2021-10-12 2021-10-12 Face recognition base screening method based on deep learning graph convolution network

Country Status (1)

Country Link
CN (1) CN113936312B (en)

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109711384A (en) * 2019-01-09 2019-05-03 江苏星云网格信息技术有限公司 A kind of face identification method based on depth convolutional neural networks
CN110070010A (en) * 2019-04-10 2019-07-30 武汉大学 A kind of face character correlating method identified again based on pedestrian
CN110472495A (en) * 2019-07-08 2019-11-19 南京邮电大学盐城大数据研究院有限公司 A kind of deep learning face identification method based on graphical inference global characteristics
CN111103891A (en) * 2019-12-30 2020-05-05 西安交通大学 Unmanned aerial vehicle rapid posture control system and method based on skeleton point detection
CN111339983A (en) * 2020-03-05 2020-06-26 四川长虹电器股份有限公司 Method for fine-tuning face recognition model
WO2020155873A1 (en) * 2019-02-02 2020-08-06 福州大学 Deep apparent features and adaptive aggregation network-based multi-face tracking method
CN111753884A (en) * 2020-06-08 2020-10-09 浙江工业大学 Depth map convolution model defense method and device based on network feature reinforcement
US20210000404A1 (en) * 2019-07-05 2021-01-07 The Penn State Research Foundation Systems and methods for automated recognition of bodily expression of emotion
CN112215822A (en) * 2020-10-13 2021-01-12 北京中电兴发科技有限公司 Face image quality evaluation method based on lightweight regression network
CN112381987A (en) * 2020-11-10 2021-02-19 中国人民解放军国防科技大学 Intelligent entrance guard epidemic prevention system based on face recognition
CN112488034A (en) * 2020-12-14 2021-03-12 上海交通大学 Video processing method based on lightweight face mask detection model
CN112613385A (en) * 2020-12-18 2021-04-06 成都三零凯天通信实业有限公司 Face recognition method based on monitoring video
WO2021134871A1 (en) * 2019-12-30 2021-07-08 深圳市爱协生科技有限公司 Forensics method for synthesized face image based on local binary pattern and deep learning

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109711384A (en) * 2019-01-09 2019-05-03 江苏星云网格信息技术有限公司 A kind of face identification method based on depth convolutional neural networks
WO2020155873A1 (en) * 2019-02-02 2020-08-06 福州大学 Deep apparent features and adaptive aggregation network-based multi-face tracking method
CN110070010A (en) * 2019-04-10 2019-07-30 武汉大学 A kind of face character correlating method identified again based on pedestrian
US20210000404A1 (en) * 2019-07-05 2021-01-07 The Penn State Research Foundation Systems and methods for automated recognition of bodily expression of emotion
CN110472495A (en) * 2019-07-08 2019-11-19 南京邮电大学盐城大数据研究院有限公司 A kind of deep learning face identification method based on graphical inference global characteristics
WO2021134871A1 (en) * 2019-12-30 2021-07-08 深圳市爱协生科技有限公司 Forensics method for synthesized face image based on local binary pattern and deep learning
CN111103891A (en) * 2019-12-30 2020-05-05 西安交通大学 Unmanned aerial vehicle rapid posture control system and method based on skeleton point detection
CN111339983A (en) * 2020-03-05 2020-06-26 四川长虹电器股份有限公司 Method for fine-tuning face recognition model
CN111753884A (en) * 2020-06-08 2020-10-09 浙江工业大学 Depth map convolution model defense method and device based on network feature reinforcement
CN112215822A (en) * 2020-10-13 2021-01-12 北京中电兴发科技有限公司 Face image quality evaluation method based on lightweight regression network
CN112381987A (en) * 2020-11-10 2021-02-19 中国人民解放军国防科技大学 Intelligent entrance guard epidemic prevention system based on face recognition
CN112488034A (en) * 2020-12-14 2021-03-12 上海交通大学 Video processing method based on lightweight face mask detection model
CN112613385A (en) * 2020-12-18 2021-04-06 成都三零凯天通信实业有限公司 Face recognition method based on monitoring video

Also Published As

Publication number Publication date
CN113936312B (en) 2024-06-07

Similar Documents

Publication Publication Date Title
CN104063883B (en) A kind of monitor video abstraction generating method being combined based on object and key frame
US10198657B2 (en) All-weather thermal-image pedestrian detection method
CN110910420B (en) Moving target detection tracking method based on image stream
WO2022099598A1 (en) Video dynamic target detection method based on relative statistical features of image pixels
JP5036580B2 (en) How to adapt the boost classifier to new samples
Liu et al. Pose-guided R-CNN for jersey number recognition in sports
JP2006209755A (en) Method for tracing moving object inside frame sequence acquired from scene
CN113537099A (en) Dynamic detection method for fire smoke in highway tunnel
CN110969078A (en) Abnormal behavior identification method based on human body key points
CN110956158A (en) Pedestrian shielding re-identification method based on teacher and student learning frame
CN114842553A (en) Behavior detection method based on residual shrinkage structure and non-local attention
CN109858351B (en) Gait recognition method based on hierarchy real-time memory
CN109359530B (en) Intelligent video monitoring method and device
CN111160115B (en) Video pedestrian re-identification method based on twin double-flow 3D convolutional neural network
Almomani et al. Segtrack: A novel tracking system with improved object segmentation
Jin et al. Fusing Canny operator with vibe algorithm for target detection
CN112487926A (en) Scenic spot feeding behavior identification method based on space-time diagram convolutional network
CN110795995B (en) Data processing method, device and computer readable storage medium
CN112633179A (en) Farmer market aisle object occupying channel detection method based on video analysis
CN110046601B (en) Pedestrian detection method for crossroad scene
Wibowo et al. Tracking failures detection and correction for face tracking by detection approach based on fuzzy coding histogram and point representation
CN113936312A (en) Face recognition base screening method based on deep learning graph convolution network
CN112070048B (en) Vehicle attribute identification method based on RDSNet
CN107229936B (en) Sequence classification method based on spherical robust sequence local binarization mode
JP2011096072A (en) Device and program for detecting object

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant