WO2019100608A1 - Dispositif de capture vidéo, procédé de reconnaissance de visage, système, et support d'informations lisible par ordinateur - Google Patents

Dispositif de capture vidéo, procédé de reconnaissance de visage, système, et support d'informations lisible par ordinateur Download PDF

Info

Publication number
WO2019100608A1
WO2019100608A1 PCT/CN2018/076140 CN2018076140W WO2019100608A1 WO 2019100608 A1 WO2019100608 A1 WO 2019100608A1 CN 2018076140 W CN2018076140 W CN 2018076140W WO 2019100608 A1 WO2019100608 A1 WO 2019100608A1
Authority
WO
WIPO (PCT)
Prior art keywords
face
point
video data
image
face image
Prior art date
Application number
PCT/CN2018/076140
Other languages
English (en)
Chinese (zh)
Inventor
陈林
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2019100608A1 publication Critical patent/WO2019100608A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • G06V40/166Detection; Localisation; Normalisation using acquisition arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships

Definitions

  • the present application relates to the field of image processing technologies, and in particular, to a camera device, a method and system for recognizing a face, and a computer readable storage medium.
  • the existing 1:N dynamic face recognition system generally uses one server to connect one or more network cameras, and the server collects video data from the camera through the network, and performs face recognition on the video data, such a centralized analysis scheme
  • the computing pressure of the server is very large, especially when the number of cameras is large, usually one server can not meet the demand, the server array needs to be built, and there are high requirements in terms of power consumption and heat dissipation; in addition, since the video data needs to be transmitted from the camera To the server, the pressure on the network is also relatively large, and this pressure will rise as the resolution and quality of the camera increase.
  • the purpose of the present application is to provide an image capturing apparatus, a method and system for recognizing a face, and a computer readable storage medium, which aim to alleviate the calculation pressure of the server in face recognition and reduce the network transmission pressure.
  • an image pickup apparatus including a memory and a processor connected to the memory, wherein the memory stores a processing system operable on the processor, the processing The system implements the following steps when executed by the processor:
  • Detection step performing face detection on the video data to obtain a face image
  • Tracking step tracking the face image to obtain a sequence of face images
  • Image quality scoring step performing image quality scoring on the sequence of face images, and obtaining a preset number of face images with the highest score;
  • Feature point positioning step performing feature point positioning on a preset number of face images with a higher score, and correcting based on the positioned face image;
  • Feature vector output step inputting the corrected face image into a depth neural network model generated by pre-training, and acquiring the output face feature vector;
  • Transmission step transmitting the face feature vector to the server to perform a step of performing a comparison operation with the face image in the sample in the face image sample library.
  • the present application further provides a method for face recognition, and the method for face recognition includes:
  • the corrected face image is input into a depth neural network model generated by pre-training, and the output face feature vector is obtained;
  • S6 Send the face feature vector to the server to perform a step of performing a comparison operation with the face image in the sample in the face image sample library.
  • the present application further provides a system for face recognition, the system for face recognition comprising:
  • a detecting module configured to perform face detection on the video data to obtain a face image
  • a tracking module for tracking a face image to obtain a sequence of face images
  • a scoring module configured to perform image quality scoring on the sequence of face images, and obtain a preset number of face images with a higher score
  • a correction module configured to perform feature point positioning on a preset number of face images that are scored first, and perform correction based on the positioned face image
  • An input module configured to input the corrected face image into a depth neural network model generated by pre-training, and obtain an output face feature vector
  • the sending module is configured to send the face feature vector to the server to trigger an operation of comparing the face image in the sample in the face image sample library.
  • the application further provides a computer readable storage medium having a processing system stored thereon, the processing system being implemented by a processor to implement the steps:
  • Detection step performing face detection on the video data to obtain a face image
  • Tracking step tracking the face image to obtain a sequence of face images
  • Image quality scoring step performing image quality scoring on the sequence of face images, and obtaining a preset number of face images with the highest score;
  • Feature point positioning step performing feature point positioning on a preset number of face images with a higher score, and correcting based on the positioned face image;
  • Feature vector output step inputting the corrected face image into a depth neural network model generated by pre-training, and acquiring the output face feature vector;
  • Transmission step transmitting the face feature vector to the server to perform a step of performing a comparison operation with the face image in the sample in the face image sample library.
  • the beneficial effects of the present application are: the processing of one video data per camera device of the present application, in addition to collecting video, the camera device can perform face detection, tracking, image quality scoring, feature point localization and input depth neural network model on the video. In the middle, the face feature vector is obtained, and finally only the face feature vector is transmitted to the server.
  • the calculation pressure of the server can be greatly reduced, and the server array does not need to be built, and the network transmission can be reduced to a large extent. Pressure, and network transmission pressure does not rise with the resolution and image quality of the camera.
  • FIG. 1 is a schematic diagram of an optional application environment of each embodiment of the present application.
  • FIG. 2 is a schematic diagram of a hardware architecture of an embodiment of the camera device of FIG. 1;
  • FIG. 3 is a schematic flowchart diagram of an embodiment of a method for recognizing a face of an applicant.
  • FIG. 1 it is a schematic diagram of an application environment of a preferred embodiment of the method for face recognition of the present applicant.
  • the application environment diagram includes a camera device 1 and a server 2.
  • the plurality of imaging apparatuses 1 can perform data interaction with the server 2 through suitable technologies such as a network and a near field communication technology, respectively.
  • the server 2 may be a single network server, a server group composed of a plurality of network servers, or a cloud-based cloud composed of a large number of hosts or network servers, wherein the cloud computing is a kind of distributed computing, and is a group of loosely coupled computers. A set of super virtual computers.
  • the camera device 1 is a common electronic product that includes a camera and can dynamically acquire images, and can automatically perform numerical calculation and/or information processing according to preset or stored instructions.
  • the image capturing apparatus 1 may include, but is not limited to, a memory 11, a processor 12, a network interface 13, and a camera 14 that are communicably connected to each other through a system bus.
  • the memory 11 is stored and processed.
  • Each camera device 1 includes a processor (the processor is an nvidia tx2 chip for processing images), and the nvidia tx2 chip can be connected to the camera device 1 via a usb or csi or a network interface to run the processing system.
  • the imaging device 1 and the server 2 are connected by a network, and the server 2 stores a database of human face image samples.
  • the camera device 1 is installed in a specific place (for example, an office place and a monitoring area), and a video is captured in real time for a target entering the specific place, and the processor processes the video to obtain a face feature vector, and then only sends the face feature vector through the network.
  • the server 2 performs comparison based on the face image sample library to implement face recognition.
  • the memory 11 includes a memory and at least one type of readable storage medium.
  • the memory provides a cache for the operation of the camera device 1;
  • the readable storage medium may be, for example, a flash memory, a hard disk, a multimedia card, a card type memory (eg, SD or DX memory, etc.), a random access memory (RAM), a static random access memory (SRAM).
  • a non-volatile storage medium such as a read only memory (ROM), an electrically erasable programmable read only memory (EEPROM), a programmable read only memory (PROM), a magnetic memory, a magnetic disk, an optical disk, or the like.
  • the readable storage medium may be an internal storage unit of the camera 1, such as a hard disk of the camera 1; in other embodiments, the non-volatile storage medium may also be external to the camera 1.
  • the storage device is, for example, a plug-in hard disk provided on the camera 1, a smart memory card (SMC), a Secure Digital (SD) card, a flash card, or the like.
  • the readable storage medium of the memory 11 is generally used to store an operating system installed in the image pickup apparatus 1 and various types of application software, such as program codes of the processing system in an embodiment of the present application. Further, the memory 11 can also be used to temporarily store various types of data that have been output or are to be output.
  • the processor 12 is configured to run program code or process data stored in the memory 11, such as running a processing system or the like.
  • the network interface 13 may comprise a wireless network interface or a wired network interface, which is typically used to establish a communication connection between the camera device 1 and other electronic devices.
  • the network interface 13 is mainly used to connect the camera device 1 to the server 2, and establish a data transmission channel and a communication connection between the camera device 1 and the server 2.
  • the processing system is stored in the memory 11 and includes at least one computer readable instruction stored in the memory 11, the at least one computer readable instruction being executable by the processor 12 to implement the methods of various embodiments of the present application;
  • the at least one computer readable instruction can be classified into different logic modules depending on the functions implemented by its various parts.
  • Detection step performing face detection on the video data to obtain a face image
  • Face detection is performed on each frame image in the video data based on the features of the face, and there may be one or more faces in each frame image, or no face, and after face detection, the image may be extracted from the image. Face image.
  • the face image is an image including only a face region (no other background), and the face region can be large or small.
  • the face region is small, and for a close-up shot of a face image,
  • the face area is large.
  • the face area is a minimum area including a human face, and is preferably a rectangular area including a human face. Of course, it may be an area including a human face of other shapes, such as a circular area, and is not limited thereto.
  • Tracking step tracking the face image to obtain a sequence of face images
  • the similarity of the adjacent two frames of the face image can be calculated to implement face tracking.
  • the similarity of the face may be calculated according to the X and Y coordinate values of the center point of the face region in the adjacent two frames of the face image; in other embodiments, the face image of the adjacent two frames may be used.
  • the X, Y coordinate values of the center point of the face region in the face region, and the height H and the width W value of the face region are calculated, and the similarity of the faces in the adjacent two frames of the face image is calculated.
  • the face tracking is performed based on the similarity of the faces in the adjacent two frames of the face image, and a sequence of face images of the same person is obtained, and two or more persons appearing in the face image may also be respectively obtained.
  • Image quality scoring step performing image quality scoring on the sequence of face images, and obtaining a preset number of face images with the highest score;
  • the quality of each face image in the series is scored according to the gradient values and coordinate values of predetermined points in the face image.
  • the predetermined points include an eye point, a nose point and a mouth point
  • the gradient value of the predetermined point is a mean gradient
  • the average gradient refers to a boundary of a predetermined point of the face image or a gray near the sides of the shadow line.
  • the coordinate values of the predetermined points include at least the x-axis of the eye point and the nose point.
  • Feature point positioning step performing feature point positioning on a preset number of face images with a higher score, and correcting based on the positioned face image;
  • the scoring results are arranged in descending order, that is, the face image is arranged in front of the face image, and the sequence is selected from the arranged sequence.
  • the preset number of face images of the top score for example, 7 face images are selected.
  • the feature points include at least an eye feature point, a mouth feature point, and a nose feature point, and are corrected based on the face image after the feature point is positioned.
  • Feature vector output step inputting the corrected face image into a depth neural network model generated by pre-training, and acquiring the output face feature vector;
  • Transmission step transmitting the face feature vector to the server to perform a step of performing a comparison operation with the face image in the sample in the face image sample library.
  • the corrected face image is input into a depth neural network model generated by pre-training, and is calculated by a deep neural network model, and then the face feature vector of each face image is output, and then the camera device only The face feature vector is transmitted to the server for 1:N dynamic recognition.
  • each camera device in the embodiment processes one channel of video data, and the camera device can perform face detection, tracking, image quality scoring, feature point location, and input depth neural network model in addition to video capture.
  • the face feature vector is obtained, and finally only the face feature vector is transmitted to the server.
  • the calculation pressure of the server can be greatly reduced, and the server array does not need to be built, and the network transmission can be reduced to a large extent. Pressure, and network transmission pressure does not rise with the resolution and image quality of the camera.
  • the method further includes:
  • the format of the video data is converted into a format capable of face detection
  • the video data is decoded and the format of the video data is converted into a format capable of face detection.
  • the camera device may compress the video data after the video data is collected.
  • the compressed video data may be non-real-time compressed or compressed in real time according to the real-time performance.
  • This embodiment is preferably real-time compression.
  • the captured video data may be subjected to lossy compression according to actual conditions, and the compression ratio is a predetermined ratio, preferably 5:1.
  • Video compression algorithms include M-JPEG (Motion-Join Photographic Experts Group), Mpeg (Moving Pictures Experts Group), H.264, Wavelet (Wavelet Compression), JPEG 2000, AVS compression, etc., through the above compression algorithm to obtain compressed data. Before the face detection, it can be analyzed whether the video data is compressed.
  • the format is a compressed format, and if it is further processed, for example, the camera is compressed by M-JPEG, and the format is YCrCB. , you need to convert the video data in YCrCB format to RGB format so that face detection can be performed.
  • the tracking step specifically includes:
  • Face tracking is performed based on the similarity of faces in adjacent two frames of face images.
  • the similarity calculation step includes:
  • the S i,j is a similarity, and the w x , w y , w w , w h are the x-direction distance, the y-direction distance, the width difference, and the height difference of the adjacent two frames of the face i and the face j, respectively.
  • the face in the adjacent two frames of the face image is determined to be the same person's face.
  • the image quality scoring step specifically includes: each of the series according to a gradient value and a coordinate value of a predetermined point in the face image. The quality of the face image is scored.
  • the predetermined point includes an eye point, a nose point, and a mouth point
  • the gradient value is an average gradient of an eye point, a nose point, and a mouth point
  • the eye point includes a left eye point and The right eye point
  • the mouth point includes a left mouth corner point and a right mouth corner point
  • the image quality score is calculated as:
  • x_LeftEye represents the X coordinate of the left and right eyeballs
  • x_Nose represents the X coordinate of the tip of the nose
  • grad is the average gradient of the eye point, nose point and mouth point.
  • the coordinates of the eye point, the nose point, and the mouth point in the face are selected to score the quality of the face image, and the face image can be objectively and accurately evaluated to obtain a person with a high score.
  • the face image is convenient for subsequent correction and other processing.
  • FIG. 3 is a schematic flowchart of an embodiment of a method for recognizing a face of an applicant, where the method includes the following steps:
  • Step S1 performing face detection on the video data to obtain a face image
  • Face detection is performed on each frame image in the video data based on the features of the face, and there may be one or more faces in each frame image, or no face, and after face detection, the image may be extracted from the image. Face image.
  • the face image is an image including only a face region (no other background), and the face region can be large or small.
  • the face region is small, and for a close-up shot of a face image,
  • the face area is large.
  • the face area is a minimum area including a human face, and is preferably a rectangular area including a human face. Of course, it may be an area including a human face of other shapes, such as a circular area, and is not limited thereto.
  • Step S2 tracking the face image to obtain a sequence of face images
  • the similarity of the adjacent two frames of the face image can be calculated to implement face tracking.
  • the similarity of the face may be calculated according to the X and Y coordinate values of the center point of the face region in the adjacent two frames of the face image; in other embodiments, the face image of the adjacent two frames may be used.
  • the X, Y coordinate values of the center point of the face region in the face region, and the height H and the width W value of the face region are calculated, and the similarity of the faces in the adjacent two frames of the face image is calculated.
  • the face tracking is performed based on the similarity of the faces in the adjacent two frames of the face image, and a sequence of face images of the same person is obtained, and two or more persons appearing in the face image may also be respectively obtained.
  • Step S3 performing image quality scoring on the sequence of face images, and obtaining a preset number of face images with a higher score
  • the quality of each face image in the series is scored according to the gradient values and coordinate values of predetermined points in the face image.
  • the predetermined points include an eye point, a nose point and a mouth point
  • the gradient value of the predetermined point is a mean gradient
  • the average gradient refers to a boundary of a predetermined point of the face image or a gray near the sides of the shadow line.
  • the coordinate values of the predetermined points include at least the x-axis of the eye point and the nose point.
  • Step S4 performing feature point positioning on the preset number of face images that are scored first, and correcting based on the positioned face image;
  • the scoring results are arranged in descending order, that is, the face image is arranged in front of the face image, and the sequence is selected from the arranged sequence.
  • the preset number of face images of the top score for example, 7 face images are selected.
  • the feature points include at least an eye feature point, a mouth feature point, and a nose feature point, and are corrected based on the face image after the feature point is positioned.
  • Step S5 the corrected face image is input into the depth neural network model generated by the pre-training, and the output face feature vector is obtained;
  • step S6 the face feature vector is sent to the server to perform a step of performing a comparison operation with the face image in the sample in the face image sample library.
  • the corrected face image is input into a depth neural network model generated by pre-training, and is calculated by a deep neural network model, and then the face feature vector of each face image is output, and then the camera device only The face feature vector is transmitted to the server for 1:N dynamic recognition.
  • each camera device in the embodiment processes one channel of video data, and the camera device can perform face detection, tracking, image quality scoring, feature point location, and input depth neural network model in addition to video capture.
  • the face feature vector is obtained, and finally only the face feature vector is transmitted to the server.
  • the calculation pressure of the server can be greatly reduced, and the server array does not need to be built, and the network transmission can be reduced to a large extent. Pressure, and network transmission pressure does not rise with the resolution and image quality of the camera.
  • the method further includes:
  • the format of the video data is converted into a format capable of face detection
  • the video data is decoded and the format of the video data is converted into a format capable of face detection.
  • the camera device may compress the video data after the video data is collected.
  • the compressed video data may be non-real-time compressed or compressed in real time according to the real-time performance.
  • This embodiment is preferably real-time compression.
  • the captured video data may be subjected to lossy compression according to actual conditions, and the compression ratio is a predetermined ratio, preferably 5:1.
  • Video compression algorithms include M-JPEG (Motion-Join Photographic Experts Group), Mpeg (Moving Pictures Experts Group), H.264, Wavelet (Wavelet Compression), JPEG 2000, AVS compression, etc., obtains compressed video data through the above compression algorithm. Before the face detection, it can be analyzed whether the video data is compressed.
  • the format is a compressed format, and if it is further processed, for example, the camera is compressed by M-JPEG, and the format is YCrCB. , you need to convert the video data in YCrCB format to RGB format so that face detection can be performed.
  • the step S2 specifically includes:
  • Face tracking is performed based on the similarity of faces in adjacent two frames of face images.
  • the similarity calculation step includes:
  • the S i,j is a similarity, and the w x , w y , w w , w h are the x-direction distance, the y-direction distance, the width difference, and the height difference of the adjacent two frames of the face i and the face j, respectively.
  • the face in the adjacent two frames of the face image is determined to be the same person's face.
  • the step S3 specifically includes:
  • the quality of each face image in the series is scored according to the gradient values and coordinate values of the predetermined points in the face image.
  • the predetermined point includes an eye point, a nose point, and a mouth point
  • the gradient value is an average gradient of an eye point, a nose point, and a mouth point
  • the eye point includes a left eye point and The right eye point
  • the mouth point includes a left mouth corner point and a right mouth corner point
  • the image quality score is calculated as:
  • x_LeftEye represents the X coordinate of the left and right eyeballs
  • x_Nose represents the X coordinate of the tip of the nose
  • grad is the average gradient of the eye point, nose point and mouth point.
  • the coordinates of the eye point, the nose point, and the mouth point in the face are selected to score the quality of the face image, and the face image can be objectively and accurately evaluated to obtain a person with a high score.
  • the face image is convenient for subsequent correction and other processing.
  • the present application also provides a computer readable storage medium having stored thereon a processing system, the processing system being executed by a processor to implement the steps of the method of face recognition described above.
  • the foregoing embodiment method can be implemented by means of software plus a necessary general hardware platform, and of course, can also be through hardware, but in many cases, the former is better.
  • Implementation Based on such understanding, the technical solution of the present application, which is essential or contributes to the prior art, may be embodied in the form of a software product stored in a storage medium (such as ROM/RAM, disk,
  • the optical disc includes a number of instructions for causing a terminal device (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to perform the methods described in various embodiments of the present application.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Library & Information Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Databases & Information Systems (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

La présente invention concerne un dispositif de capture vidéo, un procédé de reconnaissance de visage, un système, et un support d'informations lisible par ordinateur. Le procédé de reconnaissance de visage comprend : l'exécution d'une détection de visage sur des données vidéo pour obtenir une image de visage ; le suivi de l'image de visage et l'acquisition d'une séquence d'images de visage ; la réalisation d'une évaluation de qualité d'image sur la séquence d'images de visage, et l'acquisition d'un nombre prédéfini d'images de visage de première qualité ; l'exécution d'un positionnement de point caractéristique sur le nombre prédéfini d'images de visage de première qualité et l'exécution d'un étalonnage sur la base des images de visage positionnées ; l'entrée des images de visage étalonnées dans un modèle de réseau neuronal profond qui est généré par apprentissage préalable, et l'acquisition d'un vecteur de caractéristique de visage sorti ; et l'envoi du vecteur de caractéristique de visage à un serveur afin d'exécuter une opération de comparaison avec des images de visage dans des échantillons dans une bibliothèque d'échantillons d'images de visage. Selon la présente invention, la charge de calculs du serveur durant une reconnaissance de visage peut être réduite et la charge de transmissions dans le réseau peut être diminuée.
PCT/CN2018/076140 2017-11-21 2018-02-10 Dispositif de capture vidéo, procédé de reconnaissance de visage, système, et support d'informations lisible par ordinateur WO2019100608A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201711166813.8 2017-11-21
CN201711166813.8A CN108038422B (zh) 2017-11-21 2017-11-21 摄像装置、人脸识别的方法及计算机可读存储介质

Publications (1)

Publication Number Publication Date
WO2019100608A1 true WO2019100608A1 (fr) 2019-05-31

Family

ID=62094093

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/076140 WO2019100608A1 (fr) 2017-11-21 2018-02-10 Dispositif de capture vidéo, procédé de reconnaissance de visage, système, et support d'informations lisible par ordinateur

Country Status (2)

Country Link
CN (1) CN108038422B (fr)
WO (1) WO2019100608A1 (fr)

Cited By (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110245630A (zh) * 2019-06-18 2019-09-17 广东中安金狮科创有限公司 监控数据处理方法、装置及可读存储介质
CN110610533A (zh) * 2019-08-07 2019-12-24 重庆灵翎互娱科技有限公司 一种基于人脸三维网格模型进行表情捕捉的方法和设备
CN110659564A (zh) * 2019-08-12 2020-01-07 万翼科技有限公司 区域内用户追踪方法、装置、计算机设备以及存储介质
CN111008582A (zh) * 2019-11-28 2020-04-14 厦门瑞为信息技术有限公司 一种头部照片分析方法、***和设备
CN111126219A (zh) * 2019-12-16 2020-05-08 国网浙江省电力有限公司电力科学研究院 一种基于人工智能的变电站人员的身份识别***和方法
CN111222433A (zh) * 2019-12-30 2020-06-02 新大陆数字技术股份有限公司 自动人脸稽核方法、***、设备及可读存储介质
CN111401170A (zh) * 2020-03-06 2020-07-10 西安奥卡云数据科技有限公司 一种人脸检测方法及装置
CN111427448A (zh) * 2020-03-05 2020-07-17 融信信息科技有限公司 一种人像标注方法、装置及计算机可读存储介质
CN111783674A (zh) * 2020-07-02 2020-10-16 厦门市美亚柏科信息股份有限公司 一种基于ar眼镜的人脸识别方法和***
CN111797797A (zh) * 2020-07-13 2020-10-20 深圳大学 基于网格形变优化的人脸图像处理方法、终端及存储介质
CN111898408A (zh) * 2020-06-09 2020-11-06 广州杰赛科技股份有限公司 一种快速人脸识别方法及装置
CN112052729A (zh) * 2020-07-30 2020-12-08 广州市标准化研究院 一种基于人脸识别的智能动态高清视频检测方法及***
CN112215156A (zh) * 2020-10-13 2021-01-12 北京中电兴发科技有限公司 一种视频监控中的人脸抓拍方法及***
CN112241670A (zh) * 2019-07-18 2021-01-19 杭州海康威视数字技术股份有限公司 图像处理方法及装置
CN112347849A (zh) * 2020-09-29 2021-02-09 咪咕视讯科技有限公司 视频会议处理方法、电子设备及存储介质
CN112487396A (zh) * 2020-12-08 2021-03-12 平安国际智慧城市科技股份有限公司 图片的处理方法、装置、计算机设备及存储介质
CN113033587A (zh) * 2019-12-24 2021-06-25 深圳云天励飞技术有限公司 图像识别结果评估方法、装置、电子设备及存储介质
CN113099150A (zh) * 2020-01-08 2021-07-09 华为技术有限公司 图像处理的方法、设备及***
CN113283305A (zh) * 2021-04-29 2021-08-20 百度在线网络技术(北京)有限公司 人脸识别方法、装置、电子设备及计算机可读存储介质
WO2023041963A1 (fr) * 2021-09-20 2023-03-23 Sensetime International Pte. Ltd. Procédés et appareils d'identification de visage
CN115985007A (zh) * 2022-12-06 2023-04-18 杭州未兰石云信息科技有限公司 一种基于低功耗的5g记录仪视频巡检方法与***

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110634116B (zh) * 2018-05-30 2022-04-05 杭州海康威视数字技术股份有限公司 一种面部图像评分方法及摄像机
CN110580808B (zh) * 2018-06-08 2021-03-23 杭州海康威视数字技术股份有限公司 一种信息处理方法、装置、电子设备及智能交通***
CN111199165B (zh) * 2018-10-31 2024-02-06 浙江宇视科技有限公司 图像处理方法及装置
CN109858328B (zh) * 2018-12-14 2023-06-02 上海集成电路研发中心有限公司 一种基于视频的人脸识别的方法及装置
CN111332305A (zh) * 2018-12-18 2020-06-26 朱向雷 一种主动预警型交通道路感知辅助驾驶预警***
CN109714597A (zh) * 2019-01-22 2019-05-03 成都神州数码索贝科技有限公司 一种无损视频压缩方法
CN110784628B (zh) * 2019-08-14 2022-04-05 腾讯科技(深圳)有限公司 图像数据的采集处理方法、***及智能摄像头、服务器
CN112911385B (zh) * 2021-01-12 2021-12-07 平安科技(深圳)有限公司 待识别图片的提取方法、装置、设备以及存储介质
AU2021240278A1 (en) * 2021-09-20 2023-04-06 Sensetime International Pte. Ltd. Face identification methods and apparatuses
CN116863640A (zh) * 2023-07-03 2023-10-10 河南大学 基于多目标的行为识别和远程监测的警报***及方法

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102360421A (zh) * 2011-10-19 2012-02-22 苏州大学 一种基于视频流的人脸识别方法及***
US20120106790A1 (en) * 2010-10-26 2012-05-03 DigitalOptics Corporation Europe Limited Face or Other Object Detection Including Template Matching
CN105787478A (zh) * 2016-04-14 2016-07-20 中南大学 基于神经网络和灵敏度参数的人脸转向识别方法
CN106022317A (zh) * 2016-06-27 2016-10-12 北京小米移动软件有限公司 人脸识别方法及装置
CN106503682A (zh) * 2016-10-31 2017-03-15 北京小米移动软件有限公司 视频数据中的关键点定位方法及装置

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
RU2007102021A (ru) * 2007-01-19 2008-07-27 Корпораци "Самсунг Электроникс Ко., Лтд." (KR) Способ и система распознавания личности
CN102201061B (zh) * 2011-06-24 2012-10-31 常州锐驰电子科技有限公司 基于多阶层过滤人脸识别的智能安全监控***及方法
CN105701486B (zh) * 2014-11-26 2019-11-19 上海骏聿数码科技有限公司 一种在摄像机内实现人脸信息分析及提取的方法
CN205451095U (zh) * 2015-12-02 2016-08-10 深圳市商汤科技有限公司 一种人脸识别装置
CN105488478B (zh) * 2015-12-02 2020-04-07 深圳市商汤科技有限公司 一种人脸识别***和方法

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120106790A1 (en) * 2010-10-26 2012-05-03 DigitalOptics Corporation Europe Limited Face or Other Object Detection Including Template Matching
CN102360421A (zh) * 2011-10-19 2012-02-22 苏州大学 一种基于视频流的人脸识别方法及***
CN105787478A (zh) * 2016-04-14 2016-07-20 中南大学 基于神经网络和灵敏度参数的人脸转向识别方法
CN106022317A (zh) * 2016-06-27 2016-10-12 北京小米移动软件有限公司 人脸识别方法及装置
CN106503682A (zh) * 2016-10-31 2017-03-15 北京小米移动软件有限公司 视频数据中的关键点定位方法及装置

Cited By (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110245630A (zh) * 2019-06-18 2019-09-17 广东中安金狮科创有限公司 监控数据处理方法、装置及可读存储介质
CN112241670A (zh) * 2019-07-18 2021-01-19 杭州海康威视数字技术股份有限公司 图像处理方法及装置
CN112241670B (zh) * 2019-07-18 2024-03-01 杭州海康威视数字技术股份有限公司 图像处理方法及装置
CN110610533A (zh) * 2019-08-07 2019-12-24 重庆灵翎互娱科技有限公司 一种基于人脸三维网格模型进行表情捕捉的方法和设备
CN110610533B (zh) * 2019-08-07 2023-06-16 重庆灵翎互娱科技有限公司 一种基于人脸三维网格模型进行表情捕捉的方法和设备
CN110659564A (zh) * 2019-08-12 2020-01-07 万翼科技有限公司 区域内用户追踪方法、装置、计算机设备以及存储介质
CN111008582A (zh) * 2019-11-28 2020-04-14 厦门瑞为信息技术有限公司 一种头部照片分析方法、***和设备
CN111008582B (zh) * 2019-11-28 2023-04-07 厦门瑞为信息技术有限公司 一种头部照片分析方法、***和设备
CN111126219A (zh) * 2019-12-16 2020-05-08 国网浙江省电力有限公司电力科学研究院 一种基于人工智能的变电站人员的身份识别***和方法
CN113033587A (zh) * 2019-12-24 2021-06-25 深圳云天励飞技术有限公司 图像识别结果评估方法、装置、电子设备及存储介质
CN113033587B (zh) * 2019-12-24 2024-06-11 深圳云天励飞技术有限公司 图像识别结果评估方法、装置、电子设备及存储介质
CN111222433A (zh) * 2019-12-30 2020-06-02 新大陆数字技术股份有限公司 自动人脸稽核方法、***、设备及可读存储介质
CN111222433B (zh) * 2019-12-30 2023-06-20 新大陆数字技术股份有限公司 自动人脸稽核方法、***、设备及可读存储介质
CN113099150A (zh) * 2020-01-08 2021-07-09 华为技术有限公司 图像处理的方法、设备及***
CN111427448A (zh) * 2020-03-05 2020-07-17 融信信息科技有限公司 一种人像标注方法、装置及计算机可读存储介质
CN111427448B (zh) * 2020-03-05 2023-07-28 融信信息科技有限公司 一种人像标注方法、装置及计算机可读存储介质
CN111401170A (zh) * 2020-03-06 2020-07-10 西安奥卡云数据科技有限公司 一种人脸检测方法及装置
CN111401170B (zh) * 2020-03-06 2023-06-06 西安奥卡云数据科技有限公司 一种人脸检测方法及装置
CN111898408B (zh) * 2020-06-09 2023-09-19 广州杰赛科技股份有限公司 一种快速人脸识别方法及装置
CN111898408A (zh) * 2020-06-09 2020-11-06 广州杰赛科技股份有限公司 一种快速人脸识别方法及装置
CN111783674A (zh) * 2020-07-02 2020-10-16 厦门市美亚柏科信息股份有限公司 一种基于ar眼镜的人脸识别方法和***
CN111797797B (zh) * 2020-07-13 2023-09-15 深圳大学 基于网格形变优化的人脸图像处理方法、终端及存储介质
CN111797797A (zh) * 2020-07-13 2020-10-20 深圳大学 基于网格形变优化的人脸图像处理方法、终端及存储介质
CN112052729A (zh) * 2020-07-30 2020-12-08 广州市标准化研究院 一种基于人脸识别的智能动态高清视频检测方法及***
CN112052729B (zh) * 2020-07-30 2024-04-16 广州市标准化研究院 一种基于人脸识别的智能动态高清视频检测方法及***
CN112347849B (zh) * 2020-09-29 2024-03-26 咪咕视讯科技有限公司 视频会议处理方法、电子设备及存储介质
CN112347849A (zh) * 2020-09-29 2021-02-09 咪咕视讯科技有限公司 视频会议处理方法、电子设备及存储介质
CN112215156B (zh) * 2020-10-13 2022-10-14 北京中电兴发科技有限公司 一种视频监控中的人脸抓拍方法及***
CN112215156A (zh) * 2020-10-13 2021-01-12 北京中电兴发科技有限公司 一种视频监控中的人脸抓拍方法及***
CN112487396A (zh) * 2020-12-08 2021-03-12 平安国际智慧城市科技股份有限公司 图片的处理方法、装置、计算机设备及存储介质
CN113283305A (zh) * 2021-04-29 2021-08-20 百度在线网络技术(北京)有限公司 人脸识别方法、装置、电子设备及计算机可读存储介质
CN113283305B (zh) * 2021-04-29 2024-03-26 百度在线网络技术(北京)有限公司 人脸识别方法、装置、电子设备及计算机可读存储介质
WO2023041963A1 (fr) * 2021-09-20 2023-03-23 Sensetime International Pte. Ltd. Procédés et appareils d'identification de visage
CN115985007A (zh) * 2022-12-06 2023-04-18 杭州未兰石云信息科技有限公司 一种基于低功耗的5g记录仪视频巡检方法与***

Also Published As

Publication number Publication date
CN108038422A (zh) 2018-05-15
CN108038422B (zh) 2021-12-21

Similar Documents

Publication Publication Date Title
WO2019100608A1 (fr) Dispositif de capture vidéo, procédé de reconnaissance de visage, système, et support d'informations lisible par ordinateur
KR102319177B1 (ko) 이미지 내의 객체 자세를 결정하는 방법 및 장치, 장비, 및 저장 매체
WO2018219180A1 (fr) Procédé et appareil pour déterminer une qualité d'image faciale, dispositif électronique et support d'informations d'ordinateur
WO2019033574A1 (fr) Dispositif électronique, procédé et système de reconnaissance dynamique de visage dans une vidéo, et support d'informations
CN108764071B (zh) 一种基于红外和可见光图像的真实人脸检测方法及装置
WO2020094091A1 (fr) Procédé de capture d'image, caméra de surveillance et système de surveillance
CN110008806B (zh) 信息处理装置、学习处理方法、学习装置及物体识别装置
CN113228626B (zh) 视频监控***和方法
WO2020184207A1 (fr) Dispositif de suivi d'objet et procédé de suivi d'objet
WO2019033575A1 (fr) Dispositif électronique, procédé et système de suivi de visage, et support d'informations
KR20140045897A (ko) 시각적 이미지 매칭을 기반으로 한 미디어 스트림 식별 장치 및 방법
WO2021008205A1 (fr) Traitement d'images
CN111160202A (zh) 基于ar设备的身份核验方法、装置、设备及存储介质
CN113228105A (zh) 一种图像处理方法、装置和电子设备
US11605220B2 (en) Systems and methods for video surveillance
CN113822927B (zh) 一种适用弱质量图像的人脸检测方法、装置、介质及设备
CN109711287B (zh) 人脸采集方法及相关产品
CN113158773B (zh) 一种活体检测模型的训练方法及训练装置
WO2021196925A1 (fr) Procédé et appareil de détection et de suivi d'objet mobile
US20230260211A1 (en) Three-Dimensional Point Cloud Generation Method, Apparatus and Electronic Device
WO2019071663A1 (fr) Appareil électronique, procédé de production d'échantillon virtuel et support de stockage
CN112561795A (zh) 基于Spark和OpenCV的实时生成全景图实现***及方法
EP3699865B1 (fr) Dispositif de calcul de forme tridimensionnelle de visage, procédé de calcul de forme tridimensionnelle de visage et support non transitoire lisible par ordinateur
CN113486788A (zh) 视频相似度确定方法、装置、电子设备及存储介质
CN113379853B (zh) 相机内参的获取方法、装置、设备及可读存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18882113

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 05/10/2020)

122 Ep: pct application non-entry in european phase

Ref document number: 18882113

Country of ref document: EP

Kind code of ref document: A1