WO2020185948A1 - Detecting spoofing of facial recognition with mobile devices - Google Patents
Detecting spoofing of facial recognition with mobile devices Download PDFInfo
- Publication number
- WO2020185948A1 WO2020185948A1 PCT/US2020/022168 US2020022168W WO2020185948A1 WO 2020185948 A1 WO2020185948 A1 WO 2020185948A1 US 2020022168 W US2020022168 W US 2020022168W WO 2020185948 A1 WO2020185948 A1 WO 2020185948A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- data representation
- data
- user
- predictive model
- match
- Prior art date
Links
- 230000001815 facial effect Effects 0.000 title description 5
- 238000000034 method Methods 0.000 claims abstract description 139
- 238000012545 processing Methods 0.000 claims abstract description 21
- 230000008859 change Effects 0.000 claims description 34
- 230000008569 process Effects 0.000 claims description 31
- 238000004590 computer program Methods 0.000 claims description 24
- 230000003247 decreasing effect Effects 0.000 claims description 9
- 238000004891 communication Methods 0.000 description 6
- 238000011161 development Methods 0.000 description 5
- 230000008901 benefit Effects 0.000 description 4
- 238000013135 deep learning Methods 0.000 description 4
- 230000006870 function Effects 0.000 description 4
- 238000010191 image analysis Methods 0.000 description 4
- 238000010801 machine learning Methods 0.000 description 4
- 230000002093 peripheral effect Effects 0.000 description 4
- 239000004973 liquid crystal related substance Substances 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 230000004044 response Effects 0.000 description 3
- 229920001621 AMOLED Polymers 0.000 description 2
- 238000004883 computer application Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000014509 gene expression Effects 0.000 description 2
- APTZNLHMIGJTEW-UHFFFAOYSA-N pyraflufen-ethyl Chemical compound C1=C(Cl)C(OCC(=O)OCC)=CC(C=2C(=C(OC(F)F)N(C)N=2)Cl)=C1F APTZNLHMIGJTEW-UHFFFAOYSA-N 0.000 description 2
- 238000009877 rendering Methods 0.000 description 2
- 239000010979 ruby Substances 0.000 description 2
- 229910001750 ruby Inorganic materials 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- VYZAMTAEIAYCRO-UHFFFAOYSA-N Chromium Chemical compound [Cr] VYZAMTAEIAYCRO-UHFFFAOYSA-N 0.000 description 1
- 240000000594 Heliconia bihai Species 0.000 description 1
- 241000700605 Viruses Species 0.000 description 1
- 238000012790 confirmation Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- VJYFKVYYMZPMAB-UHFFFAOYSA-N ethoprophos Chemical compound CCCSP(=O)(OCC)SCCC VJYFKVYYMZPMAB-UHFFFAOYSA-N 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 230000003252 repetitive effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 239000010454 slate Substances 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- WZWYJBNHTWCXIM-UHFFFAOYSA-N tenoxicam Chemical compound O=C1C=2SC=CC=2S(=O)(=O)N(C)C1=C(O)NC1=CC=CC=N1 WZWYJBNHTWCXIM-UHFFFAOYSA-N 0.000 description 1
- 229960002871 tenoxicam Drugs 0.000 description 1
- 239000010409 thin film Substances 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/22—Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
- G06V10/235—Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition based on user input or interaction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/166—Detection; Localisation; Normalisation using acquisition arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/40—Spoof detection, e.g. liveness detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/40—Spoof detection, e.g. liveness detection
- G06V40/45—Detection of the body part being alive
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L63/00—Network architectures or network communication protocols for network security
- H04L63/08—Network architectures or network communication protocols for network security for authentication of entities
- H04L63/0861—Network architectures or network communication protocols for network security for authentication of entities using biometrical features, e.g. fingerprint, retina-scan
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L63/00—Network architectures or network communication protocols for network security
- H04L63/14—Network architectures or network communication protocols for network security for detecting or protecting against malicious traffic
- H04L63/1441—Countermeasures against malicious traffic
- H04L63/1483—Countermeasures against malicious traffic service impersonation, e.g. phishing, pharming or web spoofing
Definitions
- “Spoofing” a security system is generally defined as an act of masquerading as an authenticated user, by submitting false data.
- methods of liveness detection may be employed to determine whether a biometric modality, such as a face, a palm (palm print), a finger (fingerprint), or an ear, carries the unique structural qualities of the original three-dimensional biometric modality, or is a two-dimensional replicate.
- the method further comprises rejecting the identity recognition match if, after comparing the second data representation with the predictive model generated from the first data representation, the second data representation does not match the predictive model generated from the first data representation. In some embodiments, the method further comprises validating an additional data representation identity recognition match if the one or more additional data representations match the predictive model generated from the first data representation. In some embodiments, the method further comprises rejecting an additional data representation identity recognition match if the one or more additional data representations does not match the predictive model generated from the first data representation.
- the method further comprises comparing additional data to determine if the user presenting the face image or set of images at the first distance is a match to the user presenting the face image or set of images at the second distance if the second data representation does not match the predictive model generated from the first data representation.
- additional data comprises a name, password, identity number, address, geo-location, device ID, unique data characteristic of the user’s software environment on the mobile device, other biometric data, predictive models of user data or biometric data, other data, or any combination thereof.
- the method further comprising: (a) processing the second data representation into a predictive model that estimates the data representations of the user’s face image or set of images at other distances or orientations between the user’s face and the camera; (b) comparing the first data representation with the predictive model generated from the second data representation to determine if they match; and (c) validating an identity recognition match if the first data representation matches the predictive model generated from the second data representation.
- the comparison of a predictive model to a data representation comprises configuring a matching architecture.
- the comparison of a predictive model to a data representation comprises comparing a predictive model generated from a first data representation with a second data representation, comparing a predictive model generated from a second data representation with a first data representation, or any combination thereof.
- the configuration of matching architecture changes upon successive matching exercises, or changes upon certain successive matching exercise and not others, wherein changes may be randomized between matching exercises, or be based on non-randomized determinate data or protocols, or which may not change.
- changes to the configuration of matching architecture are based on changes being randomized between matching exercises, changes being based on non-randomized determinate data or protocols, or any combination thereof.
- the configuration of matching architecture does not change.
- a guided user interface is used to capture the first and/or second data representations.
- information captured from the guided user interface is used in matching exercises.
- the method further comprises: (a) recording, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the first distance; (b) recording, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the second distance; (c) comparing the motion and/or location data recorded at the first distance and motion and/or location data from the second distance with the predictive model generated from the first data representation, and the second data representation; and (d) validating an identity recognition match if (I) the second data representation matches the predictive model generated from the first data representation predictive modeling; and (II) the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation.
- the method further comprising comparing (I) the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation with (II) the predictive model generated from the first data representation, and the second data representation.
- the method further comprising validating an identity recognition match if (I) the second data representation matches the predictive model generated from the first data representation; and (II) the motion and/or location data recorded continuously or at a plurality of intervals match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- the method further comprising (a) recording, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the first distance; (b) recording, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the second distance; (c) comparing the motion and/or location data recorded at the first distance and motion and/or location data from the second distance with the predictive model generated from the second data representation, and the first data representation; and; and (d) validating an identity recognition match if (I) the first data representation matches the predictive model generated from the second data representation predictive modeling; and (II) the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation.
- the method further comprises comparing the (I) the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation with (II) the predictive model generated from the second data representation, and the first data representation.
- the method further comprises validating an identity recognition match if (I) the first data representation matches the predictive model generated from the second data representation; and (II) the motion and/or location data recorded continuously or at a plurality of intervals match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- a computing device comprising at least one processor, an operating system configured to perform executable instructions, a memory, and a computer program including instructions executable by the computing device to create an application for detecting spoofing of biometric identity recognition using the camera of a mobile device
- the application comprising: (a) a software module configured to record a user’s face image or set of images at a first distance; (b) a software module configured to process the user’s face image or set of images at the first distance with algorithms to generate a first data representation of the user’s image or set of images; (c) a software module configured to process the first data representation into a predictive model that estimates the data representations of the user’s face image or set of images at other distances or orientations between the user’s face and the camera; (d) a software module configured to change the distance between the user’s face and the camera, wherein changing the distance comprises increasing the distance between the user’s face and the camera, decreasing the
- the system further comprises a software module configured to reject the identity recognition match if, after the software module configured to compare the second data representation with the predictive model generated from the first data representation, the second data representation does not match the predictive model generated from the first data representation.
- the system further comprising a software module configured to capture one or more additional data representations from the user’s face image or set of images from the first or second distance, and compare the captured one or more additional data representations with the predictive model generated from the first data representation to determine if they match if, after comparing the second data representation with the predictive model generated from the first data representation, the second data representation does not match the predictive model generated from the first data representation.
- system further comprising a software module configured to validate an additional data representation identity recognition match if the one or more additional data representations match the predictive model generated from the first data representation. In some embodiments, the system further comprising a software module configured to reject an additional data representation identity recognition match if the one or more additional data representations does not match the predictive model generated from the first data representation. In some embodiments, the system further comprising a software module configured to compare additional data to determine if the user presenting the face image or set of images at the first distance is a match to the user presenting the face image or set of images at the second distance if the second data representation does not match the predictive model generated from the first data representation.
- additional data comprises a name, password, identity number, address, geo-location, device ID, unique data characteristic of the user’s software environment on the mobile device, other biometric data, predictive models of user data or biometric data, other data, or any combination thereof.
- the system further comprises: (a) a software module configured to process the second data representation into a predictive model that estimates the data representations of the user’s face image or set of images at other distances or orientations between the user’s face and the camera; (b) a software module configured to compare the first data representation with the predictive model generated from the second data representation to determine if they match; and (c) a software module configured to validate an identity recognition match if the first data representation matches the predictive model generated from the second data representation.
- the software module configured to compare the first data representation with the predictive model generated from the second data representation comprises a matching architecture configuration.
- any software module configured to compare a predictive model to a data representation comprises comparing a predictive model generated from a first data representation with a second data representation, comparing a predictive model generated from a second data representation with a first data representation, or any combination thereof.
- the configuration of matching architecture changes upon successive matching exercises, or changes upon certain successive matching exercise and not others, wherein changes may be randomized between matching exercises, or be based on non-randomized determinate data or protocols, or which may not change.
- changes to the configuration of matching architecture are based on changes being randomized between matching exercises, changes being based on non-randomized determinate data or protocols, or any combination thereof.
- the matching architecture configuration does not change.
- the system further comprises a guided user interface used to capture the first and/or second data representations. In some embodiments, information captured from the guided user interface is used in matching exercises.
- the system further comprises: (a) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the first distance; (b) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’ s face image or set of images is recorded by the camera at the second distance; (c) a software module configured to compare the motion and/or location data recorded at the first distance and motion and/or location data from the second distance with the predictive model generated from the first data representation, and the second data representation; and (d) a software module configured to validate an identity recognition match if (I) the second data representation matches the predictive model generated from the first data representation predictive modeling; and (II) the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- a software module configured to record, from one or more sensors on the mobile device, motion and/
- the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation.
- the system further comprising a software module configured to compare (I) the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation with (II) the predictive model generated from the first data representation, and the second data representation.
- system further comprising a software module configured to validate an identity recognition match if (I) the second data representation matches the predictive model generated from the first data representation; and (II) the motion and/or location data recorded continuously or at a plurality of intervals match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- the system further comprising: (a) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the first distance; (b) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the second distance; (c) a software module configured to compare the motion and/or location data recorded at the first distance and motion and/or location data from the second distance with the predictive model generated from the second data representation, and the first data representation; and (d) a software module configured to validate an identity recognition match if (I) the first data representation matches the predictive model generated from the second data representation predictive modeling; and (II) the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- a software module configured to record, from one or more sensors on the mobile device, motion and/
- the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation.
- the system further comprising a software module configured to compare the (I) the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation with (II) the predictive model generated from the second data representation, and the first data representation.
- system further comprising a software module configured to validate an identity recognition match if (I) the first data representation matches the predictive model generated from the second data representation; and (II) the motion and/or location data recorded continuously or at a plurality of intervals match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- the media further comprises a software module configured to reject the identity recognition match if, after the software module configured to compare the second data representation with the predictive model generated from the first data representation, the second data representation does not match the predictive model generated from the first data representation.
- the media further comprising a software module configured to capture one or more additional data representations from the user’s face image or set of images from the first or second distance, and compare the captured one or more additional data representations with the predictive model generated from the first data representation to determine if they match if, after comparing the second data representation with the predictive model generated from the first data representation, the second data representation does not match the predictive model generated from the first data representation.
- the media further comprising a software module configured to validate an additional data representation identity recognition match if the one or more additional data representations match the predictive model generated from the first data representation. In some embodiments, the media further comprising a software module configured to reject an additional data representation identity recognition match if the one or more additional data representations does not match the predictive model generated from the first data representation. In some embodiments, the media further comprising a software module configured to compare additional data to determine if the user presenting the face image or set of images at the first distance is a match to the user presenting the face image or set of images at the second distance if the second data representation does not match the predictive model generated from the first data representation.
- additional data comprises a name, password, identity number, address, geo location, device ID, unique data characteristic of the user’s software environment on the mobile device, other biometric data, predictive models of user data or biometric data, other data, or any combination thereof.
- the media further comprises: (a) a software module configured to process the second data representation into a predictive model that estimates the data representations of the user’s face image or set of images at other distances or orientations between the user’s face and the camera; (b) a software module configured to compare the first data representation with the predictive model generated from the second data representation to determine if they match; and (c) a software module configured to validate an identity recognition match if the first data representation matches the predictive model generated from the second data representation.
- the software module configured to compare the first data representation with the predictive model generated from the second data representation comprises a matching architecture configuration.
- a software module configured to compare a predictive model to a data representation comprises comparing a predictive model generated from a first data representation with a second data representation, comparing a predictive model generated from a second data representation with a first data representation, or any combination thereof.
- the configuration of matching architecture changes upon successive matching exercises, or changes upon certain successive matching exercise and not others, wherein changes may be randomized between matching exercises, or be based on non-randomized determinate data or protocols, or which may not change.
- changes to the configuration of matching architecture are based on changes being randomized between matching exercises, changes being based on non-randomized determinate data or protocols, or any combination thereof.
- the configuration of matching architecture does not change.
- the media further comprises a guided user interface used to capture the first and/or second data representations.
- information captured from the guided user interface is used in matching exercises.
- the media further comprising: (a) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the first distance; (b) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’ s face image or set of images is recorded by the camera at the second distance; (c) a software module configured to compare the motion and/or location data recorded at the first distance and motion and/or location data from the second distance with the predictive model generated from the first data representation, and the second data representation; and (d) a software module configured to validate an identity recognition match if (I) the second data representation matches the predictive model generated from the first data representation predictive modeling; and (II) the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- a software module configured to record, from one or more sensors on the mobile device, motion and
- the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation.
- the media further comprising a software module configured to compare (I) the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation with (II) the predictive model generated from the first data representation, and the second data representation.
- the media further comprising a software module configured to validate an identity recognition match if (I) the second data representation matches the predictive model generated from the first data representation; and (II) the motion and/or location data recorded continuously or at a plurality of intervals match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- the media further comprising: (a) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the first distance; (b) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the second distance; (c) a software module configured to compare the motion and/or location data recorded at the first distance and motion and/or location data from the second distance with the predictive model generated from the second data representation, and the first data representation; and (d) a software module configured to validate an identity recognition match if (I) the first data representation matches the predictive model generated from the second data representation predictive modeling; and (II) the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- a software module configured to record, from one or more sensors on the mobile device, motion and/
- the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation.
- the media further comprising a software module configured to compare the (I) the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation with (II) the predictive model generated from the second data representation, and the first data representation.
- the media further comprising a software module configured to validate an identity recognition match if (I) the first data representation matches the predictive model generated from the second data representation; and (II) the motion and/or location data recorded continuously or at a plurality of intervals match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- FIG. 1 shows a non-limiting example of a computing device; in this case, a device with one or more processors, memory, storage, and a network interface, per an embodiment herein;
- FIG. 2 shows a non-limiting example of a web/mobile application provision system; in this case, a system providing browser-based and/or native mobile user interfaces, per an embodiment herein;
- FIG. 3 shows a non-limiting example of a cloud-based web/mobile application provision system; in this case, a system comprising an elastically load balanced, auto-scaling web server and application server resources as well synchronously replicated databases, per an embodiment herein;
- FIG. 4 shows a non-limiting example of a first and second data predictive matching architecture, per an embodiment herein;
- FIG. 5 shows a non-limiting example of a continuous data predictive matching architecture, per an embodiment herein.
- FIG. 6 shows a non-limiting example of a continuous data predictive matching architecture; in this case an architecture using multiple past frames, per an embodiment herein.
- the method further comprises rejecting the identity recognition match if, after comparing the second data representation with the predictive model generated from the first data representation, the second data representation does not match the predictive model generated from the first data representation. In some embodiments, the method further comprises validating an additional data representation identity recognition match if the one or more additional data representations match the predictive model generated from the first data representation. In some embodiments, the method further comprises rejecting an additional data representation identity recognition match if the one or more additional data representations does not match the predictive model generated from the first data representation.
- the method further comprises comparing additional data to determine if the user presenting the face image or set of images at the first distance is a match to the user presenting the face image or set of images at the second distance if the second data representation does not match the predictive model generated from the first data representation.
- additional data comprises a name, password, identity number, address, geo-location, device ID, unique data characteristic of the user’s software environment on the mobile device, other biometric data, predictive models of user data or biometric data, other data, or any combination thereof.
- the method further comprising: (a) processing the second data representation into a predictive model that estimates the data representations of the user’s face image or set of images at other distances or orientations between the user’s face and the camera; (b) comparing the first data representation with the predictive model generated from the second data representation to determine if they match; and (c) validating an identity recognition match if the first data representation matches the predictive model generated from the second data representation.
- the comparison of a predictive model to a data representation comprises configuring a matching architecture.
- the comparison of a predictive model to a data representation comprises comparing a predictive model generated from a first data representation with a second data representation, comparing a predictive model generated from a second data representation with a first data representation, or any combination thereof.
- the configuration of matching architecture changes upon successive matching exercises, or changes upon certain successive matching exercise and not others, wherein changes may be randomized between matching exercises, or be based on non-randomized determinate data or protocols, or which may not change.
- changes to the configuration of matching architecture are based on changes being randomized between matching exercises, changes being based on non-randomized determinate data or protocols, or any combination thereof.
- the configuration of matching architecture does not change.
- a guided user interface is used to capture the first and/or second data representations.
- information captured from the guided user interface is used in matching exercises.
- the method further comprises: (a) recording, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the first distance; (b) recording, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the second distance; (c) comparing the motion and/or location data recorded at the first distance and motion and/or location data from the second distance with the predictive model generated from the first data representation, and the second data representation; and (d) validating an identity recognition match if (I) the second data representation matches the predictive model generated from the first data representation predictive modeling; and (II) the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation.
- the method further comprising comparing (I) the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation with (II) the predictive model generated from the first data representation, and the second data representation.
- the method further comprising validating an identity recognition match if (I) the second data representation matches the predictive model generated from the first data representation; and (II) the motion and/or location data recorded continuously or at a plurality of intervals match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- the method further comprising (a) recording, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the first distance; (b) recording, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the second distance; (c) comparing the motion and/or location data recorded at the first distance and motion and/or location data from the second distance with the predictive model generated from the second data representation, and the first data representation; and; and (d) validating an identity recognition match if (I) the first data representation matches the predictive model generated from the second data representation predictive modeling; and (II) the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation.
- the method further comprises comparing the (I) the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation with (II) the predictive model generated from the second data representation, and the first data representation.
- the method further comprises validating an identity recognition match if (I) the first data representation matches the predictive model generated from the second data representation; and (II) the motion and/or location data recorded continuously or at a plurality of intervals match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- a computing device comprising at least one processor, an operating system configured to perform executable instructions, a memory, and a computer program including instructions executable by the computing device to create an application for detecting spoofing of biometric identity recognition using the camera of a mobile device
- the application comprising: (a) a software module configured to record a user’s face image or set of images at a first distance; (b) a software module configured to process the user’s face image or set of images at the first distance with algorithms to generate a first data representation of the user’s image or set of images; (c) a software module configured to process the first data representation into a predictive model that estimates the data representations of the user’s face image or set of images at other distances or orientations between the user’s face and the camera; (d) a software module configured to change the distance between the user’s face and the camera, wherein changing the distance comprises increasing the distance between the user’s face and the camera, decreasing the
- the system further comprises a software module configured to reject the identity recognition match if, after the software module configured to compare the second data representation with the predictive model generated from the first data representation, the second data representation does not match the predictive model generated from the first data representation.
- the system further comprising a software module configured to capture one or more additional data representations from the user’s face image or set of images from the first or second distance, and compare the captured one or more additional data representations with the predictive model generated from the first data representation to determine if they match if, after comparing the second data representation with the predictive model generated from the first data representation, the second data representation does not match the predictive model generated from the first data representation.
- system further comprising a software module configured to validate an additional data representation identity recognition match if the one or more additional data representations match the predictive model generated from the first data representation. In some embodiments, the system further comprising a software module configured to reject an additional data representation identity recognition match if the one or more additional data representations does not match the predictive model generated from the first data representation. In some embodiments, the system further comprising a software module configured to compare additional data to determine if the user presenting the face image or set of images at the first distance is a match to the user presenting the face image or set of images at the second distance if the second data representation does not match the predictive model generated from the first data representation.
- additional data comprises a name, password, identity number, address, geo-location, device ID, unique data characteristic of the user’s software environment on the mobile device, other biometric data, predictive models of user data or biometric data, other data, or any combination thereof.
- the system further comprises: (a) a software module configured to process the second data representation into a predictive model that estimates the data representations of the user’s face image or set of images at other distances or orientations between the user’s face and the camera; (b) a software module configured to compare the first data representation with the predictive model generated from the second data representation to determine if they match; and (c) a software module configured to validate an identity recognition match if the first data representation matches the predictive model generated from the second data representation.
- the software module configured to compare the first data representation with the predictive model generated from the second data representation comprises a matching architecture configuration.
- any software module configured to compare a predictive model to a data representation comprises comparing a predictive model generated from a first data representation with a second data representation, comparing a predictive model generated from a second data representation with a first data representation, or any combination thereof.
- the configuration of matching architecture changes upon successive matching exercises, or changes upon certain successive matching exercise and not others, wherein changes may be randomized between matching exercises, or be based on non-randomized determinate data or protocols, or which may not change.
- changes to the configuration of matching architecture are based on changes being randomized between matching exercises, changes being based on non-randomized determinate data or protocols, or any combination thereof.
- the matching architecture configuration does not change.
- the system further comprises a guided user interface used to capture the first and/or second data representations. In some embodiments, information captured from the guided user interface is used in matching exercises.
- the system further comprises: (a) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the first distance; (b) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’ s face image or set of images is recorded by the camera at the second distance; (c) a software module configured to compare the motion and/or location data recorded at the first distance and motion and/or location data from the second distance with the predictive model generated from the first data representation, and the second data representation; and (d) a software module configured to validate an identity recognition match if (I) the second data representation matches the predictive model generated from the first data representation predictive modeling; and (II) the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- a software module configured to record, from one or more sensors on the mobile device, motion and/
- the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation.
- the system further comprising a software module configured to compare (I) the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation with (II) the predictive model generated from the first data representation, and the second data representation.
- system further comprising a software module configured to validate an identity recognition match if (I) the second data representation matches the predictive model generated from the first data representation; and (II) the motion and/or location data recorded continuously or at a plurality of intervals match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- the system further comprising: (a) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the first distance; (b) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the second distance; (c) a software module configured to compare the motion and/or location data recorded at the first distance and motion and/or location data from the second distance with the predictive model generated from the second data representation, and the first data representation; and (d) a software module configured to validate an identity recognition match if (I) the first data representation matches the predictive model generated from the second data representation predictive modeling; and (II) the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- a software module configured to record, from one or more sensors on the mobile device, motion and/
- the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation.
- the system further comprising a software module configured to compare the (I) the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation with (II) the predictive model generated from the second data representation, and the first data representation.
- system further comprising a software module configured to validate an identity recognition match if (I) the first data representation matches the predictive model generated from the second data representation; and (II) the motion and/or location data recorded continuously or at a plurality of intervals match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- the media further comprises a software module configured to reject the identity recognition match if, after the software module configured to compare the second data representation with the predictive model generated from the first data representation, the second data representation does not match the predictive model generated from the first data representation.
- the media further comprising a software module configured to capture one or more additional data representations from the user’s face image or set of images from the first or second distance, and compare the captured one or more additional data representations with the predictive model generated from the first data representation to determine if they match if, after comparing the second data representation with the predictive model generated from the first data representation, the second data representation does not match the predictive model generated from the first data representation.
- the media further comprising a software module configured to validate an additional data representation identity recognition match if the one or more additional data representations match the predictive model generated from the first data representation. In some embodiments, the media further comprising a software module configured to reject an additional data representation identity recognition match if the one or more additional data representations does not match the predictive model generated from the first data representation. In some embodiments, the media further comprising a software module configured to compare additional data to determine if the user presenting the face image or set of images at the first distance is a match to the user presenting the face image or set of images at the second distance if the second data representation does not match the predictive model generated from the first data representation.
- additional data comprises a name, password, identity number, address, geo location, device ID, unique data characteristic of the user’s software environment on the mobile device, other biometric data, predictive models of user data or biometric data, other data, or any combination thereof.
- the media further comprises: (a) a software module configured to process the second data representation into a predictive model that estimates the data representations of the user’s face image or set of images at other distances or orientations between the user’s face and the camera; (b) a software module configured to compare the first data representation with the predictive model generated from the second data representation to determine if they match; and (c) a software module configured to validate an identity recognition match if the first data representation matches the predictive model generated from the second data representation.
- the software module configured to compare the first data representation with the predictive model generated from the second data representation comprises a matching architecture configuration.
- a software module configured to compare a predictive model to a data representation comprises comparing a predictive model generated from a first data representation with a second data representation, comparing a predictive model generated from a second data representation with a first data representation, or any combination thereof.
- the configuration of matching architecture changes upon successive matching exercises, or changes upon certain successive matching exercise and not others, wherein changes may be randomized between matching exercises, or be based on non-randomized determinate data or protocols, or which may not change.
- changes to the configuration of matching architecture are based on changes being randomized between matching exercises, changes being based on non-randomized determinate data or protocols, or any combination thereof.
- the configuration of matching architecture does not change.
- the media further comprises a guided user interface used to capture the first and/or second data representations.
- information captured from the guided user interface is used in matching exercises.
- the media further comprising: (a) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the first distance; (b) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’ s face image or set of images is recorded by the camera at the second distance; (c) a software module configured to compare the motion and/or location data recorded at the first distance and motion and/or location data from the second distance with the predictive model generated from the first data representation, and the second data representation; and (d) a software module configured to validate an identity recognition match if (I) the second data representation matches the predictive model generated from the first data representation predictive modeling; and (II) the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- a software module configured to record, from one or more sensors on the mobile device, motion and
- the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation.
- the media further comprising a software module configured to compare (I) the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation with (II) the predictive model generated from the first data representation, and the second data representation.
- the media further comprising a software module configured to validate an identity recognition match if (I) the second data representation matches the predictive model generated from the first data representation; and (II) the motion and/or location data recorded continuously or at a plurality of intervals match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- the media further comprising: (a) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the first distance; (b) a software module configured to record, from one or more sensors on the mobile device, motion and/or location data at the time the user’s face image or set of images is recorded by the camera at the second distance; (c) a software module configured to compare the motion and/or location data recorded at the first distance and motion and/or location data from the second distance with the predictive model generated from the second data representation, and the first data representation; and (d) a software module configured to validate an identity recognition match if (I) the first data representation matches the predictive model generated from the second data representation predictive modeling; and (II) the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- a software module configured to record, from one or more sensors on the mobile device, motion and/
- the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation.
- the media further comprising a software module configured to compare the (I) the motion and/or location data is recorded continuously or at a plurality of intervals between the time of recording of the first data representation and the time of recording of the second data representation with (II) the predictive model generated from the second data representation, and the first data representation.
- the media further comprising a software module configured to validate an identity recognition match if (I) the first data representation matches the predictive model generated from the second data representation; and (II) the motion and/or location data recorded continuously or at a plurality of intervals match the expected motion and/or location data attributing to the position of the mobile device to the user’s face.
- a user’s face image or set of images is recorded by the camera at a first distance; the user’s face image or set of images at the first distance is processed with algorithms to generate a first data representation of the user’s image or set of images; the first data representation is processed into a predictive model that estimates the data representations of the user’s face image or set of images at other distances or orientations between the user’s face and the camera; the distance between the user’s face and the camera changes, whether moving closer together, or further away, and/or with changes in orientation of the face to the camera; the user’s face image or set of images is recorded by the camera at a second distance; the user’s face image or set of images at the second distance is processed with algorithms to generate a second data representation of the user’s second image or set of images; the second data representation is compared with the predictive model generated from the first data representation to determine if they
- a predictive model is generated from the second data representation, and the validating of the identity recognition match is determined from a comparison of the predictive model generated from the second data representation and the first data representation.
- Changing the data and software architecture for matching provides security benefits, where a dedicated attacker that comprises the system and gains access to some portion or all of the software and/or data, is less able to not able to understand how the matching is being performed, and is less able or not able to compromise the system and/or replay attack the user identity matching exercises on behalf of a user or users, which may be done as part of an effort to access information or resources, or cause changes to information or resources.
- the configuration of matching architecture be it from (a) a comparison of a predictive model generated from a first data representation with a second data representation, or (b) a comparison of a predictive model generated from a second data representation with a first data representation, changes upon successive matching exercises, or changes upon certain successive matching exercise and not others, which changes may be randomized between matching exercises, or be based on non-randomized determinate data or protocols, or which may not change.
- a guided user interface is used to capture the first and/or second data representations.
- the guidance information used in the matching exercise is used in some embodiments of the method.
- the configuration of the predictive modeling architecture can be from (a) predictive modeling from a given data or data representation to generate and match another data or data representation, or (b) predictive modeling from more than one captured data or data representations to generate multiple predictions to match multiple other data or data representations, with the first data, second data and/or additional data used for predictive modeling for matching other data or data representations.
- the configuration of the predictive modeling architecture can be optimized based on security level and speed of execution, for example in addressing data consumption, file sizes, processing steps, and other data and software architecture characteristics to change the speed of execution, which in some cases may be decided in connection with associated security specifications of the implementation.
- the predictive model to generate one or more predictive data is trained using methods including machine learning, image analysis, deep learning, and other methods.
- the predictive model is trained on both real data collected from users and/or synthetic data generated by image rendering or other techniques of image data representation known in the art.
- motion and/or location data is recorded from the sensors on the mobile device at the time the user’s face image or set of images is recorded by the camera at the first distance; motion and/or location data is recorded from the sensors on the mobile device at the time the user’s face image or set of images is recorded by the camera at the second distance; the motion and/or location data recorded at the first distance and the motion and/or location data recorded at the second distance are compared with the predictive model generated from the first data representation, and the second data representation; validating the identity recognition match if the second data representation matches the predictive model generated from the first data representation and if the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face, as said position may change with movement of the mobile device and/or the movement of the user’s face between the time of recording of the first data representation and the time of recording of the second data representation.
- motion and/or location data is recorded continuously, or at a plurality of intervals, between the time of recording of the first data representation and the time of recording of the second data representation; the motion and/or location data recorded continuously, or at a plurality of intervals, between the time of recording of the first data representation and the time of recording of the second data representation, is compared with the predictive model generated from the first data representation, and the second data representation; validating the identity recognition match if the second data representation matches the predictive model generated from the first data representation and if the motion and/or location data recorded continuously, or at a plurality of intervals, match the expected motion and/or location data attributing to the position of the mobile device to the user’s face, as said position may change with movement of the mobile device and/or the movement of the user’s face between the time of recording of the first data representation and the time of recording of the second data representation.
- motion and/or location data is recorded from the sensors on the mobile device at the time the user’s face image or set of images is recorded by the camera at the first distance; motion and/or location data is recorded from the sensors on the mobile device at the time the user’s face image or set of images is recorded by the camera at the second distance; the motion and/or location data recorded at the first distance and the motion and/or location data recorded at the second distance are compared with the predictive model generated from the second data representation, and the first data representation; validating the identity recognition match if the first data representation matches the predictive model generated from the second data representation and if the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face, as said position may change with movement of the mobile device and/or the movement of the user’s face between the time of recording of the first data representation and the time of recording of the second data representation.
- motion and/or location data is recorded continuously, or at a plurality of intervals, between the time of recording of the first data representation and the time of recording of the second data representation; the motion and/or location data recorded continuously, or at a plurality of intervals, between the time of recording of the first data representation and the time of recording of the second data representation, is compared with the predictive model generated from the second data representation, and the first data representation; validating the identity recognition match if the first data representation matches the predictive model generated from the second data representation and if the motion and/or location data recorded continuously, or at a plurality of intervals, match the expected motion and/or location data attributing to the position of the mobile device to the user’s face, as said position may change with movement of the mobile device and/or the movement of the user’s face between the time of recording of the first data representation and the time of recording of the second data representation.
- the term“about” refers to an amount that is near the stated amount by 10%, 5%, or 1%, including increments therein.
- the term“about” in reference to a percentage refers to an amount that is greater or less the stated percentage by 10%, 5%, or 1%, including increments therein.
- each of the expressions“at least one of A, B and C,”“at least one of A, B, or C,”“one or more of A, B, and C,” “one or more of A, B, or C” and“A, B, and/or C” means A alone, B alone, C alone, A and B together, A and C together, B and C together, or A, B and C together.
- FIG. 1 a block diagram is shown depicting an exemplary machine that includes a computer system 100 (e.g., a processing or computing system) within which a set of instructions can execute for causing a device to perform or execute any one or more of the aspects and/or methodologies for static code scheduling of the present disclosure.
- a computer system 100 e.g., a processing or computing system
- the components in Fig. 1 are examples only and do not limit the scope of use or functionality of any hardware, software, embedded logic component, or a combination of two or more such components implementing particular embodiments.
- Computer system 100 may include one or more processors 101, a memory 103, and a storage 108 that communicate with each other, and with other components, via a bus 140.
- the bus 140 may also link a display 132, one or more input devices 133 (which may, for example, include a keypad, a keyboard, a mouse, a stylus, etc.), one or more output devices 134, one or more storage devices 135, and various tangible storage media 136. All of these elements may interface directly or via one or more interfaces or adaptors to the bus 140.
- the various tangible storage media 136 can interface with the bus 140 via storage medium interface 126.
- Computer system 100 may have any suitable physical form, including but not limited to one or more integrated circuits (ICs), printed circuit boards (PCBs), mobile handheld devices (such as mobile telephones or PDAs), laptop or notebook computers, distributed computer systems, computing grids, or servers.
- ICs integrated circuits
- PCBs printed circuit boards
- mobile handheld devices such as mobile telephones or PDAs
- Computer system 100 includes one or more processor(s) 101 (e.g., central processing units (CPUs) or general purpose graphics processing units (GPGPUs)) that carry out functions.
- processor(s) 101 optionally contains a cache memory unit 102 for temporary local storage of instructions, data, or computer addresses.
- Processor(s) 101 are configured to assist in execution of computer readable instructions.
- Computer system 100 may provide functionality for the components depicted in Fig. 1 as a result of the processor(s) 101 executing non-transitory, processor-executable instructions embodied in one or more tangible computer-readable storage media, such as memory 103, storage 108, storage devices 135, and/or storage medium 136.
- the computer-readable media may store software that implements particular embodiments, and processor(s) 101 may execute the software.
- Memory 103 may read the software from one or more other computer-readable media (such as mass storage device(s) 135, 136) or from one or more other sources through a suitable interface, such as network interface 120.
- the software may cause processor(s) 101 to carry out one or more processes or one or more steps of one or more processes described or illustrated herein. Carrying out such processes or steps may include defining data structures stored in memory 103 and modifying the data structures as directed by the software.
- the memory 103 may include various components (e.g., machine readable media) including, but not limited to, a random access memory component (e.g., RAM 104) (e.g., static RAM (SRAM), dynamic RAM (DRAM), ferroelectric random access memory (FRAM), phase-change random access memory (PRAM), etc.), a read-only memory component (e.g., ROM 105), and any combinations thereof.
- ROM 105 may act to communicate data and instructions unidirectionally to processor(s) 101
- RAM 104 may act to communicate data and instructions bidirectionally with processor(s) 101.
- ROM 105 and RAM 104 may include any suitable tangible computer-readable media described below.
- a basic input/output system 106 (BIOS) including basic routines that help to transfer information between elements within computer system 100, such as during start-up, may be stored in the memory 103.
- Fixed storage 108 is connected bidirectionally to processor(s) 101, optionally through storage control unit 107.
- Fixed storage 108 provides additional data storage capacity and may also include any suitable tangible computer-readable media described herein.
- Storage 108 may be used to store operating system 109, executable(s) 110, data 111, applications 112 (application programs), and the like.
- Storage 108 can also include an optical disk drive, a solid-state memory device (e.g., flash-based systems), or a combination of any of the above.
- Information in storage 108 may, in appropriate cases, be incorporated as virtual memory in memory 103.
- storage device(s) 135 may be removably interfaced with computer system 100 (e.g., via an external port connector (not shown)) via a storage device interface 125.
- storage device(s) 135 and an associated machine-readable medium may provide non-volatile and/or volatile storage of machine-readable instructions, data structures, program modules, and/or other data for the computer system 100.
- software may reside, completely or partially, within a machine-readable medium on storage device(s) 135.
- software may reside, completely or partially, within processor(s) 101.
- Bus 140 connects a wide variety of subsystems.
- reference to a bus may encompass one or more digital signal lines serving a common function, where appropriate.
- Bus 140 may be any of several types of bus structures including, but not limited to, a memory bus, a memory controller, a peripheral bus, a local bus, and any combinations thereof, using any of a variety of bus architectures.
- bus architectures include an Industry Standard Architecture (ISA) bus, an Enhanced ISA (EISA) bus, a Micro Channel Architecture (MCA) bus, a Video Electronics Standards Association local bus (VLB), a Peripheral Component Interconnect (PCI) bus, a PCI-Express (PCI-X) bus, an Accelerated Graphics Port (AGP) bus, HyperTransport (HTX) bus, serial advanced technology attachment (SATA) bus, and any combinations thereof.
- ISA Industry Standard Architecture
- EISA Enhanced ISA
- MCA Micro Channel Architecture
- VLB Video Electronics Standards Association local bus
- PCI Peripheral Component Interconnect
- PCI-X PCI-Express
- AGP Accelerated Graphics Port
- HTX HyperTransport
- SATA serial advanced technology attachment
- Computer system 100 may also include an input device 133.
- a user of computer system 100 may enter commands and/or other information into computer system 100 via input device(s) 133.
- Examples of an input device(s) 133 include, but are not limited to, an alpha numeric input device (e.g., a keyboard), a pointing device (e.g., a mouse or touchpad), a touchpad, a touch screen, a multi -touch screen, a joystick, a stylus, a gamepad, an audio input device (e.g., a microphone, a voice response system, etc.), an optical scanner, a video or still image capture device (e.g., a camera), and any combinations thereof.
- an alpha numeric input device e.g., a keyboard
- a pointing device e.g., a mouse or touchpad
- a touchpad e.g., a touch screen
- a multi -touch screen e.g., a joystick,
- the input device is a Kinect, Leap Motion, or the like.
- Input device(s) 133 may be interfaced to bus 140 via any of a variety of input interfaces 123 (e.g., input interface 123) including, but not limited to, serial, parallel, game port, USB, FIREWIRE, THUNDERBOLT, or any combination of the above.
- computer system 100 when computer system 100 is connected to network 130, computer system 100 may communicate with other devices, specifically mobile devices and enterprise systems, distributed computing systems, cloud storage systems, cloud computing systems, and the like, connected to network 130. Communications to and from computer system 100 may be sent through network interface 120.
- network interface 120 may receive incoming communications (such as requests or responses from other devices) in the form of one or more packets (such as Internet Protocol (IP) packets) from network 130, and computer system 100 may store the incoming communications in memory 103 for processing.
- Computer system 100 may similarly store outgoing communications (such as requests or responses to other devices) in the form of one or more packets in memory 103 and communicated to network 130 from network interface 120.
- Processor(s) 101 may access these communication packets stored in memory 103 for processing.
- Examples of the network interface 120 include, but are not limited to, a network interface card, a modem, and any combination thereof.
- Examples of a network 130 or network segment 130 include, but are not limited to, a distributed computing system, a cloud computing system, a wide area network (WAN) (e.g., the Internet, an enterprise network), a local area network (LAN) (e.g., a network associated with an office, a building, a campus or other relatively small geographic space), a telephone network, a direct connection between two computing devices, a peer-to-peer network, and any combinations thereof.
- a network, such as network 130 may employ a wired and/or a wireless mode of communication. In general, any network topology may be used.
- Information and data can be displayed through a display 132.
- a display 132 include, but are not limited to, a cathode ray tube (CRT), a liquid crystal display (LCD), a thin film transistor liquid crystal display (TFT-LCD), an organic liquid crystal display (OLED) such as a passive-matrix OLED (PMOLED) or active-matrix OLED (AMOLED) display, a plasma display, and any combinations thereof.
- the display 132 can interface to the processor(s) 101, memory 103, and fixed storage 108, as well as other devices, such as input device(s) 133, via the bus 140.
- the display 132 is linked to the bus 140 via a video interface 122, and transport of data between the display 132 and the bus 140 can be controlled via the graphics control 121.
- the display is a video projector.
- the display is a head-mounted display (HMD) such as a VR headset.
- suitable VR headsets include, by way of non-limiting examples, HTC Vive, Oculus Rift, Samsung Gear VR, Microsoft HoloLens, Razer OSVR, FOVE VR, Zeiss VR One, Avegant Glyph, Freefly VR headset, and the like.
- the display is a combination of devices such as those disclosed herein.
- computer system 100 may include one or more other peripheral output devices 134 including, but not limited to, an audio speaker, a printer, a storage device, and any combinations thereof.
- peripheral output devices may be connected to the bus 140 via an output interface 124.
- Examples of an output interface 124 include, but are not limited to, a serial port, a parallel connection, a USB port, a FIREWIRE port, a THUNDERBOLT port, and any combinations thereof.
- computer system 100 may provide functionality as a result of logic hardwired or otherwise embodied in a circuit, which may operate in place of or together with software to execute one or more processes or one or more steps of one or more processes described or illustrated herein.
- Reference to software in this disclosure may encompass logic, and reference to logic may encompass software.
- reference to a computer-readable medium may encompass a circuit (such as an IC) storing software for execution, a circuit embodying logic for execution, or both, where appropriate.
- the present disclosure encompasses any suitable combination of hardware, software, or both.
- DSP digital signal processor
- ASIC application specific integrated circuit
- FPGA field programmable gate array
- a general purpose processor may be a microprocessor, but in the alternative, the processor may be any conventional processor, controller, microcontroller, or state machine.
- a processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.
- a software module may reside in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art.
- An exemplary storage medium is coupled to the processor such the processor can read information from, and write information to, the storage medium.
- the storage medium may be integral to the processor.
- the processor and the storage medium may reside in an ASIC.
- the ASIC may reside in a user terminal.
- the processor and the storage medium may reside as discrete components in a user terminal.
- suitable computing devices include, by way of non-limiting examples, server computers, desktop computers, laptop computers, notebook computers, sub-notebook computers, netbook computers, netpad computers, set-top computers, media streaming devices, handheld computers, Internet appliances, mobile smartphones, tablet computers, personal digital assistants, video game consoles, and vehicles.
- server computers desktop computers, laptop computers, notebook computers, sub-notebook computers, netbook computers, netpad computers, set-top computers, media streaming devices, handheld computers, Internet appliances, mobile smartphones, tablet computers, personal digital assistants, video game consoles, and vehicles.
- Suitable tablet computers include those with booklet, slate, and convertible configurations, known to those of skill in the art.
- the computing device includes an operating system configured to perform executable instructions.
- the operating system is, for example, software, including programs and data, which manages the device’s hardware and provides services for execution of applications.
- suitable server operating systems include, by way of non limiting examples, FreeBSD, OpenBSD, NetBSD ® , Linux, Apple ® Mac OS X Server ® , Oracle ® Solaris ® , Windows Server ® , and Novell ® NetWare ® .
- suitable personal computer operating systems include, by way of non-limiting examples, Microsoft ® Windows ® , Apple ® Mac OS X ® , UNIX ® , and UNIX-like operating systems such as GNU/Linux ® .
- the operating system is provided by cloud computing.
- suitable mobile smartphone operating systems include, by way of non limiting examples, Nokia ® Symbian ® OS, Apple ® iOS ® , Research In Motion ® BlackBerry OS ® , Google ® Android ® , Microsoft ® Windows Phone ® OS, Microsoft ® Windows Mobile ® OS, Linux ® , and Palm ® WebOS ® .
- suitable media streaming device operating systems include, by way of non-limiting examples, Apple TV ® , Roku ® , Boxee ® , Google TV ® , Google Chromecast ® , Amazon Fire ® , and Samsung ® HomeSync ® .
- suitable video game console operating systems include, by way of non-limiting examples, Sony ® PS3 ® , Sony ® PS4 ® , Microsoft ® Xbox 360 ® , Microsoft Xbox One, Nintendo ® Wii ® , Nintendo ® Wii U ® , and Ouya ® .
- Non-transitory computer readable storage medium
- the platforms, systems, media, and methods disclosed herein include one or more non-transitory computer readable storage media encoded with a program including instructions executable by the operating system of an optionally networked computing device.
- a computer readable storage medium is a tangible component of a computing device.
- a computer readable storage medium is optionally removable from a computing device.
- a computer readable storage medium includes, by way of non-limiting examples, CD-ROMs, DVDs, flash memory devices, solid state memory, magnetic disk drives, magnetic tape drives, optical disk drives, distributed computing systems including cloud computing systems and services, and the like.
- the program and instructions are permanently, substantially permanently, semi-permanently, or non-transitorily encoded on the media.
- the platforms, systems, media, and methods disclosed herein include at least one computer program, or use of the same.
- a computer program includes a sequence of instructions, executable by one or more processor(s) of the computing device’s CPU, written to perform a specified task.
- Computer readable instructions may be implemented as program modules, such as functions, objects, Application Programming Interfaces (APIs), computing data structures, and the like, that perform particular tasks or implement particular abstract data types.
- APIs Application Programming Interfaces
- a computer program comprises one sequence of instructions. In some embodiments, a computer program comprises a plurality of sequences of instructions. In some embodiments, a computer program is provided from one location. In other embodiments, a computer program is provided from a plurality of locations. In various embodiments, a computer program includes one or more software modules. In various embodiments, a computer program includes, in part or in whole, one or more web applications, one or more mobile applications, one or more standalone applications, one or more web browser plug-ins, extensions, add-ins, or add ons, or combinations thereof.
- a computer program includes a web application.
- a web application in various embodiments, utilizes one or more software frameworks and one or more database systems.
- a web application is created upon a software framework such as Microsoft ® .NET or Ruby on Rails (RoR).
- a web application utilizes one or more database systems including, by way of non -limiting examples, relational, non-relational, object oriented, associative, and XML database systems.
- suitable relational database systems include, by way of non-limiting examples, Microsoft ® SQL Server, mySQLTM, and Oracle ® .
- a web application in various embodiments, is written in one or more versions of one or more languages.
- a web application may be written in one or more markup languages, presentation definition languages, client-side scripting languages, server-side coding languages, database query languages, or combinations thereof.
- a web application is written to some extent in a markup language such as Hypertext Markup Language (HTML), Extensible Hypertext Markup Language (XHTML), or extensible Markup Language (XML).
- a web application is written to some extent in a presentation definition language such as Cascading Style Sheets (CSS).
- CSS Cascading Style Sheets
- a web application is written to some extent in a client-side scripting language such as Asynchronous JavaScript and XML (AJAX), Flash ® ActionScript, JavaScript, or Silverlight ® .
- AJAX Asynchronous JavaScript and XML
- Flash ® ActionScript JavaScript
- Silverlight ® a web application is written to some extent in a server-side coding language such as Active Server Pages (ASP), ColdFusion ® , Perl, JavaTM, JavaServer Pages (JSP), Hypertext Preprocessor (PHP), PythonTM, Ruby, Tel, Smalltalk, WebDNA ® , or Groovy.
- a web application is written to some extent in a database query language such as Structured Query Language (SQL).
- SQL Structured Query Language
- a web application integrates enterprise server products such as IBM ® Lotus Domino ® .
- a web application includes a media player element.
- a media player element utilizes one or more of many suitable multimedia technologies including, by way of non-limiting examples, Adobe ® Flash ® , HTML 5, Apple ® QuickTime ® , Microsoft ® Silverlight ® , JavaTM, and Unity ® .
- an application provision system comprises one or more databases 200 accessed by a relational database management system (RDBMS) 210.
- RDBMSs include Firebird, MySQL, PostgreSQL, SQLite, Oracle Database, Microsoft SQL Server, IBM DB2, IBM Informix, SAP Sybase, SAP Sybase, Teradata, and the like.
- the application provision system further comprises one or more application severs 220 (such as Java servers, .NET servers, PHP servers, and the like) and one or more web servers 230 (such as Apache, IIS, GWS and the like).
- the web server(s) optionally expose one or more web services via app application programming interfaces (APIs) 240.
- APIs app application programming interfaces
- an application provision system alternatively has a distributed, cloud-based architecture 300 and comprises elastically load balanced, auto-scaling web server resources 310 and application server resources 320 as well synchronously replicated databases 330.
- a computer program includes a mobile application provided to a mobile computing device.
- the mobile application is provided to a mobile computing device at the time it is manufactured.
- the mobile application is provided to a mobile computing device via the computer network described herein.
- a mobile application is created by techniques known to those of skill in the art using hardware, languages, and development environments known to the art. Those of skill in the art will recognize that mobile applications are written in several languages.
- Suitable programming languages include, by way of non-limiting examples, C, C++, C#, Objective-C, JavaTM, Javascript, Pascal, Object Pascal, PythonTM, Ruby, VB.NET, WML, and XHTML/HTML with or without CSS, or combinations thereof.
- Suitable mobile application development environments are available from several sources. Commercially available development environments include, by way of non-limiting examples, AirplaySDK, alcheMo, Appcelerator ® , Celsius, Bedrock, Flash Lite, .NET Compact Framework, Rhomobile, and WorkLight Mobile Platform. Other development environments are available without cost including, by way of non-limiting examples, Lazarus, MobiFlex, MoSync, and Phonegap. Also, mobile device manufacturers distribute software developer kits including, by way of non-limiting examples, iPhone and iPad (iOS) SDK, AndroidTM SDK, BlackBerry ® SDK, BREW SDK, Palm ® OS SDK, Symbian SDK, webOS SDK, and Windows ® Mobile SDK.
- iOS iPhone and iPad
- a computer program includes a standalone application, which is a program that is run as an independent computer process, not an add-on to an existing process, e.g., not a plug-in.
- standalone applications are often compiled.
- a compiler is a computer program(s) that transforms source code written in a programming language into binary object code such as assembly language or machine code. Suitable compiled programming languages include, by way of non-limiting examples, C, C++, Objective-C, COBOL, Delphi, Eiffel, JavaTM, Lisp, PythonTM, Visual Basic, and VB .NET, or combinations thereof. Compilation is often performed, at least in part, to create an executable program.
- a computer program includes one or more executable complied applications.
- the computer program includes a web browser plug-in (e.g., extension, etc.).
- a plug-in is one or more software components that add specific functionality to a larger software application. Makers of software applications support plug-ins to enable third-party developers to create abilities which extend an application, to support easily adding new features, and to reduce the size of an application. When supported, plug-ins enable customizing the functionality of a software application. For example, plug-ins are commonly used in web browsers to play video, generate interactivity, scan for viruses, and display particular file types.
- the toolbar comprises one or more web browser extensions, add-ins, or add-ons.
- the toolbar comprises one or more explorer bars, tool bands, or desk bands.
- plug-in frameworks are available that enable development of plug-ins in various programming languages, including, by way of non-limiting examples, C++, Delphi, JavaTM, PHP, PythonTM, and VB .NET, or combinations thereof.
- Web browsers are software applications, designed for use with network-connected computing devices, for retrieving, presenting, and traversing information resources on the World Wide Web. Suitable web browsers include, by way of non-limiting examples, Microsoft ® Internet Explorer ® , Mozilla ® Firefox ® , Google ® Chrome, Apple ® Safari ® , Opera Software ® Opera ® , and KDE Konqueror. In some embodiments, the web browser is a mobile web browser.
- Mobile web browsers are designed for use on mobile computing devices including, by way of non-limiting examples, handheld computers, tablet computers, netbook computers, subnotebook computers, smartphones, music players, personal digital assistants (PDAs), and handheld video game systems.
- Suitable mobile web browsers include, by way of non-limiting examples, Google ® Android ® browser, RIM BlackBerry ® Browser, Apple ® Safari ® , Palm ® Blazer, Palm ® WebOS ® Browser, Mozilla ® Firefox ® for mobile, Microsoft ® Internet Explorer ® Mobile, Amazon ® Kindle ® Basic Web, Nokia ® Browser, Opera Software ® Opera ® Mobile, and Sony ® PSPTM browser.
- the platforms, systems, media, and methods disclosed herein include software, server, and/or database modules, or use of the same.
- software modules are created by techniques known to those of skill in the art using machines, software, and languages known to the art.
- the software modules disclosed herein are implemented in a multitude of ways.
- a software module comprises a file, a section of code, a programming object, a programming structure, or combinations thereof.
- a software module comprises a plurality of files, a plurality of sections of code, a plurality of programming objects, a plurality of programming structures, or combinations thereof.
- the one or more software modules comprise, by way of non-limiting examples, a web application, a mobile application, and a standalone application.
- software modules are in one computer program or application.
- software modules are in more than one computer program or application. In some embodiments, software modules are hosted on one machine. In other embodiments, software modules are hosted on more than one machine. In further embodiments, software modules are hosted on a distributed computing platform such as a cloud computing platform. In some embodiments, software modules are hosted on one or more machines in one location. In other embodiments, software modules are hosted on one or more machines in more than one location.
- the platforms, systems, media, and methods disclosed herein include one or more databases, or use of the same.
- suitable databases include, by way of non-limiting examples, relational databases, non-relational databases, object oriented databases, object databases, entity-relationship model databases, associative databases, and XML databases. Further non-limiting examples include SQL, PostgreSQL, MySQL, Oracle, DB2, and Sybase.
- a database is internet- based.
- a database is web-based.
- a database is cloud computing-based.
- a database is a distributed database.
- a database is based on one or more local computer storage devices.
- Novel and unique data and software architectures for detecting spoofing attempts provides may security benefits, such as, for instance, to deter a dedicated attacker.
- Such systems and methods herein include specific identification matching and confirmation algorithms that cannot be directly accessed and/or tampered by even any such attacker who has obtained access to at least a portion of a security software and/or data.
- the method comprises: recording, by the camera of the mobile device, a first image of a face of a user; generating a first data representation 401A of the face of the user from the first image; forming 403 a predictive model 402A from the first data representation 401A; recording, by the camera of the mobile device, a second image of the face of the user; generating a second data representation 401B of the face of the user from the second image; and determining 404 if the second data representation 401B matches the predictive model 402A.
- the first image of the face of the user may be captured when the camera is a first distance from the user.
- the second image of the face of the user may be captured when the camera is a second distance from the user.
- the first distance may be greater than the second distance.
- the second distance can be greater than the first distance.
- the first distance may be greater than the second distance by 5%, 10%, 15%, 20%, 25%, 30%, 35%, 40%, 45%, 50%, 60%, 70%, 80%, 90%, 100%, or more, including increments therein.
- the second distance may be greater than the first distance by 5%, 10%, 15%, 20%, 25%, 30%, 35%, 40%, 45%, 50%, 60%, 70%, 80%, 90%, 100%, or more, including increments therein.
- At least one of the first distance and the second distance may be measured as a minimal distance to the face of the user, a normal distance to the face of the user, a minimal distance to the nose of the user, or a normal distance to the nose of the user.
- the first image of the face of the user may be captured when the camera is at a first orientation relative to the user.
- the second image of the face of the user may be captured when the camera is at a second orientation relative to the user.
- the first orientation may comprise a first pitch angle, a first yaw angle, a first roll angle, or any combination thereof, with respect to the user.
- the second orientation may comprise a second pitch angle, a second yaw angle, a second roll angle, or any combination thereof, with respect to the user.
- the first pitch angle, the second pitch angle, or both may be measured about an axis parallel to the ground and parallel to a forward direction of the user.
- the first yaw angle, the second yaw angle, or both may be measured about an axis perpendicular to the ground.
- the first roll angle, the second roll angle, or both may be measured about an axis parallel to the ground and perpendicular to a forward direction of the user.
- At least one of the first pitch angle, the first yaw angle, and the first roll angle may be greater than one or more of the second pitch angle, the second yaw angle, and the second roll angle.
- At least one of the first pitch angle, the first yaw angle, and the first roll angle may be greater than one or more of the second pitch angle, the second yaw angle, and the second roll angle by 5%, 10%, 15%, 20%, 25%, 30%, 35%, 40%, 45%, 50%, 60%, 70%, 80%, 90%, 100%, or more, including increments therein. At least one of the second pitch angle, the second yaw angle, and the second roll angle may be greater than one or more of the first pitch angle, the first yaw angle, and the first roll angle.
- At least one of the second pitch angle, the second yaw angle, and the second roll angle may be greater than one or more of the first pitch angle, the first yaw angle, and the first roll angle by 5%, 10%, 15%, 20%, 25%, 30%, 35%, 40%, 45%, 50%, 60%, 70%, 80%, 90%, 100%, or more, including increments therein.
- Determining 404 if the second data representation 401B matches the predictive model 402A can be performed by comparing the second data representation 401B to the predictive model 402A. If the second data representation 401B does not match the predictive model 402A an identity match of the user can be rejected. An identity match that is rejected can indicate a spoofing attempt. If the second data representation 401B does match the predictive model 402A an identity match of the user can be confirmed. An identity match that is confirm can indicate a lack of a spoofing attempt.
- the method may further comprise capturing a third image of the face of the user, generating a third data representation 401C of the face of the user from the third image, and determining 404 if the third data representation 401C matches the predictive model 402A. Additionally or alternatively, if the second data representation 401B does not match the predictive model 402A the method may further comprise capturing a third image of the face of the user, generating a third data representation 401C of the face of the user from the third image, and determining 404 if the third data representation 401C matches the predictive model 402A.
- the method may further comprise accepting an input provided by the user matches an additional data associated with the user.
- the method may further comprise accepting an input provided by the user matches an additional data associated with the user.
- the method may further comprise capturing a motion data of the user, a location data of the user, or both. In such cases, the method may further comprise accepting the user if the motion data of the user is within a predetermined range from a set motion, accepting the user if the location data of the user is within a predetermined range from a set location, or both. In such cases, the method may further comprise rejecting the user if the motion data of the user is not within a predetermined range from a set motion, rejecting the user if the location data of the user is not within a predetermined range from a set location, or both. [0078] The user may capture the first image, the second image, the third image, or any combination thereof via a guided user interface on the mobile device.
- Validating the identity recognition match may be performed if the second data representation 401B matches the predictive model 402A generated from the first data representation 401A and/or if the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device to the user’s face. Said position can change with movement of the mobile device and/or the movement of the user’s face between the time of recording of the first data representation 401A and the time of recording of the second data representation 401B.
- the method may further comprise accepting an additional data provided by the user, and determining 404 if the additional data matches an additional data associated with the user. Additionally, or alternatively, if the third data representation 401C does not match the predictive model 402A, the method may further comprise accepting an additional data provided by the user, and determining 404 if the additional data matches an additional data associated with the user.
- validating of the identity recognition match may occur if the first data representation 401A matches the predictive model 402A generated from the second data representation 401B, and if the motion and/or location data match the expected motion and/or location data attributing to the position of the mobile device. Said position may change with movement of the mobile device and/or the movement of the user’s face between the time of recording of the first data representation 401A and the time of recording of the second data representation 401B.
- the methods and systems herein record, by the camera of the mobile device, a first image of a face of a user, and generate a first data representation of the face of the user from the first image.
- the first data representation may comprise: a three dimensional map of at least a portion of the face of the user, a distance between a first feature and a second feature of the face of the user, a location of a pluralities of facial features, a calculated volume of at least a portion of the face of the user, a profile curve of a portion of the face of the user, a slope map of the surface of at least a portion of the face of the user, or any combination thereof.
- the methods and systems herein record, by the camera of the mobile device, a second image of a face of a user, and generate a second data representation of the face of the user from the second image.
- the second data representation may comprise: a three dimensional map of at least a portion of the face of the user, a distance between a second feature and a second feature of the face of the user, a location of a pluralities of facial features, a calculated volume of at least a portion of the face of the user, a profile curve of a portion of the face of the user, a slope map of the surface of at least a portion of the face of the user, or any combination thereof.
- the motion data, the location data, or both can be recorded from a sensor on the mobile device at the time of the recording of the first image, at the time of the recording of the first image, or both.
- the motion the location data or both may be recorded at a single point in time.
- the motion the location data or both may be recorded continuously.
- a sequence of the motion the location data or both may be recorded.
- a repetitive sequence of the motion the location data or both may be recorded at a set repetition interval sequence of the motion the location data or both may begin at the time of the recording of the first image, at the time of the recording of the first image, or both.
- the set repetition interval may be 0.1 seconds, 0.5 seconds, 1 second, 2 seconds, 3 seconds, 4 seconds, 5 seconds, or more, including increments therein.
- the predictive model is generated from the first data representation. In some cases, the predictive model is generated from the second data representation. In some cases, the predictive model is generated from the first data representation and the second data representation.
- the predictive model may be generated by a machine learning algorithm, an image analysis algorithm, a deep learning algorithm, or any combination thereof. At least one of the machine learning algorithm, the image analysis algorithm, or the deep learning algorithm, may be trained using a plurality of user images. At least one of the machine learning algorithm, the image analysis algorithm, or the deep learning algorithm, may be trained using a plurality of user images and a plurality of spoofed user images. The user images may comprise public user images.
- the predictive model can be trained on both real data collected from users and/or synthetic data generated by image rendering or other techniques of image data representation known in the art. [0090]
- the predictive model may be formed from the first data representation and at least one of the location data and the motion data.
- Motion and/or location data is recorded continuously, or at a plurality of intervals, between the time of recording of the first data representation and the time of recording of the second data representation.
- the configuration of the predictive modeling architecture may comprise: (a) a predictive modeling architecture derived from a given data set or data representation set to generate and match another data or data representation; or (b) a predictive modeling architecture derived from more than one captured data or data representations to generate multiple predictions that match multiple other data or data representations.
- a plurality of data sets can be used for predictive modeling to match matching data or data representations.
- the configuration of the predictive modeling architecture may comprise: (a) a comparison between a predictive model generated from a first data representation and a predictive model generated from a second data representation; or (b) a comparison of a predictive model generated from a second data representation with a predictive model generated from a first data representation.
- changes between successive matching exercises, or changes between certain successive matching exercise and not others may be randomized between matching exercises
- changes between successive matching exercises, or changes between certain successive matching exercise and not others may be based on non-randomized determinate data or protocols, or which may not change.
- the configuration of the predictive modeling architecture can be optimized based on security level and speed of execution.
- the predictive modeling architecture can be optimized based on data consumption, file sizes, processing steps, or any combination thereof. Such architecture characteristics can change the speed of execution depending on associated security specifications.
- the method may further comprise comparing additional data to a data submitted by the user to the mobile device.
- the additional data may comprise a name, a password, an identity number, an address, a geo-location, a device ID, unique data characteristic of the user’s software environment on the mobile device, a biometric data, a predictive biometric data, or any combination thereof.
- the method may further comprise accepting an input provided by the user matches an additional data associated with the user.
- the third data representation does not match the predictive model the method may further comprise accepting an input provided by the user matches an additional data associated with the user.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Computer Security & Cryptography (AREA)
- Human Computer Interaction (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Computer Networks & Wireless Communication (AREA)
- Computer Hardware Design (AREA)
- Signal Processing (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- Medical Informatics (AREA)
- Biomedical Technology (AREA)
- Databases & Information Systems (AREA)
- Data Mining & Analysis (AREA)
- Mathematical Physics (AREA)
- Collating Specific Patterns (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
Description
Claims
Priority Applications (8)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AU2020237108A AU2020237108B2 (en) | 2019-03-12 | 2020-03-11 | Detecting spoofing of facial recognition with mobile devices |
CN202080035642.9A CN113826110B (en) | 2019-03-12 | 2020-03-11 | Method, system and medium for detecting spoofing of biometric identification with a mobile device |
KR1020217032304A KR20220004628A (en) | 2019-03-12 | 2020-03-11 | Detection of facial recognition spoofing using mobile devices |
SG11202109983R SG11202109983RA (en) | 2019-03-12 | 2020-03-11 | Detecting spoofing of facial recognition with mobile devices |
EP20770199.6A EP3938953A4 (en) | 2019-03-12 | 2020-03-11 | Detecting spoofing of facial recognition with mobile devices |
JP2021555104A JP7464619B2 (en) | 2019-03-12 | 2020-03-11 | Detecting spoofing using facial recognition on mobile devices |
BR112021018149-0A BR112021018149B1 (en) | 2019-03-12 | 2020-03-11 | COMPUTER-IMPLEMENTED METHOD FOR DETECTING BIOMETRIC IDENTITY RECOGNITION FORGERY USING THE CAMERA OF A MOBILE DEVICE, COMPUTER-IMPLEMENTED SYSTEM, AND NON-TRAINER COMPUTER-READABLE STORAGE MEDIUM |
CA3133229A CA3133229C (en) | 2019-03-12 | 2020-03-11 | Detecting spoofing of facial recognition with mobile devices |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201962817554P | 2019-03-12 | 2019-03-12 | |
US62/817,554 | 2019-03-12 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020185948A1 true WO2020185948A1 (en) | 2020-09-17 |
Family
ID=72424323
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2020/022168 WO2020185948A1 (en) | 2019-03-12 | 2020-03-11 | Detecting spoofing of facial recognition with mobile devices |
Country Status (11)
Country | Link |
---|---|
US (1) | US11343277B2 (en) |
EP (1) | EP3938953A4 (en) |
JP (1) | JP7464619B2 (en) |
KR (1) | KR20220004628A (en) |
CN (1) | CN113826110B (en) |
AU (1) | AU2020237108B2 (en) |
BR (1) | BR112021018149B1 (en) |
CA (1) | CA3133229C (en) |
SG (1) | SG11202109983RA (en) |
TW (1) | TWI781383B (en) |
WO (1) | WO2020185948A1 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11343277B2 (en) | 2019-03-12 | 2022-05-24 | Element Inc. | Methods and systems for detecting spoofing of facial recognition in connection with mobile devices |
US11425562B2 (en) | 2017-09-18 | 2022-08-23 | Element Inc. | Methods, systems, and media for detecting spoofing in mobile authentication |
US11507248B2 (en) | 2019-12-16 | 2022-11-22 | Element Inc. | Methods, systems, and media for anti-spoofing using eye-tracking |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2023511242A (en) * | 2020-12-31 | 2023-03-17 | 商▲湯▼国▲際▼私人有限公司 | METHOD, APPARATUS, DEVICE AND RECORDING MEDIUM FOR RELATED OBJECT DETECTION IN IMAGE |
US11922731B1 (en) | 2021-06-30 | 2024-03-05 | Jumio Corporation | Liveness detection |
FR3127610A1 (en) * | 2021-09-27 | 2023-03-31 | Idemia Identity & Security France | Method for imaging a body part of a user, anti-replay method of a displacement of a body part and associated computer program product |
ES2957496A1 (en) * | 2022-06-09 | 2024-01-19 | Fenomatch S L | PROCEDURE FOR SELECTION AMONG A GROUP OF DONORS THROUGH FACIAL MATCHING, THROUGH A 2D FACIAL IMAGE OF THE DONOR SUBJECT (Machine-translation by Google Translate, not legally binding) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120288166A1 (en) * | 2011-05-13 | 2012-11-15 | Microsoft Corporation | Association and prediction in facial recognition |
US20170124385A1 (en) * | 2007-12-31 | 2017-05-04 | Applied Recognition Inc. | Face authentication to mitigate spoofing |
US20170304732A1 (en) * | 2014-11-10 | 2017-10-26 | Lego A/S | System and method for toy recognition |
US20180181737A1 (en) * | 2014-08-28 | 2018-06-28 | Facetec, Inc. | Facial Recognition Authentication System Including Path Parameters |
Family Cites Families (181)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3733861A (en) | 1972-01-19 | 1973-05-22 | Recognition Devices | Electronic recognition door lock |
US4371951A (en) | 1980-09-29 | 1983-02-01 | Control Data Corporation | Apparatus for converting serial input sparse vector format to parallel unpacked format for input to tandem arithmetic logic units |
US5067164A (en) | 1989-11-30 | 1991-11-19 | At&T Bell Laboratories | Hierarchical constrained automatic learning neural network for character recognition |
US5058179A (en) | 1990-01-31 | 1991-10-15 | At&T Bell Laboratories | Hierarchical constrained automatic learning network for character recognition |
US5450523A (en) | 1990-11-15 | 1995-09-12 | Matsushita Electric Industrial Co., Ltd. | Training module for estimating mixture Gaussian densities for speech unit models in speech recognition systems |
US5774059A (en) | 1995-07-20 | 1998-06-30 | Vindicator Corporation | Programmable electronic lock |
CA2156236C (en) | 1995-08-16 | 1999-07-20 | Stephen J. Borza | Biometrically secured control system for preventing the unauthorized use of a vehicle |
JP2815045B2 (en) | 1996-12-16 | 1998-10-27 | 日本電気株式会社 | Image feature extraction device, image feature analysis device, and image matching system |
US6152563A (en) | 1998-02-20 | 2000-11-28 | Hutchinson; Thomas E. | Eye gaze direction tracker |
US6178255B1 (en) | 1998-04-28 | 2001-01-23 | Cross Match Technologies, Inc. | Individualized fingerprint scanner |
US6274745B1 (en) | 1999-03-24 | 2001-08-14 | Tosoh Corporation | Binaphthol monophosphoric acid derivative and its use |
JP4727065B2 (en) | 2000-05-11 | 2011-07-20 | 株式会社半導体エネルギー研究所 | Authentication apparatus and communication system |
ES2241598T3 (en) | 2000-05-16 | 2005-11-01 | Swisscom Mobile Ag | BIOMETRIC PROCEDURE OF IDENTIFICATION AND AUTHENTICATION. |
US6956608B1 (en) | 2000-08-11 | 2005-10-18 | Identix Incorporated | Fingerprint imaging device including an optical plate having microreflectors |
US6819219B1 (en) | 2000-10-13 | 2004-11-16 | International Business Machines Corporation | Method for biometric-based authentication in wireless communication for access control |
JP2002259345A (en) | 2001-02-27 | 2002-09-13 | Nec Corp | Method/device for authentication for preventing unauthorized use of physical feature data, and program |
US6633090B2 (en) | 2001-09-07 | 2003-10-14 | Delphi Technologies, Inc. | Starting system for an automotive vehicle using fingerprint recognition |
JP2003148017A (en) | 2001-11-08 | 2003-05-21 | Sharp Corp | Lock device, lock control system and method for controlling lock |
JP2003178306A (en) | 2001-12-12 | 2003-06-27 | Toshiba Corp | Personal identification device and personal identification method |
US6873714B2 (en) | 2002-02-19 | 2005-03-29 | Delphi Technologies, Inc. | Auto calibration and personalization of eye tracking system using larger field of view imager with higher resolution |
US8590013B2 (en) | 2002-02-25 | 2013-11-19 | C. S. Lee Crawford | Method of managing and communicating data pertaining to software applications for processor-based devices comprising wireless communication circuitry |
US7616784B2 (en) | 2002-07-29 | 2009-11-10 | Robert William Kocher | Method and apparatus for contactless hand recognition |
GB2396001B (en) | 2002-10-09 | 2005-10-26 | Canon Kk | Gaze tracking system |
US7321671B2 (en) | 2002-12-06 | 2008-01-22 | Cross Match Technologies, Inc. | System and method for generating a preview display in a print capturing system using a non-planar prism |
US20060133651A1 (en) | 2002-12-31 | 2006-06-22 | Polcha Andrew J | Recoverable biometric identity system and method |
EP1594085A4 (en) | 2003-02-13 | 2008-02-27 | Nec Corp | Unauthorized person detection device and unauthorized person detection method |
DE10315923A1 (en) | 2003-04-08 | 2004-10-28 | Tbs Holding Ag | Procedure to detect data of uneven surfaces for biometric data, using non-contact optical sensing of surface |
US6923370B2 (en) | 2003-05-20 | 2005-08-02 | Bradley L. Gotfried | Access system |
US6992562B2 (en) | 2003-06-10 | 2006-01-31 | Visteon Global Technologies, Inc. | Biometric keyless entry system |
US8948468B2 (en) * | 2003-06-26 | 2015-02-03 | Fotonation Limited | Modification of viewing parameters for digital images using face detection information |
US7317815B2 (en) * | 2003-06-26 | 2008-01-08 | Fotonation Vision Limited | Digital image processing composition using face detection information |
US8181017B2 (en) | 2004-10-22 | 2012-05-15 | Nds Limited | Certificate renewal |
US20060120568A1 (en) | 2004-12-06 | 2006-06-08 | Mcconville Patrick J | System and method for tracking individuals |
WO2006090322A1 (en) | 2005-02-22 | 2006-08-31 | Koninklijke Philips Electronics, N.V. | System and method for transferring media rights under predetermined conditions |
JP4696610B2 (en) | 2005-03-15 | 2011-06-08 | オムロン株式会社 | Subject authentication device, face authentication device, mobile phone, and subject authentication method |
US8151358B1 (en) | 2005-04-07 | 2012-04-03 | Aol Inc. | Annotation of digital items in a shared list |
US20060294393A1 (en) | 2005-06-24 | 2006-12-28 | Mc Call Clark E | Remote biometric registration for vehicles |
WO2007000504A1 (en) | 2005-06-27 | 2007-01-04 | France Telecom | Biometric hand recognition method and associated system and device |
US8026840B2 (en) | 2005-10-28 | 2011-09-27 | Raytheon Company | Biometric radar system and method for identifying persons and positional states of persons |
JP4793179B2 (en) * | 2005-11-14 | 2011-10-12 | オムロン株式会社 | Authentication device and portable terminal |
US8290220B2 (en) | 2006-03-01 | 2012-10-16 | Nec Corporation | Face authenticating apparatus, face authenticating method, and program |
DE102006018956A1 (en) | 2006-04-24 | 2007-10-25 | Robert Bosch Gmbh | Particle`s mass or mass flow determining method for internal-combustion engine, involves arranging sensor in exhaust tract of engine, and comparing measured signal change of sensor with predicted signal change of sensor |
US7983451B2 (en) | 2006-06-30 | 2011-07-19 | Motorola Mobility, Inc. | Recognition method using hand biometrics with anti-counterfeiting |
JP4807167B2 (en) | 2006-07-06 | 2011-11-02 | オムロン株式会社 | Impersonation detection device |
US7660442B2 (en) | 2006-09-01 | 2010-02-09 | Handshot, Llc | Method and system for capturing fingerprints, palm prints and hand geometry |
US7986816B1 (en) | 2006-09-27 | 2011-07-26 | University Of Alaska | Methods and systems for multiple factor authentication using gaze tracking and iris scanning |
KR101309594B1 (en) | 2006-09-29 | 2013-11-21 | 스캠멜, 댄 | A system and method for verifying a user's identity in electronic transactions |
JP2008242631A (en) | 2007-03-26 | 2008-10-09 | Oki Electric Ind Co Ltd | Iris registration apparatus and iris authentication apparatus |
US20080284726A1 (en) | 2007-05-17 | 2008-11-20 | Marc Boillot | System and Method for Sensory Based Media Control |
US8126788B2 (en) | 2007-05-29 | 2012-02-28 | Exaktime Innovations, Inc. | Method for tracking time attendance of either a dedicated user or multiple non-dedicated users, interchangeably, using a single multi-function electronic hand-held device |
EP2009575B1 (en) | 2007-06-27 | 2013-06-12 | Research In Motion Limited | System and method for improving smart card reader reconnections |
JP4693818B2 (en) | 2007-07-09 | 2011-06-01 | 株式会社エヌ・ティ・ティ・ドコモ | Authentication system and authentication method |
GB0714344D0 (en) | 2007-07-24 | 2007-09-05 | Univ Wales Swansea | Biometric attendance verification |
WO2009082814A1 (en) * | 2007-12-31 | 2009-07-09 | Ray Ganong | Method, system, and computer program for identification and sharing of digital images with face signatures |
JP5186929B2 (en) | 2008-01-21 | 2013-04-24 | 日本電気株式会社 | Authentication imaging device |
EP2263190A2 (en) * | 2008-02-13 | 2010-12-22 | Ubisoft Entertainment S.A. | Live-action image capture |
US9286742B2 (en) | 2008-03-31 | 2016-03-15 | Plantronics, Inc. | User authentication system and method |
US8358856B2 (en) | 2008-06-02 | 2013-01-22 | Eastman Kodak Company | Semantic event detection for digital content records |
US20100042940A1 (en) | 2008-08-14 | 2010-02-18 | Caterpillar Inc. | Geofence system with integrated user interface |
US8175379B2 (en) | 2008-08-22 | 2012-05-08 | Adobe Systems Incorporated | Automatic video image segmentation |
CN102197412B (en) | 2008-10-28 | 2014-01-08 | 日本电气株式会社 | Spoofing detection system, spoofing detection method and spoofing detection program |
US8345932B2 (en) | 2008-11-24 | 2013-01-01 | International Business Machines Corporation | Support vector machine for biometric data processing |
JP5098973B2 (en) | 2008-11-27 | 2012-12-12 | 富士通株式会社 | Biometric authentication device, biometric authentication method, and biometric authentication program |
JP4636171B2 (en) | 2008-12-17 | 2011-02-23 | トヨタ自動車株式会社 | Biometric authentication system for vehicles |
JP2010146502A (en) | 2008-12-22 | 2010-07-01 | Toshiba Corp | Authentication processor and authentication processing method |
EP2386097A4 (en) | 2009-01-06 | 2013-02-20 | Xped Holdings Pty Ltd | An arrangement for managing mobile device access to precinct regions containing services and products and information |
US20100191551A1 (en) | 2009-01-26 | 2010-07-29 | Apple Inc. | Systems and methods for accessing hotel services using a portable electronic device |
JP5436876B2 (en) | 2009-02-02 | 2014-03-05 | 株式会社ディスコ | Grinding method |
US20100246902A1 (en) | 2009-02-26 | 2010-09-30 | Lumidigm, Inc. | Method and apparatus to combine biometric sensing and other functionality |
US8194938B2 (en) | 2009-06-02 | 2012-06-05 | George Mason Intellectual Properties, Inc. | Face authentication using recognition-by-parts, boosting, and transduction |
US8472681B2 (en) | 2009-06-15 | 2013-06-25 | Honeywell International Inc. | Iris and ocular recognition system using trace transforms |
US8638939B1 (en) | 2009-08-20 | 2014-01-28 | Apple Inc. | User authentication on an electronic device |
US8447119B2 (en) | 2010-03-16 | 2013-05-21 | Nec Laboratories America, Inc. | Method and system for image classification |
US8326001B2 (en) | 2010-06-29 | 2012-12-04 | Apple Inc. | Low threshold face recognition |
WO2012020591A1 (en) | 2010-08-09 | 2012-02-16 | 日本電気株式会社 | System for identifying individuals, feature value specification device, feature specification method, and recording medium |
US8670935B2 (en) | 2010-08-17 | 2014-03-11 | Blackberry Limited | Tagging a location by pairing devices |
JP5565285B2 (en) | 2010-11-19 | 2014-08-06 | コニカミノルタ株式会社 | Manufacturing method of glass optical element |
US20120137137A1 (en) | 2010-11-30 | 2012-05-31 | Brickell Ernest F | Method and apparatus for key provisioning of hardware devices |
CN102013011B (en) * | 2010-12-16 | 2013-09-04 | 重庆大学 | Front-face-compensation-operator-based multi-pose human face recognition method |
CN102034097B (en) * | 2010-12-21 | 2012-07-04 | 中国科学院半导体研究所 | Method for recognizing human face by comprehensively utilizing front and lateral images |
KR101816170B1 (en) | 2010-12-22 | 2018-01-09 | 한국전자통신연구원 | Apparatus and method for obtaining 3D depth information |
US8457370B2 (en) | 2011-01-20 | 2013-06-04 | Daon Holdings Limited | Methods and systems for authenticating users with captured palm biometric data |
US8675543B2 (en) | 2011-02-24 | 2014-03-18 | Verizon Patent And Licensing Inc. | Route limiting in border gateway protocol over satellite networks |
CN102103689B (en) * | 2011-03-07 | 2012-07-18 | 北京大学 | Frontal face image synthesis-based face recognition method |
WO2012135861A1 (en) | 2011-04-01 | 2012-10-04 | Tony Lam | Battery powered passive keyless entry system for premise entry |
CN103477352A (en) | 2011-04-11 | 2013-12-25 | 英特尔公司 | Gesture recognition using depth images |
US9082235B2 (en) | 2011-07-12 | 2015-07-14 | Microsoft Technology Licensing, Llc | Using facial data for device authentication or subject identification |
US8548207B2 (en) | 2011-08-15 | 2013-10-01 | Daon Holdings Limited | Method of host-directed illumination and system for conducting host-directed illumination |
WO2013035150A1 (en) | 2011-09-05 | 2013-03-14 | 株式会社モルフォ | Facial authentication system, facial authentication method, and facial authentication program |
CN102426715A (en) | 2011-09-30 | 2012-04-25 | 华为技术有限公司 | Unlocking method for electronic door lock, electronic door lock and electronic door lock system |
US8947202B2 (en) | 2011-10-20 | 2015-02-03 | Apple Inc. | Accessing a vehicle using portable devices |
US9111402B1 (en) | 2011-10-31 | 2015-08-18 | Replicon, Inc. | Systems and methods for capturing employee time for time and attendance management |
US8235529B1 (en) | 2011-11-30 | 2012-08-07 | Google Inc. | Unlocking a screen using eye tracking information |
CN104054091B (en) * | 2011-12-09 | 2018-01-26 | 视德公司 | The facial marking for image recognition generates |
TW201327413A (en) | 2011-12-21 | 2013-07-01 | Xid Technologies Pte Ltd | Systems and methods for face authentication or recognition using spectrally and/or temporally filtered flash illumination |
CN104106080B (en) | 2011-12-27 | 2017-10-03 | 英特尔公司 | User authentication and user checking system, apparatus and method on the scene based on turing test |
KR101242390B1 (en) | 2011-12-29 | 2013-03-12 | 인텔 코오퍼레이션 | Method, apparatus and computer-readable recording medium for identifying user |
US9519769B2 (en) | 2012-01-09 | 2016-12-13 | Sensible Vision, Inc. | System and method for disabling secure access to an electronic device using detection of a predetermined device orientation |
JP5866216B2 (en) | 2012-01-31 | 2016-02-17 | 株式会社東海理化電機製作所 | Electronic key registration system |
US8705070B2 (en) | 2012-02-24 | 2014-04-22 | Canon Kabushiki Kaisha | Systems and methods for managing use of an imaging device |
US9323912B2 (en) | 2012-02-28 | 2016-04-26 | Verizon Patent And Licensing Inc. | Method and system for multi-factor biometric authentication |
US20130268418A1 (en) | 2012-04-04 | 2013-10-10 | Accu-Time Systems, Inc. | Methods and apparatus for wireless communication of time and attendance information |
HK1160574A2 (en) | 2012-04-13 | 2012-07-13 | King Hei Francis Kwong | Secure electronic payment system and process |
US20130286161A1 (en) | 2012-04-25 | 2013-10-31 | Futurewei Technologies, Inc. | Three-dimensional face recognition for mobile devices |
US9070162B2 (en) | 2012-04-25 | 2015-06-30 | ZR Investments, LLC | Time tracking device and method |
US9047376B2 (en) | 2012-05-01 | 2015-06-02 | Hulu, LLC | Augmenting video with facial recognition |
JP5780361B2 (en) | 2012-05-29 | 2015-09-16 | 株式会社村田製作所 | Electronic key system and electronic equipment |
US20130342672A1 (en) | 2012-06-25 | 2013-12-26 | Amazon Technologies, Inc. | Using gaze determination with device input |
US8396265B1 (en) * | 2012-06-26 | 2013-03-12 | Google Inc. | Facial recognition |
US8542879B1 (en) * | 2012-06-26 | 2013-09-24 | Google Inc. | Facial recognition |
US8869053B2 (en) | 2012-07-06 | 2014-10-21 | Sap Ag | Organizer for managing employee time and attendance |
US8437513B1 (en) | 2012-08-10 | 2013-05-07 | EyeVerify LLC | Spoof detection for biometric authentication |
US20140195974A1 (en) | 2012-08-29 | 2014-07-10 | Identity Validation Products, Llc | Method and apparatus for using a finger swipe interface to control a system |
SG11201501691VA (en) | 2012-09-05 | 2015-04-29 | Element Inc | Biometric authentication in connection with camera-equipped devices |
US9740917B2 (en) | 2012-09-07 | 2017-08-22 | Stone Lock Global, Inc. | Biometric identification systems and methods |
US9002586B2 (en) | 2012-12-03 | 2015-04-07 | Honda Motor Co., Ltd. | Integrated biometric switch |
US8856541B1 (en) | 2013-01-10 | 2014-10-07 | Google Inc. | Liveness detection |
US9355612B1 (en) | 2013-01-22 | 2016-05-31 | Amazon Technologies, Inc. | Display security using gaze tracking |
US9244529B2 (en) | 2013-01-27 | 2016-01-26 | Dmitri Model | Point-of-gaze estimation robust to head rotations and/or device rotations |
US20140294257A1 (en) | 2013-03-28 | 2014-10-02 | Kevin Alan Tussy | Methods and Systems for Obtaining Information Based on Facial Identification |
US9294475B2 (en) | 2013-05-13 | 2016-03-22 | Hoyos Labs Ip, Ltd. | System and method for generating a biometric identifier |
US9003196B2 (en) | 2013-05-13 | 2015-04-07 | Hoyos Labs Corp. | System and method for authorizing access to access-controlled environments |
CN104166835A (en) | 2013-05-17 | 2014-11-26 | 诺基亚公司 | Method and device for identifying living user |
US20140351163A1 (en) | 2013-05-21 | 2014-11-27 | Kevin Alan Tussy | System and method for personalized delivery verification |
US9014452B2 (en) | 2013-08-21 | 2015-04-21 | Seiko Epson Corporation | Orientation-aware average intensity histogram to indicate object boundary depth in ultrasound images |
AR097974A1 (en) | 2013-10-11 | 2016-04-20 | Element Inc | SYSTEM AND METHOD FOR BIOMETRIC AUTHENTICATION IN CONNECTION WITH DEVICES EQUIPPED WITH CAMERA |
KR101556599B1 (en) | 2013-10-30 | 2015-10-02 | 연세대학교 산학협력단 | Pattern Inputting Apparatus and Method, and Recording Medium Using the Same |
US20150227735A1 (en) | 2014-02-13 | 2015-08-13 | Robert Chappell | System and method for eye tracking authentication |
US20150227780A1 (en) | 2014-02-13 | 2015-08-13 | FacialNetwork, Inc. | Method and apparatus for determining identity and programing based on image features |
US10027884B2 (en) | 2014-03-05 | 2018-07-17 | Disney Enterprises, Inc. | Method for capturing photographs and videos on a handheld client device without continually observing the device's screen |
KR20170005847A (en) | 2014-05-13 | 2017-01-16 | 엘리먼트, 인크. | System and method for electronic key provisioning and access management in connection with mobile devices |
US9408076B2 (en) * | 2014-05-14 | 2016-08-02 | The Regents Of The University Of California | Sensor-assisted biometric authentication for smartphones |
US20150348214A1 (en) | 2014-05-28 | 2015-12-03 | Shailendra Jain | Messaging service for geofence-based automatic time clocking |
US20150347833A1 (en) | 2014-06-03 | 2015-12-03 | Mark Ries Robinson | Noncontact Biometrics with Small Footprint |
US9965728B2 (en) | 2014-06-03 | 2018-05-08 | Element, Inc. | Attendance authentication and management in connection with mobile devices |
US10803160B2 (en) | 2014-08-28 | 2020-10-13 | Facetec, Inc. | Method to verify and identify blockchain with user question data |
CA2902093C (en) | 2014-08-28 | 2023-03-07 | Kevin Alan Tussy | Facial recognition authentication system including path parameters |
US10915618B2 (en) * | 2014-08-28 | 2021-02-09 | Facetec, Inc. | Method to add remotely collected biometric images / templates to a database record of personal information |
US10698995B2 (en) * | 2014-08-28 | 2020-06-30 | Facetec, Inc. | Method to verify identity using a previously collected biometric image/data |
US11256792B2 (en) * | 2014-08-28 | 2022-02-22 | Facetec, Inc. | Method and apparatus for creation and use of digital identification |
US9405967B2 (en) * | 2014-09-03 | 2016-08-02 | Samet Privacy Llc | Image processing apparatus for facial recognition |
JP6809226B2 (en) | 2014-10-15 | 2021-01-06 | 日本電気株式会社 | Biometric device, biometric detection method, and biometric detection program |
EP3218845A4 (en) | 2014-11-13 | 2018-07-11 | Intel Corporation | Facial liveness detection in image biometrics |
US9928603B2 (en) | 2014-12-31 | 2018-03-27 | Morphotrust Usa, Llc | Detecting facial liveliness |
USD763872S1 (en) | 2015-01-07 | 2016-08-16 | Facetec, Inc. | Display screen or portion thereof with graphical user interface |
EP3284016B1 (en) | 2015-04-16 | 2023-11-22 | Tobii AB | Authentication of a user of a device |
US10678897B2 (en) | 2015-04-16 | 2020-06-09 | Tobii Ab | Identification, authentication, and/or guiding of a user using gaze information |
MY182294A (en) | 2015-06-16 | 2021-01-18 | Eyeverify Inc | Systems and methods for spoof detection and liveness analysis |
JP6785305B2 (en) * | 2015-07-15 | 2020-11-18 | 15 セカンズ オブ フェイム,インコーポレイテッド | Equipment and methods for video analysis techniques to identify individuals with face recognition and contextual video streams |
US20170053175A1 (en) | 2015-08-17 | 2017-02-23 | Facetec, Inc. | System and method for validating three-dimensional objects |
CN105023010B (en) * | 2015-08-17 | 2018-11-06 | 中国科学院半导体研究所 | A kind of human face in-vivo detection method and system |
CN105184246B (en) | 2015-08-28 | 2020-05-19 | 北京旷视科技有限公司 | Living body detection method and living body detection system |
KR102377182B1 (en) | 2015-09-01 | 2022-03-22 | 엘지전자 주식회사 | Mobile terminal and control method for the mobile terminal |
US20170186170A1 (en) | 2015-12-24 | 2017-06-29 | Thomas A. Nugraha | Facial contour recognition for identification |
US10102358B2 (en) | 2015-12-29 | 2018-10-16 | Sensory, Incorporated | Face-controlled liveness verification |
US10733275B1 (en) | 2016-04-01 | 2020-08-04 | Massachusetts Mutual Life Insurance Company | Access control through head imaging and biometric authentication |
US10063560B2 (en) | 2016-04-29 | 2018-08-28 | Microsoft Technology Licensing, Llc | Gaze-based authentication |
KR102471916B1 (en) | 2016-06-03 | 2022-11-29 | 엘지전자 주식회사 | Mobile device and method for controlling thereof |
KR20180006133A (en) | 2016-07-08 | 2018-01-17 | 삼성전자주식회사 | Electronic device and operating method thereof |
US10089521B2 (en) * | 2016-09-02 | 2018-10-02 | VeriHelp, Inc. | Identity verification via validated facial recognition and graph database |
CN107886032B (en) | 2016-09-30 | 2021-12-14 | 阿里巴巴集团控股有限公司 | Terminal device, smart phone, authentication method and system based on face recognition |
CN106503687B (en) * | 2016-11-09 | 2019-04-05 | 合肥工业大学 | Merge the monitor video system for identifying figures and its method of face multi-angle feature |
CN110114777B (en) | 2016-12-30 | 2023-10-20 | 托比股份公司 | Identification, authentication and/or guidance of a user using gaze information |
US9983687B1 (en) | 2017-01-06 | 2018-05-29 | Adtile Technologies Inc. | Gesture-controlled augmented reality experience using a mobile communications device |
US10747859B2 (en) | 2017-01-06 | 2020-08-18 | International Business Machines Corporation | System, method and computer program product for stateful instruction-based dynamic man-machine interactions for humanness validation |
GB2560340A (en) | 2017-03-07 | 2018-09-12 | Eyn Ltd | Verification method and system |
CN107609471A (en) * | 2017-08-02 | 2018-01-19 | 深圳元见智能科技有限公司 | A kind of human face in-vivo detection method |
CN107622227B (en) * | 2017-08-25 | 2021-04-13 | 深圳依偎控股有限公司 | 3D face recognition method, terminal device and readable storage medium |
KR102185854B1 (en) * | 2017-09-09 | 2020-12-02 | 애플 인크. | Implementation of biometric authentication |
EP4156129A1 (en) | 2017-09-09 | 2023-03-29 | Apple Inc. | Implementation of biometric enrollment |
TWI781226B (en) | 2017-09-18 | 2022-10-21 | 美商艾勒門公司 | Methods, systems, and media for detecting spoofing in mobile authentication |
CN109583285B (en) * | 2017-09-29 | 2023-03-07 | 阿里巴巴集团控股有限公司 | Object recognition method |
CN107800964A (en) * | 2017-10-26 | 2018-03-13 | 武汉大学 | It is a kind of that method of the face automatic detection with capturing is realized based on dual camera |
US11513205B2 (en) * | 2017-10-30 | 2022-11-29 | The Research Foundation For The State University Of New York | System and method associated with user authentication based on an acoustic-based echo-signature |
CN108154099B (en) * | 2017-12-20 | 2021-04-30 | 北京奇艺世纪科技有限公司 | Figure identification method and device and electronic equipment |
WO2019151368A1 (en) | 2018-02-01 | 2019-08-08 | 日本電気株式会社 | Biometric authentication device, system, method and recording medium |
CN108596135A (en) * | 2018-04-26 | 2018-09-28 | 上海诚数信息科技有限公司 | Personal identification method and system |
CA3045819C (en) * | 2018-06-11 | 2022-09-06 | Laurence Hamid | Liveness detection |
JP7464619B2 (en) | 2019-03-12 | 2024-04-09 | エレメント インク. | Detecting spoofing using facial recognition on mobile devices |
US10997396B2 (en) * | 2019-04-05 | 2021-05-04 | Realnetworks, Inc. | Face liveness detection systems and methods |
US11507248B2 (en) | 2019-12-16 | 2022-11-22 | Element Inc. | Methods, systems, and media for anti-spoofing using eye-tracking |
-
2020
- 2020-03-11 JP JP2021555104A patent/JP7464619B2/en active Active
- 2020-03-11 KR KR1020217032304A patent/KR20220004628A/en active IP Right Grant
- 2020-03-11 EP EP20770199.6A patent/EP3938953A4/en active Pending
- 2020-03-11 CN CN202080035642.9A patent/CN113826110B/en active Active
- 2020-03-11 US US16/815,700 patent/US11343277B2/en active Active
- 2020-03-11 SG SG11202109983R patent/SG11202109983RA/en unknown
- 2020-03-11 CA CA3133229A patent/CA3133229C/en active Active
- 2020-03-11 WO PCT/US2020/022168 patent/WO2020185948A1/en unknown
- 2020-03-11 BR BR112021018149-0A patent/BR112021018149B1/en active IP Right Grant
- 2020-03-11 AU AU2020237108A patent/AU2020237108B2/en active Active
- 2020-03-12 TW TW109108240A patent/TWI781383B/en active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170124385A1 (en) * | 2007-12-31 | 2017-05-04 | Applied Recognition Inc. | Face authentication to mitigate spoofing |
US20120288166A1 (en) * | 2011-05-13 | 2012-11-15 | Microsoft Corporation | Association and prediction in facial recognition |
US20180181737A1 (en) * | 2014-08-28 | 2018-06-28 | Facetec, Inc. | Facial Recognition Authentication System Including Path Parameters |
US20170304732A1 (en) * | 2014-11-10 | 2017-10-26 | Lego A/S | System and method for toy recognition |
Non-Patent Citations (1)
Title |
---|
See also references of EP3938953A4 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11425562B2 (en) | 2017-09-18 | 2022-08-23 | Element Inc. | Methods, systems, and media for detecting spoofing in mobile authentication |
US11343277B2 (en) | 2019-03-12 | 2022-05-24 | Element Inc. | Methods and systems for detecting spoofing of facial recognition in connection with mobile devices |
US11507248B2 (en) | 2019-12-16 | 2022-11-22 | Element Inc. | Methods, systems, and media for anti-spoofing using eye-tracking |
Also Published As
Publication number | Publication date |
---|---|
BR112021018149B1 (en) | 2023-12-26 |
KR20220004628A (en) | 2022-01-11 |
EP3938953A1 (en) | 2022-01-19 |
JP7464619B2 (en) | 2024-04-09 |
AU2020237108A1 (en) | 2021-11-04 |
US20200296132A1 (en) | 2020-09-17 |
AU2020237108B2 (en) | 2022-12-15 |
US11343277B2 (en) | 2022-05-24 |
JP2022525138A (en) | 2022-05-11 |
CA3133229A1 (en) | 2020-09-17 |
SG11202109983RA (en) | 2021-10-28 |
BR112021018149A2 (en) | 2022-01-18 |
TW202040427A (en) | 2020-11-01 |
CN113826110B (en) | 2023-06-20 |
CA3133229C (en) | 2023-04-04 |
EP3938953A4 (en) | 2022-12-28 |
CN113826110A (en) | 2021-12-21 |
TWI781383B (en) | 2022-10-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
AU2020237108B2 (en) | Detecting spoofing of facial recognition with mobile devices | |
US10902034B2 (en) | Method for populating a map with a plurality of avatars through the use of a mobile technology platform | |
TWI781226B (en) | Methods, systems, and media for detecting spoofing in mobile authentication | |
US11507248B2 (en) | Methods, systems, and media for anti-spoofing using eye-tracking | |
US11645873B1 (en) | Systems, media, and methods providing a golf swing coach | |
US10706087B1 (en) | Delegated decision tree evaluation | |
US11058948B1 (en) | Predictive modeling for multiplayer gaming | |
US10430926B1 (en) | Systems, media, and methods for improving responsiveness of spotlight tours in a dynamic viewport environment | |
US11150789B2 (en) | Method, systems, and media to arrange a plurality of digital images within an image display section of a graphical user inteface (GUI) | |
US20230325798A1 (en) | Systems and methods for regulating access and ticketing with non-fungible tokens | |
WO2023034597A1 (en) | Methods and systems for improving blockchain technology |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20770199 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2021555104 Country of ref document: JP Kind code of ref document: A Ref document number: 3133229 Country of ref document: CA |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112021018149 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 2020770199 Country of ref document: EP Effective date: 20211012 |
|
ENP | Entry into the national phase |
Ref document number: 2020237108 Country of ref document: AU Date of ref document: 20200311 Kind code of ref document: A |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01E Ref document number: 112021018149 Country of ref document: BR Free format text: APRESENTE O RELATORIO DESCRITIVO E OS DESENHOS. |
|
ENP | Entry into the national phase |
Ref document number: 112021018149 Country of ref document: BR Kind code of ref document: A2 Effective date: 20210913 |