WO2011104833A1 - 認証装置、認証プログラム、および認証方法 - Google Patents
認証装置、認証プログラム、および認証方法 Download PDFInfo
- Publication number
- WO2011104833A1 WO2011104833A1 PCT/JP2010/052882 JP2010052882W WO2011104833A1 WO 2011104833 A1 WO2011104833 A1 WO 2011104833A1 JP 2010052882 W JP2010052882 W JP 2010052882W WO 2011104833 A1 WO2011104833 A1 WO 2011104833A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- display
- palette
- color
- authentication
- colors
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 70
- 239000003086 colorant Substances 0.000 claims description 48
- 238000003384 imaging method Methods 0.000 claims description 40
- 238000003491 array Methods 0.000 claims description 5
- 230000000295 complement effect Effects 0.000 claims description 4
- 238000012545 processing Methods 0.000 abstract description 80
- 210000003462 vein Anatomy 0.000 abstract description 36
- 238000010586 diagram Methods 0.000 description 14
- 230000006870 function Effects 0.000 description 11
- 238000012795 verification Methods 0.000 description 10
- 238000001514 detection method Methods 0.000 description 8
- 210000004204 blood vessel Anatomy 0.000 description 7
- 238000006243 chemical reaction Methods 0.000 description 5
- 238000004891 communication Methods 0.000 description 4
- 238000000605 extraction Methods 0.000 description 4
- 102000001554 Hemoglobins Human genes 0.000 description 3
- 108010054147 Hemoglobins Proteins 0.000 description 3
- 238000005259 measurement Methods 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 102000053602 DNA Human genes 0.000 description 2
- 108020004414 DNA Proteins 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- QVGXLLKOCUKJST-UHFFFAOYSA-N atomic oxygen Chemical compound [O] QVGXLLKOCUKJST-UHFFFAOYSA-N 0.000 description 1
- 230000017531 blood circulation Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 210000003743 erythrocyte Anatomy 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 210000000887 face Anatomy 0.000 description 1
- 210000000554 iris Anatomy 0.000 description 1
- 230000001678 irradiating effect Effects 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 229910052760 oxygen Inorganic materials 0.000 description 1
- 239000001301 oxygen Substances 0.000 description 1
- 210000001525 retina Anatomy 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/60—Static or dynamic means for assisting the user to position a body part for biometric acquisition
- G06V40/67—Static or dynamic means for assisting the user to position a body part for biometric acquisition by interactive indications to the user
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T1/00—General purpose image data processing
Definitions
- the present invention relates to an authentication device, an authentication program, and an authentication method for performing personal authentication using characteristics of a living body.
- the human body includes biological information that can identify an individual, and some of them are used as information for identifying and authenticating the individual.
- biometric information that can be used for authentication includes fingerprints, eye retinas and irises, faces, blood vessels, and DNA (DeoxyriboNucleic Acid).
- biometric authentication is performed by comparing the biometric information collected at the time of registration with the biometric information acquired at the time of authentication.
- the authentication apparatus announces to the user (authenticated person) so that the positional relationship of the biometric feature detection site with respect to the sensor is appropriate, and guides the biometric feature detection site to an appropriate positional relationship.
- the authentication apparatus announces to the user (authenticated person) so that the positional relationship of the biometric feature detection site with respect to the sensor is appropriate, and guides the biometric feature detection site to an appropriate positional relationship.
- an authentication device that displays a captured image superimposed on the silhouette of a registered hand in order to efficiently guide to an imaging state at the time of registration (see, for example, Patent Document 2).
- an authentication apparatus that guides a biological feature detection site to an appropriate positional relationship performs a conventional authentication control process as shown in FIG.
- the authentication control process 90 acquires an image from the sensor 95 and outputs a verification result to the user application 97.
- the authentication control process 90 acquires an image from the sensor 95 by the image acquisition process 91 and passes the acquired image to the position / orientation determination process 92.
- the position / orientation determination processing 92 determines the position and orientation of the feature detection part, and passes the determination result together with the acquired image to a guidance GUI (Graphical ⁇ User Interface) display processing 93.
- the position / orientation determination process 92 requests an image from the image acquisition process 91 when determining that the position and orientation of the feature detection part are not appropriate.
- the guidance GUI display processing 93 displays a guidance GUI display image 96 on a monitor or the like, and guides the feature detection site of the living body to have an appropriate positional relationship with the sensor 95.
- the position / orientation determination process 92 makes a verification request to the verification process 94 and passes the acquired image.
- the collation process 94 collates the acquired image with an image or feature registered in advance, and outputs a collation result.
- JP 2006-42880 A JP 2008-71158 A JP 2009-5273 A
- the captured image displayed to guide the biological feature detection site to an appropriate positional relationship may not be sufficiently visible to the user, and may not function properly as a guide silhouette as it is.
- the display image may be delayed from the actual operation, and the display image may not function properly as a guide silhouette.
- the captured image when displayed including biometric information, there may be a problem with privacy or security, and depending on the authentication environment, it may not be suitable as a display image.
- the present invention has been made in view of these points, and is capable of displaying a guidance display screen with good visibility and reduced risk of causing privacy or security problems without applying processing load.
- An object is to provide an apparatus, an authentication program, and an authentication method.
- the imaging data generating means generates imaging data having a single color palette in which an index color in which a living body is imaged and single colors are assigned in order of density is arranged.
- the display data generation means arranges one of the arrays divided by a predetermined background boundary threshold as a background display array in which index colors to which a background color is assigned is arranged, and arranges an index color in which specific colors are assigned in order of density.
- Display data is generated by replacing the single color palette of the imaging data with the color palette used as the living body display arrangement.
- the display means displays the living body as display data.
- an authentication program for causing a computer to execute a personal authentication process using biometric features causes the computer to function as an imaging data generation unit, a display data generation unit, and a display unit. .
- an authentication method for performing personal authentication using characteristics of a living body captures a living body, generates imaging data having a single color palette in which index colors are arranged in which the single colors are assigned in order of density, One of the arrays divided by a predetermined background boundary threshold is used as a background display array in which index colors assigned with a background color are arranged, and the other is a biological display array in which index colors assigned with specific colors in order of density are arranged.
- the display color is generated by replacing the single color palette included in the imaging data, and the living body is displayed as the display data.
- FIG. 1 is a diagram illustrating a configuration of an authentication system according to an embodiment.
- a system in which the authentication system 1 performs authentication using a palm vein is exemplified, but the present invention is not limited to this, and the present invention can also be applied to a system that performs authentication at another feature detection site of a living body.
- the authentication system 1 is a system that recognizes the characteristics of a living body and identifies and authenticates an individual, and can be used for information system logon, entrance / exit management, and the like.
- the authentication system 1 includes an authentication device 10 and a database server 40 connected to the authentication device 10 via the network 2.
- the database server 40 associates and stores identification information for identifying an individual and biometric information registered in advance before biometric authentication.
- the identification information for identifying an individual is a unique ID (IDentification) given to a user directly (for example, an employee number) or indirectly (for example, an account number).
- the biometric information registered in advance is feature information obtained by extracting a feature portion from image information, encoded information obtained by encoding image information or feature information, and the like.
- the authentication device 10 includes a processing device 11, a display 12, and a sensor unit 20, and includes a keyboard 13, a mouse 14, an IC (Integrated Circuit) card reader / writer 30 as necessary.
- the sensor unit 20 has a built-in imaging device, images the palm of the user, and outputs a captured image to the processing device 11.
- the IC card reader / writer 30 reads and writes information on the IC card 31 of the user.
- the keyboard 13 and the mouse 14 accept user input operations.
- a user who seeks authentication inputs identification information (for example, a user ID) for identifying the user through the keyboard 13, mouse 14, or IC card reader / writer 30.
- the authentication device 10 uses the display 12 to obtain input of biometric information for authenticating the user by display.
- the user inputs biometric information by holding his hand over the sensor unit 20.
- the authentication apparatus 10 that has input the palm image as the biometric information collates with the biometric information recorded in advance in the storage unit of the database server 40 or the user's IC card 31.
- FIG. 2 is a diagram illustrating conversion from a captured image to a display image performed by the authentication apparatus according to the embodiment.
- FIG. 3 is a diagram illustrating an image display example of the embodiment.
- the captured image 50 is acquired when the user holds the palm of the hand over the sensor unit 20.
- the photographed image 50 is an image obtained by irradiating the palm with near infrared rays and photographing reflected light. Since hemoglobin in red blood cells flowing in the veins has lost oxygen, this hemoglobin (reduced hemoglobin) has a property of absorbing near infrared rays in the vicinity of 760 nanometers. Therefore, when near infrared rays are applied to the palm, only a portion where the vein is present is less reflected, and the position of the vein can be recognized by the intensity of reflected light of the near infrared ray.
- the photographed image 50 is an achromatic image although it is easy to extract characteristic information by using a specific light source.
- the captured image 50 is grayscale data represented by a value of 0 to 255 (8 bits) per pixel, it is possible to use a 256-level grayscale color palette (grayscale palette).
- the captured image palette 51 is a 256-level gray scale palette in which light white to dark black are arranged in order of density.
- the captured image palette 51 can display the captured image 50 with a sufficient color resolution as a single color, for example, by displaying on the display.
- the captured image 50 has been used as a silhouette image for guiding the position of the hand in the palm vein system.
- the vein image is conspicuous and unpleasant, such as “scary and uncomfortable”
- the captured image 50 becomes dark and the visibility of the outline is deteriorated, and the captured image 50 may not function sufficiently as a silhouette image.
- the authentication device 10 uses the captured image 50 as a silhouette image after converting the captured image 50 to the display image 52 by the guidance GUI instead of using the captured image 50 as it is. At this time, if an advanced image conversion process is performed, a time lag occurs between the movement of the hand and the display of the silhouette image. Therefore, the burden of the process of converting the captured image 50 to the display image 52 may not be excessive. Desired.
- the authentication device 10 displays the display image 52 by replacing the captured image palette 51 used for displaying the captured image 50 with the display image palette 53.
- the display image palette 53 includes a background palette 55 and a palm palette 56.
- the background palette 55 and the palm palette 56 are not independent color palettes but are part of the display image palette 53.
- the background palette 55 is from the index color [0] of the display image palette 53 to the index color [35] that becomes the background boundary threshold value 54.
- the background palette 55 is a color palette in which a single background color is assigned to all index colors from [0] to [35].
- the palm palette 56 is from the index color [36] next to the background boundary threshold 54 to the index color [255] of the display image palette 53.
- the palm palette 56 is a 220-stage color palette in which specific colors (for example, light orange) are arranged from light colors to dark colors.
- the palm palette 56 arranges the index colors in order of density in the reverse direction from light color to dark color.
- the vein image that is displayed in a dark color in the captured image 50 is displayed in a light color in the display image 52, making visual identification difficult.
- the palm palette 56 is set so that the density range of the color to be used is smaller than the density range of the captured image palette 51.
- the vein image has a smaller density difference from the other palm portions, making it more difficult to visually recognize the vein image.
- the display image 52 when the display image 52 is a silhouette image, it reduces the discomfort caused by the user's vision. Further, when the display image 52 is a silhouette image, it is difficult to visually recognize a vein image of a third party, so that the possibility of causing a privacy or security problem is reduced.
- the background color assigned to all index colors by the background palette 55 can be a complementary color of the boundary color (color assigned to the index color [36]) on the captured image palette 51 side. In that case, the background color emphasizes the outline of the palm.
- the palm palette 56 arranges the index colors in the order of density while being in the opposite direction to the captured image palette 51, it is possible to express the shade of the hand expressed in gray scale, the degree of inclination of the hand, etc. Can be grasped well.
- images such as the display image 60 and the display image 61 can be displayed.
- the display image 60 and the display image 61 have clear contrast between the palm and the background, and the outline of the palm can be easily grasped.
- the display image 60 and the display image 61 are difficult to identify by visual observation of a vein image, and in addition to the discomfort caused by the user's vision, the possibility of causing privacy and security problems is reduced.
- the display image 60 and the display image 61 sufficiently express the color density difference due to the tilt of the palm, the tilt of the hand and the sense of distance from the sensor unit 20 can be well understood.
- FIG. 4 is a block diagram of the authentication system according to the embodiment.
- the sensor unit 20 is equipped with an imaging device 24.
- the imaging device 24 includes an infrared sensor (for example, a complementary metal oxide semiconductor (CMOS) sensor), a condenser lens 23, and a distance sensor 22 that are not shown in the center, and a plurality of near-infrared light emitting elements in the imaging direction. (LED) 21 is provided.
- the near-infrared light emitting elements 21 are provided, for example, at eight locations around the infrared sensor, and emit near-infrared light upward (in the imaging direction).
- the sensor unit 20 outputs the captured image and measurement distance information measured by the distance sensor 22 to the processing device 11.
- the processing device 11 executes a captured image acquisition process 111, a guidance GUI display process 112, a blood vessel image extraction process 113, an imaging state determination process 114, a registration process 115, a feature matching process 116, and a search process 117.
- the captured image acquisition process 111 acquires measurement distance information from the sensor unit 20 and a captured image.
- the captured image acquisition process 111 notifies the acquired measurement distance information and the captured image to the imaging state determination process 114, the guidance GUI display process 112, the blood vessel image extraction process 113, and the registration process 115.
- the guidance GUI display processing 112 displays a guidance GUI display for guiding the palm to an appropriate position on the display 12.
- the guidance GUI display processing 112 when displaying the guidance GUI display on the display 12, the color palette used for displaying the captured image is replaced from the captured image palette 51 to the display image palette 53 as described above.
- the guidance GUI display process 112 acquires information on whether or not the palm position is appropriate from an imaging state determination process 114 described later.
- the blood vessel image extraction process 113 extracts a blood vessel image (biological feature data) of a vein from the captured image and notifies the registration process 115 of it.
- the registration process 115 registers captured images, imaging conditions, biometric feature data, and the like in the storage unit 118 of the database server 40 when performing user registration. Note that the storage unit 118 may be included in the IC card 31.
- the search process 117 searches the storage unit 118 for a captured image, a shooting condition, and biometric feature data corresponding to the user ID based on the user ID input at the time of collation.
- the imaging state determination process 114 determines whether or not the captured image and imaging conditions acquired from the storage unit 118 by the search process 117 are appropriate by comparing the captured image and imaging conditions input from the sensor unit 20.
- the palm is at an appropriate distance from the sensor unit 20, or the contour (position, posture) of the palm detected from the captured image is at an appropriate position compared to the standard condition or the imaging condition at the time of registration. Determine whether. If the imaging state determination processing 114 determines that the captured image and the imaging conditions at the time of authentication are appropriate, the imaging state determination processing 114 notifies the feature verification processing 116 that the feature verification can be executed. In addition, the imaging state determination process 114 notifies the guidance GUI display process 112 of the evaluation of the captured image and the imaging condition at the time of authentication.
- the feature matching process 116 determines the degree of coincidence (similarity) between the biometric feature data acquired from the storage unit 118 by the search process 117 and the biometric feature data (blood vessel image data) extracted by the blood vessel image extraction process 113. As a result, the success or failure of authentication is output to the user application.
- FIG. 5 is a diagram illustrating a hardware configuration example of the authentication device according to the embodiment.
- the authentication device 10 includes a processing device 11, a display 12, a keyboard 13, a mouse 14, a sensor unit 20, and an IC card reader / writer 30.
- the entire processing apparatus 11 is controlled by a CPU (Central Processing Unit) 101.
- a RAM Random Access Memory
- HDD Hard Disk Drive
- a communication interface 104 a graphic processing device 105, and an input / output interface 106 are connected to the CPU 101 via a bus 107.
- the RAM 102 temporarily stores at least part of an OS (Operating System) program and application programs to be executed by the CPU 101.
- the RAM 102 stores various data necessary for processing by the CPU 101.
- the HDD 103 stores an OS and application programs.
- a display 12 is connected to the graphic processing device 105.
- the graphic processing device 105 displays an image on the screen of the display 12 in accordance with a command from the CPU 101.
- the input / output interface 106 is connected with a keyboard 13, a mouse 14, a sensor unit 20, and an IC card reader / writer 30.
- the input / output interface 106 can be connected to a portable recording medium interface that can write information to the portable recording medium 110 and read information from the portable recording medium 110.
- the input / output interface 106 transmits signals sent from the keyboard 13, mouse 14, sensor unit 20, IC card reader / writer 30, and portable recording medium interface to the CPU 101 via the bus 107.
- the communication interface 104 is connected to the network 2.
- the communication interface 104 transmits / receives data to / from the database server 40.
- the database server 40 can also be realized with the same hardware configuration.
- the processing device 11 can be configured to include modules each composed of an FPGA (Field Programmable Gate Array), a DSP (Digital Signal Processor), or the like, or can be configured without the CPU 101.
- each of the processing devices 11 includes a nonvolatile memory (for example, EEPROM (Electrically Erasable and Programmable Read Only Memory), flash memory, flash memory type memory card, etc.) and stores module firmware.
- the nonvolatile memory can write firmware via the portable recording medium 110 or the communication interface 104. In this way, the processing device 11 can also update the firmware by rewriting the firmware stored in the nonvolatile memory.
- FIG. 6 is a diagram illustrating a data structure of an 8-bit (256 colors) uncompressed DIB according to the embodiment.
- the captured image data has an 8-bit (256 colors) uncompressed DIB (Device-Independent Bitmap) data structure 300.
- the data structure 300 of 8-bit (256 colors) uncompressed DIB includes a bitmap information header, a color palette, and pixel data.
- the bitmap information header includes information such as the number of horizontal pixels of the image, the number of vertical pixels of the image, the number of planes, the number of bits per pixel, the compression method, and the size of the entire pixel data.
- the color palette has an array structure from index color [0] to index color [255], and each array includes array elements “B”, “G”, “R”. "And"-"as data.
- the array element “B” stores data representing B (blue) luminance of a pixel with 256 gradations using 1-byte data.
- the array element “G” stores data representing G (green) luminance of a pixel with 256 gradations using 1-byte data.
- the array element “R” stores data representing R (red) luminance of a pixel with 256 gradations using 1-byte data.
- the array element “-” stores “0” as 1-byte reservation data.
- FIG. 7 is a diagram illustrating a data configuration example of the original image DIB of the embodiment.
- FIG. 8 is a diagram illustrating an RGB gradation graph example of the 256 gradation gray scale palette according to the embodiment.
- the photographed image data has an 8-bit (256 colors) uncompressed DIB data structure 300, but more specifically, is an original image DIB301.
- the original image DIB301 has a 256 gray scale palette.
- the 256 gradation gray scale palette is a gray scale color palette in which an index color [0] for specifying dark black and an index color [255] for specifying light white are arranged in 256 levels in order of density. .
- the values in parentheses are the aforementioned B luminance, G luminance, R luminance, and reserved data 0 in order.
- FIG. 9 is a diagram illustrating a data configuration example of the display image DIB of the embodiment.
- FIG. 10 is a diagram illustrating an RGB gradation graph example of a color palette dedicated to the palm vein GUI according to the embodiment.
- the display image data has an 8-bit (256 colors) uncompressed DIB data structure 300, but more specifically, is a display image DIB302.
- the display image DIB 302 has a palm vein GUI dedicated color palette.
- the palm vein GUI dedicated color palette includes a background palette from index color [0] to index color [35] that specifies the background color, and a dark specific color from index color [36] that specifies the light specific color.
- a color palette including a palm palette in which up to index colors [255] for designating colors are arranged in order of density.
- the background palette represents the background.
- the index color [m] (175, 110, 100, 0 ).
- the palm palette is used to express the palm.
- n is an integer from 36 to 255
- the index color [n] (f1 (n), f2 ( n), f3 (n), 0).
- f1 (n), f2 (n), and f3 (n) are predetermined functions of n.
- f1 (n) 210 ⁇ int ((n ⁇ 35) / 2) ⁇ mod (n ⁇ 35, 2)
- f2 (n) 250 ⁇ int ((n ⁇ 35) / 2) ⁇ mod (n ⁇ 35, 2)
- f3 (n) 255.
- the function int (x) is the maximum integer that does not exceed x
- the function mod (x, y) is a remainder when x is divided by y.
- the pixel data stores an index color array number for each pixel corresponding to the photographed image for the number of pixels. Note that the pixel data of the display image data is the same as the pixel data of the captured image data.
- the captured image data and the display image data have an 8-bit (256 colors) uncompressed DIB data structure 300, and the pixel data is also the same.
- the captured image data and the display image data can display different images by using different palettes.
- the processing device 11 can perform high-speed image conversion since it does not process the pixel data.
- FIG. 11 is a diagram illustrating a processing flow of the guidance GUI display processing according to the embodiment.
- the guidance GUI display process 112 is a process of displaying a guidance GUI display for guiding the palm to an appropriate position on the display 12 based on the captured image acquired from the captured image acquisition process 111.
- the guidance GUI display process 112 is a process for acquiring the guidance information 320 from the imaging state determination process 114.
- the guidance information 320 includes position information 321 and a captured image 322.
- the processing device 11 executes a dedicated color palette (palm vein GUI dedicated color palette) creation process as an initialization process, and creates a palm vein GUI dedicated color palette.
- the dedicated color palette creation process will be described later in detail with reference to FIG.
- Step S12 The processing device 11 waits for acquisition of the guidance information 320.
- Step S13 The processing device 11 acquires the guidance information 320.
- the acquired captured image 322 has the data structure of the original image DIB301, and a 256 gray scale palette is set.
- Step S14 The processing device 11 replaces the 256 gradation gray scale palette set in the captured image 322 with a palm vein GUI dedicated color palette.
- Step S15 The processing device 11 performs a guidance GUI display for guiding the palm to an appropriate position by using a display image in which a palm vein GUI dedicated color palette is set on the captured image acquired from the captured image acquisition processing 111, and step S12. Proceed.
- FIG. 12 is a flowchart of dedicated color palette creation processing according to the embodiment.
- the dedicated color palette creation process is a process for creating a palm vein GUI dedicated color palette.
- the creation of the palm vein GUI dedicated color palette is performed only once when the authentication apparatus 10 is activated, but may be performed each time authentication is performed.
- the processing device 11 initializes the values of the elements “B”, “G”, and “R” of the index color of the color palette.
- This color is a light orange color, and it is thought that there is little sense of incongruity as a color that expresses the palm, but in some cases it may cause a sense of incongruity depending on the country or region. Therefore, the initial value may be arbitrarily set according to the environment, personal preference, and the like.
- the processing device 11 initializes the background boundary threshold.
- Step S24 The processing apparatus 11 determines whether the array element number is equal to or smaller than the background boundary threshold. If the array element number is equal to or smaller than the background boundary threshold, the processing device 11 proceeds to step S25. On the other hand, if the array element number is not less than or equal to the background boundary threshold, the processing device 11 proceeds to step S26.
- the selection of the background color is a complementary color of light pale orange, clarifying the boundary between the background color and the light pale orange adjacent to the background color, and improving the visibility of the palm contour.
- the processing apparatus 11 determines whether the array element number has reached the upper limit value “255” in step S30. If the array element number has not reached the upper limit value “255”, the processing apparatus 11 proceeds to step S31, increments the array element number by “1”, and proceeds to step S24.
- the processing device 11 sets a background color for each element for each index color with an array number of “0” to “35”.
- the processing apparatus 11 determines whether or not the remainder obtained by dividing “n” of the array element number by “2” is “0”. In other words, the processing device 11 determines whether or not “n” of the array element number is divisible by “2”. When the remainder obtained by dividing the array element number “n” by “2” is “0”, the processing device 11 proceeds to step S27. On the other hand, if the remainder obtained by dividing “n” of the array element number by “2” is not “0”, the processing device 11 proceeds to step S29.
- the processing device 11 sets the index color element “B”, “G”, and “R” value update processing executed in step S27, which will be described later, once every two times, and uses it to display the palm.
- the color gradation width is reduced.
- the vein display image becomes inconspicuous in the palm display image.
- the number excluding “n” in the sequence number can be arbitrarily set. As the number excluding the array number “n” is set to a larger value, the gradation width of the color used for displaying the palm can be reduced, and the vein image becomes less conspicuous as the gradation width becomes smaller.
- Step S28 The processing device 11 corrects the values set in the index color elements “B”, “G”, and “R” to values of 0 or more and 255 or less. For example, the processing device 11 corrects a calculated value that has become negative due to decrement to “0”.
- Step S29 The processing device 11 sets the values calculated in Step S27 and Step S28 to the elements “B”, “G”, and “R” of the index color of the array number [n].
- step S30 the processing device 11 determines whether the array element number has reached the upper limit “255”. If the array element number has not reached the upper limit value “255”, the processing apparatus 11 proceeds to step S31, increments the array element number by “1”, and proceeds to step S24.
- the processing device 11 sets the color used for displaying the palm to each element for each index color having the array element numbers “36” to “255”.
- the processing apparatus 11 ends the dedicated color palette creation process.
- the dedicated color palette creation process can be executed at an arbitrary timing such as when the authentication device 10 is activated or at the time of authentication.
- a dedicated color palette creation process may be executed when the authentication device 10 is activated so that all users use the same palm vein GUI dedicated color palette.
- a dedicated color palette creation process may be executed at the time of user authentication to use different palm vein GUI dedicated color palettes for each user.
- a color palette dedicated to the palm vein GUI that varies from user to user is, for example, user-specific information such as the user's skin color, blood circulation, sex, and age, or the external environment (time, weather, region, ambient brightness), etc. Create a color palette with a different color array using as a parameter.
- the dedicated color palette creation process enables generation of an optimal palette according to the conditions at the time of authentication.
- FIG. 13 is a diagram illustrating display examples (1) and (2) of the guidance GUI display window according to the embodiment.
- FIG. 14 is a diagram illustrating display examples (3) and (4) of the guidance GUI display window according to the embodiment.
- the guidance GUI display window 70 is a display example (1) of the guidance GUI display window displayed on the display 12 by the authentication device 10.
- the guidance GUI display window 70 includes a guidance GUI display image 701, a normal position image 702, a guidance display 703, a guidance message 704, and a result message 705 in the window.
- the guidance GUI display image 701 is a display image that is displayed by replacing the captured image from the captured image palette to the display image palette.
- the guidance GUI display image 701 displays a background, a palm image, and a guidance display 703 indicating the normal position of the palm.
- the guidance display 703 displays, for example, a perfect circle indicating the normal position of the palm and an ellipse indicating that the user's palm is tilted.
- the normal position image 702 is a display for guiding the user's palm to the normal position.
- the guidance message 704 is a message display for guiding the user's palm to the normal position. For example, the guidance message 704 displays “Please keep your hand level” to notify that the user's palm is tilted.
- the result message 705 is a message display for notifying the result of user verification by the authentication device 10. For example, the result message 705 does not guide anything because the user's palm is tilted and cannot be authenticated.
- the guidance GUI display window 71 is a display example (2) of the guidance GUI display window displayed on the display 12 by the authentication device 10.
- the guidance GUI display window 71 includes a guidance GUI display image 711, a normal position image 712, a guidance display 713, a guidance message 714, and a result message 715 in the window.
- the guidance GUI display image 711 is a display image that is displayed by replacing the captured image from the captured image palette to the display image palette.
- the guidance GUI display image 711 displays a background, a palm image, and a guidance display 713 indicating the normal position of the palm.
- the guidance display 713 displays, for example, a perfect circle indicating the normal position of the palm and an arrow indicating that the user's palm is displaced from the center position.
- the normal position image 712 is a display for guiding the user's palm to the normal position.
- the normal position image 712 may be a palm registered image of the user.
- the normal position image 712 is displayed for guiding the normal position or by replacing the user's palm registration image from the captured image palette to the display image palette.
- the guidance message 714 is a message display for guiding the user's palm to the normal position. For example, the guidance message 714 displays “Please pull your hand slightly behind” to notify that the user's palm is displaced from the center position.
- the result message 715 is a message display for notifying the result of user verification by the authentication device 10. For example, the result message 715 does not guide anything because the user's palm is out of the center position and cannot be authenticated.
- the guidance GUI display window 72 is a display example (3) of the guidance GUI display window displayed on the display 12 by the authentication device 10.
- the guidance GUI display window 72 includes a guidance GUI display image 721, a normal position image 722, a guidance display 723, a guidance message 724, and a result message 725 in the window.
- the guidance GUI display image 721 is a display image that is displayed by replacing the captured image from the captured image palette to the display image palette.
- the guidance GUI display image 721 displays a background, a palm image, and a guidance display 723 indicating the normal position of the palm.
- the guidance display 723 displays, for example, a perfect circle indicating the normal position of the palm and an arrow indicating that the user's palm is shifted counterclockwise.
- the normal position image 722 is a display for guiding the user's palm to the normal position.
- the guidance message 724 is a message display for guiding the user's palm to the normal position. For example, the guidance message 724 displays “Please align your hand with the direction of the sensor” to notify that the palm of the user is deviating counterclockwise.
- the result message 725 is a message display for notifying the result of user verification by the authentication device 10. For example, the result message 725 does not guide anything because the user's palm is shifted counterclockwise and cannot be authenticated.
- the guidance GUI display window 73 is a display example (4) of the guidance GUI display window displayed on the display 12 by the authentication device 10.
- the guidance GUI display window 73 includes a guidance GUI display image 731, a normal position image 732, a guidance display 733, a guidance message 734, and a result message 735 within the window.
- the guidance GUI display image 731 is a display image that is displayed by replacing the captured image with the display image palette from the captured image palette.
- the guidance GUI display image 731 displays a background, a palm image, and a guidance display 733 indicating the normal position of the palm.
- the guidance display 733 displays only the perfect circle indicating the correct position of the palm because the palm of the user is closer to the sensor unit 20 than the correct position but the inclination and the position in the front and rear, right and left directions are correct.
- the normal position image 732 is a display for guiding the user's palm to the normal position.
- the guidance message 734 is a message display for guiding the user's palm to the normal position. For example, the guidance message 734 displays “Please release your hand” to notify that the palm of the user is closer to the sensor unit 20 than the normal position.
- the result message 735 is a message display for notifying the result of user verification by the authentication device 10. For example, the result message 735 displays “authenticated.”, Which is a result of the authentication device 10 succeeding in user verification.
- the above processing functions can be realized by a computer.
- a program describing the processing contents of the functions that the authentication device 10 and the database server 40 should have is provided.
- the program describing the processing contents can be recorded on a computer-readable recording medium (including a portable recording medium).
- the computer-readable recording medium include a magnetic recording device, an optical disk, a magneto-optical recording medium, and a semiconductor memory.
- the magnetic recording device include a hard disk device (HDD), a flexible disk (FD), and a magnetic tape.
- Optical discs include DVD (Digital Versatile Disc), DVD-RAM, CD-ROM, CD-R (Recordable) / RW (ReWritable), and the like.
- Magneto-optical recording media include MO (Magneto-Optical disk).
- a portable recording medium such as a DVD or CD-ROM in which the program is recorded is sold. It is also possible to store the program in a storage device of a server computer and transfer the program from the server computer to another computer via a network.
- the computer that executes the program stores, for example, the program recorded on the portable recording medium or the program transferred from the server computer in its own storage device. Then, the computer reads the program from its own storage device and executes processing according to the program. The computer can also read the program directly from the portable recording medium and execute processing according to the program. Further, each time the program is transferred from the server computer, the computer can sequentially execute processing according to the received program.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
- Image Input (AREA)
- Collating Specific Patterns (AREA)
Abstract
Description
たとえば、生体の特徴検出部位を適切な位置関係へと誘導する認証装置は、図15に示すような従来の認証制御処理をおこなう。
撮像データ生成手段は、生体を撮像し、単色を濃度順に割り当てたインデックスカラーを配列した単色パレットを有する撮像データを生成する。表示データ生成手段は、所定の背景境界閾値で分割した配列のうち一方を、背景色を割り当てたインデックスカラーを配列した背景表示用配列とし、他方を、特定色を濃度順に割り当てたインデックスカラーを配列した生体表示用配列とするカラーパレットで、撮像データが有する単色パレットを差し替えて表示データを生成する。表示手段は、生体を表示データで表示する。
図1は、実施形態の認証システムの構成を示す図である。実施形態として、認証システム1が手のひらの静脈を用いて認証をおこなうシステムを例示するが、これに限らず生体のその他の特徴検出部位で認証をおこなうシステムにも適用可能である。
センサユニット20は、撮像装置24を搭載する。撮像装置24は、中央部に図示しない赤外センサ(たとえば、CMOS(Complementary Metal Oxide Semiconductor)センサ)と集光レンズ23と、距離センサ22とを備え、撮像方向に向かって複数の近赤外線発光素子(LED)21が備えられる。近赤外線発光素子21は、たとえば、赤外センサの周囲8ケ所に設けられ、近赤外線を上方(撮像方向)に向けて発光する。
登録処理115は、ユーザ登録をおこなう場合に、データベースサーバ40の記憶部118に撮影画像、撮影条件、生体特徴データなどを登録する。なお、記憶部118は、ICカード31が備えるものであってもよい。
認証装置10は、処理装置11、ディスプレイ12、キーボード13、マウス14、センサユニット20、ICカードリーダライタ30を備える。
以上のようなハードウェア構成によって、本実施の形態の処理機能を実現することができる。なお、データベースサーバ40も同様のハードウェア構成で実現できる。
撮影画像データは、8ビット(256色)非圧縮DIB(Device-Independent Bitmap)のデータ構造300を有する。8ビット(256色)非圧縮DIBのデータ構造300は、ビットマップ情報ヘッダと、カラーパレットと、画素データとから構成される。
カラーパレットは、たとえば、256色が指定されている場合、インデックスカラー[0]からインデックスカラー[255]までの配列構造を有し、各配列は、配列要素「B」、「G」、「R」、「-」をデータとして持つ。配列要素「B」は、1バイトのデータでピクセルのB(青)輝度を256階調で表現するデータを格納する。配列要素「G」は、1バイトのデータでピクセルのG(緑)輝度を256階調で表現するデータを格納する。配列要素「R」は、1バイトのデータでピクセルのR(赤)輝度を256階調で表現するデータを格納する。配列要素「-」は、1バイトの予約データで「0」を格納する。
次に、元画像DIBのデータ構成について図7、図8を用いて説明する。図7は、実施形態の元画像DIBのデータ構成例を示す図である。図8は、実施形態の256階調グレースケールパレットのRGB階調グラフ例を示す図である。
次に、表示画像DIBのデータ構成について図9、図10を用いて説明する。図9は、実施形態の表示画像DIBのデータ構成例を示す図である。図10は、実施形態の手のひら静脈GUI専用カラーパレットのRGB階調グラフ例を示す図である。
誘導GUI表示処理112は、撮影画像取得処理111から取得した撮影画像にもとづいて、手のひらを適正位置に誘導するための誘導GUI表示をディスプレイ12に表示する処理である。
[ステップS11]処理装置11は、初期化処理として専用カラーパレット(手のひら静脈GUI専用カラーパレット)作成処理を実行し、手のひら静脈GUI専用カラーパレットを作成する。専用カラーパレット作成処理は、後で図12を用いて詳細に説明する。
[ステップS13]処理装置11は、誘導情報320を取得する。ここで、取得される撮影画像322は、元画像DIB301のデータ構造を有し、256階調グレースケールパレットが設定されている。
[ステップS15]処理装置11は、撮影画像取得処理111から取得した撮影画像に手のひら静脈GUI専用カラーパレットを設定した表示画像により、手のひらを適正位置に誘導するための誘導GUI表示をおこない、ステップS12にすすむ。
専用カラーパレット作成処理は、手のひら静脈GUI専用カラーパレットを作成する処理である。
[ステップS21]処理装置11は、カラーパレットのインデックスカラーの各要素「B」、「G」、「R」の値を初期化する。各要素「B」、「G」、「R」は、あらかじめ設定した最大階調値に初期化する。たとえば、処理装置11は、要素「B」=「210」、要素「G」=「250」、要素「R」=「255」を初期値に設定する。この色は、薄橙色であり、手のひらを表現する色として違和感が少ないと考えられるが、国や地域により違和感を生じる場合もある。そのため、初期値は、環境や個人の嗜好などによって任意に設定可能としてよい。
[ステップS24]処理装置11は、配列番号が背景境界閾値以下か判定する。処理装置11は、配列番号が背景境界閾値以下である場合、ステップS25にすすむ。一方、処理装置11は、配列番号が背景境界閾値以下でない場合、ステップS26にすすむ。
[ステップS26]処理装置11は、配列番号の「n」を「2」で除した剰余が「0」か否かを判定する。言い換えれば、処理装置11は、配列番号の「n」が「2」で割り切れるか否かを判定する。処理装置11は、配列番号の「n」を「2」で除した剰余が「0」の場合、ステップS27にすすむ。一方、処理装置11は、配列番号の「n」を「2」で除した剰余が「0」でない場合、ステップS29にすすむ。
さらに、上述の実施の形態は、多数の変形、変更が当業者にとって可能であり、説明した正確な構成および応用例に限定されるものではない。
2 ネットワーク
10 認証装置
11 処理装置
20 センサユニット
40 データベースサーバ
50 撮影画像
51 撮影画像用パレット
52 表示画像
53 表示画像用パレット
54 背景境界閾値
55 背景用パレット
56 手のひら用パレット
Claims (9)
- 生体の特徴を利用して個人認証する処理をコンピュータに実行させるための認証プログラムであって、
前記コンピュータを、
前記生体を撮像し、単色を濃度順に割り当てたインデックスカラーを配列した単色パレットを有する撮像データを生成する撮像データ生成手段、
所定の背景境界閾値で分割した前記配列のうち一方を、背景色を割り当てたインデックスカラーを配列した背景表示用配列とし、他方を、特定色を濃度順に割り当てたインデックスカラーを配列した生体表示用配列とするカラーパレットで、前記撮像データが有する前記単色パレットを差し替えて表示データを生成する表示データ生成手段、
前記生体を前記表示データで表示する表示手段、
として機能させることを特徴とする認証プログラム。 - 前記単色パレットが配列するインデックスカラーの濃度順と、前記生体表示用配列が配列するインデックスカラーの濃度順とは、逆向きであることを特徴とする請求の範囲第1項記載の認証プログラム。
- 前記生体表示用配列が配列するインデックスカラーの濃度幅は、前記単色パレットが配列するインデックスカラーの濃度幅より小さいことを特徴とする請求の範囲第2項記載の認証プログラム。
- 前記背景表示用配列のインデックスカラーのうち前記生体表示用配列と隣接するインデックスカラーと、前記生体表示用配列のインデックスカラーのうち前記背景表示用配列と隣接するインデックスカラーとは、互いに補色の関係にあることを特徴とする請求の範囲第2項記載の認証プログラム。
- 前記背景色は、単色であることを特徴とする請求の範囲第4項記載の認証プログラム。
- 前記単色パレットは、グレースケールパレットであることを特徴とする請求の範囲第1項記載の認証プログラム。
- 前記生体表示用配列が配列する前記特定色の濃度順とは、逆向きであることを特徴とする請求の範囲第1項乃至請求の範囲第6項のいずれか1項に記載の認証プログラム。
- 生体の特徴を利用して個人認証する認証装置において、
前記生体を撮像し、単色を濃度順に割り当てたインデックスカラーを配列した単色パレットを有する撮像データを生成する撮像データ生成手段と、
所定の背景境界閾値で分割した前記配列のうち一方を、背景色を割り当てたインデックスカラーを配列した背景表示用配列とし、他方を、特定色を濃度順に割り当てたインデックスカラーを配列した生体表示用配列とするカラーパレットで、前記撮像データが有する前記単色パレットを差し替えて表示データを生成する表示データ生成手段と、
前記生体を前記表示データで表示する表示手段と、
を備えることを特徴とする認証装置。 - 生体の特徴を利用して個人認証する認証方法において、
前記生体を撮像し、単色を濃度順に割り当てたインデックスカラーを配列した単色パレットを有する撮像データを生成し、
所定の背景境界閾値で分割した前記配列のうち一方を、背景色を割り当てたインデックスカラーを配列した背景表示用配列とし、他方を、特定色を濃度順に割り当てたインデックスカラーを配列した生体表示用配列とするカラーパレットで、前記撮像データが有する前記単色パレットを差し替えて表示データを生成し、
前記生体を前記表示データで表示することを特徴とする認証方法。
Priority Applications (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201080064134XA CN102782718A (zh) | 2010-02-24 | 2010-02-24 | 认证装置、认证程序以及认证方法 |
KR1020127021355A KR101425293B1 (ko) | 2010-02-24 | 2010-02-24 | 인증 장치, 인증 프로그램을 기록한 컴퓨터 판독 가능한 기록 매체, 및 인증 방법 |
PCT/JP2010/052882 WO2011104833A1 (ja) | 2010-02-24 | 2010-02-24 | 認証装置、認証プログラム、および認証方法 |
EP10846498A EP2541492A1 (en) | 2010-02-24 | 2010-02-24 | Authentication device, authentication program, and authentication method |
JP2012501570A JP5503725B2 (ja) | 2010-02-24 | 2010-02-24 | 認証装置、認証プログラム、および認証方法 |
BR112012020872A BR112012020872A2 (pt) | 2010-02-24 | 2010-02-24 | sistema de autenticação, programa de autenticação e método de autenticação |
US13/568,686 US8514240B2 (en) | 2010-02-24 | 2012-08-07 | Authentication system, authentication program, and method of authentication |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2010/052882 WO2011104833A1 (ja) | 2010-02-24 | 2010-02-24 | 認証装置、認証プログラム、および認証方法 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/568,686 Continuation US8514240B2 (en) | 2010-02-24 | 2012-08-07 | Authentication system, authentication program, and method of authentication |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2011104833A1 true WO2011104833A1 (ja) | 2011-09-01 |
Family
ID=44506280
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2010/052882 WO2011104833A1 (ja) | 2010-02-24 | 2010-02-24 | 認証装置、認証プログラム、および認証方法 |
Country Status (7)
Country | Link |
---|---|
US (1) | US8514240B2 (ja) |
EP (1) | EP2541492A1 (ja) |
JP (1) | JP5503725B2 (ja) |
KR (1) | KR101425293B1 (ja) |
CN (1) | CN102782718A (ja) |
BR (1) | BR112012020872A2 (ja) |
WO (1) | WO2011104833A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2019050024A (ja) * | 2012-11-14 | 2019-03-28 | ゴラン ウェイス | 登録および認証のための生体認証方法およびシステム |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8548206B2 (en) | 2011-01-20 | 2013-10-01 | Daon Holdings Limited | Methods and systems for capturing biometric data |
US8823726B2 (en) | 2011-02-16 | 2014-09-02 | Apple Inc. | Color balance |
US8854370B2 (en) * | 2011-02-16 | 2014-10-07 | Apple Inc. | Color waveform |
CN104969543B (zh) * | 2013-02-14 | 2019-04-02 | 松下知识产权经营株式会社 | 电子镜装置 |
US9269022B2 (en) | 2013-04-11 | 2016-02-23 | Digimarc Corporation | Methods for object recognition and related arrangements |
TW201445358A (zh) * | 2013-05-27 | 2014-12-01 | Pixart Imaging Inc | 導引裝置及其觸發安全設定之方法 |
US9158959B2 (en) * | 2013-07-17 | 2015-10-13 | Motorola Solutions, Inc. | Palm identification and in-place personalized interactive display |
US9489765B2 (en) | 2013-11-18 | 2016-11-08 | Nant Holdings Ip, Llc | Silhouette-based object and texture alignment, systems and methods |
KR102212064B1 (ko) * | 2013-12-06 | 2021-02-05 | 삼성전자주식회사 | 냉장고에 의해 헬스 서비스를 제공하는 방법 및 이를 위한 냉장고 |
CN110279402A (zh) * | 2019-07-31 | 2019-09-27 | 杭州泽铭睿股权投资有限公司 | 一种皮下血管光学视频图像的成像方法 |
JP2022156047A (ja) * | 2021-03-31 | 2022-10-14 | 京セラドキュメントソリューションズ株式会社 | 画像形成装置 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11203347A (ja) * | 1998-01-20 | 1999-07-30 | Toyobo Co Ltd | 試着状況を示す画像の合成方法および該方法を用いた仮想試着システム |
JP2001216497A (ja) * | 2000-01-31 | 2001-08-10 | Fuji Photo Film Co Ltd | 画像解析装置 |
JP2008071158A (ja) * | 2006-09-14 | 2008-03-27 | Fujitsu Ltd | 生体認証装置の生体誘導制御方法及び生体認証装置 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6310627B1 (en) | 1998-01-20 | 2001-10-30 | Toyo Boseki Kabushiki Kaisha | Method and system for generating a stereoscopic image of a garment |
US7065242B2 (en) * | 2000-03-28 | 2006-06-20 | Viewpoint Corporation | System and method of three-dimensional image capture and modeling |
JP3967650B2 (ja) | 2002-09-20 | 2007-08-29 | 朝日レントゲン工業株式会社 | 頭部x線規格画像の計測処理方法および計測処理装置ならびに計測処理プログラムを記録したコンピュータ読み取り可能な記録媒体 |
DE10347123B4 (de) * | 2003-10-02 | 2005-07-28 | Petra Dr.-Ing. Perner | Verfahren zur automatischen Bestimmung auf einem Träger gesammelter luftgetragener Partikel insbesondere Pilzsporen |
JP4515850B2 (ja) | 2004-07-30 | 2010-08-04 | 富士通株式会社 | 生体認証装置の誘導画面制御方法、生体認証装置及びそのプログラム |
US20080247649A1 (en) * | 2005-07-07 | 2008-10-09 | Chun Hing Cheng | Methods For Silhouette Extraction |
JP5231759B2 (ja) | 2007-06-25 | 2013-07-10 | 富士フイルム株式会社 | 撮像装置及び撮像方法 |
JP2011085536A (ja) * | 2009-10-16 | 2011-04-28 | Nuflare Technology Inc | レビュー装置および検査装置システム |
-
2010
- 2010-02-24 JP JP2012501570A patent/JP5503725B2/ja active Active
- 2010-02-24 BR BR112012020872A patent/BR112012020872A2/pt not_active IP Right Cessation
- 2010-02-24 WO PCT/JP2010/052882 patent/WO2011104833A1/ja active Application Filing
- 2010-02-24 EP EP10846498A patent/EP2541492A1/en not_active Withdrawn
- 2010-02-24 CN CN201080064134XA patent/CN102782718A/zh active Pending
- 2010-02-24 KR KR1020127021355A patent/KR101425293B1/ko active IP Right Grant
-
2012
- 2012-08-07 US US13/568,686 patent/US8514240B2/en not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11203347A (ja) * | 1998-01-20 | 1999-07-30 | Toyobo Co Ltd | 試着状況を示す画像の合成方法および該方法を用いた仮想試着システム |
JP2001216497A (ja) * | 2000-01-31 | 2001-08-10 | Fuji Photo Film Co Ltd | 画像解析装置 |
JP2008071158A (ja) * | 2006-09-14 | 2008-03-27 | Fujitsu Ltd | 生体認証装置の生体誘導制御方法及び生体認証装置 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2019050024A (ja) * | 2012-11-14 | 2019-03-28 | ゴラン ウェイス | 登録および認証のための生体認証方法およびシステム |
Also Published As
Publication number | Publication date |
---|---|
EP2541492A1 (en) | 2013-01-02 |
CN102782718A (zh) | 2012-11-14 |
US8514240B2 (en) | 2013-08-20 |
JP5503725B2 (ja) | 2014-05-28 |
KR20120124066A (ko) | 2012-11-12 |
BR112012020872A2 (pt) | 2016-05-03 |
JPWO2011104833A1 (ja) | 2013-06-17 |
US20120299949A1 (en) | 2012-11-29 |
KR101425293B1 (ko) | 2014-07-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5503725B2 (ja) | 認証装置、認証プログラム、および認証方法 | |
EP2866170B1 (en) | Image processing device and image processing method | |
US9294475B2 (en) | System and method for generating a biometric identifier | |
US8275174B2 (en) | Vein pattern management system, vein pattern registration apparatus, vein pattern authentication apparatus, vein pattern registration method, vein pattern authentication method, program, and vein data configuration | |
WO2012111664A1 (ja) | 認証装置、認証プログラム、および認証方法 | |
US8320997B2 (en) | Vein authentication apparatus and template registration method | |
JP5622928B2 (ja) | 照合装置、照合プログラム、および照合方法 | |
EP3362942B1 (en) | Electronic devices with improved iris recognition and methods thereof | |
KR101014325B1 (ko) | 적외선을 이용한 얼굴인식 시스템 및 방법 | |
JP5509335B2 (ja) | 登録プログラム、登録装置、および登録方法 | |
KR20040059313A (ko) | 치아영상으로부터 치아영역 추출방법 및 치아영상을이용한 신원확인방법 및 장치 | |
JP6799525B2 (ja) | 生体情報解析装置及び手肌解析方法 | |
US20220043895A1 (en) | Biometric authentication system, biometric authentication method, and program | |
US11315360B2 (en) | Live facial recognition system and method | |
EP2148295A1 (en) | Vein pattern management system, vein pattern registration device, vein pattern authentication device, vein pattern registration method, vein pattern authentication method, program, and vein data structure | |
WO2019163066A1 (ja) | なりすまし検知装置、なりすまし検知方法、及びコンピュータ読み取り可能な記録媒体 | |
US20090290763A1 (en) | Vein authentication apparatus and vein authentication method | |
CN111259757B (zh) | 一种基于图像的活体识别方法、装置及设备 | |
KR101336834B1 (ko) | Usb 홍채 인식기 | |
JP6629150B2 (ja) | 手のひら検知装置、掌紋認証装置、手のひら検知方法、及びプログラム | |
JP6621383B2 (ja) | 個人認証装置 | |
KR100973832B1 (ko) | 영상 선택 방법, 장치 및 그 방법을 실행하는 프로그램이기록된 기록매체 | |
JP2018092272A (ja) | 生体認証装置、生体認証方法及びプログラム | |
KR20210050649A (ko) | 모바일 기기의 페이스 인증 방법 | |
JP5685272B2 (ja) | 認証装置、認証プログラム、および認証方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201080064134.X Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10846498 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012501570 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2059/KOLNP/2012 Country of ref document: IN |
|
ENP | Entry into the national phase |
Ref document number: 20127021355 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010846498 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112012020872 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 112012020872 Country of ref document: BR Kind code of ref document: A2 Effective date: 20120820 |