WO2014125831A1 - 電子ミラー装置 - Google Patents
電子ミラー装置 Download PDFInfo
- Publication number
- WO2014125831A1 WO2014125831A1 PCT/JP2014/000758 JP2014000758W WO2014125831A1 WO 2014125831 A1 WO2014125831 A1 WO 2014125831A1 JP 2014000758 W JP2014000758 W JP 2014000758W WO 2014125831 A1 WO2014125831 A1 WO 2014125831A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- user
- display
- camera
- live video
- Prior art date
Links
- 238000000034 method Methods 0.000 claims description 24
- 230000002194 synthesizing effect Effects 0.000 claims description 6
- 230000008569 process Effects 0.000 claims description 5
- 239000000284 extract Substances 0.000 claims description 4
- 238000003384 imaging method Methods 0.000 abstract description 2
- 238000010586 diagram Methods 0.000 description 11
- 238000012545 processing Methods 0.000 description 10
- 239000000470 constituent Substances 0.000 description 6
- 230000001815 facial effect Effects 0.000 description 5
- 238000004590 computer program Methods 0.000 description 4
- 239000002537 cosmetic Substances 0.000 description 4
- 210000003128 head Anatomy 0.000 description 3
- 238000005286 illumination Methods 0.000 description 3
- 239000011148 porous material Substances 0.000 description 3
- 238000003672 processing method Methods 0.000 description 3
- 150000001875 compounds Chemical class 0.000 description 2
- 238000012937 correction Methods 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 150000004032 porphyrins Chemical class 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 208000002874 Acne Vulgaris Diseases 0.000 description 1
- 241001465754 Metazoa Species 0.000 description 1
- 206010000496 acne Diseases 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 210000000795 conjunctiva Anatomy 0.000 description 1
- 210000004087 cornea Anatomy 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 210000005069 ears Anatomy 0.000 description 1
- 238000003708 edge detection Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 210000000887 face Anatomy 0.000 description 1
- 230000001678 irradiating effect Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000007665 sagging Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000011282 treatment Methods 0.000 description 1
- 230000037303 wrinkles Effects 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/74—Details of notification to user or communication with user or patient ; user input means
- A61B5/742—Details of notification to user or communication with user or patient ; user input means using visual displays
- A61B5/7425—Displaying combinations of multiple images regardless of image source, e.g. displaying a reference anatomical image with a live image
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0059—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
- A61B5/0077—Devices for viewing the surface of the body, e.g. camera, magnifying lens
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7246—Details of waveform analysis using correlation, e.g. template matching or determination of similarity
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/60—Rotation of whole images or parts thereof
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/24—Aligning, centring, orientation detection or correction of the image
- G06V10/242—Aligning, centring, orientation detection or correction of the image by image rotation, e.g. by 90 degrees
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
- H04N23/611—Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/64—Computer-aided capture of images, e.g. transfer from script file into camera, check of taken image quality, advice or proposal for image composition or decision on when to take image
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/95—Computational photography systems, e.g. light-field imaging systems
- H04N23/951—Computational photography systems, e.g. light-field imaging systems by using two or more images to influence resolution, frame rate or aspect ratio
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/2621—Cameras specially adapted for the electronic generation of special effects during image pickup, e.g. digital cameras, camcorders, video cameras having integrated special effects capability
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2503/00—Evaluating a particular growth phase or type of persons or animals
- A61B2503/12—Healthy persons not otherwise provided for, e.g. subjects of a marketing survey
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0059—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
- A61B5/0075—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence by spectroscopy, i.e. measuring spectra, e.g. Raman spectroscopy, infrared absorption spectroscopy
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/117—Identification of persons
- A61B5/1171—Identification of persons based on the shapes or appearances of their bodies or parts thereof
- A61B5/1176—Recognition of faces
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/44—Detecting, measuring or recording for evaluating the integumentary system, e.g. skin, hair or nails
- A61B5/441—Skin evaluation, e.g. for skin disorder diagnosis
- A61B5/445—Evaluating skin irritation or skin trauma, e.g. rash, eczema, wound, bed sore
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7264—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
Definitions
- the present invention relates to an electronic mirror device.
- Face information is detailed information such as moisture content, oil content, pores, texture, stains, wrinkles, porphyrins, sagging, skin color irregularities, face lines, and overall facial information such as makeup finish including.
- the past image and the newly captured image should be captured under the same conditions as much as possible. preferable. If the color and brightness of the illumination at the time of shooting, the positional relationship between the illumination and the face, and the positional relationship between the camera and the face are different, accurate comparison and evaluation may not be possible.
- Patent Document 1 In order to capture a face image under the same conditions as possible, in Patent Document 1, the position of the face and the imaging camera is always kept constant by a face holder, the face is surrounded by a light diffusing plate to uniformly illuminate the face, and disturbance A method for eliminating the influence of light has been proposed.
- a method for determining whether or not the head is at the reference position by detecting the top of the head for example, see Patent Document 2
- detecting facial features such as eyes, ears, and nose
- Patent Document 3 a method for detecting a shift amount compared with reference waveform data in a face image
- Patent Document 4 a method for detecting a shift amount compared with reference waveform data in a face image
- the present invention provides an electronic mirror device that performs simple and good photographing suitable for comparing a previously photographed image with a newly photographed image.
- An electronic mirror device includes a camera that captures a user, a display that has a display surface at a location that can be visually recognized by the user that the camera captures, and an image of the user that is captured by the camera.
- a controller that outputs to a mirror image of the user's live video captured by the camera, and reads a reference image contrasted with the mirror image of the user's live video,
- a display image for displaying a mirror image of a live image and the reference image in an overlapping area of the display is synthesized and output to the display.
- An electronic mirror device includes a camera that captures a user, a display that has a display surface at a position that can be visually recognized by the user that the camera captures, and the user's image captured by the camera.
- a controller that outputs to a mirror image of the user's live video captured by the camera, and reads a reference image that is contrasted with the mirror image of the user's live video, The degree of coincidence between the live video and the reference image is extracted, and the mirror image of the live video and the extracted image showing the degree of coincidence are combined and output to the display.
- An electronic mirror device includes a camera that captures a user, a display that has a display surface at a position that can be visually recognized by the user that the camera captures, and the user's image captured by the camera. And a controller that outputs the mirror image of the user's live image captured by the camera to a mirror image, and based on the user's live image captured by the camera.
- First information related to at least one of the position and the face orientation of the user is continuously acquired, and the position and the face orientation of the user in the reference image compared with the mirror image of the live video of the user
- Second information related to at least one of the first information and the first information is compared with the second information to extract a degree of coincidence.
- An electronic mirror device includes a camera that captures an image from which parallax can be extracted, a display having a display surface at a position that can be visually recognized by a user who captures the camera, and an image captured by the camera.
- a controller for outputting to a display; and a memory, wherein the controller outputs the live video of the user captured by the camera to the display as a mirror image that is horizontally reversed, and the camera captures the image at a predetermined timing.
- an electronic mirror device that performs simple and good photographing suitable for comparing a previously photographed image with a newly photographed image.
- FIG. 1 is a diagram showing an example of the appearance of an electronic mirror device according to Embodiment 1 of the present invention.
- FIG. 2 is a block diagram showing an example of a functional configuration of the electronic mirror device according to Embodiment 1 of the present invention.
- FIG. 3 is a conceptual diagram illustrating an example of use of the electronic mirror device according to Embodiment 1 of the present invention.
- FIG. 4 is a flowchart showing an example of a basic operation of the electronic mirror device according to Embodiment 1 of the present invention.
- FIG. 5 is a flowchart showing an example of detailed operation of the electronic mirror device according to Embodiment 1 of the present invention.
- FIG. 6 is a diagram showing a display example of the electronic mirror device according to Embodiment 1 of the present invention.
- FIG. 1 is a diagram showing an example of the appearance of an electronic mirror device according to Embodiment 1 of the present invention.
- FIG. 2 is a block diagram showing an example of a functional configuration of the electronic mirror device
- FIG. 7 is an external view of the electronic mirror device according to Embodiment 1 of the present invention.
- FIG. 8 is a diagram showing a display example of the electronic mirror device according to Embodiment 2 of the present invention.
- FIG. 9 is a diagram showing a display example of the electronic mirror device according to Embodiment 3 of the present invention.
- FIG. 10 is a diagram showing a display example of the electronic mirror device according to Embodiment 4 of the present invention.
- FIG. 11 is a flowchart showing an example of detailed operation of the electronic mirror device according to Embodiment 5 of the present invention.
- FIG. 12 is a diagram showing an example of detailed operation of the electronic mirror device according to Embodiment 5 of the present invention.
- FIG. 13 is a flowchart showing an example of detailed operation of the electronic mirror device according to Embodiment 6 of the present invention.
- FIG. 14 is a flowchart showing an example of detailed operation of the electronic mirror device according to Embodiment 7 of the present invention.
- An electronic mirror device includes a camera that captures a user, a display that has a display surface at a location that can be visually recognized by the user that the camera captures, and an image of the user that is captured by the camera.
- a controller that outputs to a mirror image of the user's live video captured by the camera, and reads a reference image contrasted with the mirror image of the user's live video,
- a display image for displaying a mirror image of a live image and the reference image in an overlapping area of the display is synthesized and output to the display.
- the image processing apparatus further includes a memory, wherein the memory records one frame of a live video previously captured by the camera as the reference image, and the controller reads the reference image from the memory,
- the display image may be synthesized from the mirror image of the reference image read from the memory and the mirror image of the live image and output to the display.
- the reference image may be one frame of the user's live video taken in the past by the camera.
- the controller may read the reference image from an external medium, synthesize the display image from a mirror image of the reference image read from the external medium and a mirror image of the live image, and output the synthesized image to the display. .
- controller may process the reference image to be translucent, synthesize the display video from the translucent reference image and the live video, and output the synthesized video to the display.
- controller may extract an outline of the reference image, synthesize the display video from the image representing the outline and the live video, and output the synthesized video to the display.
- the controller may detect a feature point of the reference image, synthesize the display image from the image representing the position of the feature point and the live image, and output the synthesized image to the display.
- image information for supporting alignment can be provided to the user using various display modes.
- An electronic mirror device includes a camera that captures a user, a display that has a display surface at a position that can be visually recognized by the user that the camera captures, and an image of the user that is captured by the camera.
- a controller that outputs to the display, and the controller reverses the live video of the user captured by the camera to a mirror image, and compares the reference image with the mirror image of the user live video. Reading, extracting the degree of coincidence between the live video and the reference image, synthesizing the mirror image of the live video and the image indicating the degree of coincidence extracted and outputting to the display.
- the user can check the live video together with the matching degree with the reference image. Therefore, it is possible to release the shutter at a timing that is closer to the reference image.
- the controller is represented by the reference image and a pixel that is inside the face contour represented by the live video.
- the degree of coincidence may be extracted based on the sum of luminance differences for each pixel with respect to the pixels inside the face outline.
- an electronic mirror device includes a camera that captures a user, a display that has a display surface at a location that can be visually recognized by the user that the camera captures, and an image of the user that the camera captures.
- a controller that outputs to the display, and the controller reverses the live video of the user captured by the camera to a mirror image, and is based on the live image of the user captured by the camera.
- First information related to at least one of the user's position and face orientation is continuously acquired, and the user's position and face orientation in a reference image compared with the mirror image of the user's live video
- Second information related to at least one of the two is acquired, and the degree of coincidence is extracted by comparing the first information with the second information , By synthesizing the image showing the degree of coincidence the mirror image and the extracted of the live video output to the display.
- the user can check the live video together with the matching degree with the reference image. Therefore, it is possible to release the shutter at a timing that is closer to the reference image.
- the memory further includes a memory, and the memory is related to at least one of the position and the face orientation of the user in the reference image, with one frame of a live video taken in the past by the camera as the reference image.
- the second information is recorded, and the controller reads the second information from the memory, compares the first information with the second information read from the memory, and matches the second information. The degree may be extracted.
- an electronic mirror device includes a camera that captures an image from which parallax can be extracted, a display that has a display surface at a position that can be viewed by a user that the camera captures, and an image that is captured by the camera. And a memory, and the controller outputs the live video of the user captured by the camera to the display as a mirror image that is reversed left and right.
- One frame of the user's live video being shot is recorded in the memory, the parallax is extracted based on the user image of the one frame, and the distance between the camera and the user is calculated based on the parallax. The calculated distance and the image of the one frame are recorded in the memory in association with each other.
- the memory records definition information that defines a gesture, and the controller determines whether a user in the live video has performed a gesture defined by the definition information recorded in the storage unit. When it is determined that the user has made the gesture, one frame in the live video may be recorded in the memory.
- the memory may record color sample data, and the controller may adjust the color balance of the live image based on the color sample data recorded in the storage unit. .
- FIG. 1 is a diagram illustrating an example of an appearance of the electronic mirror device 1 according to the first embodiment.
- the electronic mirror device 1 according to the first embodiment includes a camera 2 that captures a user and a display 3.
- the user's live video captured by the camera 2 can be displayed on the display 3.
- a live video is a moving image represented by a sequence of still image frames (also simply referred to as frames or pictures).
- one frame of the live video may be referred to as a live image.
- the display 3 has a display surface at a position where a user who takes a picture with the camera 2 can visually recognize.
- the electronic mirror device 1 has a memory 4 and a controller 5 inside the housing.
- FIG. 2 is a block diagram illustrating an example of a functional configuration of the electronic mirror device 1 according to the first embodiment.
- the memory 4 records a reference image that is contrasted with a user image captured by the camera 2.
- the reference image may be, for example, one frame of a live video taken in the past by the camera 2.
- the reference image can be called from the memory 4 and displayed on the display 3.
- the controller 5 synthesizes the display image from the live image acquired by the camera 2 and the reference image recorded in the memory 4 using image processing such as inversion, rotation, and color correction.
- the synthesized display image can be displayed on the display 3.
- controller 5 reads the reference image from the memory 4.
- the reference image may be read from the controller 5 or an external medium.
- the controller 5 synthesizes the display image from the mirror image of the reference image and the mirror image of the live image read from the memory 4 or an external medium.
- the synthesized display image is output and displayed on the display 3.
- FIG. 3 is a schematic diagram showing the positional relationship between the electronic mirror device 1 and the user.
- the electronic mirror device 1 is arranged in front of the paper and the back of the paper.
- FIG. 4 is a flowchart showing an example of a basic operation of the electronic mirror device 1.
- the user's image A is acquired by the camera 2 (S01, S02), and the controller 5 inverts the image A horizontally to generate an image B and outputs it to the display 3 (S03, S04).
- the display 3 displays the image B (S05).
- the images A and B constitute one frame of the user's live video.
- the image A and the image B can be recorded and stored in the memory 4 and can be printed, transmitted to data, or moved to an external memory.
- the user can record and save in the memory 4 as a reference image one frame in a state that he / she prefers while viewing the live video visually.
- Reference images stored in the past can be recalled to contrast the user in the live video currently being captured.
- the characteristic support operation performed by the electronic mirror device 1 will be described with reference to the flowchart of FIG. This will be described based on a display example.
- the controller 5 acquires the reference image recorded in the memory 4 (S11).
- the reference image may be, for example, an image captured in the past by the camera 2 and recorded in the memory 4.
- the controller 5 processes the acquired reference image (S12).
- the processing of the reference image may include reduction, left / right inversion, and the like.
- the controller 5 acquires a live video of the user currently captured by the camera 2 (S13).
- the frame rate of live video is not limited, but as a practical example, it may be several frames to several tens of frames per second.
- the controller 5 processes the acquired live video (S14). Live video processing may include reduction, left-right reversal, color correction, and the like.
- the controller 5 combines the display image from the reference image and the mirror image of the live image (S15), and displays the combined display image on the display 3 (S16).
- the method for synthesizing the display video is not limited, as an example, the controller 5 may synthesize a display video for simultaneously displaying the reference image and the live video in a non-overlapping area of the display 3.
- the reference image A ⁇ b> 1 is horizontally reversed and reduced and displayed in a small area in the upper left of the display 3, and the live image B ⁇ b> 1 is horizontally reversed and the rest of the display 3 excluding the small area. It may be displayed in the area.
- the user adjusts the face direction while viewing such a display image displayed on the display 3, and determines that the reference image A1 and the live image B1 have the same face direction.
- the frame can be recorded and stored in the memory 4 (S17).
- the electronic mirror device 1 can support simple and good alignment when comparing the reference image and the live video.
- the operation panel (not shown) of the electronic mirror device 1 may be a contact type such as a button or a touch panel, or may be a non-contact type.
- a human sensor or the like may be used, or an image processing technique may be used.
- the definition information that defines the user's gesture and the operation of the electronic mirror device 1 are recorded in advance in the memory 4 and the recorded definition information is used.
- the electronic mirror device 1 may perform an associated operation.
- the non-contact operation method using gestures has the following advantages, for example.
- an ultraviolet light source may be turned on and an image may be recorded and stored after a few seconds have elapsed since the user performed a gesture to close the eyes. What is necessary is just to set the time which closes the eye used as a trigger in the time longer than blink.
- the user can be notified by voice of a countdown until recording and saving and a cue to notify the end of shooting such as a shutter sound.
- the non-contact type operation method by gesture is also useful in such a case.
- the influence of the color and brightness of the illumination may be reduced, and the past and present skin colors may be easily compared.
- the color balance may be adjusted by, for example, storing color sample data in the electronic mirror device 1 and photographing the actual color sample corresponding to the color sample data with the camera 2.
- a dedicated touch pen 6 with a color sample 7 attached is attached to the electronic mirror device 1, the color sample 7 can be easily photographed to adjust the color balance.
- a color sample data corresponding to the color sample 7 is recorded in the memory 4 in advance. Compared to the case where a color sample such as a color chart is prepared and photographed, it is possible to reduce the troublesomeness that makes you feel that you are doing something special.
- the reference image is not necessarily read from the memory 4 provided in the electronic mirror device 1.
- the controller 5 may read a reference image from an external medium.
- the reference image indicates the position of at least one feature point among an image photographed by a camera other than the camera 2 of the electronic mirror device 1 and a face outline, eyes, nose, and mouth imitating a human face. Illustrations, sample images of makeup published in magazines and Internet sites, etc. can be considered.
- a structure for physically connecting to the external medium may be provided at the end of the electronic mirror device 1 in order to read the reference image from the external medium.
- the reference image may be read wirelessly from an external medium.
- the past reference image A2 that has been reversed in the left and right direction obtained by the method as described above is called and processed into a translucent reference image B2.
- a display image D2 obtained by superimposing the translucent reference image B2 and the live image C2 captured by the camera 2 is synthesized and displayed in an overlapping area of the display 3.
- the synthesized display video D2 is an example of a display video for displaying a mirror image of the live video and a reference image in an overlapping area of the display, and more specifically, is represented by the mirror image of the live video.
- 5 is an example of a display video in which a user and the reference image are arranged in an overlapping area of the display.
- the user uses the translucent reference image B2 included in the display video D2 as a guide for face orientation, adjusts the distance and positional relationship between the camera 2 and himself so as to be the same face orientation, Can be recorded as a new reference image E2.
- the third embodiment is different from the first and second embodiments in the image processing method.
- the third embodiment will be described with reference to FIG.
- the past reference image A3 that has been horizontally reversed is called out, and the contour is extracted to process the reference image B3 with only the contour.
- a display image D3 obtained by superimposing the reference image B3 with only the outline and the live image C3 captured by the camera 2 is synthesized and displayed in an overlapping area of the display 3.
- the composited display video D3 is an example of a display video for displaying a mirror image of the live video and a reference image in an overlapping area of the display. More specifically, the display video D3 is represented by the mirror image of the live video. 5 is an example of a display video in which a user and the reference image are arranged in an overlapping area of the display.
- the user uses the reference image B3 of only the outline included in the display video D3 as a guide for face orientation, adjusts the distance and positional relationship between the camera 2 and himself so that the face orientation is the same, When it is determined that the orientations match, one frame of the live video C3 can be recorded as a new reference image E3.
- the contour means not only the outer contour of the face but also the contour of the components such as eyes, nose and mouth.
- the contour may be extracted by using an edge detection method or by simply binarizing the image and adjusting the contrast and brightness.
- the threshold value and the like can be freely set by the user.
- the contour may be extracted by tracing the image by the user. What is important in the present invention is the ease of use of the user, and the accuracy of image processing is not particularly required.
- the fourth embodiment is different from the first, second, and third embodiments in the image processing method.
- the fourth embodiment will be described with reference to FIG.
- the past reference image A4 that has been horizontally reversed is called and the feature points are extracted, thereby processing the reference image B4 with only the feature points.
- a display image D4 obtained by superimposing the reference image B4 including only the feature points and the live image C4 captured by the camera 2 is synthesized and displayed in an overlapping area of the display 3.
- the composited display video D4 is an example of a display video for displaying a mirror image of the live video and a reference image in an overlapping area of the display. More specifically, the display video D4 is represented by the mirror image of the live video. 5 is an example of a display video in which a user and the reference image are arranged in an overlapping area of the display.
- the user uses the reference image B4 of only the feature points included in the display video D4 as a guide for face orientation, adjusts the distance and positional relationship between the camera 2 and itself so that the face orientation is the same, When it is determined that the face orientations match, one frame of the live video C4 can be recorded as a new reference image E4.
- the feature points include not only the eyes, nose, mouth, and contour, but also the portions that the user himself / herself wants to observe over time, such as spots and acne.
- FIG. 9 shows an example in which feature points are marked with black circles as pigmented portions of the top of the head, jaw, eyes, corners of the eyes, nose, mouth corner, and right eye area, but the definition of the feature points is not limited to this.
- the extraction of the feature points may be automatically detected by determining an algorithm, but may be specified by the user himself tracing the image.
- a reference image captured by the camera 2 in the past is compared with a live image currently captured to extract the degree of coincidence.
- An image indicating the extracted degree of coincidence is displayed on the display 3 together with the live video. That is, the image displayed on the display 3 together with the live video is an image showing the degree of coincidence between the reference image and the live video, and is different from Embodiments 1 to 4.
- the memory 4 records a reference image that is contrasted with the user's live video captured by the camera 2.
- the reference image may be, for example, one frame of a live video taken in the past by the camera 2.
- the controller 5 extracts the contour of the user's face based on the user's live video captured by the camera 2 (S51, S52). Also, a reference image to be compared with the live video is called from the memory 4, and the contour of the user's face in the reference image is extracted (S53, S54). Then, with respect to the pixels inside the live video and the face outline extracted in the reference image, the difference in luminance value is calculated for each pixel, and the sum of them is calculated (S55). Based on the calculated sum of differences in luminance values, an image indicating the degree of coincidence between the user's live image and the reference image is generated (S56). The generated image indicating the degree of coincidence and the mirror image of the live video captured by the camera 2 are combined and output (S57).
- This configuration allows the user to check the live video together with the degree of coincidence with the reference image. Therefore, the user can release the shutter at a shooting timing at which a live image is obtained under conditions closer to the reference image (that is, more suitable for comparison with the reference image) (S58).
- the reference image or the live image is enlarged or enlarged so that the size of the face outline to be extracted and the arrangement of the feature points are the same.
- the reference image or the live image is enlarged or enlarged so that the size of the face outline to be extracted and the arrangement of the feature points are the same.
- the image showing the degree of coincidence based on the luminance value difference is obtained by, for example, quantifying the sum of the luminance value differences for each pixel calculated for pixels inside the facial image extracted in the live image and the reference image, for example. It may be displayed or graphed and displayed. It is not limited to these as long as the user can check the degree of matching between the frame and the reference image in the live video.
- the difference in luminance value is calculated for each pixel for the pixels located inside the outline of the face extracted in the live video and the reference image.
- the present invention is not limited to this.
- the sum of the luminance values of the pixels inside the face outline in the live image is calculated
- the sum of the luminance values of the pixels inside the face outline in the reference image is calculated
- the calculated luminance values You may take the difference of the sum total.
- the position or face orientation of the user in the reference image previously captured by the camera 2 is compared with the position or face orientation of the user in the live image currently being captured. Extract the degree of match.
- An image indicating the extracted degree of coincidence is displayed on the display 3 together with the live video. That is, the image displayed on the display 3 together with the live video is an image showing the degree of coincidence between the reference image and the live video, and is different from Embodiments 1 to 4.
- the memory 4 records second information related to the position or face orientation of the user in the reference image contrasted with the user's live video captured by the camera 2.
- the reference image may be, for example, one frame of a live video taken in the past by the camera 2.
- the memory 4 may record the reference image itself that is contrasted with the user's live video.
- the controller 5 may read the reference image from the memory 4 and calculate the second information related to the position or face orientation of the user in the reference image.
- the controller 5 can continuously acquire the first information related to the user's position or face direction based on the user's live image taken by the camera 2 (S61, S62).
- the controller 5 can call (or calculate) the second information from the memory 4 and compare it with the first information (S63, S64). Then, the first information and the second information are compared to extract the degree of coincidence between them. The extracted image indicating the degree of coincidence and the mirror image of the live video taken by the camera 2 are combined and output (S65, S66).
- This configuration allows the user to check the live video together with the degree of coincidence with the reference image. Therefore, the user can release the shutter at a shooting timing at which a live image is obtained under conditions closer to the reference image (that is, more suitable for comparison with the reference image) (S67).
- the first information and the second information are described as information related to the user's position or face orientation.
- the first information and the second information are the user's position and face orientation. Information including both of them may be used.
- Information related to the position of the user is obtained by measuring the distance from the camera 2 to the user.
- a distance measuring technique using parallax such as a compound eye camera can be used.
- the user's face orientation can be acquired by estimating the direction of the face connecting the feature points based on the feature point positions of the faces in the captured image, for example. In addition, it can be obtained using a known technique.
- the image indicating the degree of coincidence between the first information and the second information may be displayed by quantifying the difference between the first information and the second information, or may be displayed as a graph. It is not limited to these as long as the user can check the degree of coincidence between the first information and the second information.
- the camera 2 can capture an image from which parallax can be extracted, such as a compound eye camera. The distance between the camera 2 and the user is calculated and recorded based on the parallax of the image captured by the camera 2.
- the controller 5 records one frame of the user's live video captured by the camera at the shooting timing in the memory 4 (S71, S72). Then, the parallax is extracted based on the image of one frame recorded in the memory 4 (S73). The distance between the camera 2 and the user is calculated based on the extracted parallax, and is recorded in the memory 4 in association with the captured image (S74, S75). The above operation is performed for each shooting. Thereby, for example, based on the number of pixels corresponding to spots and pores in the captured image and the recorded distance information, the actual area of the spots and pores can be estimated. Therefore, even if the shooting conditions differ between an image shot in the past and a newly shot image, comparison and evaluation can be performed based on the distance information.
- each component may be configured by dedicated hardware or may be realized by executing a software program suitable for each component.
- Each component may be realized by a program execution unit such as a CPU (Central Processing Unit) or a processor reading and executing a software program recorded on a recording medium such as a hard disk or a semiconductor memory.
- a program execution unit such as a CPU (Central Processing Unit) or a processor reading and executing a software program recorded on a recording medium such as a hard disk or a semiconductor memory.
- the user of the electronic mirror device of the present invention may be other than a human face, and is effective for observing a change over time even if it is a part other than the face of the body, another animal or plant, or even a non-living object.
- the electronic mirror device according to the present invention can be applied to a skin diagnosis system and the like.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Pathology (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Medical Informatics (AREA)
- Molecular Biology (AREA)
- Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- Biophysics (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Radiology & Medical Imaging (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physiology (AREA)
- Psychiatry (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Human Computer Interaction (AREA)
- Computing Systems (AREA)
- Studio Devices (AREA)
- Closed-Circuit Television Systems (AREA)
- Image Processing (AREA)
Abstract
Description
図1は、実施の形態1の電子ミラー装置1の外観の一例を示す図である。本実施形態1の電子ミラー装置1は、ユーザを撮影するカメラ2と、ディスプレイ3を備える。カメラ2で撮影されたユーザのライブ映像は、ディスプレイ3に表示することができる。ライブ映像は、静止画フレーム(単にフレーム、又はピクチャとも言う)の連なりで表される動画像である。以下では、ライブ映像の1つのフレームをライブ画像と称することがある。ディスプレイ3は、カメラ2が撮影するユーザが視認できる位置にその表示面を有する。電子ミラー装置1は、筐体内部に、メモリ4及びコントローラ5を有している。
実施の形態2は、画像の加工方法と表示方法が実施の形態1と異なる。本実施の形態2について、図8を参照しながら説明する。
実施の形態3は、画像の加工方法が実施の形態1および実施の形態2と異なる。本実施の形態3について、図9を参照しながら説明する。
実施の形態4は、画像の加工方法が実施の形態1、実施の形態2および実施の形態3と異なる。本実施の形態4について、図10を参照しながら説明する。
実施の形態5に係る電子ミラー装置では、過去にカメラ2で撮影した参照画像と、現在撮影しているライブ映像とを比較して一致度合いを抽出する。抽出した一致度合いを示す画像を、ライブ映像とともにディスプレイ3に表示する。すなわち、ライブ映像とともにディスプレイ3に表示させる画像が、参照画像とライブ映像との一致度合いを示す画像である点で、実施の形態1~4と異なる。
実施の形態6に係る電子ミラー装置では、過去にカメラ2で撮影した参照画像中のユーザの位置または顔向きと、現在撮影しているライブ映像中のユーザの位置または顔向きとを比較して一致度合いを抽出する。抽出した一致度合いを示す画像を、ライブ映像とともにディスプレイ3に表示する。すなわち、ライブ映像とともにディスプレイ3に表示させる画像が、参照画像とライブ映像との一致度合いを示す画像である点で、実施の形態1~4と異なる。
実施の形態7に係る電子ミラー装置では、カメラ2が、複眼カメラなどの視差を抽出できる画像を撮影することができる。また、カメラ2で撮影した画像の視差に基づきカメラ2とユーザとの距離を算出し記録する。
2 カメラ
3 ディスプレイ
4 メモリ
5 コントローラ
6 タッチペン
7 色見本
Claims (12)
- ユーザを撮影するカメラと、
前記カメラが撮影するユーザが視認できる位置に表示面を有するディスプレイと、
前記カメラが撮影した前記ユーザの画像を前記ディスプレイに出力するコントローラと、を備え、
前記コントローラは、
前記カメラで撮影している前記ユーザのライブ映像を左右反転して鏡像とし、
前記ユーザのライブ映像の鏡像と対照される参照画像を読み込み、
前記ライブ映像の鏡像と前記参照画像とを前記ディスプレイの重複する領域に表示するための表示映像を合成して前記ディスプレイに出力する、
電子ミラー装置。 - さらにメモリを備え、
前記メモリは、前記カメラで過去に撮影されたライブ映像の1つのフレームを、前記参照画像として記録しており、
前記コントローラは、前記メモリから前記参照画像を読み込み、
前記メモリから読み込まれた前記参照画像の鏡像と前記ライブ映像の鏡像とから前記表示映像を合成して前記ディスプレイに出力する、
請求項1に記載の電子ミラー装置。 - 前記参照画像が、前記カメラで過去に撮影された前記ユーザのライブ映像の1つのフレームである、
請求項2に記載の電子ミラー装置。 - 前記コントローラは、外部媒体から前記参照画像を読み込み、前記外部媒体から読み込まれた前記参照画像の鏡像と前記ライブ映像の鏡像とから前記表示映像を合成して前記ディスプレイに出力する、
請求項1に記載の電子ミラー装置。 - 前記コントローラは、前記参照画像を半透明に加工し、半透明の前記参照画像と前記ライブ映像とから前記表示映像を合成して前記ディスプレイに出力する、
請求項1から4のいずれか1項に記載の電子ミラー装置。 - 前記コントローラは、前記参照画像の輪郭を抽出し、前記輪郭を表す画像と前記ライブ映像とから前記表示映像を合成して前記ディスプレイに出力する、
請求項1から4のいずれか1項に記載の電子ミラー装置。 - 前記コントローラは、前記参照画像の特徴点を検出し、前記特徴点の位置を表す画像と前記ライブ映像とから前記表示映像を合成して前記ディスプレイに出力する、
請求項1から4のいずれか1項に記載の電子ミラー装置。 - ユーザを撮影するカメラと、
前記カメラが撮影するユーザが視認できる位置に表示面を有するディスプレイと、
前記カメラが撮影した前記ユーザの画像を前記ディスプレイに出力するコントローラと、を備え、
前記コントローラは、前記カメラで撮影している前記ユーザのライブ映像を左右反転して鏡像とし、前記ユーザのライブ映像の前記鏡像と対照される参照画像を読み込み、前記ライブ映像と前記参照画像との一致度合いを抽出し、前記ライブ映像の前記鏡像と前記抽出した一致度合いを示す画像とを合成して前記ディスプレイに出力する、
電子ミラー装置。 - 前記コントローラは、
前記ライブ映像及び前記参照画像が前記ユーザの顔をそれぞれ表す場合に、前記ライブ映像で表される前記顔の輪郭よりも内側にある画素と、前記参照画像で表される前記顔の輪郭よりも内側にある画素とについて、画素ごとの輝度の差分の総和に基づいて、前記一致度合いを抽出する、
請求項8に記載の電子ミラー装置。 - ユーザを撮影するカメラと、
前記カメラが撮影するユーザが視認できる位置に表示面を有するディスプレイと、
前記カメラが撮影した前記ユーザの画像を前記ディスプレイに出力するコントローラと、を備え、
前記コントローラは、
前記カメラが撮影している前記ユーザのライブ映像を左右反転して鏡像とし、
前記カメラが撮影している前記ユーザのライブ映像に基づいて前記ユーザの位置及び顔向きのうちの少なくとも何れか一方に関連する第1の情報を連続的に取得し、
前記ユーザのライブ映像の前記鏡像と対照される参照画像中のユーザの位置及び顔向きのうちの少なくとも前記一方に関連する第2の情報を取得し、
前記第1の情報と前記第2の情報とを比較して一致度合いを抽出し、
前記ライブ映像の前記鏡像と前記抽出した一致度合いを示す画像とを合成して前記ディスプレイに出力する、
電子ミラー装置。 - さらにメモリを備え、
前記メモリは、前記カメラで過去に撮影されたライブ映像の1つのフレームを前記参照画像として、前記参照画像中のユーザの位置及び顔向きのうちの少なくとも前記一方に関連する前記第2の情報を記録しており、
前記コントローラは、前記メモリから前記第2の情報を読み込み、
前記第1の情報と前記メモリから読み込まれた前記第2の情報とを比較して一致度合いを抽出する、
請求項10に記載の電子ミラー装置。 - 視差を抽出できる画像を撮影するカメラと、
前記カメラが撮影するユーザが視認できる位置に表示面を有するディスプレイと、
前記カメラが撮影した画像を前記ディスプレイに出力するコントローラと、
メモリと、を備え、
前記コントローラは、
前記カメラが撮影している前記ユーザのライブ映像を左右反転した鏡像として前記ディスプレイに出力し、
撮影のタイミングで前記カメラが撮影している前記ユーザのライブ映像の1フレームを前記メモリに記録し、
前記1フレームの前記ユーザの画像に基づいて視差を抽出し、
前記視差に基づいて前記カメラと前記ユーザとの距離を算出し、
前記算出した距離と前記1フレームの前記画像とを対応づけて前記メモリに記録する、
電子ミラー装置。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP14751139.8A EP2945372A4 (en) | 2013-02-14 | 2014-02-14 | ELECTRONIC MIRROR DEVICE |
CN201480007478.5A CN104969543B (zh) | 2013-02-14 | 2014-02-14 | 电子镜装置 |
US14/765,587 US9992409B2 (en) | 2013-02-14 | 2014-02-14 | Digital mirror apparatus |
JP2015500151A JP6090679B2 (ja) | 2013-02-14 | 2014-02-14 | 電子ミラー装置 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013026504 | 2013-02-14 | ||
JP2013-026504 | 2013-02-14 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014125831A1 true WO2014125831A1 (ja) | 2014-08-21 |
Family
ID=51353842
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2014/000758 WO2014125831A1 (ja) | 2013-02-14 | 2014-02-14 | 電子ミラー装置 |
Country Status (5)
Country | Link |
---|---|
US (1) | US9992409B2 (ja) |
EP (2) | EP2945372A4 (ja) |
JP (1) | JP6090679B2 (ja) |
CN (1) | CN104969543B (ja) |
WO (1) | WO2014125831A1 (ja) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3021569A1 (en) * | 2014-11-17 | 2016-05-18 | Ricoh Company, Ltd. | Image acquisition and management |
US9558553B2 (en) | 2014-11-17 | 2017-01-31 | Ricoh Company, Ltd. | Image acquisition and management using a reference image |
WO2017149778A1 (ja) * | 2016-03-04 | 2017-09-08 | 株式会社オプティム | 鏡、画像表示方法及びプログラム |
US10136050B2 (en) | 2015-03-06 | 2018-11-20 | Ricoh Company, Ltd. | Image acquisition and management using a reference image |
US10217022B2 (en) | 2015-03-06 | 2019-02-26 | Ricoh Company, Ltd. | Image acquisition and management |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10063780B2 (en) * | 2010-06-02 | 2018-08-28 | Shan-Le Shih | Electronic imaging system for capturing and displaying images in real time |
TW201540264A (zh) * | 2014-04-18 | 2015-11-01 | Sony Corp | 資訊處理裝置、資訊處理方法、及程式 |
TWI669103B (zh) * | 2014-11-14 | 2019-08-21 | 日商新力股份有限公司 | 資訊處理裝置、資訊處理方法及程式 |
US10410412B2 (en) | 2015-05-29 | 2019-09-10 | Hover Inc. | Real-time processing of captured building imagery |
CN106650217B (zh) * | 2015-10-29 | 2020-06-30 | 佳能市场营销日本株式会社 | 信息处理装置以及信息处理方法 |
CN107968912B (zh) * | 2016-10-20 | 2020-09-25 | 时善乐 | 即时摄取影像并加以显示的电子影像*** |
JP2018069861A (ja) * | 2016-10-26 | 2018-05-10 | パナソニックIpマネジメント株式会社 | 電子ミラー装置 |
JP6723547B2 (ja) * | 2016-10-31 | 2020-07-15 | 株式会社ディー・ディー・エス | 皮膚情報処理プログラム及び皮膚情報処理装置 |
CN106388441B (zh) * | 2016-11-09 | 2018-01-26 | 广州视源电子科技股份有限公司 | 一种用于显示的方法、装置及智能镜子 |
US10382673B2 (en) | 2017-01-11 | 2019-08-13 | Hover Inc. | Automated guide for image capturing for 3D model creation |
CN108683851B (zh) * | 2018-05-15 | 2019-11-19 | 维沃移动通信有限公司 | 一种拍照方法及移动终端 |
US11779222B2 (en) | 2019-07-10 | 2023-10-10 | Compal Electronics, Inc. | Method of and imaging system for clinical sign detection |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2000013775A (ja) * | 1998-06-19 | 2000-01-14 | Yotaro Murase | 画像再生方法、画像再生装置及び画像データ作成装置 |
JP2003189161A (ja) | 2001-12-14 | 2003-07-04 | Sony Corp | 画像処理装置および方法、記録媒体、並びにプログラム |
JP2004302424A (ja) | 2003-03-20 | 2004-10-28 | Moritex Corp | 顔面撮像装置 |
JP2005117316A (ja) | 2003-10-07 | 2005-04-28 | Matsushita Electric Ind Co Ltd | 撮影装置、撮影方法、およびプログラム |
JP2009053328A (ja) * | 2007-08-24 | 2009-03-12 | Nec Corp | 画像表示装置 |
JP2009061178A (ja) * | 2007-09-07 | 2009-03-26 | Noritsu Koki Co Ltd | 顔撮影装置及び顔撮影方法 |
JP2009176208A (ja) * | 2008-01-28 | 2009-08-06 | Nec Corp | 顔認証装置、システム、方法及びプログラム |
JP2009188528A (ja) | 2008-02-04 | 2009-08-20 | Noritsu Koki Co Ltd | 顔撮影装置 |
Family Cites Families (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003021859A (ja) | 2001-07-09 | 2003-01-24 | Konica Corp | 撮影用ガイド、人物写真撮影装置、人物写真プリント装置 |
JP4174404B2 (ja) * | 2003-10-01 | 2008-10-29 | キヤノン株式会社 | 撮像装置、画像表示方法、プログラムおよび記憶媒体 |
JP4482021B2 (ja) * | 2004-01-23 | 2010-06-16 | オリンパス株式会社 | カメラ |
JP4153444B2 (ja) * | 2004-02-18 | 2008-09-24 | 富士フイルム株式会社 | デジタルカメラ |
US7782384B2 (en) * | 2004-11-05 | 2010-08-24 | Kelly Douglas J | Digital camera having system for digital image composition and related method |
JP4635739B2 (ja) * | 2005-06-24 | 2011-02-23 | ソニー株式会社 | 撮像装置 |
JP2007013768A (ja) * | 2005-07-01 | 2007-01-18 | Konica Minolta Photo Imaging Inc | 撮像装置 |
JP5034359B2 (ja) | 2006-08-07 | 2012-09-26 | 富士通株式会社 | 画像認証装置、画像認証方法、画像認証プログラム、記録媒体及び電子機器 |
JP2008054031A (ja) | 2006-08-24 | 2008-03-06 | Fujifilm Corp | デジタルカメラ及び表示制御方法 |
GB2448221B (en) * | 2007-04-02 | 2012-02-01 | Samsung Electronics Co Ltd | Method and apparatus for providing composition information in digital image processing device |
JP4702418B2 (ja) * | 2008-09-09 | 2011-06-15 | カシオ計算機株式会社 | 撮影装置、画像領域の存否判定方法及びプログラム |
KR101618759B1 (ko) * | 2008-12-24 | 2016-05-09 | 삼성전자주식회사 | 영상 촬영 장치 및 방법 |
JP2010171797A (ja) * | 2009-01-23 | 2010-08-05 | Nikon Corp | 撮像装置及びプログラム |
JP5388611B2 (ja) * | 2009-02-03 | 2014-01-15 | キヤノン株式会社 | 撮像装置およびその制御方法、プログラム |
JP2010183317A (ja) * | 2009-02-05 | 2010-08-19 | Olympus Imaging Corp | 撮像装置、画像合成表示装置、画像合成表示方法、およびプログラム |
JP5206516B2 (ja) | 2009-03-13 | 2013-06-12 | 株式会社デンソー | 撮影補助装置、プログラムおよび撮影システム |
JP5620651B2 (ja) | 2009-06-26 | 2014-11-05 | キヤノン株式会社 | 再生装置、撮像装置、及びその制御方法 |
JP2011017800A (ja) * | 2009-07-07 | 2011-01-27 | Canon Inc | 焦点検出装置 |
JP2011070415A (ja) * | 2009-09-25 | 2011-04-07 | Fujifilm Corp | 画像処理装置及び画像処理方法 |
EP2323102A1 (en) * | 2009-10-23 | 2011-05-18 | ST-Ericsson (France) SAS | Image capturing aid |
EP2339534A1 (en) * | 2009-11-18 | 2011-06-29 | Panasonic Corporation | Specular reflection compensation |
KR101604846B1 (ko) * | 2009-11-27 | 2016-03-21 | 엘지전자 주식회사 | 휴대 단말기 및 그 동작 제어방법 |
BR112012020872A2 (pt) * | 2010-02-24 | 2016-05-03 | Fujitsu Frontech Ltd | sistema de autenticação, programa de autenticação e método de autenticação |
JP5327211B2 (ja) | 2010-12-28 | 2013-10-30 | カシオ計算機株式会社 | 撮像装置、撮像制御方法及びプログラム |
JP2013042250A (ja) * | 2011-08-12 | 2013-02-28 | Sony Corp | 撮像装置および撮像方法 |
JP6112419B2 (ja) * | 2012-02-03 | 2017-04-12 | パナソニックIpマネジメント株式会社 | 撮像装置および測距装置 |
KR101197863B1 (ko) * | 2012-02-27 | 2012-11-05 | 경일대학교산학협력단 | 피부 진단을 위한 단말 및 이를 위한 방법이 기록된 컴퓨터 판독 가능한 기록매체 |
KR102000536B1 (ko) * | 2012-12-28 | 2019-07-16 | 삼성전자주식회사 | 합성 이미지를 촬영하는 촬영 장치 및 그 방법 |
JP6143469B2 (ja) * | 2013-01-17 | 2017-06-07 | キヤノン株式会社 | 情報処理装置、情報処理方法及びプログラム |
US9589205B2 (en) * | 2014-05-15 | 2017-03-07 | Fuji Xerox Co., Ltd. | Systems and methods for identifying a user's demographic characteristics based on the user's social media photographs |
-
2014
- 2014-02-14 EP EP14751139.8A patent/EP2945372A4/en not_active Withdrawn
- 2014-02-14 EP EP16176569.8A patent/EP3091737A3/en not_active Withdrawn
- 2014-02-14 WO PCT/JP2014/000758 patent/WO2014125831A1/ja active Application Filing
- 2014-02-14 JP JP2015500151A patent/JP6090679B2/ja active Active
- 2014-02-14 US US14/765,587 patent/US9992409B2/en active Active
- 2014-02-14 CN CN201480007478.5A patent/CN104969543B/zh active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2000013775A (ja) * | 1998-06-19 | 2000-01-14 | Yotaro Murase | 画像再生方法、画像再生装置及び画像データ作成装置 |
JP2003189161A (ja) | 2001-12-14 | 2003-07-04 | Sony Corp | 画像処理装置および方法、記録媒体、並びにプログラム |
JP2004302424A (ja) | 2003-03-20 | 2004-10-28 | Moritex Corp | 顔面撮像装置 |
JP2005117316A (ja) | 2003-10-07 | 2005-04-28 | Matsushita Electric Ind Co Ltd | 撮影装置、撮影方法、およびプログラム |
JP2009053328A (ja) * | 2007-08-24 | 2009-03-12 | Nec Corp | 画像表示装置 |
JP2009061178A (ja) * | 2007-09-07 | 2009-03-26 | Noritsu Koki Co Ltd | 顔撮影装置及び顔撮影方法 |
JP2009176208A (ja) * | 2008-01-28 | 2009-08-06 | Nec Corp | 顔認証装置、システム、方法及びプログラム |
JP2009188528A (ja) | 2008-02-04 | 2009-08-20 | Noritsu Koki Co Ltd | 顔撮影装置 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2945372A4 |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3021569A1 (en) * | 2014-11-17 | 2016-05-18 | Ricoh Company, Ltd. | Image acquisition and management |
US9451170B2 (en) | 2014-11-17 | 2016-09-20 | Ricoh Company, Ltd. | Image acquisition and management using a reference image |
US9558553B2 (en) | 2014-11-17 | 2017-01-31 | Ricoh Company, Ltd. | Image acquisition and management using a reference image |
US10136050B2 (en) | 2015-03-06 | 2018-11-20 | Ricoh Company, Ltd. | Image acquisition and management using a reference image |
US10217022B2 (en) | 2015-03-06 | 2019-02-26 | Ricoh Company, Ltd. | Image acquisition and management |
WO2017149778A1 (ja) * | 2016-03-04 | 2017-09-08 | 株式会社オプティム | 鏡、画像表示方法及びプログラム |
JPWO2017149778A1 (ja) * | 2016-03-04 | 2018-10-04 | 株式会社オプティム | 鏡、画像表示方法及びプログラム |
Also Published As
Publication number | Publication date |
---|---|
JPWO2014125831A1 (ja) | 2017-02-02 |
US20150373264A1 (en) | 2015-12-24 |
EP3091737A3 (en) | 2017-02-15 |
CN104969543A (zh) | 2015-10-07 |
CN104969543B (zh) | 2019-04-02 |
EP2945372A1 (en) | 2015-11-18 |
EP3091737A2 (en) | 2016-11-09 |
US9992409B2 (en) | 2018-06-05 |
EP2945372A4 (en) | 2016-03-30 |
JP6090679B2 (ja) | 2017-03-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6090679B2 (ja) | 電子ミラー装置 | |
US10587795B2 (en) | System for producing compliant facial images for selected identification documents | |
CN105488782B (zh) | 光泽判定装置以及光泽判定方法 | |
EP3241151A1 (en) | An image face processing method and apparatus | |
WO2017203913A1 (ja) | 肌診断装置および肌診断方法 | |
JP6859611B2 (ja) | 画像処理装置、画像処理方法及びプログラム | |
JP5949030B2 (ja) | 画像生成装置、画像生成方法及びプログラム | |
JP2020526735A (ja) | 瞳孔距離測定方法、装着型眼用機器及び記憶媒体 | |
JP5771647B2 (ja) | 肌解析装置、肌解析システム、肌解析方法および肌解析プログラム | |
JP2015232746A (ja) | 皺検出装置および皺検出方法 | |
JP2015080647A (ja) | 撮影画像表示装置 | |
JP2008167853A (ja) | 試験シート、物体診断装置および方法並びにプログラム | |
JP2013182062A (ja) | 表示装置および投影装置 | |
JP6406966B2 (ja) | 撮像装置、撮像装置の制御方法、撮像装置の制御プログラム | |
WO2013187282A1 (ja) | 撮像画像表示装置、撮像画像表示方法、記録媒体 | |
JP2006081846A (ja) | 顔面溝の評価方法及び装置 | |
CA3187876A1 (en) | System and method for automatic personalized assessment of human body surface conditions | |
JP2009302910A (ja) | 撮像装置及び画像再生装置 | |
JP2008171264A (ja) | 顔撮影装置および顔撮影方法 | |
WO2020240989A1 (ja) | 撮像装置、撮像制御方法および撮像制御プログラム | |
KR20190097494A (ko) | 전자 장치 및 그의 제어 방법 | |
US20220196394A1 (en) | Imaging device, control method therefor, measuring device, and storage medium | |
JP2008243119A (ja) | 顔撮影装置 | |
JP5962268B2 (ja) | 画像処理装置、画像処理方法、画像生成方法及びプログラム | |
GB2612364A (en) | Method and system for determining user-screen distance |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14751139 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2015500151 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14765587 Country of ref document: US |
|
REEP | Request for entry into the european phase |
Ref document number: 2014751139 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2014751139 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |