US20190387165A1 - Image device for generating depth images and related electronic device - Google Patents
Image device for generating depth images and related electronic device Download PDFInfo
- Publication number
- US20190387165A1 US20190387165A1 US16/432,911 US201916432911A US2019387165A1 US 20190387165 A1 US20190387165 A1 US 20190387165A1 US 201916432911 A US201916432911 A US 201916432911A US 2019387165 A1 US2019387165 A1 US 2019387165A1
- Authority
- US
- United States
- Prior art keywords
- image
- depth
- capturer
- color
- view angle
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000003287 optical effect Effects 0.000 claims description 23
- 238000000034 method Methods 0.000 claims description 13
- 238000010586 diagram Methods 0.000 description 13
- 230000008859 change Effects 0.000 description 3
- 101100272626 Drosophila melanogaster BomBc1 gene Proteins 0.000 description 2
- 101100444028 Drosophila melanogaster Dso2 gene Proteins 0.000 description 2
- 230000003190 augmentative effect Effects 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 238000009877 rendering Methods 0.000 description 2
- 101000969688 Homo sapiens Macrophage-expressed gene 1 protein Proteins 0.000 description 1
- 102100021285 Macrophage-expressed gene 1 protein Human genes 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000008878 coupling Effects 0.000 description 1
- 238000010168 coupling process Methods 0.000 description 1
- 238000005859 coupling reaction Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/2628—Alteration of picture size, shape, position or orientation, e.g. zooming, rotation, rolling, perspective, translation
-
- H04N5/23238—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/02—Constructional features of telephone sets
- H04M1/0202—Portable telephone sets, e.g. cordless phones, mobile phones or bar type handsets
- H04M1/026—Details of the structure or mounting of specific components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/80—Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
- G06T7/85—Stereo camera calibration
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/97—Determining parameters from multiple pictures
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/02—Constructional features of telephone sets
- H04M1/0202—Portable telephone sets, e.g. cordless phones, mobile phones or bar type handsets
- H04M1/026—Details of the structure or mounting of specific components
- H04M1/0264—Details of the structure or mounting of specific components for a camera module assembly
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/02—Constructional features of telephone sets
- H04M1/0202—Portable telephone sets, e.g. cordless phones, mobile phones or bar type handsets
- H04M1/026—Details of the structure or mounting of specific components
- H04M1/0266—Details of the structure or mounting of specific components for a display module assembly
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72403—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
- H04M1/7243—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
- H04M1/72439—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for image or video messaging
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/15—Processing image signals for colour aspects of image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/161—Encoding, multiplexing or demultiplexing different image signal components
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/239—Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/254—Image signal generators using stereoscopic image cameras in combination with electromagnetic radiation sources for illuminating objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/296—Synchronisation thereof; Control thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N17/00—Diagnosis, testing or measuring for television systems or their details
- H04N17/002—Diagnosis, testing or measuring for television systems or their details for television cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/50—Constructional details
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/50—Constructional details
- H04N23/54—Mounting of pick-up tubes, electronic image sensors, deviation or focusing coils
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/56—Cameras or camera modules comprising electronic image sensors; Control thereof provided with illuminating means
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/57—Mechanical or electrical details of cameras or camera modules specially adapted for being embedded in other devices
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/698—Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/95—Computational photography systems, e.g. light-field imaging systems
- H04N23/957—Light-field or plenoptic cameras or camera modules
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N2013/0074—Stereoscopic image analysis
- H04N2013/0081—Depth or disparity estimation from stereoscopic image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2201/00—Indexing scheme relating to scanning, transmission or reproduction of documents or the like, and to details thereof
- H04N2201/32—Circuits or arrangements for control or supervision between transmitter and receiver or between image input and image output device, e.g. between a still-image camera and its memory or between a still-image camera and a printer device
- H04N2201/3201—Display, printing, storage or transmission of additional information, e.g. ID code, date and time or title
- H04N2201/3225—Display, printing, storage or transmission of additional information, e.g. ID code, date and time or title of data relating to an image, a page or a document
- H04N2201/3254—Orientation, e.g. landscape or portrait; Location or order of the image data, e.g. in memory
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2201/00—Indexing scheme relating to scanning, transmission or reproduction of documents or the like, and to details thereof
- H04N2201/32—Circuits or arrangements for control or supervision between transmitter and receiver or between image input and image output device, e.g. between a still-image camera and its memory or between a still-image camera and a printer device
- H04N2201/3201—Display, printing, storage or transmission of additional information, e.g. ID code, date and time or title
- H04N2201/3274—Storage or retrieval of prestored additional information
- H04N2201/3277—The additional information being stored in the same storage device as the image data
Definitions
- the present invention relates to an image device and an electronic device, and particularly to an image device that has a simpler mechanism to generate depth images, and an electronic device that can display a corresponding depth image according to motion or rotation of the electronic device.
- an image device for generating panoramic depth images usually includes two image capturing groups, and each image capturing group includes more than two image capturers, wherein one image capturing group of the two image capturing group is used for capturing a plurality of first images, and the other image capturing group of the two image capturing group is used for capturing a plurality of second images. Then, a processor of the image device utilizes the plurality of first images and the plurality of second images to generate multiple depth images corresponding to different view angles, and stitches the multiple depth images as a 360 degree panoramic depth image.
- the image device first needs to execute image processing on a plurality of images captured by the plurality of image capturers (e.g. integrating the optical centers of the plurality of image capturers into a virtual optical center) to generate the 360 degree panoramic depth image, otherwise a problem of blind spots will occur at stitching regions of the 360 degree panoramic depth image.
- an electronic device provided by the prior art displays a 360 degree panoramic color image
- the electronic device takes the electronic device as a center to display the 360 degree panoramic color image. That is, when a user controls the electronic device to view the 360 degree panoramic color image through a touch mode, the electronic device can only change a view angle for displaying the 360 degree panoramic color image. Therefore, when the electronic device is moved, the electronic device cannot change an image displayed by the electronic device according to motion of the electronic device.
- An embodiment of the present invention provides an image device for generating depth images.
- the image device includes at least two image capturers and a rotating device. When the rotating device rotates the at least two image capturers, multiple images captured by the at least two image capturers are utilized to generate a depth image, wherein a view angle corresponding to the depth image is not less than a view angle of each image capturer of the at least two image capturers.
- the image device includes at least one image capturer, at least one light source, and a rotating device.
- the at least one light source is used for emitting emission light.
- the rotating device rotates the at least one image capturer and the at least one light source, multiple images captured by the at least one image capturer and the emission light are used for generating a depth image, wherein a view angle corresponding to the depth image is not less than a view angle of each image capturer of the at least one image capturer.
- the electronic device for viewing a panoramic color and depth image.
- the electronic device includes a display and a processor.
- the processor is used for reading a 360 degree panoramic color image and a 360 degree panoramic depth image, wherein the 360 degree panoramic color image corresponds to the 360 degree panoramic depth image.
- the processor When the electronic device is moved or rotated, the processor further converts a part of the 360 degree panoramic color image corresponding to a view angle corresponding to motion or rotation of the electronic device into a planar color image and a part of the 360 degree panoramic depth image corresponding to the view angle into a planar depth image according to the view angle, combines the planar color image with the planar depth image to generate a planar color and depth image corresponding to the view angle, and makes the display display the planar color image according to the planar color and depth image.
- the present invention provides an image device for generating depth images and related electronic device. Because the image device utilizes a rotating device to rotate a plurality of image capturers included in the image device, and utilizes a processor to generate a depth image according to plurality of images captured by the plurality of image capturers, compared to the prior art, the image device does not have a problem that optical centers in the prior art cannot overlap. In addition, because a 360 degree panoramic color and depth image generated by the present invention has depth information of a 360 degree panoramic depth image, compared to the prior art, when the electronic device displays the 360 degree panoramic color and depth image, the electronic device can display a corresponding planar color and depth image according to motion or rotation of the electronic device
- FIG. 1 is a diagram illustrating an image device for generating depth images according to a first embodiment of the present invention.
- FIG. 2 is a diagram illustrating a top view of the image device.
- FIG. 3A is a diagram illustrating an image device for generating depth images according to a second embodiment of the present invention
- FIG. 3B is a diagram illustrating an image device for generating depth images according to another embodiment of the present invention
- FIG. 3C is a diagram illustrating an image device 370 for generating depth images according to another embodiment of the present invention.
- FIG. 4 is a diagram illustrating an electronic device for viewing a panoramic color and depth image according to a third embodiment of the present invention.
- FIG. 5 is a diagram illustrating the electronic device being moved from a position to another position.
- FIG. 6 is a diagram illustrating a block of the display displaying the corresponding left eye color image and another block of the display displaying the corresponding right eye color image.
- FIG. 1 is a diagram illustrating an image device 100 for generating depth images according to a first embodiment of the present invention, wherein as shown in FIG. 1 , the image device 100 includes two image capturers 102 , 104 , a supporting unit 106 , a rotating device 108 , and a processor 110 , the supporting unit 106 is coupled to the rotating device 108 , the processor 110 is coupled to the image capturers 102 , 104 , the image capturers 102 , 104 are installed on the supporting unit 106 , and the image capturers 102 , 104 are non-fisheye image capturers.
- the image capturers 102 , 104 are fisheye image capturers.
- the present invention is not limited to the image device 100 including the image capturers 102 , 104 . That is, the image device 100 can include more than two image capturers.
- each image capturer of the image capturers 102 , 104 at least includes a lens and an image sensor (e.g. a charge-coupled device (CCD) image sensor or a complementary metal-oxide-semiconductor image sensor).
- the processor 110 is installed in a host outside the image device 100 .
- a rotating axis RA of the rotating device 108 passes through an optical center C 1 of the image capturer 102 and an optical center C 2 of the image capturer 104 , and the rotating device 108 can rotate the image capturers 102 , 104 by the supporting unit 106 , wherein the rotating device 108 is driven by a gear 112 .
- the rotating device 108 is driven by a motor.
- the processor 110 can stitch the plurality of first images to generate a first 360 degree panoramic color image and stitch the plurality of second images to generate a second 360 degree panoramic color image according to a feature point matching method or a fixed angle method provided by the prior art.
- the image capturer 102 can capture 4 first images IM 11 -IM 14 and the image capturer 104 can capture 4 second images IM 21 -IM 24 , wherein the first image IM 11 and the second image IM 21 correspond to a first view angle FOV 1 , the first image IM 12 and the second image IM 22 correspond to a second view angle FOV 2 , the first image IM 13 and the second image IM 23 correspond to a third view angle FOV 3 , the first image IM 14 and the second image IM 24 correspond to a fourth view angle FOV 4 , the first view angle FOV 1 , the second view angle FOV 2 , the third view angle FOV 3 , and the fourth view angle FOV 4 can be referred to FIG.
- FIG. 2 is a top view of the image device 100 , FIG. 2 does not show the image capturer 104 .
- the processor 110 can stitch the first images IM 11 -IM 14 to generate the first 360 degree panoramic color image and stitch the second images IM 21 -IM 24 to generate the second 360 degree panoramic color image according to the feature point matching method (e.g. feature points of the first images IM 11 -IM 14 and feature points of the second images IM 21 -IM 24 ) or the fixed angle method (e.g. the first view angle FOV 1 , the second view angle FOV 2 , the third view angle FOV 3 , and the fourth view angle FOV 4 ).
- the feature point matching method e.g. feature points of the first images IM 11 -IM 14 and feature points of the second images IM 21 -IM 24
- the fixed angle method e.g. the first view angle FOV 1 , the second view angle FOV 2 , the third view angle FOV 3 , and the fourth view angle FOV 4 .
- the processor 110 can utilize the first 360 degree panoramic color image and the second 360 degree panoramic color image to generate a 360 degree panoramic depth image, wherein in one embodiment of the present invention, each depth value of the 360 degree panoramic depth image corresponds to the optical center C 1 or the optical center C 2 . But, in another embodiment of the present invention, the each depth value of the 360 degree panoramic depth image corresponds to a plane where the optical center C 1 and the optical center C 2 are located.
- the processor 110 can generate a first depth image corresponding to the first view angle FOV 1 according to the first image IM 11 and the second image IM 21 , generate a second depth image corresponding to the second view angle FOV 2 according to the first image IM 12 and the second image IM 22 , generate a third depth image corresponding to the third view angle FOV 3 according to the first image IM 13 and the second image IM 23 , and generate a fourth depth image corresponding to the fourth view angle FOV 4 according to the first image IM 14 and the second image IM 24 .
- the processor 110 stitches the first depth image, the second depth image, the third depth image, and the fourth depth image to generate the 360 degree panoramic depth image according to the feature point matching method or the fixed angle method.
- the image capturer 102 can capture the plurality of first images and the image capturer 104 can capture the plurality of second images, wherein the plurality of first images correspond to the plurality of second images, a view angle corresponding to the plurality of first images is between 360 degrees and a view angle of the image capturer 102 (or a view angle of the image capturer 104 ).
- a view angle corresponding to a color image generated by the processor 110 stitching the plurality of first images (or stitching the plurality of second images) is also between 360 degrees and the view angle of the image capturer 102 (or the view angle of the image capturer 104 ).
- a view angle corresponding to a depth image generated by the processor 110 according to the plurality of first images and the plurality of second images is also between 360 degrees and the view angle of the image capturer 102 (or the view angle of the image capturer 104 ).
- the processor 110 can combine the first 360 degree panoramic color image (or the second 360 degree panoramic color image) with the 360 degree panoramic depth image to generate a 360 degree panoramic color and depth image, and the processor 110 compresses the first 360 degree panoramic color image (or the second 360 degree panoramic color image) according to a standard compression format (e.g. a Joint Photographic Experts Group (JPEG) format) and compresses the 360 degree panoramic depth image according to a lossy format (or a lossless format) to integrate the 360 degree panoramic color and depth image into a file, wherein a compressed 360 degree panoramic depth image is stored in a header of the file. That is, the compressed 360 degree panoramic depth image is usually stored in a user defined header of the file.
- JPEG Joint Photographic Experts Group
- the standard compression format is not limited to JPEG, that is, the first 360 degree panoramic color image (or the second 360 degree panoramic color image) can also be compressed according to PNG, MPEG1, MPEG2, MPEG4, H.264, H.265, and so on.
- the 360 degree panoramic depth image is not compressed
- FIG. 3A is a diagram illustrating an image device 300 for generating depth images according to a second embodiment of the present invention, wherein as shown in FIG. 3A , differences between the image device 300 and the image device 100 are that the image device 300 further includes a light source 302 , an image capturer 304 are used for capturing a plurality of second images including emission light generated by the light source 302 , the processor 110 is coupled to the light source 302 , the image capturer 304 , and the image capturer 102 , and the light source 302 , the image capturer 304 , and the image capturer 102 are installed on the supporting unit 106 .
- the emission light is structured light (or the emission light includes a random pattern), and the light source 302 utilizes a collimator, at least one laser beam (e.g. at least one infrared laser beam), and at least one optical element (e.g. at least one diffraction optical element (DOE) to generate the emission light.
- the light source 302 does not need the collimator. As shown in FIG.
- the rotating device 108 rotates the light source 302 , the image capturer 304 , and the image capturer 102 , the plurality of first images captured by the image capturer 102 are used for generating a 360 degree panoramic color image, the plurality of second images including the emission light captured by the image capturer 304 are used for generating a 360 degree panoramic depth image, wherein the sum of the view angles corresponding to the plurality of first images is not less than 360 degrees, and a sum of view angles corresponding to the plurality of second images is also not less than 360 degrees.
- the processor 110 can turn the light source 302 according to at least one of luminance of an environment which the image device 300 is located at, a quality of the 360 degree panoramic depth image, and a difference of the 360 degree panoramic depth image corresponding to turning-on and turning-off of the light source 302 .
- the light source 302 is a vertical-cavity surface-emitting Laser (VCSEL)
- the light source 302 optionally includes the collimator and the at least one optical element. That is, the collimator and the at least one are not necessary for the light source 302 .
- the light source 302 can be a light source with other configurations and wavelengths.
- the processor 110 can optionally adjust intensity of the light source 302 according to luminance corresponding to each second image of the plurality of second images and a target value, wherein the target value is set according to reflection coefficient of a human skin of a user corresponding to the structured light emitted by the light source 302 .
- the processor 110 can generate a luminance distribution map corresponding to the each second image according to the each second image, and optionally adjust the intensity of the light source 302 according to a percentage of the each second image occupied by an area corresponding to a maximum luminance value of at least one luminance value within the luminance distribution map greater than the target value.
- the processor 110 can optionally adjust the intensity of the light source 302 according to average luminance of the each second image and the target value.
- the processor 110 can generate a luminance histogram corresponding to a plurality of pixels of the each second image according to the each second image, and optionally adjust the intensity of the light source 302 according to a median of the luminance histogram and the target value, or according to a predetermined quantile of the luminance histogram and the target value.
- the processor 110 can optionally dynamically adjust the intensity of the light source 302 according to a distance between at least one predetermined object within the each second image and the image capturer 304 (or the image capturer 102 ) and a first lookup table, wherein the first lookup table stores relationships between a distance corresponding to an object and the intensity of the light source 302 .
- the processor 110 continuously detects the luminance of the environment which the image device 300 is located at under the light source 302 being turned off.
- the processor 110 increases the intensity of the light source 302 (when the light source 302 is turned on) according to a second lookup table, wherein the second lookup table stores relationships between the intensity of the light source 302 (when the light source 302 is turned on) and the luminance of the environment.
- the processor 110 when the processor 110 utilizes a first pulse width modulation signal in a continuous mode to adjust the intensity of the emitted light, the processor 110 can adjust the intensity of the emitted light by changing a duty cycle of the first pulse width modulation signal; in the above-mentioned embodiments, when the processor 110 utilizes a second pulse width modulation signal in a burst mode to adjust the intensity of the emitted light, the processor 110 can adjust the intensity of the emitted light by changing an enabling time of the second pulse width modulation signal; in the above-mentioned embodiments, when the processor 110 utilizes the first pulse width modulation signal and the second pulse width modulation signal to adjust the intensity of the emitted light, the processor 110 can adjust the intensity of the emitted light by simultaneously changing the enabling time of the second pulse width modulation signal and the duty cycle of the first pulse width modulation signal.
- the light source 302 is applied to a time of flight (TOF), wherein when the light source 302 is applied to the time of flight, the emission light is diffused light, that is, the emission light is uniform light. Therefore, the processor 110 can generate the 360 degree panoramic depth image according to a difference between a receiving time for the image capturer 304 receiving reflected light and a generating time corresponding to the emission light, wherein the reflected light is generated by at least one object reflecting the emitted light, and meanwhile the image capturer 304 is a time of flight sensor.
- operational principles of the processor 110 determining whether to adjust the intensity of the emitted light can be referred to the above-mentioned corresponding descriptions, so further description thereof is omitted for simplicity.
- processor 110 the above-mentioned functions of the processor 110 are fully disclosed, one of ordinary skill in the art should easily utilize a field programmable gate array (FPGA) with the above-mentioned functions of the processor 110 , or an application-specific integrated circuit (ASIC) with the above-mentioned functions of the processor 110 , or a software module with the above-mentioned functions of the processor 110 , or an analog integrated circuit with the above-mentioned functions of the processor 110 to realize the processor 110 according to corresponding descriptions of the above-mentioned functions of processor 110 . Therefore, a corresponding structure of the processor 110 is omitted for simplicity.
- FPGA field programmable gate array
- ASIC application-specific integrated circuit
- the image capturer 102 and the image capturer 304 can act as a stereo camera (or a depth camera), and the emission light generated by the light source 302 is used for assisting the stereo camera.
- the processor 110 can utilize a time division multiplexing method to control the light source 302 .
- the stereo camera acts as the stereo camera, and when the processor 110 turns off the light source 302 , images captured by the image capturer 102 and the image capturer 304 are used for generating a color image.
- FIG. 3B is a diagram illustrating an image device 350 for generating depth images according to another embodiment of the present invention, wherein as shown in FIG. 3B , a difference between the image device 350 and the image device 300 is that the image device 350 further includes an image capturer 306 , wherein the image capturer 102 and the image capturer 304 can act as a stereo camera (or a depth camera), the emission light generated by the light source 302 is used for assisting the stereo camera, and images captured by the image capturer 306 are used for generating a color image (that is, the image capturer 306 only acts as a color sensor and has an infrared light filter).
- operational principles of the image capturer 102 , the image capturer 304 , and the image capturer 306 can be referred to the above-mentioned corresponding descriptions, so further description thereof is omitted for simplicity.
- FIG. 3C is a diagram illustrating an image device 370 for generating depth images according to another embodiment of the present invention, wherein as shown in FIG. 3C , a difference between the image device 370 and the image device 300 is that the image device 370 does not include the image capturer 304 .
- the image capturer 102 when the emission light is structured light, the image capturer 102 is a structured light depth sensor.
- the image capturer 102 can be a time of flight sensor, meanwhile the emission light is diffused light, and the operational principles of the image capturer 102 can be referred to the above-mentioned corresponding descriptions, so further description thereof is omitted for simplicity.
- FIG. 4 is a diagram illustrating an electronic device 400 for viewing a panoramic color and depth image according to a third embodiment of the present invention, wherein as shown in FIG. 4 , the electronic device 400 includes a display 402 , a processor 404 , and an inertial sensor 406 (e.g.
- the processor 404 is coupled to the display 402 and the inertial sensor 406 (because the processor 404 and the inertial sensor 406 are installed within the electronic device 400 , related coupling lines corresponding to the processor 404 and the inertial sensor 406 are represented by dotted lines), and the inertial sensor 406 can be a gyro, an accelerator, a microelectromechanical systems (MEMS), and so on.
- the electronic device 400 can be a smart phone or a tablet computer. But, the present invention is not limited to the electronic device 400 being a smart phone or a tablet computer. As shown in FIG.
- the processor 404 can first read a 360 degree panoramic color image and a 360 degree panoramic depth image of the 360 degree panoramic color and depth image, or respectively read a 360 degree panoramic color image and a 360 degree panoramic depth image generated from other image devices, wherein each depth value of the 360 degree panoramic depth image corresponds to a virtual optical center which the electronic device 400 acts as.
- the inertial sensor 406 can determine a view angle corresponding to motion or rotation of the electronic device 400 according to the motion or the rotation of the electronic device 400 (that is, the view angle corresponds to the position B).
- the processor 404 can convert a part of the 360 degree panoramic color image corresponding to the view angle into a planar color image and convert a part of the 360 degree panoramic depth image corresponding to the view angle into a planar depth image according to the view angle, and utilize a depth-image-based rendering (DIBR) provided by the prior art to combine the planar color image with the planar depth image to generate a planar color and depth image corresponding to the view angle, wherein when the display 402 displays the planar color and depth image, because the planar color and depth image has depth information of the planar depth image, the planar color and depth image can make a user watching the planar color and depth image have a three-dimensional feeling.
- DIBR depth-image-based rendering
- each depth value of the planar depth image corresponds to a virtual optical plane where the electronic device 400 is located.
- the user of the electronic device 400 utilizes a mouse or touches the display 402 to change a view angle of the display 402 for displaying the 360 degree panoramic color and depth image.
- the 360 degree panoramic color image and the 360 degree panoramic depth image correspond to two different optical centers, the processor 404 can utilize the prior art to compensate a shift between the two different optical centers when the display 402 displays the planar color and depth image.
- the processor 404 can convert the planar color and depth image into a corresponding left eye color image and a corresponding right eye color image according to the depth information of the planar depth image and the depth-image-based rendering, wherein as shown in FIG. 6 , a block 4022 of the display 402 displays the corresponding left eye color image and a block 4024 of the display 402 displays the corresponding right eye color image.
- the electronic device 400 when the electronic device 400 is mounted in a head mounted display, the user can view contents of the virtual reality (VR), the augmented reality (AR), the substitutional reality (SR), and the mixed reality (MR) through the head mounted display and the electronic device 400 .
- the block 4022 of the display 402 is independent from the block 4024 of the display 402 . That is, the block 4022 and the block 4024 are two independent displays.
- processor 404 the above-mentioned functions of the processor 404 are fully disclosed, one of ordinary skill in the art should easily utilize a field programmable gate array with the above-mentioned functions of the processor 404 , or an application-specific integrated circuit with the above-mentioned functions of the processor 404 , or a software module with the above-mentioned functions of the processor 404 , or an analog integrated circuit with the above-mentioned functions of the processor 404 to realize the processor 404 according to corresponding descriptions of the above-mentioned functions of processor 404 . Therefore, a corresponding structure of the processor 404 is omitted for simplicity.
- the image device utilizes the rotating device to rotate the image capturers, and utilizes the processor to generate a 360 degree panoramic depth image according to multiple images captured by the image capturers, compared to the prior art, the image device does not have a problem that optical centers in the prior art cannot overlap.
- a 360 degree panoramic color and depth image generated by the present invention has depth information of a 360 degree panoramic depth image, compared to the prior art, when the electronic device displays the 360 degree panoramic color and depth image, the electronic device can display a corresponding planar color and depth image according to motion or rotation of the electronic device.
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Electromagnetism (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- Business, Economics & Management (AREA)
- General Business, Economics & Management (AREA)
- Human Computer Interaction (AREA)
- Computer Networks & Wireless Communication (AREA)
- Image Processing (AREA)
- Studio Devices (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Length Measuring Devices By Optical Means (AREA)
Abstract
An image device for generating depth images includes at least two image capturers and a rotating device. When the rotating device rotates the at least two image capturers, multiple images captured by the at least two image capturers are utilized to generate a depth image, wherein a view angle corresponding to the depth image is not less than a view angle of each image capturer of the at least two image capturers.
Description
- This application claims the benefit of U.S. Provisional Application No. 62/681,683, filed on Jun. 7, 2018 and entitled “IMAGE PROCESSING METHOD AND SELF-CALIBRATION/LIGHT CONTROL METHOD FOR CAMERA DEVICE,” the contents of which are incorporated herein by reference.
- The present invention relates to an image device and an electronic device, and particularly to an image device that has a simpler mechanism to generate depth images, and an electronic device that can display a corresponding depth image according to motion or rotation of the electronic device.
- In the prior art, an image device for generating panoramic depth images usually includes two image capturing groups, and each image capturing group includes more than two image capturers, wherein one image capturing group of the two image capturing group is used for capturing a plurality of first images, and the other image capturing group of the two image capturing group is used for capturing a plurality of second images. Then, a processor of the image device utilizes the plurality of first images and the plurality of second images to generate multiple depth images corresponding to different view angles, and stitches the multiple depth images as a 360 degree panoramic depth image. However, because optical centers of a plurality of image capturers included in the each image capturing group cannot completely overlap (that is, a mechanism of the image device has an intrinsic defect), the image device first needs to execute image processing on a plurality of images captured by the plurality of image capturers (e.g. integrating the optical centers of the plurality of image capturers into a virtual optical center) to generate the 360 degree panoramic depth image, otherwise a problem of blind spots will occur at stitching regions of the 360 degree panoramic depth image.
- In addition, when an electronic device provided by the prior art displays a 360 degree panoramic color image, the electronic device takes the electronic device as a center to display the 360 degree panoramic color image. That is, when a user controls the electronic device to view the 360 degree panoramic color image through a touch mode, the electronic device can only change a view angle for displaying the 360 degree panoramic color image. Therefore, when the electronic device is moved, the electronic device cannot change an image displayed by the electronic device according to motion of the electronic device.
- Therefore, how to solve the above-mentioned problems of the prior art becomes an important issue.
- An embodiment of the present invention provides an image device for generating depth images. The image device includes at least two image capturers and a rotating device. When the rotating device rotates the at least two image capturers, multiple images captured by the at least two image capturers are utilized to generate a depth image, wherein a view angle corresponding to the depth image is not less than a view angle of each image capturer of the at least two image capturers.
- Another embodiment of the present invention provides an image device for generating depth images. The image device includes at least one image capturer, at least one light source, and a rotating device. The at least one light source is used for emitting emission light. When the rotating device rotates the at least one image capturer and the at least one light source, multiple images captured by the at least one image capturer and the emission light are used for generating a depth image, wherein a view angle corresponding to the depth image is not less than a view angle of each image capturer of the at least one image capturer.
- Another embodiment of the present invention provides an electronic device for viewing a panoramic color and depth image. The electronic device includes a display and a processor. The processor is used for reading a 360 degree panoramic color image and a 360 degree panoramic depth image, wherein the 360 degree panoramic color image corresponds to the 360 degree panoramic depth image. When the electronic device is moved or rotated, the processor further converts a part of the 360 degree panoramic color image corresponding to a view angle corresponding to motion or rotation of the electronic device into a planar color image and a part of the 360 degree panoramic depth image corresponding to the view angle into a planar depth image according to the view angle, combines the planar color image with the planar depth image to generate a planar color and depth image corresponding to the view angle, and makes the display display the planar color image according to the planar color and depth image.
- The present invention provides an image device for generating depth images and related electronic device. Because the image device utilizes a rotating device to rotate a plurality of image capturers included in the image device, and utilizes a processor to generate a depth image according to plurality of images captured by the plurality of image capturers, compared to the prior art, the image device does not have a problem that optical centers in the prior art cannot overlap. In addition, because a 360 degree panoramic color and depth image generated by the present invention has depth information of a 360 degree panoramic depth image, compared to the prior art, when the electronic device displays the 360 degree panoramic color and depth image, the electronic device can display a corresponding planar color and depth image according to motion or rotation of the electronic device
- These and other objectives of the present invention will no doubt become obvious to those of ordinary skill in the art after reading the following detailed description of the preferred embodiment that is illustrated in the various figures and drawings.
-
FIG. 1 is a diagram illustrating an image device for generating depth images according to a first embodiment of the present invention. -
FIG. 2 is a diagram illustrating a top view of the image device. -
FIG. 3A is a diagram illustrating an image device for generating depth images according to a second embodiment of the present invention,FIG. 3B is a diagram illustrating an image device for generating depth images according to another embodiment of the present invention, andFIG. 3C is a diagram illustrating animage device 370 for generating depth images according to another embodiment of the present invention. -
FIG. 4 is a diagram illustrating an electronic device for viewing a panoramic color and depth image according to a third embodiment of the present invention. -
FIG. 5 is a diagram illustrating the electronic device being moved from a position to another position. -
FIG. 6 is a diagram illustrating a block of the display displaying the corresponding left eye color image and another block of the display displaying the corresponding right eye color image. - Please refer to
FIG. 1 .FIG. 1 is a diagram illustrating animage device 100 for generating depth images according to a first embodiment of the present invention, wherein as shown inFIG. 1 , theimage device 100 includes twoimage capturers unit 106, arotating device 108, and aprocessor 110, the supportingunit 106 is coupled to therotating device 108, theprocessor 110 is coupled to theimage capturers image capturers unit 106, and theimage capturers image capturers image device 100 including theimage capturers image device 100 can include more than two image capturers. In addition, one of ordinary skill in the art should well know that each image capturer of theimage capturers processor 110 is installed in a host outside theimage device 100. - As shown in
FIG. 1 , a rotating axis RA of therotating device 108 passes through an optical center C1 of the image capturer 102 and an optical center C2 of the image capturer 104, and therotating device 108 can rotate theimage capturers unit 106, wherein therotating device 108 is driven by agear 112. But, in another embodiment of the present invention, therotating device 108 is driven by a motor. When therotating device 108 rotates the image capturers 102, 104, if a sum of view angles corresponding to a plurality of first images captured by the image capturer 102 is not less than 360 degrees (a sum of view angles corresponding to a plurality of second images captured by the image capturer 104 is not less than 360 degrees), theprocessor 110 can stitch the plurality of first images to generate a first 360 degree panoramic color image and stitch the plurality of second images to generate a second 360 degree panoramic color image according to a feature point matching method or a fixed angle method provided by the prior art. For example, when therotating device 108 rotates theimage capturers FIG. 2 , and all of the first view angle FOV1, the second view angle FOV2, the third view angle FOV3, and the fourth view angle FOV4 are equal to 90 degrees. But, in another embodiment of the present invention, the first view angle FOV1, the second view angle FOV2, the third view angle FOV3, and the fourth view angle FOV4 can be different from each other. In addition, becauseFIG. 2 is a top view of theimage device 100,FIG. 2 does not show the image capturer 104. Therefore, after the image capturer 102 captures the first images IM11-IM14 and the image capturer 104 captures the second images IM21-IM24, theprocessor 110 can stitch the first images IM11-IM14 to generate the first 360 degree panoramic color image and stitch the second images IM21-IM24 to generate the second 360 degree panoramic color image according to the feature point matching method (e.g. feature points of the first images IM11-IM14 and feature points of the second images IM21-IM24) or the fixed angle method (e.g. the first view angle FOV1, the second view angle FOV2, the third view angle FOV3, and the fourth view angle FOV4). In addition, after theprocessor 110 generates the first 360 degree panoramic color image and the second 360 degree panoramic color image, theprocessor 110 can utilize the first 360 degree panoramic color image and the second 360 degree panoramic color image to generate a 360 degree panoramic depth image, wherein in one embodiment of the present invention, each depth value of the 360 degree panoramic depth image corresponds to the optical center C1 or the optical center C2. But, in another embodiment of the present invention, the each depth value of the 360 degree panoramic depth image corresponds to a plane where the optical center C1 and the optical center C2 are located. In addition, in another embodiment of the present invention, theprocessor 110 can generate a first depth image corresponding to the first view angle FOV1 according to the first image IM11 and the second image IM21, generate a second depth image corresponding to the second view angle FOV2 according to the first image IM12 and the second image IM22, generate a third depth image corresponding to the third view angle FOV3 according to the first image IM13 and the second image IM23, and generate a fourth depth image corresponding to the fourth view angle FOV4 according to the first image IM14 and the second image IM24. Then, theprocessor 110 stitches the first depth image, the second depth image, the third depth image, and the fourth depth image to generate the 360 degree panoramic depth image according to the feature point matching method or the fixed angle method. In addition, in another embodiment of the present invention, when therotating device 108 rotates theimage capturers processor 110 stitching the plurality of first images (or stitching the plurality of second images) is also between 360 degrees and the view angle of the image capturer 102 (or the view angle of the image capturer 104). Meanwhile, a view angle corresponding to a depth image generated by theprocessor 110 according to the plurality of first images and the plurality of second images is also between 360 degrees and the view angle of the image capturer 102 (or the view angle of the image capturer 104). - After the
processor 110 generates the 360 degree panoramic depth image, theprocessor 110 can combine the first 360 degree panoramic color image (or the second 360 degree panoramic color image) with the 360 degree panoramic depth image to generate a 360 degree panoramic color and depth image, and theprocessor 110 compresses the first 360 degree panoramic color image (or the second 360 degree panoramic color image) according to a standard compression format (e.g. a Joint Photographic Experts Group (JPEG) format) and compresses the 360 degree panoramic depth image according to a lossy format (or a lossless format) to integrate the 360 degree panoramic color and depth image into a file, wherein a compressed 360 degree panoramic depth image is stored in a header of the file. That is, the compressed 360 degree panoramic depth image is usually stored in a user defined header of the file. In addition, the standard compression format is not limited to JPEG, that is, the first 360 degree panoramic color image (or the second 360 degree panoramic color image) can also be compressed according to PNG, MPEG1, MPEG2, MPEG4, H.264, H.265, and so on. In addition, in another embodiment of the present invention, when the 360 degree panoramic color and depth image is integrated into the file, the 360 degree panoramic depth image is not compressed - Please refer to
FIG. 3A .FIG. 3A is a diagram illustrating animage device 300 for generating depth images according to a second embodiment of the present invention, wherein as shown inFIG. 3A , differences between theimage device 300 and theimage device 100 are that theimage device 300 further includes alight source 302, animage capturer 304 are used for capturing a plurality of second images including emission light generated by thelight source 302, theprocessor 110 is coupled to thelight source 302, theimage capturer 304, and theimage capturer 102, and thelight source 302, theimage capturer 304, and theimage capturer 102 are installed on the supportingunit 106. In addition, the emission light is structured light (or the emission light includes a random pattern), and thelight source 302 utilizes a collimator, at least one laser beam (e.g. at least one infrared laser beam), and at least one optical element (e.g. at least one diffraction optical element (DOE) to generate the emission light. But, in another embodiment of the present invention, thelight source 302 does not need the collimator. As shown inFIG. 3A , when therotating device 108 rotates thelight source 302, theimage capturer 304, and theimage capturer 102, the plurality of first images captured by theimage capturer 102 are used for generating a 360 degree panoramic color image, the plurality of second images including the emission light captured by theimage capturer 304 are used for generating a 360 degree panoramic depth image, wherein the sum of the view angles corresponding to the plurality of first images is not less than 360 degrees, and a sum of view angles corresponding to the plurality of second images is also not less than 360 degrees. In addition, in one embodiment of the present invention, theprocessor 110 can turn thelight source 302 according to at least one of luminance of an environment which theimage device 300 is located at, a quality of the 360 degree panoramic depth image, and a difference of the 360 degree panoramic depth image corresponding to turning-on and turning-off of thelight source 302. In addition, in another embodiment of the present invention, when thelight source 302 is a vertical-cavity surface-emitting Laser (VCSEL), thelight source 302 optionally includes the collimator and the at least one optical element. That is, the collimator and the at least one are not necessary for thelight source 302. In addition, in another embodiment of the present invention, thelight source 302 can be a light source with other configurations and wavelengths. - In addition, in another embodiment of the present invention, the
processor 110 can optionally adjust intensity of thelight source 302 according to luminance corresponding to each second image of the plurality of second images and a target value, wherein the target value is set according to reflection coefficient of a human skin of a user corresponding to the structured light emitted by thelight source 302. For example, theprocessor 110 can generate a luminance distribution map corresponding to the each second image according to the each second image, and optionally adjust the intensity of thelight source 302 according to a percentage of the each second image occupied by an area corresponding to a maximum luminance value of at least one luminance value within the luminance distribution map greater than the target value. In addition, in another embodiment of the present invention, theprocessor 110 can optionally adjust the intensity of thelight source 302 according to average luminance of the each second image and the target value. In addition, in another embodiment of the present invention, theprocessor 110 can generate a luminance histogram corresponding to a plurality of pixels of the each second image according to the each second image, and optionally adjust the intensity of thelight source 302 according to a median of the luminance histogram and the target value, or according to a predetermined quantile of the luminance histogram and the target value. - In addition, in another embodiment of the present invention, after the
light source 302 is turned on, theprocessor 110 can optionally dynamically adjust the intensity of thelight source 302 according to a distance between at least one predetermined object within the each second image and the image capturer 304 (or the image capturer 102) and a first lookup table, wherein the first lookup table stores relationships between a distance corresponding to an object and the intensity of thelight source 302. In addition, in another embodiment of the present invention, theprocessor 110 continuously detects the luminance of the environment which theimage device 300 is located at under thelight source 302 being turned off. When the luminance of the environment is brighter, theprocessor 110 increases the intensity of the light source 302 (when thelight source 302 is turned on) according to a second lookup table, wherein the second lookup table stores relationships between the intensity of the light source 302 (when thelight source 302 is turned on) and the luminance of the environment. - In addition, in the above-mentioned embodiments, when the
processor 110 utilizes a first pulse width modulation signal in a continuous mode to adjust the intensity of the emitted light, theprocessor 110 can adjust the intensity of the emitted light by changing a duty cycle of the first pulse width modulation signal; in the above-mentioned embodiments, when theprocessor 110 utilizes a second pulse width modulation signal in a burst mode to adjust the intensity of the emitted light, theprocessor 110 can adjust the intensity of the emitted light by changing an enabling time of the second pulse width modulation signal; in the above-mentioned embodiments, when theprocessor 110 utilizes the first pulse width modulation signal and the second pulse width modulation signal to adjust the intensity of the emitted light, theprocessor 110 can adjust the intensity of the emitted light by simultaneously changing the enabling time of the second pulse width modulation signal and the duty cycle of the first pulse width modulation signal. - In addition, subsequent operational principles of the
image device 300 can be referred to those of theimage device 100, so further description thereof is omitted for simplicity. - In addition, in another embodiment of the present invention, the
light source 302 is applied to a time of flight (TOF), wherein when thelight source 302 is applied to the time of flight, the emission light is diffused light, that is, the emission light is uniform light. Therefore, theprocessor 110 can generate the 360 degree panoramic depth image according to a difference between a receiving time for theimage capturer 304 receiving reflected light and a generating time corresponding to the emission light, wherein the reflected light is generated by at least one object reflecting the emitted light, and meanwhile theimage capturer 304 is a time of flight sensor. In addition, when thelight source 302 is applied to the time of flight, operational principles of theprocessor 110 determining whether to adjust the intensity of the emitted light can be referred to the above-mentioned corresponding descriptions, so further description thereof is omitted for simplicity. - In addition, because the above-mentioned functions of the
processor 110 are fully disclosed, one of ordinary skill in the art should easily utilize a field programmable gate array (FPGA) with the above-mentioned functions of theprocessor 110, or an application-specific integrated circuit (ASIC) with the above-mentioned functions of theprocessor 110, or a software module with the above-mentioned functions of theprocessor 110, or an analog integrated circuit with the above-mentioned functions of theprocessor 110 to realize theprocessor 110 according to corresponding descriptions of the above-mentioned functions ofprocessor 110. Therefore, a corresponding structure of theprocessor 110 is omitted for simplicity. - In addition, in another embodiment of the present invention, the
image capturer 102 and theimage capturer 304 can act as a stereo camera (or a depth camera), and the emission light generated by thelight source 302 is used for assisting the stereo camera. Theprocessor 110 can utilize a time division multiplexing method to control thelight source 302. When theprocessor 110 control thelight source 302 to generate the emission light, the stereo camera acts as the stereo camera, and when theprocessor 110 turns off thelight source 302, images captured by theimage capturer 102 and theimage capturer 304 are used for generating a color image. - In addition, please refer to
FIG. 3B .FIG. 3B is a diagram illustrating animage device 350 for generating depth images according to another embodiment of the present invention, wherein as shown inFIG. 3B , a difference between theimage device 350 and theimage device 300 is that theimage device 350 further includes animage capturer 306, wherein theimage capturer 102 and theimage capturer 304 can act as a stereo camera (or a depth camera), the emission light generated by thelight source 302 is used for assisting the stereo camera, and images captured by theimage capturer 306 are used for generating a color image (that is, theimage capturer 306 only acts as a color sensor and has an infrared light filter). In addition, operational principles of theimage capturer 102, theimage capturer 304, and theimage capturer 306 can be referred to the above-mentioned corresponding descriptions, so further description thereof is omitted for simplicity. - In addition, please refer to
FIG. 3C .FIG. 3C is a diagram illustrating animage device 370 for generating depth images according to another embodiment of the present invention, wherein as shown inFIG. 3C , a difference between theimage device 370 and theimage device 300 is that theimage device 370 does not include theimage capturer 304. As shown inFIG. 3C , when the emission light is structured light, theimage capturer 102 is a structured light depth sensor. In addition, theimage capturer 102 can be a time of flight sensor, meanwhile the emission light is diffused light, and the operational principles of theimage capturer 102 can be referred to the above-mentioned corresponding descriptions, so further description thereof is omitted for simplicity. - Please refer to
FIG. 4 .FIG. 4 is a diagram illustrating anelectronic device 400 for viewing a panoramic color and depth image according to a third embodiment of the present invention, wherein as shown inFIG. 4 , theelectronic device 400 includes adisplay 402, aprocessor 404, and an inertial sensor 406 (e.g. an inertial measurement unit, IMU), theprocessor 404 is coupled to thedisplay 402 and the inertial sensor 406 (because theprocessor 404 and theinertial sensor 406 are installed within theelectronic device 400, related coupling lines corresponding to theprocessor 404 and theinertial sensor 406 are represented by dotted lines), and theinertial sensor 406 can be a gyro, an accelerator, a microelectromechanical systems (MEMS), and so on. In addition, theelectronic device 400 can be a smart phone or a tablet computer. But, the present invention is not limited to theelectronic device 400 being a smart phone or a tablet computer. As shown inFIG. 4 , after theprocessor 404 receives a 360 degree panoramic color and depth image generated from the image device 100 (or the image device 300), theprocessor 404 can first read a 360 degree panoramic color image and a 360 degree panoramic depth image of the 360 degree panoramic color and depth image, or respectively read a 360 degree panoramic color image and a 360 degree panoramic depth image generated from other image devices, wherein each depth value of the 360 degree panoramic depth image corresponds to a virtual optical center which theelectronic device 400 acts as. - When the
electronic device 400 is moved or rotated (e.g. as shown inFIG. 5 , theelectronic device 400 is moved from a position A to a position B), theinertial sensor 406 can determine a view angle corresponding to motion or rotation of theelectronic device 400 according to the motion or the rotation of the electronic device 400 (that is, the view angle corresponds to the position B). After theinertial sensor 406 determines the view angle, theprocessor 404 can convert a part of the 360 degree panoramic color image corresponding to the view angle into a planar color image and convert a part of the 360 degree panoramic depth image corresponding to the view angle into a planar depth image according to the view angle, and utilize a depth-image-based rendering (DIBR) provided by the prior art to combine the planar color image with the planar depth image to generate a planar color and depth image corresponding to the view angle, wherein when thedisplay 402 displays the planar color and depth image, because the planar color and depth image has depth information of the planar depth image, the planar color and depth image can make a user watching the planar color and depth image have a three-dimensional feeling. In addition, each depth value of the planar depth image corresponds to a virtual optical plane where theelectronic device 400 is located. In addition, in another embodiment of the present invention, the user of theelectronic device 400 utilizes a mouse or touches thedisplay 402 to change a view angle of thedisplay 402 for displaying the 360 degree panoramic color and depth image. In addition, in another embodiment of the present invention, the 360 degree panoramic color image and the 360 degree panoramic depth image correspond to two different optical centers, theprocessor 404 can utilize the prior art to compensate a shift between the two different optical centers when thedisplay 402 displays the planar color and depth image. - In addition, because the planar color and depth image has the depth information of the planar depth image, when the
electronic device 400 is applied to virtual reality (VR), augmented reality (AR), substitutional reality (SR), and mixed reality (MR), theprocessor 404 can convert the planar color and depth image into a corresponding left eye color image and a corresponding right eye color image according to the depth information of the planar depth image and the depth-image-based rendering, wherein as shown inFIG. 6 , ablock 4022 of thedisplay 402 displays the corresponding left eye color image and a block 4024 of thedisplay 402 displays the corresponding right eye color image. Therefore, when theelectronic device 400 is mounted in a head mounted display, the user can view contents of the virtual reality (VR), the augmented reality (AR), the substitutional reality (SR), and the mixed reality (MR) through the head mounted display and theelectronic device 400. In addition, in another embodiment of the present invention, theblock 4022 of thedisplay 402 is independent from the block 4024 of thedisplay 402. That is, theblock 4022 and the block 4024 are two independent displays. - In addition, because the above-mentioned functions of the
processor 404 are fully disclosed, one of ordinary skill in the art should easily utilize a field programmable gate array with the above-mentioned functions of theprocessor 404, or an application-specific integrated circuit with the above-mentioned functions of theprocessor 404, or a software module with the above-mentioned functions of theprocessor 404, or an analog integrated circuit with the above-mentioned functions of theprocessor 404 to realize theprocessor 404 according to corresponding descriptions of the above-mentioned functions ofprocessor 404. Therefore, a corresponding structure of theprocessor 404 is omitted for simplicity. - To sum up, because the image device utilizes the rotating device to rotate the image capturers, and utilizes the processor to generate a 360 degree panoramic depth image according to multiple images captured by the image capturers, compared to the prior art, the image device does not have a problem that optical centers in the prior art cannot overlap. In addition, because a 360 degree panoramic color and depth image generated by the present invention has depth information of a 360 degree panoramic depth image, compared to the prior art, when the electronic device displays the 360 degree panoramic color and depth image, the electronic device can display a corresponding planar color and depth image according to motion or rotation of the electronic device.
- Those skilled in the art will readily observe that numerous modifications and alterations of the device and method may be made while retaining the teachings of the invention. Accordingly, the above disclosure should be construed as limited only by the metes and bounds of the appended claims.
Claims (16)
1. An image device for generating depth images, comprising:
at least two image capturers; and
a rotating device, wherein when the rotating device rotates the at least two image capturers, multiple images captured by the at least two image capturers are utilized to generate a depth image, wherein a view angle corresponding to the depth image is not less than a view angle of each image capturer of the at least two image capturers.
2. The image device of claim 1 , further comprising:
a supporting unit coupled to the rotating device, wherein the at least two image capturers are installed on the supporting unit.
3. The image device of claim 1 , wherein a rotation axis of the rotating device passes through an optical center of the each image capturer.
4. The image device of claim 1 , wherein a processor stitches a plurality of first images captured by an image capturer of the at least two image capturers to generate a color image according to a feature point matching method or a fixed angle method, and stitches a plurality of second images captured by another image capturer of the at least two image capturers to generate the depth image according to the feature point matching method or the fixed angle method, wherein the another image capturer is different from the image capturer.
5. The image device of claim 4 , wherein the processor is installed in the image device or outside the image device.
6. The image device of claim 4 , wherein the color image and the depth image are integrated into a color and depth image, and the color and depth image is stored as a file.
7. The image device of claim 6 , wherein when the color and depth image is stored as the file, the color image is compressed according to a Joint Photographic Experts Group (JPEG) format, the depth image is compressed according to a lossless format, and a compressed depth image is stored in a header of the file.
8. An image device for generating depth images, comprising:
at least one image capturer;
at least one light source emitting emission light; and
a rotating device;
wherein when the rotating device rotates the at least one image capturer and the at least one light source, multiple images captured by the at least one image capturer and the emission light are used for generating a depth image, wherein a view angle corresponding to the depth image is not less than a view angle of each image capturer of the at least one image capturer.
9. The image device of claim 8 , wherein when the rotating device rotates the at least one image capturer and the at least one light source, a plurality of images comprising the emission light captured by an image capturer of the at least one image capturer or flight time corresponding to the emission light passing from the at least one light source to each object and passing from the each object to the image capturer after the emission light is reflected by the each object is used for generating the depth map.
10. The image device of claim 9 , wherein when the plurality of images are used for generating the depth image, the emission light is structured light.
11. The image device of claim 8 , further comprising:
a supporting unit coupled to the rotating device, wherein the at least one image capturer and the at least one light source are installed on the supporting unit.
12. The image device of claim 8 , wherein a rotating axis of the rotating device passes through an optical center of the each image capturer of the at least one image capturer.
13. An electronic device for viewing a panoramic color and depth image, comprising:
a display; and
a processor reading a 360 degree panoramic color image and a 360 degree panoramic depth image, wherein the 360 degree panoramic color image corresponds to the 360 degree panoramic depth image;
wherein when the electronic device is moved or rotated, the processor further converts a part of the 360 degree panoramic color image corresponding to a view angle corresponding to motion or rotation of the electronic device into a planar color image and a part of the 360 degree panoramic depth image corresponding to the view angle into a planar depth image according to the view angle, combines the planar color image with the planar depth image to generate a planar color and depth image corresponding to the view angle, and makes the display display the planar color image according to the planar color and depth image.
14. The electronic device of claim 13 , further comprising:
an inertial sensor determining the view angle corresponding to the motion or the rotation of the electronic device according to the motion or the rotation of the electronic device.
15. The electronic device of claim 13 , wherein each depth value of the 360 degree panoramic depth image corresponds to a virtual optical center which the electronic device acts as, and each depth value of the planar depth image corresponds to a virtual optical plane where the electronic device is located.
16. The electronic device of claim 13 , wherein when the processor combines the planar color image with the planar depth image to generate the planar color and depth image, the processor converts the planar depth image into depth information corresponding to a virtual optical plane to make the planar color and depth image have the depth information.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/432,911 US20190387165A1 (en) | 2018-06-07 | 2019-06-05 | Image device for generating depth images and related electronic device |
US17/142,251 US11438525B2 (en) | 2018-06-07 | 2021-01-06 | Image device for generating depth images and related electronic device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201862681683P | 2018-06-07 | 2018-06-07 | |
US16/432,911 US20190387165A1 (en) | 2018-06-07 | 2019-06-05 | Image device for generating depth images and related electronic device |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/142,251 Division US11438525B2 (en) | 2018-06-07 | 2021-01-06 | Image device for generating depth images and related electronic device |
Publications (1)
Publication Number | Publication Date |
---|---|
US20190387165A1 true US20190387165A1 (en) | 2019-12-19 |
Family
ID=68764333
Family Applications (5)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/432,911 Abandoned US20190387165A1 (en) | 2018-06-07 | 2019-06-05 | Image device for generating depth images and related electronic device |
US16/432,888 Active US11582402B2 (en) | 2018-06-07 | 2019-06-05 | Image processing device |
US16/432,935 Active US10708520B2 (en) | 2018-06-07 | 2019-06-06 | Calibration method of an image device and related image device and operational device thereof |
US17/142,251 Active US11438525B2 (en) | 2018-06-07 | 2021-01-06 | Image device for generating depth images and related electronic device |
US18/092,947 Pending US20230231965A1 (en) | 2018-06-07 | 2023-01-04 | Image processing device |
Family Applications After (4)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/432,888 Active US11582402B2 (en) | 2018-06-07 | 2019-06-05 | Image processing device |
US16/432,935 Active US10708520B2 (en) | 2018-06-07 | 2019-06-06 | Calibration method of an image device and related image device and operational device thereof |
US17/142,251 Active US11438525B2 (en) | 2018-06-07 | 2021-01-06 | Image device for generating depth images and related electronic device |
US18/092,947 Pending US20230231965A1 (en) | 2018-06-07 | 2023-01-04 | Image processing device |
Country Status (3)
Country | Link |
---|---|
US (5) | US20190387165A1 (en) |
CN (2) | CN110580718B (en) |
TW (2) | TWI723419B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220114298A1 (en) * | 2020-10-13 | 2022-04-14 | Flyreel, Inc. | Generating measurements of physical structures and environments through automated analysis of sensor data |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190387165A1 (en) * | 2018-06-07 | 2019-12-19 | Eys3D Microelectronics, Co. | Image device for generating depth images and related electronic device |
CN112105312A (en) * | 2018-07-03 | 2020-12-18 | 柯惠Lp公司 | Systems, methods, and computer-readable media for detecting image degradation during a surgical procedure |
US11215828B1 (en) * | 2018-08-03 | 2022-01-04 | Rockwell Collins, Inc. | In field visor characterization for visor projected displays |
WO2021216037A1 (en) * | 2020-04-20 | 2021-10-28 | Hewlett-Packard Development Company, L.P. | Image rendering |
CN115278067A (en) * | 2022-07-19 | 2022-11-01 | Oppo广东移动通信有限公司 | Camera, electronic device, photographing method, and storage medium |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130077882A1 (en) * | 2011-09-28 | 2013-03-28 | Pelican Imaging Corporation | Systems and methods for decoding light field image files |
US20140225985A1 (en) * | 2012-10-17 | 2014-08-14 | DotProduct LLC | Handheld portable optical scanner and method of using |
US20150356358A1 (en) * | 2013-02-21 | 2015-12-10 | Sharp Kabushiki Kaisha | Obstacle detection device and obstacle detection method |
US20160029009A1 (en) * | 2014-07-24 | 2016-01-28 | Etron Technology, Inc. | Attachable three-dimensional scan module |
US20170161561A1 (en) * | 2015-10-05 | 2017-06-08 | Pillar Vision, Inc. | Systems and methods for monitoring objects at sporting events |
US20180108150A1 (en) * | 2016-10-17 | 2018-04-19 | Dsi Assignments, Llc | Systems and methods for in-field stereocamera calibration |
US20180278916A1 (en) * | 2015-08-07 | 2018-09-27 | Samsung Electronics Co., Ltd. | Electronic device for generating 360-degree three-dimensional image and method therefor |
US20190082165A1 (en) * | 2014-05-20 | 2019-03-14 | Nextvr Inc. | Methods and apparatus including or for use with one or more cameras |
Family Cites Families (66)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6028795A (en) * | 1985-09-24 | 2000-02-22 | Hitachi, Ltd. | One chip semiconductor integrated circuit device having two modes of data write operation and bits setting operation |
US6552730B1 (en) * | 1984-10-05 | 2003-04-22 | Hitachi, Ltd. | Method and apparatus for bit operational process |
US4988984A (en) * | 1988-10-31 | 1991-01-29 | International Business Machines Corporation | Image interpolator for an image display system |
US5111192A (en) * | 1989-12-20 | 1992-05-05 | Xerox Corporation | Method to rotate a bitmap image 90 degrees |
US5986672A (en) * | 1997-10-14 | 1999-11-16 | Minnesota, Mining And Manufacturing 3M Center | Method and system for forming a rotated image on an imaging element using limited system resources |
US6563535B1 (en) * | 1998-05-19 | 2003-05-13 | Flashpoint Technology, Inc. | Image processing system for high performance digital imaging devices |
US6137491A (en) | 1998-06-05 | 2000-10-24 | Microsoft Corporation | Method and apparatus for reconstructing geometry using geometrically constrained structure from motion with points on planes |
EP1433335B1 (en) * | 2001-08-15 | 2010-09-29 | Koninklijke Philips Electronics N.V. | 3d video conferencing system |
JP4227468B2 (en) * | 2002-06-24 | 2009-02-18 | キヤノン株式会社 | Image forming apparatus and method, and control program |
US6958757B2 (en) * | 2003-07-18 | 2005-10-25 | Microsoft Corporation | Systems and methods for efficiently displaying graphics on a display device regardless of physical orientation |
JP4130207B2 (en) * | 2003-09-16 | 2008-08-06 | 富士通株式会社 | Image processing display device and image processing display method |
US7643038B2 (en) * | 2005-06-29 | 2010-01-05 | Qualcomm Incorporated | Virtual device buffer for embedded device |
CA2553473A1 (en) * | 2005-07-26 | 2007-01-26 | Wa James Tam | Generating a depth map from a tw0-dimensional source image for stereoscopic and multiview imaging |
US7535474B1 (en) * | 2005-12-09 | 2009-05-19 | Advanced Micro Devices, Inc. | System and method for rotating rasterized image data |
US7876977B2 (en) * | 2006-02-15 | 2011-01-25 | Conexant Systems, Inc. | Image processor and method of image rotation |
JP2009100342A (en) | 2007-10-18 | 2009-05-07 | Sanyo Electric Co Ltd | Camera calibration device, method and vehicle |
JP2009135686A (en) * | 2007-11-29 | 2009-06-18 | Mitsubishi Electric Corp | Stereoscopic video recording method, stereoscopic video recording medium, stereoscopic video reproducing method, stereoscopic video recording apparatus, and stereoscopic video reproducing apparatus |
CN101308018B (en) * | 2008-05-30 | 2010-09-15 | 汤一平 | Stereo vision measuring apparatus based on binocular omnidirectional visual sense sensor |
JP2010114576A (en) * | 2008-11-05 | 2010-05-20 | Seiko Epson Corp | Image processing apparatus |
CN101894366B (en) | 2009-05-21 | 2014-01-29 | 北京中星微电子有限公司 | Method and device for acquiring calibration parameters and video monitoring system |
CN101577795A (en) * | 2009-06-17 | 2009-11-11 | 深圳华为通信技术有限公司 | Method and device for realizing real-time viewing of panoramic picture |
US8670149B2 (en) * | 2009-08-03 | 2014-03-11 | Printable Technologies Inc. | Apparatus and methods for image processing optimization for variable data printing |
CN101710932B (en) | 2009-12-21 | 2011-06-22 | 华为终端有限公司 | Image stitching method and device |
KR101385514B1 (en) * | 2010-06-08 | 2014-04-16 | 주식회사 에스칩스 | Method And Apparatus for Transforming Stereoscopic Image by Using Depth Map Information |
KR101051509B1 (en) | 2010-06-28 | 2011-07-22 | 삼성전기주식회사 | Apparatus and method for controlling light intensity of camera |
US9020241B2 (en) * | 2011-03-03 | 2015-04-28 | Panasonic Intellectual Property Management Co., Ltd. | Image providing device, image providing method, and image providing program for providing past-experience images |
US8947551B2 (en) * | 2011-07-29 | 2015-02-03 | Aptina Imaging Corporation | Method and apparatus for frame rotation in the JPEG compressed domain |
JP2013038454A (en) * | 2011-08-03 | 2013-02-21 | Sony Corp | Image processor, method, and program |
JP2013090031A (en) * | 2011-10-14 | 2013-05-13 | Sony Corp | Image processing device, image processing method, and program |
US10848731B2 (en) * | 2012-02-24 | 2020-11-24 | Matterport, Inc. | Capturing and aligning panoramic image and depth data |
CN202838179U (en) * | 2012-09-26 | 2013-03-27 | 天津游奕科技有限公司 | System capable of achieving panoramic video interaction |
CN108401551B (en) * | 2012-10-26 | 2015-02-04 | 北京理工大学 | Twin-lens low-light stereoscopic full views imaging device and its ultra-large vision field distance measuring method |
US10237528B2 (en) * | 2013-03-14 | 2019-03-19 | Qualcomm Incorporated | System and method for real time 2D to 3D conversion of a video in a digital camera |
TWI498526B (en) | 2013-06-05 | 2015-09-01 | Nat Univ Chung Cheng | Environment depth measurement method and its image acquisition device |
CN104581136B (en) * | 2013-10-14 | 2017-05-31 | 钰立微电子股份有限公司 | The calibration method of image calibration system and stereo camera |
CN103729883B (en) | 2013-12-30 | 2016-08-24 | 浙江大学 | A kind of three-dimensional environment information gathering and reconfiguration system and method |
KR102265109B1 (en) | 2014-01-24 | 2021-06-15 | 삼성전자주식회사 | Method and apparatus for image processing |
CN104811680B (en) * | 2014-01-28 | 2017-04-12 | 聚晶半导体股份有限公司 | image obtaining device and image deformation correction method thereof |
CN103984037B (en) | 2014-04-30 | 2017-07-28 | 深圳市墨克瑞光电子研究院 | The mobile robot obstacle detection method and device of view-based access control model |
WO2015183145A1 (en) | 2014-05-27 | 2015-12-03 | Telefonaktiebolaget L M Ericsson (Publ) | Remote control of a pivotable stereoscopic camera |
CN104077585B (en) * | 2014-05-30 | 2017-09-22 | 小米科技有限责任公司 | Method for correcting image, device and terminal |
US9992483B2 (en) | 2014-09-03 | 2018-06-05 | Intel Corporation | Imaging architecture for depth camera mode with mode switching |
TWI577172B (en) * | 2014-09-10 | 2017-04-01 | 鈺立微電子股份有限公司 | Image calibration system and calibration method of a stereo camera |
JP6531367B2 (en) * | 2014-09-30 | 2019-06-19 | セイコーエプソン株式会社 | PRINTING APPARATUS, CONTROL APPARATUS, AND IMAGE PROCESSING METHOD |
US10356383B2 (en) * | 2014-12-24 | 2019-07-16 | Reald Spark, Llc | Adjustment of perceived roundness in stereoscopic image of a head |
CN105744254B (en) * | 2014-12-30 | 2017-12-29 | 钰立微电子股份有限公司 | Correct guidance system and correct the operating method of guidance system |
CN104602129B (en) * | 2015-01-27 | 2018-03-06 | 三星电子(中国)研发中心 | The player method and system of interactive multi-angle video |
CN106157358A (en) * | 2015-03-26 | 2016-11-23 | 成都理想境界科技有限公司 | Object fusion method based on video image and terminal |
WO2016154869A1 (en) | 2015-03-31 | 2016-10-06 | SZ DJI Technology Co., Ltd. | System and method for mobile platform operation |
TWI610250B (en) * | 2015-06-02 | 2018-01-01 | 鈺立微電子股份有限公司 | Monitor system and operation method thereof |
CN104966225A (en) * | 2015-07-08 | 2015-10-07 | 深圳爱布丁梦想科技有限公司 | Housing rental method and system based on mobile terminal and 3D panoramic image browsing |
US9958266B2 (en) | 2015-07-09 | 2018-05-01 | Mitutoyo Corporation | Chromatic range sensor including dynamic intensity compensation function |
CN106534828A (en) * | 2015-09-11 | 2017-03-22 | 钰立微电子股份有限公司 | Controller applied to a three-dimensional (3d) capture device and 3d image capture device |
CN105530431A (en) * | 2015-12-16 | 2016-04-27 | 景好 | Reflective panoramic imaging system and method |
CN105609087B (en) * | 2015-12-23 | 2018-11-16 | 天地伟业技术有限公司 | A kind of method of image sealer processing image |
JP7043407B2 (en) * | 2016-01-15 | 2022-03-29 | 華為技術有限公司 | Display method and terminal |
TWI584634B (en) | 2016-03-08 | 2017-05-21 | 聚晶半導體股份有限公司 | Electronic apparatus and method of generating depth map |
CN105866966A (en) * | 2016-06-14 | 2016-08-17 | 昆山罗伯斯特机器人科技有限公司 | Panoramic three-dimensional stereoscopic glasses automatically switching visual angle |
US20170372452A1 (en) * | 2016-06-22 | 2017-12-28 | Qualcomm Incorporated | Image rotation method and apparatus |
TWI592020B (en) | 2016-08-23 | 2017-07-11 | 國立臺灣科技大學 | Image correction method of projector and image correction system |
CN106921824A (en) * | 2017-05-03 | 2017-07-04 | 丁志宇 | Circulating type mixes light field imaging device and method |
CN107105219A (en) * | 2017-05-26 | 2017-08-29 | 碰海科技(北京)有限公司 | Many depth camera compact panoramic scanning equipment |
CN107167996A (en) | 2017-06-05 | 2017-09-15 | 深圳奥比中光科技有限公司 | The laser projection module and depth camera adaptively adjusted |
CN107465907B (en) * | 2017-08-16 | 2019-08-20 | 维沃移动通信有限公司 | A kind of image pickup method and mobile terminal |
EP3531376B1 (en) | 2018-02-21 | 2020-09-30 | Ficosa Adas, S.L.U. | Calibrating a camera of a vehicle |
US20190387165A1 (en) * | 2018-06-07 | 2019-12-19 | Eys3D Microelectronics, Co. | Image device for generating depth images and related electronic device |
-
2019
- 2019-06-05 US US16/432,911 patent/US20190387165A1/en not_active Abandoned
- 2019-06-05 US US16/432,888 patent/US11582402B2/en active Active
- 2019-06-06 TW TW108119668A patent/TWI723419B/en active
- 2019-06-06 CN CN201910492688.2A patent/CN110580718B/en active Active
- 2019-06-06 US US16/432,935 patent/US10708520B2/en active Active
- 2019-06-06 TW TW108119687A patent/TWI761684B/en active
- 2019-06-10 CN CN201910495324.XA patent/CN110581936A/en active Pending
-
2021
- 2021-01-06 US US17/142,251 patent/US11438525B2/en active Active
-
2023
- 2023-01-04 US US18/092,947 patent/US20230231965A1/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130077882A1 (en) * | 2011-09-28 | 2013-03-28 | Pelican Imaging Corporation | Systems and methods for decoding light field image files |
US20140225985A1 (en) * | 2012-10-17 | 2014-08-14 | DotProduct LLC | Handheld portable optical scanner and method of using |
US20150356358A1 (en) * | 2013-02-21 | 2015-12-10 | Sharp Kabushiki Kaisha | Obstacle detection device and obstacle detection method |
US20190082165A1 (en) * | 2014-05-20 | 2019-03-14 | Nextvr Inc. | Methods and apparatus including or for use with one or more cameras |
US20160029009A1 (en) * | 2014-07-24 | 2016-01-28 | Etron Technology, Inc. | Attachable three-dimensional scan module |
US20180278916A1 (en) * | 2015-08-07 | 2018-09-27 | Samsung Electronics Co., Ltd. | Electronic device for generating 360-degree three-dimensional image and method therefor |
US20170161561A1 (en) * | 2015-10-05 | 2017-06-08 | Pillar Vision, Inc. | Systems and methods for monitoring objects at sporting events |
US20180108150A1 (en) * | 2016-10-17 | 2018-04-19 | Dsi Assignments, Llc | Systems and methods for in-field stereocamera calibration |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220114298A1 (en) * | 2020-10-13 | 2022-04-14 | Flyreel, Inc. | Generating measurements of physical structures and environments through automated analysis of sensor data |
US11699001B2 (en) * | 2020-10-13 | 2023-07-11 | Flyreel, Inc. | Generating measurements of physical structures and environments through automated analysis of sensor data |
US20230259667A1 (en) * | 2020-10-13 | 2023-08-17 | Flyreel, Inc. | Generating measurements of physical structures and environments through automated analysis of sensor data |
US11960799B2 (en) * | 2020-10-13 | 2024-04-16 | Flyreel, Inc. | Generating measurements of physical structures and environments through automated analysis of sensor data |
Also Published As
Publication number | Publication date |
---|---|
US11582402B2 (en) | 2023-02-14 |
US20190378301A1 (en) | 2019-12-12 |
US10708520B2 (en) | 2020-07-07 |
TWI723419B (en) | 2021-04-01 |
TW202013956A (en) | 2020-04-01 |
US20230231965A1 (en) | 2023-07-20 |
TWI761684B (en) | 2022-04-21 |
US20210160438A1 (en) | 2021-05-27 |
US11438525B2 (en) | 2022-09-06 |
US20190379842A1 (en) | 2019-12-12 |
CN110581936A (en) | 2019-12-17 |
CN110580718A (en) | 2019-12-17 |
CN110580718B (en) | 2023-08-08 |
TW202011731A (en) | 2020-03-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11438525B2 (en) | Image device for generating depth images and related electronic device | |
CN111052727B (en) | Electronic device and control method thereof | |
US11049476B2 (en) | Minimal-latency tracking and display for matching real and virtual worlds in head-worn displays | |
US7872637B2 (en) | System and method for tracking a laser spot on a projected computer screen image | |
KR102078059B1 (en) | Prism-based Eye Tracking | |
US9325968B2 (en) | Stereo imaging using disparate imaging devices | |
TWI557491B (en) | Combined visible and non-visible projection system | |
KR102214193B1 (en) | Depth camera device, 3d image display system having the same and control methods thereof | |
US10079970B2 (en) | Controlling image focus in real-time using gestures and depth sensor data | |
EP2824923B1 (en) | Apparatus, system and method for projecting images onto predefined portions of objects | |
JP4707034B2 (en) | Image processing method and input interface device | |
CN108475180B (en) | Distributing video among multiple display areas | |
US7137707B2 (en) | Projector-camera system with laser pointers | |
US20180146186A1 (en) | Active illumination 3d imaging system | |
CN107439002B (en) | Depth imaging | |
JP2004312733A (en) | Device incorporating retina tracking and retina tracking system | |
JP2003134375A (en) | Image pickup system | |
CN111077671A (en) | Device control method and device, display device and storage medium | |
JP7314501B2 (en) | Display control device, display control method and display control program | |
US20080123956A1 (en) | Active environment scanning method and device | |
KR20190002908A (en) | Curved display apparatus and operation method thereof | |
US10712841B2 (en) | Display control device, display control system, display control method, and storage medium having stored thereon display control program | |
JP7412940B2 (en) | Imaging device, computer program and storage medium | |
US20240046584A1 (en) | Information processing apparatus | |
JP2016177121A (en) | Display system, image display device, and operation device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |