US20190387165A1 - Image device for generating depth images and related electronic device - Google Patents

Image device for generating depth images and related electronic device Download PDF

Info

Publication number
US20190387165A1
US20190387165A1 US16/432,911 US201916432911A US2019387165A1 US 20190387165 A1 US20190387165 A1 US 20190387165A1 US 201916432911 A US201916432911 A US 201916432911A US 2019387165 A1 US2019387165 A1 US 2019387165A1
Authority
US
United States
Prior art keywords
image
depth
capturer
color
view angle
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/432,911
Inventor
Chi-Feng Lee
Yung-Wei Chen
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
eYs3D Microelectronics Co
Original Assignee
eYs3D Microelectronics Co
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by eYs3D Microelectronics Co filed Critical eYs3D Microelectronics Co
Priority to US16/432,911 priority Critical patent/US20190387165A1/en
Publication of US20190387165A1 publication Critical patent/US20190387165A1/en
Priority to US17/142,251 priority patent/US11438525B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/2628Alteration of picture size, shape, position or orientation, e.g. zooming, rotation, rolling, perspective, translation
    • H04N5/23238
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/02Constructional features of telephone sets
    • H04M1/0202Portable telephone sets, e.g. cordless phones, mobile phones or bar type handsets
    • H04M1/026Details of the structure or mounting of specific components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • G06T7/85Stereo camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/97Determining parameters from multiple pictures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/02Constructional features of telephone sets
    • H04M1/0202Portable telephone sets, e.g. cordless phones, mobile phones or bar type handsets
    • H04M1/026Details of the structure or mounting of specific components
    • H04M1/0264Details of the structure or mounting of specific components for a camera module assembly
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/02Constructional features of telephone sets
    • H04M1/0202Portable telephone sets, e.g. cordless phones, mobile phones or bar type handsets
    • H04M1/026Details of the structure or mounting of specific components
    • H04M1/0266Details of the structure or mounting of specific components for a display module assembly
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72439User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for image or video messaging
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/15Processing image signals for colour aspects of image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/161Encoding, multiplexing or demultiplexing different image signal components
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/239Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/254Image signal generators using stereoscopic image cameras in combination with electromagnetic radiation sources for illuminating objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/296Synchronisation thereof; Control thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N17/00Diagnosis, testing or measuring for television systems or their details
    • H04N17/002Diagnosis, testing or measuring for television systems or their details for television cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/50Constructional details
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/50Constructional details
    • H04N23/54Mounting of pick-up tubes, electronic image sensors, deviation or focusing coils
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/56Cameras or camera modules comprising electronic image sensors; Control thereof provided with illuminating means
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/57Mechanical or electrical details of cameras or camera modules specially adapted for being embedded in other devices
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/698Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/95Computational photography systems, e.g. light-field imaging systems
    • H04N23/957Light-field or plenoptic cameras or camera modules
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0081Depth or disparity estimation from stereoscopic image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N2201/00Indexing scheme relating to scanning, transmission or reproduction of documents or the like, and to details thereof
    • H04N2201/32Circuits or arrangements for control or supervision between transmitter and receiver or between image input and image output device, e.g. between a still-image camera and its memory or between a still-image camera and a printer device
    • H04N2201/3201Display, printing, storage or transmission of additional information, e.g. ID code, date and time or title
    • H04N2201/3225Display, printing, storage or transmission of additional information, e.g. ID code, date and time or title of data relating to an image, a page or a document
    • H04N2201/3254Orientation, e.g. landscape or portrait; Location or order of the image data, e.g. in memory
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N2201/00Indexing scheme relating to scanning, transmission or reproduction of documents or the like, and to details thereof
    • H04N2201/32Circuits or arrangements for control or supervision between transmitter and receiver or between image input and image output device, e.g. between a still-image camera and its memory or between a still-image camera and a printer device
    • H04N2201/3201Display, printing, storage or transmission of additional information, e.g. ID code, date and time or title
    • H04N2201/3274Storage or retrieval of prestored additional information
    • H04N2201/3277The additional information being stored in the same storage device as the image data

Definitions

  • the present invention relates to an image device and an electronic device, and particularly to an image device that has a simpler mechanism to generate depth images, and an electronic device that can display a corresponding depth image according to motion or rotation of the electronic device.
  • an image device for generating panoramic depth images usually includes two image capturing groups, and each image capturing group includes more than two image capturers, wherein one image capturing group of the two image capturing group is used for capturing a plurality of first images, and the other image capturing group of the two image capturing group is used for capturing a plurality of second images. Then, a processor of the image device utilizes the plurality of first images and the plurality of second images to generate multiple depth images corresponding to different view angles, and stitches the multiple depth images as a 360 degree panoramic depth image.
  • the image device first needs to execute image processing on a plurality of images captured by the plurality of image capturers (e.g. integrating the optical centers of the plurality of image capturers into a virtual optical center) to generate the 360 degree panoramic depth image, otherwise a problem of blind spots will occur at stitching regions of the 360 degree panoramic depth image.
  • an electronic device provided by the prior art displays a 360 degree panoramic color image
  • the electronic device takes the electronic device as a center to display the 360 degree panoramic color image. That is, when a user controls the electronic device to view the 360 degree panoramic color image through a touch mode, the electronic device can only change a view angle for displaying the 360 degree panoramic color image. Therefore, when the electronic device is moved, the electronic device cannot change an image displayed by the electronic device according to motion of the electronic device.
  • An embodiment of the present invention provides an image device for generating depth images.
  • the image device includes at least two image capturers and a rotating device. When the rotating device rotates the at least two image capturers, multiple images captured by the at least two image capturers are utilized to generate a depth image, wherein a view angle corresponding to the depth image is not less than a view angle of each image capturer of the at least two image capturers.
  • the image device includes at least one image capturer, at least one light source, and a rotating device.
  • the at least one light source is used for emitting emission light.
  • the rotating device rotates the at least one image capturer and the at least one light source, multiple images captured by the at least one image capturer and the emission light are used for generating a depth image, wherein a view angle corresponding to the depth image is not less than a view angle of each image capturer of the at least one image capturer.
  • the electronic device for viewing a panoramic color and depth image.
  • the electronic device includes a display and a processor.
  • the processor is used for reading a 360 degree panoramic color image and a 360 degree panoramic depth image, wherein the 360 degree panoramic color image corresponds to the 360 degree panoramic depth image.
  • the processor When the electronic device is moved or rotated, the processor further converts a part of the 360 degree panoramic color image corresponding to a view angle corresponding to motion or rotation of the electronic device into a planar color image and a part of the 360 degree panoramic depth image corresponding to the view angle into a planar depth image according to the view angle, combines the planar color image with the planar depth image to generate a planar color and depth image corresponding to the view angle, and makes the display display the planar color image according to the planar color and depth image.
  • the present invention provides an image device for generating depth images and related electronic device. Because the image device utilizes a rotating device to rotate a plurality of image capturers included in the image device, and utilizes a processor to generate a depth image according to plurality of images captured by the plurality of image capturers, compared to the prior art, the image device does not have a problem that optical centers in the prior art cannot overlap. In addition, because a 360 degree panoramic color and depth image generated by the present invention has depth information of a 360 degree panoramic depth image, compared to the prior art, when the electronic device displays the 360 degree panoramic color and depth image, the electronic device can display a corresponding planar color and depth image according to motion or rotation of the electronic device
  • FIG. 1 is a diagram illustrating an image device for generating depth images according to a first embodiment of the present invention.
  • FIG. 2 is a diagram illustrating a top view of the image device.
  • FIG. 3A is a diagram illustrating an image device for generating depth images according to a second embodiment of the present invention
  • FIG. 3B is a diagram illustrating an image device for generating depth images according to another embodiment of the present invention
  • FIG. 3C is a diagram illustrating an image device 370 for generating depth images according to another embodiment of the present invention.
  • FIG. 4 is a diagram illustrating an electronic device for viewing a panoramic color and depth image according to a third embodiment of the present invention.
  • FIG. 5 is a diagram illustrating the electronic device being moved from a position to another position.
  • FIG. 6 is a diagram illustrating a block of the display displaying the corresponding left eye color image and another block of the display displaying the corresponding right eye color image.
  • FIG. 1 is a diagram illustrating an image device 100 for generating depth images according to a first embodiment of the present invention, wherein as shown in FIG. 1 , the image device 100 includes two image capturers 102 , 104 , a supporting unit 106 , a rotating device 108 , and a processor 110 , the supporting unit 106 is coupled to the rotating device 108 , the processor 110 is coupled to the image capturers 102 , 104 , the image capturers 102 , 104 are installed on the supporting unit 106 , and the image capturers 102 , 104 are non-fisheye image capturers.
  • the image capturers 102 , 104 are fisheye image capturers.
  • the present invention is not limited to the image device 100 including the image capturers 102 , 104 . That is, the image device 100 can include more than two image capturers.
  • each image capturer of the image capturers 102 , 104 at least includes a lens and an image sensor (e.g. a charge-coupled device (CCD) image sensor or a complementary metal-oxide-semiconductor image sensor).
  • the processor 110 is installed in a host outside the image device 100 .
  • a rotating axis RA of the rotating device 108 passes through an optical center C 1 of the image capturer 102 and an optical center C 2 of the image capturer 104 , and the rotating device 108 can rotate the image capturers 102 , 104 by the supporting unit 106 , wherein the rotating device 108 is driven by a gear 112 .
  • the rotating device 108 is driven by a motor.
  • the processor 110 can stitch the plurality of first images to generate a first 360 degree panoramic color image and stitch the plurality of second images to generate a second 360 degree panoramic color image according to a feature point matching method or a fixed angle method provided by the prior art.
  • the image capturer 102 can capture 4 first images IM 11 -IM 14 and the image capturer 104 can capture 4 second images IM 21 -IM 24 , wherein the first image IM 11 and the second image IM 21 correspond to a first view angle FOV 1 , the first image IM 12 and the second image IM 22 correspond to a second view angle FOV 2 , the first image IM 13 and the second image IM 23 correspond to a third view angle FOV 3 , the first image IM 14 and the second image IM 24 correspond to a fourth view angle FOV 4 , the first view angle FOV 1 , the second view angle FOV 2 , the third view angle FOV 3 , and the fourth view angle FOV 4 can be referred to FIG.
  • FIG. 2 is a top view of the image device 100 , FIG. 2 does not show the image capturer 104 .
  • the processor 110 can stitch the first images IM 11 -IM 14 to generate the first 360 degree panoramic color image and stitch the second images IM 21 -IM 24 to generate the second 360 degree panoramic color image according to the feature point matching method (e.g. feature points of the first images IM 11 -IM 14 and feature points of the second images IM 21 -IM 24 ) or the fixed angle method (e.g. the first view angle FOV 1 , the second view angle FOV 2 , the third view angle FOV 3 , and the fourth view angle FOV 4 ).
  • the feature point matching method e.g. feature points of the first images IM 11 -IM 14 and feature points of the second images IM 21 -IM 24
  • the fixed angle method e.g. the first view angle FOV 1 , the second view angle FOV 2 , the third view angle FOV 3 , and the fourth view angle FOV 4 .
  • the processor 110 can utilize the first 360 degree panoramic color image and the second 360 degree panoramic color image to generate a 360 degree panoramic depth image, wherein in one embodiment of the present invention, each depth value of the 360 degree panoramic depth image corresponds to the optical center C 1 or the optical center C 2 . But, in another embodiment of the present invention, the each depth value of the 360 degree panoramic depth image corresponds to a plane where the optical center C 1 and the optical center C 2 are located.
  • the processor 110 can generate a first depth image corresponding to the first view angle FOV 1 according to the first image IM 11 and the second image IM 21 , generate a second depth image corresponding to the second view angle FOV 2 according to the first image IM 12 and the second image IM 22 , generate a third depth image corresponding to the third view angle FOV 3 according to the first image IM 13 and the second image IM 23 , and generate a fourth depth image corresponding to the fourth view angle FOV 4 according to the first image IM 14 and the second image IM 24 .
  • the processor 110 stitches the first depth image, the second depth image, the third depth image, and the fourth depth image to generate the 360 degree panoramic depth image according to the feature point matching method or the fixed angle method.
  • the image capturer 102 can capture the plurality of first images and the image capturer 104 can capture the plurality of second images, wherein the plurality of first images correspond to the plurality of second images, a view angle corresponding to the plurality of first images is between 360 degrees and a view angle of the image capturer 102 (or a view angle of the image capturer 104 ).
  • a view angle corresponding to a color image generated by the processor 110 stitching the plurality of first images (or stitching the plurality of second images) is also between 360 degrees and the view angle of the image capturer 102 (or the view angle of the image capturer 104 ).
  • a view angle corresponding to a depth image generated by the processor 110 according to the plurality of first images and the plurality of second images is also between 360 degrees and the view angle of the image capturer 102 (or the view angle of the image capturer 104 ).
  • the processor 110 can combine the first 360 degree panoramic color image (or the second 360 degree panoramic color image) with the 360 degree panoramic depth image to generate a 360 degree panoramic color and depth image, and the processor 110 compresses the first 360 degree panoramic color image (or the second 360 degree panoramic color image) according to a standard compression format (e.g. a Joint Photographic Experts Group (JPEG) format) and compresses the 360 degree panoramic depth image according to a lossy format (or a lossless format) to integrate the 360 degree panoramic color and depth image into a file, wherein a compressed 360 degree panoramic depth image is stored in a header of the file. That is, the compressed 360 degree panoramic depth image is usually stored in a user defined header of the file.
  • JPEG Joint Photographic Experts Group
  • the standard compression format is not limited to JPEG, that is, the first 360 degree panoramic color image (or the second 360 degree panoramic color image) can also be compressed according to PNG, MPEG1, MPEG2, MPEG4, H.264, H.265, and so on.
  • the 360 degree panoramic depth image is not compressed
  • FIG. 3A is a diagram illustrating an image device 300 for generating depth images according to a second embodiment of the present invention, wherein as shown in FIG. 3A , differences between the image device 300 and the image device 100 are that the image device 300 further includes a light source 302 , an image capturer 304 are used for capturing a plurality of second images including emission light generated by the light source 302 , the processor 110 is coupled to the light source 302 , the image capturer 304 , and the image capturer 102 , and the light source 302 , the image capturer 304 , and the image capturer 102 are installed on the supporting unit 106 .
  • the emission light is structured light (or the emission light includes a random pattern), and the light source 302 utilizes a collimator, at least one laser beam (e.g. at least one infrared laser beam), and at least one optical element (e.g. at least one diffraction optical element (DOE) to generate the emission light.
  • the light source 302 does not need the collimator. As shown in FIG.
  • the rotating device 108 rotates the light source 302 , the image capturer 304 , and the image capturer 102 , the plurality of first images captured by the image capturer 102 are used for generating a 360 degree panoramic color image, the plurality of second images including the emission light captured by the image capturer 304 are used for generating a 360 degree panoramic depth image, wherein the sum of the view angles corresponding to the plurality of first images is not less than 360 degrees, and a sum of view angles corresponding to the plurality of second images is also not less than 360 degrees.
  • the processor 110 can turn the light source 302 according to at least one of luminance of an environment which the image device 300 is located at, a quality of the 360 degree panoramic depth image, and a difference of the 360 degree panoramic depth image corresponding to turning-on and turning-off of the light source 302 .
  • the light source 302 is a vertical-cavity surface-emitting Laser (VCSEL)
  • the light source 302 optionally includes the collimator and the at least one optical element. That is, the collimator and the at least one are not necessary for the light source 302 .
  • the light source 302 can be a light source with other configurations and wavelengths.
  • the processor 110 can optionally adjust intensity of the light source 302 according to luminance corresponding to each second image of the plurality of second images and a target value, wherein the target value is set according to reflection coefficient of a human skin of a user corresponding to the structured light emitted by the light source 302 .
  • the processor 110 can generate a luminance distribution map corresponding to the each second image according to the each second image, and optionally adjust the intensity of the light source 302 according to a percentage of the each second image occupied by an area corresponding to a maximum luminance value of at least one luminance value within the luminance distribution map greater than the target value.
  • the processor 110 can optionally adjust the intensity of the light source 302 according to average luminance of the each second image and the target value.
  • the processor 110 can generate a luminance histogram corresponding to a plurality of pixels of the each second image according to the each second image, and optionally adjust the intensity of the light source 302 according to a median of the luminance histogram and the target value, or according to a predetermined quantile of the luminance histogram and the target value.
  • the processor 110 can optionally dynamically adjust the intensity of the light source 302 according to a distance between at least one predetermined object within the each second image and the image capturer 304 (or the image capturer 102 ) and a first lookup table, wherein the first lookup table stores relationships between a distance corresponding to an object and the intensity of the light source 302 .
  • the processor 110 continuously detects the luminance of the environment which the image device 300 is located at under the light source 302 being turned off.
  • the processor 110 increases the intensity of the light source 302 (when the light source 302 is turned on) according to a second lookup table, wherein the second lookup table stores relationships between the intensity of the light source 302 (when the light source 302 is turned on) and the luminance of the environment.
  • the processor 110 when the processor 110 utilizes a first pulse width modulation signal in a continuous mode to adjust the intensity of the emitted light, the processor 110 can adjust the intensity of the emitted light by changing a duty cycle of the first pulse width modulation signal; in the above-mentioned embodiments, when the processor 110 utilizes a second pulse width modulation signal in a burst mode to adjust the intensity of the emitted light, the processor 110 can adjust the intensity of the emitted light by changing an enabling time of the second pulse width modulation signal; in the above-mentioned embodiments, when the processor 110 utilizes the first pulse width modulation signal and the second pulse width modulation signal to adjust the intensity of the emitted light, the processor 110 can adjust the intensity of the emitted light by simultaneously changing the enabling time of the second pulse width modulation signal and the duty cycle of the first pulse width modulation signal.
  • the light source 302 is applied to a time of flight (TOF), wherein when the light source 302 is applied to the time of flight, the emission light is diffused light, that is, the emission light is uniform light. Therefore, the processor 110 can generate the 360 degree panoramic depth image according to a difference between a receiving time for the image capturer 304 receiving reflected light and a generating time corresponding to the emission light, wherein the reflected light is generated by at least one object reflecting the emitted light, and meanwhile the image capturer 304 is a time of flight sensor.
  • operational principles of the processor 110 determining whether to adjust the intensity of the emitted light can be referred to the above-mentioned corresponding descriptions, so further description thereof is omitted for simplicity.
  • processor 110 the above-mentioned functions of the processor 110 are fully disclosed, one of ordinary skill in the art should easily utilize a field programmable gate array (FPGA) with the above-mentioned functions of the processor 110 , or an application-specific integrated circuit (ASIC) with the above-mentioned functions of the processor 110 , or a software module with the above-mentioned functions of the processor 110 , or an analog integrated circuit with the above-mentioned functions of the processor 110 to realize the processor 110 according to corresponding descriptions of the above-mentioned functions of processor 110 . Therefore, a corresponding structure of the processor 110 is omitted for simplicity.
  • FPGA field programmable gate array
  • ASIC application-specific integrated circuit
  • the image capturer 102 and the image capturer 304 can act as a stereo camera (or a depth camera), and the emission light generated by the light source 302 is used for assisting the stereo camera.
  • the processor 110 can utilize a time division multiplexing method to control the light source 302 .
  • the stereo camera acts as the stereo camera, and when the processor 110 turns off the light source 302 , images captured by the image capturer 102 and the image capturer 304 are used for generating a color image.
  • FIG. 3B is a diagram illustrating an image device 350 for generating depth images according to another embodiment of the present invention, wherein as shown in FIG. 3B , a difference between the image device 350 and the image device 300 is that the image device 350 further includes an image capturer 306 , wherein the image capturer 102 and the image capturer 304 can act as a stereo camera (or a depth camera), the emission light generated by the light source 302 is used for assisting the stereo camera, and images captured by the image capturer 306 are used for generating a color image (that is, the image capturer 306 only acts as a color sensor and has an infrared light filter).
  • operational principles of the image capturer 102 , the image capturer 304 , and the image capturer 306 can be referred to the above-mentioned corresponding descriptions, so further description thereof is omitted for simplicity.
  • FIG. 3C is a diagram illustrating an image device 370 for generating depth images according to another embodiment of the present invention, wherein as shown in FIG. 3C , a difference between the image device 370 and the image device 300 is that the image device 370 does not include the image capturer 304 .
  • the image capturer 102 when the emission light is structured light, the image capturer 102 is a structured light depth sensor.
  • the image capturer 102 can be a time of flight sensor, meanwhile the emission light is diffused light, and the operational principles of the image capturer 102 can be referred to the above-mentioned corresponding descriptions, so further description thereof is omitted for simplicity.
  • FIG. 4 is a diagram illustrating an electronic device 400 for viewing a panoramic color and depth image according to a third embodiment of the present invention, wherein as shown in FIG. 4 , the electronic device 400 includes a display 402 , a processor 404 , and an inertial sensor 406 (e.g.
  • the processor 404 is coupled to the display 402 and the inertial sensor 406 (because the processor 404 and the inertial sensor 406 are installed within the electronic device 400 , related coupling lines corresponding to the processor 404 and the inertial sensor 406 are represented by dotted lines), and the inertial sensor 406 can be a gyro, an accelerator, a microelectromechanical systems (MEMS), and so on.
  • the electronic device 400 can be a smart phone or a tablet computer. But, the present invention is not limited to the electronic device 400 being a smart phone or a tablet computer. As shown in FIG.
  • the processor 404 can first read a 360 degree panoramic color image and a 360 degree panoramic depth image of the 360 degree panoramic color and depth image, or respectively read a 360 degree panoramic color image and a 360 degree panoramic depth image generated from other image devices, wherein each depth value of the 360 degree panoramic depth image corresponds to a virtual optical center which the electronic device 400 acts as.
  • the inertial sensor 406 can determine a view angle corresponding to motion or rotation of the electronic device 400 according to the motion or the rotation of the electronic device 400 (that is, the view angle corresponds to the position B).
  • the processor 404 can convert a part of the 360 degree panoramic color image corresponding to the view angle into a planar color image and convert a part of the 360 degree panoramic depth image corresponding to the view angle into a planar depth image according to the view angle, and utilize a depth-image-based rendering (DIBR) provided by the prior art to combine the planar color image with the planar depth image to generate a planar color and depth image corresponding to the view angle, wherein when the display 402 displays the planar color and depth image, because the planar color and depth image has depth information of the planar depth image, the planar color and depth image can make a user watching the planar color and depth image have a three-dimensional feeling.
  • DIBR depth-image-based rendering
  • each depth value of the planar depth image corresponds to a virtual optical plane where the electronic device 400 is located.
  • the user of the electronic device 400 utilizes a mouse or touches the display 402 to change a view angle of the display 402 for displaying the 360 degree panoramic color and depth image.
  • the 360 degree panoramic color image and the 360 degree panoramic depth image correspond to two different optical centers, the processor 404 can utilize the prior art to compensate a shift between the two different optical centers when the display 402 displays the planar color and depth image.
  • the processor 404 can convert the planar color and depth image into a corresponding left eye color image and a corresponding right eye color image according to the depth information of the planar depth image and the depth-image-based rendering, wherein as shown in FIG. 6 , a block 4022 of the display 402 displays the corresponding left eye color image and a block 4024 of the display 402 displays the corresponding right eye color image.
  • the electronic device 400 when the electronic device 400 is mounted in a head mounted display, the user can view contents of the virtual reality (VR), the augmented reality (AR), the substitutional reality (SR), and the mixed reality (MR) through the head mounted display and the electronic device 400 .
  • the block 4022 of the display 402 is independent from the block 4024 of the display 402 . That is, the block 4022 and the block 4024 are two independent displays.
  • processor 404 the above-mentioned functions of the processor 404 are fully disclosed, one of ordinary skill in the art should easily utilize a field programmable gate array with the above-mentioned functions of the processor 404 , or an application-specific integrated circuit with the above-mentioned functions of the processor 404 , or a software module with the above-mentioned functions of the processor 404 , or an analog integrated circuit with the above-mentioned functions of the processor 404 to realize the processor 404 according to corresponding descriptions of the above-mentioned functions of processor 404 . Therefore, a corresponding structure of the processor 404 is omitted for simplicity.
  • the image device utilizes the rotating device to rotate the image capturers, and utilizes the processor to generate a 360 degree panoramic depth image according to multiple images captured by the image capturers, compared to the prior art, the image device does not have a problem that optical centers in the prior art cannot overlap.
  • a 360 degree panoramic color and depth image generated by the present invention has depth information of a 360 degree panoramic depth image, compared to the prior art, when the electronic device displays the 360 degree panoramic color and depth image, the electronic device can display a corresponding planar color and depth image according to motion or rotation of the electronic device.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Electromagnetism (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Human Computer Interaction (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Image Processing (AREA)
  • Studio Devices (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
  • Length Measuring Devices By Optical Means (AREA)

Abstract

An image device for generating depth images includes at least two image capturers and a rotating device. When the rotating device rotates the at least two image capturers, multiple images captured by the at least two image capturers are utilized to generate a depth image, wherein a view angle corresponding to the depth image is not less than a view angle of each image capturer of the at least two image capturers.

Description

    CROSS REFERENCE TO RELATED APPLICATION
  • This application claims the benefit of U.S. Provisional Application No. 62/681,683, filed on Jun. 7, 2018 and entitled “IMAGE PROCESSING METHOD AND SELF-CALIBRATION/LIGHT CONTROL METHOD FOR CAMERA DEVICE,” the contents of which are incorporated herein by reference.
  • BACKGROUND OF THE INVENTION 1. Field of the Invention
  • The present invention relates to an image device and an electronic device, and particularly to an image device that has a simpler mechanism to generate depth images, and an electronic device that can display a corresponding depth image according to motion or rotation of the electronic device.
  • 2. Description of the Prior Art
  • In the prior art, an image device for generating panoramic depth images usually includes two image capturing groups, and each image capturing group includes more than two image capturers, wherein one image capturing group of the two image capturing group is used for capturing a plurality of first images, and the other image capturing group of the two image capturing group is used for capturing a plurality of second images. Then, a processor of the image device utilizes the plurality of first images and the plurality of second images to generate multiple depth images corresponding to different view angles, and stitches the multiple depth images as a 360 degree panoramic depth image. However, because optical centers of a plurality of image capturers included in the each image capturing group cannot completely overlap (that is, a mechanism of the image device has an intrinsic defect), the image device first needs to execute image processing on a plurality of images captured by the plurality of image capturers (e.g. integrating the optical centers of the plurality of image capturers into a virtual optical center) to generate the 360 degree panoramic depth image, otherwise a problem of blind spots will occur at stitching regions of the 360 degree panoramic depth image.
  • In addition, when an electronic device provided by the prior art displays a 360 degree panoramic color image, the electronic device takes the electronic device as a center to display the 360 degree panoramic color image. That is, when a user controls the electronic device to view the 360 degree panoramic color image through a touch mode, the electronic device can only change a view angle for displaying the 360 degree panoramic color image. Therefore, when the electronic device is moved, the electronic device cannot change an image displayed by the electronic device according to motion of the electronic device.
  • Therefore, how to solve the above-mentioned problems of the prior art becomes an important issue.
  • SUMMARY OF THE INVENTION
  • An embodiment of the present invention provides an image device for generating depth images. The image device includes at least two image capturers and a rotating device. When the rotating device rotates the at least two image capturers, multiple images captured by the at least two image capturers are utilized to generate a depth image, wherein a view angle corresponding to the depth image is not less than a view angle of each image capturer of the at least two image capturers.
  • Another embodiment of the present invention provides an image device for generating depth images. The image device includes at least one image capturer, at least one light source, and a rotating device. The at least one light source is used for emitting emission light. When the rotating device rotates the at least one image capturer and the at least one light source, multiple images captured by the at least one image capturer and the emission light are used for generating a depth image, wherein a view angle corresponding to the depth image is not less than a view angle of each image capturer of the at least one image capturer.
  • Another embodiment of the present invention provides an electronic device for viewing a panoramic color and depth image. The electronic device includes a display and a processor. The processor is used for reading a 360 degree panoramic color image and a 360 degree panoramic depth image, wherein the 360 degree panoramic color image corresponds to the 360 degree panoramic depth image. When the electronic device is moved or rotated, the processor further converts a part of the 360 degree panoramic color image corresponding to a view angle corresponding to motion or rotation of the electronic device into a planar color image and a part of the 360 degree panoramic depth image corresponding to the view angle into a planar depth image according to the view angle, combines the planar color image with the planar depth image to generate a planar color and depth image corresponding to the view angle, and makes the display display the planar color image according to the planar color and depth image.
  • The present invention provides an image device for generating depth images and related electronic device. Because the image device utilizes a rotating device to rotate a plurality of image capturers included in the image device, and utilizes a processor to generate a depth image according to plurality of images captured by the plurality of image capturers, compared to the prior art, the image device does not have a problem that optical centers in the prior art cannot overlap. In addition, because a 360 degree panoramic color and depth image generated by the present invention has depth information of a 360 degree panoramic depth image, compared to the prior art, when the electronic device displays the 360 degree panoramic color and depth image, the electronic device can display a corresponding planar color and depth image according to motion or rotation of the electronic device
  • These and other objectives of the present invention will no doubt become obvious to those of ordinary skill in the art after reading the following detailed description of the preferred embodiment that is illustrated in the various figures and drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a diagram illustrating an image device for generating depth images according to a first embodiment of the present invention.
  • FIG. 2 is a diagram illustrating a top view of the image device.
  • FIG. 3A is a diagram illustrating an image device for generating depth images according to a second embodiment of the present invention, FIG. 3B is a diagram illustrating an image device for generating depth images according to another embodiment of the present invention, and FIG. 3C is a diagram illustrating an image device 370 for generating depth images according to another embodiment of the present invention.
  • FIG. 4 is a diagram illustrating an electronic device for viewing a panoramic color and depth image according to a third embodiment of the present invention.
  • FIG. 5 is a diagram illustrating the electronic device being moved from a position to another position.
  • FIG. 6 is a diagram illustrating a block of the display displaying the corresponding left eye color image and another block of the display displaying the corresponding right eye color image.
  • DETAILED DESCRIPTION
  • Please refer to FIG. 1. FIG. 1 is a diagram illustrating an image device 100 for generating depth images according to a first embodiment of the present invention, wherein as shown in FIG. 1, the image device 100 includes two image capturers 102, 104, a supporting unit 106, a rotating device 108, and a processor 110, the supporting unit 106 is coupled to the rotating device 108, the processor 110 is coupled to the image capturers 102, 104, the image capturers 102, 104 are installed on the supporting unit 106, and the image capturers 102, 104 are non-fisheye image capturers. But, in another embodiment of the present invention, the image capturers 102, 104 are fisheye image capturers. In addition, the present invention is not limited to the image device 100 including the image capturers 102, 104. That is, the image device 100 can include more than two image capturers. In addition, one of ordinary skill in the art should well know that each image capturer of the image capturers 102, 104 at least includes a lens and an image sensor (e.g. a charge-coupled device (CCD) image sensor or a complementary metal-oxide-semiconductor image sensor). In addition, in another embodiment of the present invention, the processor 110 is installed in a host outside the image device 100.
  • As shown in FIG. 1, a rotating axis RA of the rotating device 108 passes through an optical center C1 of the image capturer 102 and an optical center C2 of the image capturer 104, and the rotating device 108 can rotate the image capturers 102, 104 by the supporting unit 106, wherein the rotating device 108 is driven by a gear 112. But, in another embodiment of the present invention, the rotating device 108 is driven by a motor. When the rotating device 108 rotates the image capturers 102, 104, if a sum of view angles corresponding to a plurality of first images captured by the image capturer 102 is not less than 360 degrees (a sum of view angles corresponding to a plurality of second images captured by the image capturer 104 is not less than 360 degrees), the processor 110 can stitch the plurality of first images to generate a first 360 degree panoramic color image and stitch the plurality of second images to generate a second 360 degree panoramic color image according to a feature point matching method or a fixed angle method provided by the prior art. For example, when the rotating device 108 rotates the image capturers 102, 104, the image capturer 102 can capture 4 first images IM11-IM14 and the image capturer 104 can capture 4 second images IM21-IM24, wherein the first image IM11 and the second image IM21 correspond to a first view angle FOV1, the first image IM12 and the second image IM22 correspond to a second view angle FOV2, the first image IM13 and the second image IM23 correspond to a third view angle FOV3, the first image IM14 and the second image IM24 correspond to a fourth view angle FOV4, the first view angle FOV1, the second view angle FOV2, the third view angle FOV3, and the fourth view angle FOV4 can be referred to FIG. 2, and all of the first view angle FOV1, the second view angle FOV2, the third view angle FOV3, and the fourth view angle FOV4 are equal to 90 degrees. But, in another embodiment of the present invention, the first view angle FOV1, the second view angle FOV2, the third view angle FOV3, and the fourth view angle FOV4 can be different from each other. In addition, because FIG. 2 is a top view of the image device 100, FIG. 2 does not show the image capturer 104. Therefore, after the image capturer 102 captures the first images IM11-IM14 and the image capturer 104 captures the second images IM21-IM24, the processor 110 can stitch the first images IM11-IM14 to generate the first 360 degree panoramic color image and stitch the second images IM21-IM24 to generate the second 360 degree panoramic color image according to the feature point matching method (e.g. feature points of the first images IM11-IM14 and feature points of the second images IM21-IM24) or the fixed angle method (e.g. the first view angle FOV1, the second view angle FOV2, the third view angle FOV3, and the fourth view angle FOV4). In addition, after the processor 110 generates the first 360 degree panoramic color image and the second 360 degree panoramic color image, the processor 110 can utilize the first 360 degree panoramic color image and the second 360 degree panoramic color image to generate a 360 degree panoramic depth image, wherein in one embodiment of the present invention, each depth value of the 360 degree panoramic depth image corresponds to the optical center C1 or the optical center C2. But, in another embodiment of the present invention, the each depth value of the 360 degree panoramic depth image corresponds to a plane where the optical center C1 and the optical center C2 are located. In addition, in another embodiment of the present invention, the processor 110 can generate a first depth image corresponding to the first view angle FOV1 according to the first image IM11 and the second image IM21, generate a second depth image corresponding to the second view angle FOV2 according to the first image IM12 and the second image IM22, generate a third depth image corresponding to the third view angle FOV3 according to the first image IM13 and the second image IM23, and generate a fourth depth image corresponding to the fourth view angle FOV4 according to the first image IM14 and the second image IM24. Then, the processor 110 stitches the first depth image, the second depth image, the third depth image, and the fourth depth image to generate the 360 degree panoramic depth image according to the feature point matching method or the fixed angle method. In addition, in another embodiment of the present invention, when the rotating device 108 rotates the image capturers 102, 104, the image capturer 102 can capture the plurality of first images and the image capturer 104 can capture the plurality of second images, wherein the plurality of first images correspond to the plurality of second images, a view angle corresponding to the plurality of first images is between 360 degrees and a view angle of the image capturer 102 (or a view angle of the image capturer 104). Therefore, a view angle corresponding to a color image generated by the processor 110 stitching the plurality of first images (or stitching the plurality of second images) is also between 360 degrees and the view angle of the image capturer 102 (or the view angle of the image capturer 104). Meanwhile, a view angle corresponding to a depth image generated by the processor 110 according to the plurality of first images and the plurality of second images is also between 360 degrees and the view angle of the image capturer 102 (or the view angle of the image capturer 104).
  • After the processor 110 generates the 360 degree panoramic depth image, the processor 110 can combine the first 360 degree panoramic color image (or the second 360 degree panoramic color image) with the 360 degree panoramic depth image to generate a 360 degree panoramic color and depth image, and the processor 110 compresses the first 360 degree panoramic color image (or the second 360 degree panoramic color image) according to a standard compression format (e.g. a Joint Photographic Experts Group (JPEG) format) and compresses the 360 degree panoramic depth image according to a lossy format (or a lossless format) to integrate the 360 degree panoramic color and depth image into a file, wherein a compressed 360 degree panoramic depth image is stored in a header of the file. That is, the compressed 360 degree panoramic depth image is usually stored in a user defined header of the file. In addition, the standard compression format is not limited to JPEG, that is, the first 360 degree panoramic color image (or the second 360 degree panoramic color image) can also be compressed according to PNG, MPEG1, MPEG2, MPEG4, H.264, H.265, and so on. In addition, in another embodiment of the present invention, when the 360 degree panoramic color and depth image is integrated into the file, the 360 degree panoramic depth image is not compressed
  • Please refer to FIG. 3A. FIG. 3A is a diagram illustrating an image device 300 for generating depth images according to a second embodiment of the present invention, wherein as shown in FIG. 3A, differences between the image device 300 and the image device 100 are that the image device 300 further includes a light source 302, an image capturer 304 are used for capturing a plurality of second images including emission light generated by the light source 302, the processor 110 is coupled to the light source 302, the image capturer 304, and the image capturer 102, and the light source 302, the image capturer 304, and the image capturer 102 are installed on the supporting unit 106. In addition, the emission light is structured light (or the emission light includes a random pattern), and the light source 302 utilizes a collimator, at least one laser beam (e.g. at least one infrared laser beam), and at least one optical element (e.g. at least one diffraction optical element (DOE) to generate the emission light. But, in another embodiment of the present invention, the light source 302 does not need the collimator. As shown in FIG. 3A, when the rotating device 108 rotates the light source 302, the image capturer 304, and the image capturer 102, the plurality of first images captured by the image capturer 102 are used for generating a 360 degree panoramic color image, the plurality of second images including the emission light captured by the image capturer 304 are used for generating a 360 degree panoramic depth image, wherein the sum of the view angles corresponding to the plurality of first images is not less than 360 degrees, and a sum of view angles corresponding to the plurality of second images is also not less than 360 degrees. In addition, in one embodiment of the present invention, the processor 110 can turn the light source 302 according to at least one of luminance of an environment which the image device 300 is located at, a quality of the 360 degree panoramic depth image, and a difference of the 360 degree panoramic depth image corresponding to turning-on and turning-off of the light source 302. In addition, in another embodiment of the present invention, when the light source 302 is a vertical-cavity surface-emitting Laser (VCSEL), the light source 302 optionally includes the collimator and the at least one optical element. That is, the collimator and the at least one are not necessary for the light source 302. In addition, in another embodiment of the present invention, the light source 302 can be a light source with other configurations and wavelengths.
  • In addition, in another embodiment of the present invention, the processor 110 can optionally adjust intensity of the light source 302 according to luminance corresponding to each second image of the plurality of second images and a target value, wherein the target value is set according to reflection coefficient of a human skin of a user corresponding to the structured light emitted by the light source 302. For example, the processor 110 can generate a luminance distribution map corresponding to the each second image according to the each second image, and optionally adjust the intensity of the light source 302 according to a percentage of the each second image occupied by an area corresponding to a maximum luminance value of at least one luminance value within the luminance distribution map greater than the target value. In addition, in another embodiment of the present invention, the processor 110 can optionally adjust the intensity of the light source 302 according to average luminance of the each second image and the target value. In addition, in another embodiment of the present invention, the processor 110 can generate a luminance histogram corresponding to a plurality of pixels of the each second image according to the each second image, and optionally adjust the intensity of the light source 302 according to a median of the luminance histogram and the target value, or according to a predetermined quantile of the luminance histogram and the target value.
  • In addition, in another embodiment of the present invention, after the light source 302 is turned on, the processor 110 can optionally dynamically adjust the intensity of the light source 302 according to a distance between at least one predetermined object within the each second image and the image capturer 304 (or the image capturer 102) and a first lookup table, wherein the first lookup table stores relationships between a distance corresponding to an object and the intensity of the light source 302. In addition, in another embodiment of the present invention, the processor 110 continuously detects the luminance of the environment which the image device 300 is located at under the light source 302 being turned off. When the luminance of the environment is brighter, the processor 110 increases the intensity of the light source 302 (when the light source 302 is turned on) according to a second lookup table, wherein the second lookup table stores relationships between the intensity of the light source 302 (when the light source 302 is turned on) and the luminance of the environment.
  • In addition, in the above-mentioned embodiments, when the processor 110 utilizes a first pulse width modulation signal in a continuous mode to adjust the intensity of the emitted light, the processor 110 can adjust the intensity of the emitted light by changing a duty cycle of the first pulse width modulation signal; in the above-mentioned embodiments, when the processor 110 utilizes a second pulse width modulation signal in a burst mode to adjust the intensity of the emitted light, the processor 110 can adjust the intensity of the emitted light by changing an enabling time of the second pulse width modulation signal; in the above-mentioned embodiments, when the processor 110 utilizes the first pulse width modulation signal and the second pulse width modulation signal to adjust the intensity of the emitted light, the processor 110 can adjust the intensity of the emitted light by simultaneously changing the enabling time of the second pulse width modulation signal and the duty cycle of the first pulse width modulation signal.
  • In addition, subsequent operational principles of the image device 300 can be referred to those of the image device 100, so further description thereof is omitted for simplicity.
  • In addition, in another embodiment of the present invention, the light source 302 is applied to a time of flight (TOF), wherein when the light source 302 is applied to the time of flight, the emission light is diffused light, that is, the emission light is uniform light. Therefore, the processor 110 can generate the 360 degree panoramic depth image according to a difference between a receiving time for the image capturer 304 receiving reflected light and a generating time corresponding to the emission light, wherein the reflected light is generated by at least one object reflecting the emitted light, and meanwhile the image capturer 304 is a time of flight sensor. In addition, when the light source 302 is applied to the time of flight, operational principles of the processor 110 determining whether to adjust the intensity of the emitted light can be referred to the above-mentioned corresponding descriptions, so further description thereof is omitted for simplicity.
  • In addition, because the above-mentioned functions of the processor 110 are fully disclosed, one of ordinary skill in the art should easily utilize a field programmable gate array (FPGA) with the above-mentioned functions of the processor 110, or an application-specific integrated circuit (ASIC) with the above-mentioned functions of the processor 110, or a software module with the above-mentioned functions of the processor 110, or an analog integrated circuit with the above-mentioned functions of the processor 110 to realize the processor 110 according to corresponding descriptions of the above-mentioned functions of processor 110. Therefore, a corresponding structure of the processor 110 is omitted for simplicity.
  • In addition, in another embodiment of the present invention, the image capturer 102 and the image capturer 304 can act as a stereo camera (or a depth camera), and the emission light generated by the light source 302 is used for assisting the stereo camera. The processor 110 can utilize a time division multiplexing method to control the light source 302. When the processor 110 control the light source 302 to generate the emission light, the stereo camera acts as the stereo camera, and when the processor 110 turns off the light source 302, images captured by the image capturer 102 and the image capturer 304 are used for generating a color image.
  • In addition, please refer to FIG. 3B. FIG. 3B is a diagram illustrating an image device 350 for generating depth images according to another embodiment of the present invention, wherein as shown in FIG. 3B, a difference between the image device 350 and the image device 300 is that the image device 350 further includes an image capturer 306, wherein the image capturer 102 and the image capturer 304 can act as a stereo camera (or a depth camera), the emission light generated by the light source 302 is used for assisting the stereo camera, and images captured by the image capturer 306 are used for generating a color image (that is, the image capturer 306 only acts as a color sensor and has an infrared light filter). In addition, operational principles of the image capturer 102, the image capturer 304, and the image capturer 306 can be referred to the above-mentioned corresponding descriptions, so further description thereof is omitted for simplicity.
  • In addition, please refer to FIG. 3C. FIG. 3C is a diagram illustrating an image device 370 for generating depth images according to another embodiment of the present invention, wherein as shown in FIG. 3C, a difference between the image device 370 and the image device 300 is that the image device 370 does not include the image capturer 304. As shown in FIG. 3C, when the emission light is structured light, the image capturer 102 is a structured light depth sensor. In addition, the image capturer 102 can be a time of flight sensor, meanwhile the emission light is diffused light, and the operational principles of the image capturer 102 can be referred to the above-mentioned corresponding descriptions, so further description thereof is omitted for simplicity.
  • Please refer to FIG. 4. FIG. 4 is a diagram illustrating an electronic device 400 for viewing a panoramic color and depth image according to a third embodiment of the present invention, wherein as shown in FIG. 4, the electronic device 400 includes a display 402, a processor 404, and an inertial sensor 406 (e.g. an inertial measurement unit, IMU), the processor 404 is coupled to the display 402 and the inertial sensor 406 (because the processor 404 and the inertial sensor 406 are installed within the electronic device 400, related coupling lines corresponding to the processor 404 and the inertial sensor 406 are represented by dotted lines), and the inertial sensor 406 can be a gyro, an accelerator, a microelectromechanical systems (MEMS), and so on. In addition, the electronic device 400 can be a smart phone or a tablet computer. But, the present invention is not limited to the electronic device 400 being a smart phone or a tablet computer. As shown in FIG. 4, after the processor 404 receives a 360 degree panoramic color and depth image generated from the image device 100 (or the image device 300), the processor 404 can first read a 360 degree panoramic color image and a 360 degree panoramic depth image of the 360 degree panoramic color and depth image, or respectively read a 360 degree panoramic color image and a 360 degree panoramic depth image generated from other image devices, wherein each depth value of the 360 degree panoramic depth image corresponds to a virtual optical center which the electronic device 400 acts as.
  • When the electronic device 400 is moved or rotated (e.g. as shown in FIG. 5, the electronic device 400 is moved from a position A to a position B), the inertial sensor 406 can determine a view angle corresponding to motion or rotation of the electronic device 400 according to the motion or the rotation of the electronic device 400 (that is, the view angle corresponds to the position B). After the inertial sensor 406 determines the view angle, the processor 404 can convert a part of the 360 degree panoramic color image corresponding to the view angle into a planar color image and convert a part of the 360 degree panoramic depth image corresponding to the view angle into a planar depth image according to the view angle, and utilize a depth-image-based rendering (DIBR) provided by the prior art to combine the planar color image with the planar depth image to generate a planar color and depth image corresponding to the view angle, wherein when the display 402 displays the planar color and depth image, because the planar color and depth image has depth information of the planar depth image, the planar color and depth image can make a user watching the planar color and depth image have a three-dimensional feeling. In addition, each depth value of the planar depth image corresponds to a virtual optical plane where the electronic device 400 is located. In addition, in another embodiment of the present invention, the user of the electronic device 400 utilizes a mouse or touches the display 402 to change a view angle of the display 402 for displaying the 360 degree panoramic color and depth image. In addition, in another embodiment of the present invention, the 360 degree panoramic color image and the 360 degree panoramic depth image correspond to two different optical centers, the processor 404 can utilize the prior art to compensate a shift between the two different optical centers when the display 402 displays the planar color and depth image.
  • In addition, because the planar color and depth image has the depth information of the planar depth image, when the electronic device 400 is applied to virtual reality (VR), augmented reality (AR), substitutional reality (SR), and mixed reality (MR), the processor 404 can convert the planar color and depth image into a corresponding left eye color image and a corresponding right eye color image according to the depth information of the planar depth image and the depth-image-based rendering, wherein as shown in FIG. 6, a block 4022 of the display 402 displays the corresponding left eye color image and a block 4024 of the display 402 displays the corresponding right eye color image. Therefore, when the electronic device 400 is mounted in a head mounted display, the user can view contents of the virtual reality (VR), the augmented reality (AR), the substitutional reality (SR), and the mixed reality (MR) through the head mounted display and the electronic device 400. In addition, in another embodiment of the present invention, the block 4022 of the display 402 is independent from the block 4024 of the display 402. That is, the block 4022 and the block 4024 are two independent displays.
  • In addition, because the above-mentioned functions of the processor 404 are fully disclosed, one of ordinary skill in the art should easily utilize a field programmable gate array with the above-mentioned functions of the processor 404, or an application-specific integrated circuit with the above-mentioned functions of the processor 404, or a software module with the above-mentioned functions of the processor 404, or an analog integrated circuit with the above-mentioned functions of the processor 404 to realize the processor 404 according to corresponding descriptions of the above-mentioned functions of processor 404. Therefore, a corresponding structure of the processor 404 is omitted for simplicity.
  • To sum up, because the image device utilizes the rotating device to rotate the image capturers, and utilizes the processor to generate a 360 degree panoramic depth image according to multiple images captured by the image capturers, compared to the prior art, the image device does not have a problem that optical centers in the prior art cannot overlap. In addition, because a 360 degree panoramic color and depth image generated by the present invention has depth information of a 360 degree panoramic depth image, compared to the prior art, when the electronic device displays the 360 degree panoramic color and depth image, the electronic device can display a corresponding planar color and depth image according to motion or rotation of the electronic device.
  • Those skilled in the art will readily observe that numerous modifications and alterations of the device and method may be made while retaining the teachings of the invention. Accordingly, the above disclosure should be construed as limited only by the metes and bounds of the appended claims.

Claims (16)

What is claimed is:
1. An image device for generating depth images, comprising:
at least two image capturers; and
a rotating device, wherein when the rotating device rotates the at least two image capturers, multiple images captured by the at least two image capturers are utilized to generate a depth image, wherein a view angle corresponding to the depth image is not less than a view angle of each image capturer of the at least two image capturers.
2. The image device of claim 1, further comprising:
a supporting unit coupled to the rotating device, wherein the at least two image capturers are installed on the supporting unit.
3. The image device of claim 1, wherein a rotation axis of the rotating device passes through an optical center of the each image capturer.
4. The image device of claim 1, wherein a processor stitches a plurality of first images captured by an image capturer of the at least two image capturers to generate a color image according to a feature point matching method or a fixed angle method, and stitches a plurality of second images captured by another image capturer of the at least two image capturers to generate the depth image according to the feature point matching method or the fixed angle method, wherein the another image capturer is different from the image capturer.
5. The image device of claim 4, wherein the processor is installed in the image device or outside the image device.
6. The image device of claim 4, wherein the color image and the depth image are integrated into a color and depth image, and the color and depth image is stored as a file.
7. The image device of claim 6, wherein when the color and depth image is stored as the file, the color image is compressed according to a Joint Photographic Experts Group (JPEG) format, the depth image is compressed according to a lossless format, and a compressed depth image is stored in a header of the file.
8. An image device for generating depth images, comprising:
at least one image capturer;
at least one light source emitting emission light; and
a rotating device;
wherein when the rotating device rotates the at least one image capturer and the at least one light source, multiple images captured by the at least one image capturer and the emission light are used for generating a depth image, wherein a view angle corresponding to the depth image is not less than a view angle of each image capturer of the at least one image capturer.
9. The image device of claim 8, wherein when the rotating device rotates the at least one image capturer and the at least one light source, a plurality of images comprising the emission light captured by an image capturer of the at least one image capturer or flight time corresponding to the emission light passing from the at least one light source to each object and passing from the each object to the image capturer after the emission light is reflected by the each object is used for generating the depth map.
10. The image device of claim 9, wherein when the plurality of images are used for generating the depth image, the emission light is structured light.
11. The image device of claim 8, further comprising:
a supporting unit coupled to the rotating device, wherein the at least one image capturer and the at least one light source are installed on the supporting unit.
12. The image device of claim 8, wherein a rotating axis of the rotating device passes through an optical center of the each image capturer of the at least one image capturer.
13. An electronic device for viewing a panoramic color and depth image, comprising:
a display; and
a processor reading a 360 degree panoramic color image and a 360 degree panoramic depth image, wherein the 360 degree panoramic color image corresponds to the 360 degree panoramic depth image;
wherein when the electronic device is moved or rotated, the processor further converts a part of the 360 degree panoramic color image corresponding to a view angle corresponding to motion or rotation of the electronic device into a planar color image and a part of the 360 degree panoramic depth image corresponding to the view angle into a planar depth image according to the view angle, combines the planar color image with the planar depth image to generate a planar color and depth image corresponding to the view angle, and makes the display display the planar color image according to the planar color and depth image.
14. The electronic device of claim 13, further comprising:
an inertial sensor determining the view angle corresponding to the motion or the rotation of the electronic device according to the motion or the rotation of the electronic device.
15. The electronic device of claim 13, wherein each depth value of the 360 degree panoramic depth image corresponds to a virtual optical center which the electronic device acts as, and each depth value of the planar depth image corresponds to a virtual optical plane where the electronic device is located.
16. The electronic device of claim 13, wherein when the processor combines the planar color image with the planar depth image to generate the planar color and depth image, the processor converts the planar depth image into depth information corresponding to a virtual optical plane to make the planar color and depth image have the depth information.
US16/432,911 2018-06-07 2019-06-05 Image device for generating depth images and related electronic device Abandoned US20190387165A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US16/432,911 US20190387165A1 (en) 2018-06-07 2019-06-05 Image device for generating depth images and related electronic device
US17/142,251 US11438525B2 (en) 2018-06-07 2021-01-06 Image device for generating depth images and related electronic device

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201862681683P 2018-06-07 2018-06-07
US16/432,911 US20190387165A1 (en) 2018-06-07 2019-06-05 Image device for generating depth images and related electronic device

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/142,251 Division US11438525B2 (en) 2018-06-07 2021-01-06 Image device for generating depth images and related electronic device

Publications (1)

Publication Number Publication Date
US20190387165A1 true US20190387165A1 (en) 2019-12-19

Family

ID=68764333

Family Applications (5)

Application Number Title Priority Date Filing Date
US16/432,911 Abandoned US20190387165A1 (en) 2018-06-07 2019-06-05 Image device for generating depth images and related electronic device
US16/432,888 Active US11582402B2 (en) 2018-06-07 2019-06-05 Image processing device
US16/432,935 Active US10708520B2 (en) 2018-06-07 2019-06-06 Calibration method of an image device and related image device and operational device thereof
US17/142,251 Active US11438525B2 (en) 2018-06-07 2021-01-06 Image device for generating depth images and related electronic device
US18/092,947 Pending US20230231965A1 (en) 2018-06-07 2023-01-04 Image processing device

Family Applications After (4)

Application Number Title Priority Date Filing Date
US16/432,888 Active US11582402B2 (en) 2018-06-07 2019-06-05 Image processing device
US16/432,935 Active US10708520B2 (en) 2018-06-07 2019-06-06 Calibration method of an image device and related image device and operational device thereof
US17/142,251 Active US11438525B2 (en) 2018-06-07 2021-01-06 Image device for generating depth images and related electronic device
US18/092,947 Pending US20230231965A1 (en) 2018-06-07 2023-01-04 Image processing device

Country Status (3)

Country Link
US (5) US20190387165A1 (en)
CN (2) CN110580718B (en)
TW (2) TWI723419B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220114298A1 (en) * 2020-10-13 2022-04-14 Flyreel, Inc. Generating measurements of physical structures and environments through automated analysis of sensor data

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190387165A1 (en) * 2018-06-07 2019-12-19 Eys3D Microelectronics, Co. Image device for generating depth images and related electronic device
CN112105312A (en) * 2018-07-03 2020-12-18 柯惠Lp公司 Systems, methods, and computer-readable media for detecting image degradation during a surgical procedure
US11215828B1 (en) * 2018-08-03 2022-01-04 Rockwell Collins, Inc. In field visor characterization for visor projected displays
WO2021216037A1 (en) * 2020-04-20 2021-10-28 Hewlett-Packard Development Company, L.P. Image rendering
CN115278067A (en) * 2022-07-19 2022-11-01 Oppo广东移动通信有限公司 Camera, electronic device, photographing method, and storage medium

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130077882A1 (en) * 2011-09-28 2013-03-28 Pelican Imaging Corporation Systems and methods for decoding light field image files
US20140225985A1 (en) * 2012-10-17 2014-08-14 DotProduct LLC Handheld portable optical scanner and method of using
US20150356358A1 (en) * 2013-02-21 2015-12-10 Sharp Kabushiki Kaisha Obstacle detection device and obstacle detection method
US20160029009A1 (en) * 2014-07-24 2016-01-28 Etron Technology, Inc. Attachable three-dimensional scan module
US20170161561A1 (en) * 2015-10-05 2017-06-08 Pillar Vision, Inc. Systems and methods for monitoring objects at sporting events
US20180108150A1 (en) * 2016-10-17 2018-04-19 Dsi Assignments, Llc Systems and methods for in-field stereocamera calibration
US20180278916A1 (en) * 2015-08-07 2018-09-27 Samsung Electronics Co., Ltd. Electronic device for generating 360-degree three-dimensional image and method therefor
US20190082165A1 (en) * 2014-05-20 2019-03-14 Nextvr Inc. Methods and apparatus including or for use with one or more cameras

Family Cites Families (66)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6028795A (en) * 1985-09-24 2000-02-22 Hitachi, Ltd. One chip semiconductor integrated circuit device having two modes of data write operation and bits setting operation
US6552730B1 (en) * 1984-10-05 2003-04-22 Hitachi, Ltd. Method and apparatus for bit operational process
US4988984A (en) * 1988-10-31 1991-01-29 International Business Machines Corporation Image interpolator for an image display system
US5111192A (en) * 1989-12-20 1992-05-05 Xerox Corporation Method to rotate a bitmap image 90 degrees
US5986672A (en) * 1997-10-14 1999-11-16 Minnesota, Mining And Manufacturing 3M Center Method and system for forming a rotated image on an imaging element using limited system resources
US6563535B1 (en) * 1998-05-19 2003-05-13 Flashpoint Technology, Inc. Image processing system for high performance digital imaging devices
US6137491A (en) 1998-06-05 2000-10-24 Microsoft Corporation Method and apparatus for reconstructing geometry using geometrically constrained structure from motion with points on planes
EP1433335B1 (en) * 2001-08-15 2010-09-29 Koninklijke Philips Electronics N.V. 3d video conferencing system
JP4227468B2 (en) * 2002-06-24 2009-02-18 キヤノン株式会社 Image forming apparatus and method, and control program
US6958757B2 (en) * 2003-07-18 2005-10-25 Microsoft Corporation Systems and methods for efficiently displaying graphics on a display device regardless of physical orientation
JP4130207B2 (en) * 2003-09-16 2008-08-06 富士通株式会社 Image processing display device and image processing display method
US7643038B2 (en) * 2005-06-29 2010-01-05 Qualcomm Incorporated Virtual device buffer for embedded device
CA2553473A1 (en) * 2005-07-26 2007-01-26 Wa James Tam Generating a depth map from a tw0-dimensional source image for stereoscopic and multiview imaging
US7535474B1 (en) * 2005-12-09 2009-05-19 Advanced Micro Devices, Inc. System and method for rotating rasterized image data
US7876977B2 (en) * 2006-02-15 2011-01-25 Conexant Systems, Inc. Image processor and method of image rotation
JP2009100342A (en) 2007-10-18 2009-05-07 Sanyo Electric Co Ltd Camera calibration device, method and vehicle
JP2009135686A (en) * 2007-11-29 2009-06-18 Mitsubishi Electric Corp Stereoscopic video recording method, stereoscopic video recording medium, stereoscopic video reproducing method, stereoscopic video recording apparatus, and stereoscopic video reproducing apparatus
CN101308018B (en) * 2008-05-30 2010-09-15 汤一平 Stereo vision measuring apparatus based on binocular omnidirectional visual sense sensor
JP2010114576A (en) * 2008-11-05 2010-05-20 Seiko Epson Corp Image processing apparatus
CN101894366B (en) 2009-05-21 2014-01-29 北京中星微电子有限公司 Method and device for acquiring calibration parameters and video monitoring system
CN101577795A (en) * 2009-06-17 2009-11-11 深圳华为通信技术有限公司 Method and device for realizing real-time viewing of panoramic picture
US8670149B2 (en) * 2009-08-03 2014-03-11 Printable Technologies Inc. Apparatus and methods for image processing optimization for variable data printing
CN101710932B (en) 2009-12-21 2011-06-22 华为终端有限公司 Image stitching method and device
KR101385514B1 (en) * 2010-06-08 2014-04-16 주식회사 에스칩스 Method And Apparatus for Transforming Stereoscopic Image by Using Depth Map Information
KR101051509B1 (en) 2010-06-28 2011-07-22 삼성전기주식회사 Apparatus and method for controlling light intensity of camera
US9020241B2 (en) * 2011-03-03 2015-04-28 Panasonic Intellectual Property Management Co., Ltd. Image providing device, image providing method, and image providing program for providing past-experience images
US8947551B2 (en) * 2011-07-29 2015-02-03 Aptina Imaging Corporation Method and apparatus for frame rotation in the JPEG compressed domain
JP2013038454A (en) * 2011-08-03 2013-02-21 Sony Corp Image processor, method, and program
JP2013090031A (en) * 2011-10-14 2013-05-13 Sony Corp Image processing device, image processing method, and program
US10848731B2 (en) * 2012-02-24 2020-11-24 Matterport, Inc. Capturing and aligning panoramic image and depth data
CN202838179U (en) * 2012-09-26 2013-03-27 天津游奕科技有限公司 System capable of achieving panoramic video interaction
CN108401551B (en) * 2012-10-26 2015-02-04 北京理工大学 Twin-lens low-light stereoscopic full views imaging device and its ultra-large vision field distance measuring method
US10237528B2 (en) * 2013-03-14 2019-03-19 Qualcomm Incorporated System and method for real time 2D to 3D conversion of a video in a digital camera
TWI498526B (en) 2013-06-05 2015-09-01 Nat Univ Chung Cheng Environment depth measurement method and its image acquisition device
CN104581136B (en) * 2013-10-14 2017-05-31 钰立微电子股份有限公司 The calibration method of image calibration system and stereo camera
CN103729883B (en) 2013-12-30 2016-08-24 浙江大学 A kind of three-dimensional environment information gathering and reconfiguration system and method
KR102265109B1 (en) 2014-01-24 2021-06-15 삼성전자주식회사 Method and apparatus for image processing
CN104811680B (en) * 2014-01-28 2017-04-12 聚晶半导体股份有限公司 image obtaining device and image deformation correction method thereof
CN103984037B (en) 2014-04-30 2017-07-28 深圳市墨克瑞光电子研究院 The mobile robot obstacle detection method and device of view-based access control model
WO2015183145A1 (en) 2014-05-27 2015-12-03 Telefonaktiebolaget L M Ericsson (Publ) Remote control of a pivotable stereoscopic camera
CN104077585B (en) * 2014-05-30 2017-09-22 小米科技有限责任公司 Method for correcting image, device and terminal
US9992483B2 (en) 2014-09-03 2018-06-05 Intel Corporation Imaging architecture for depth camera mode with mode switching
TWI577172B (en) * 2014-09-10 2017-04-01 鈺立微電子股份有限公司 Image calibration system and calibration method of a stereo camera
JP6531367B2 (en) * 2014-09-30 2019-06-19 セイコーエプソン株式会社 PRINTING APPARATUS, CONTROL APPARATUS, AND IMAGE PROCESSING METHOD
US10356383B2 (en) * 2014-12-24 2019-07-16 Reald Spark, Llc Adjustment of perceived roundness in stereoscopic image of a head
CN105744254B (en) * 2014-12-30 2017-12-29 钰立微电子股份有限公司 Correct guidance system and correct the operating method of guidance system
CN104602129B (en) * 2015-01-27 2018-03-06 三星电子(中国)研发中心 The player method and system of interactive multi-angle video
CN106157358A (en) * 2015-03-26 2016-11-23 成都理想境界科技有限公司 Object fusion method based on video image and terminal
WO2016154869A1 (en) 2015-03-31 2016-10-06 SZ DJI Technology Co., Ltd. System and method for mobile platform operation
TWI610250B (en) * 2015-06-02 2018-01-01 鈺立微電子股份有限公司 Monitor system and operation method thereof
CN104966225A (en) * 2015-07-08 2015-10-07 深圳爱布丁梦想科技有限公司 Housing rental method and system based on mobile terminal and 3D panoramic image browsing
US9958266B2 (en) 2015-07-09 2018-05-01 Mitutoyo Corporation Chromatic range sensor including dynamic intensity compensation function
CN106534828A (en) * 2015-09-11 2017-03-22 钰立微电子股份有限公司 Controller applied to a three-dimensional (3d) capture device and 3d image capture device
CN105530431A (en) * 2015-12-16 2016-04-27 景好 Reflective panoramic imaging system and method
CN105609087B (en) * 2015-12-23 2018-11-16 天地伟业技术有限公司 A kind of method of image sealer processing image
JP7043407B2 (en) * 2016-01-15 2022-03-29 華為技術有限公司 Display method and terminal
TWI584634B (en) 2016-03-08 2017-05-21 聚晶半導體股份有限公司 Electronic apparatus and method of generating depth map
CN105866966A (en) * 2016-06-14 2016-08-17 昆山罗伯斯特机器人科技有限公司 Panoramic three-dimensional stereoscopic glasses automatically switching visual angle
US20170372452A1 (en) * 2016-06-22 2017-12-28 Qualcomm Incorporated Image rotation method and apparatus
TWI592020B (en) 2016-08-23 2017-07-11 國立臺灣科技大學 Image correction method of projector and image correction system
CN106921824A (en) * 2017-05-03 2017-07-04 丁志宇 Circulating type mixes light field imaging device and method
CN107105219A (en) * 2017-05-26 2017-08-29 碰海科技(北京)有限公司 Many depth camera compact panoramic scanning equipment
CN107167996A (en) 2017-06-05 2017-09-15 深圳奥比中光科技有限公司 The laser projection module and depth camera adaptively adjusted
CN107465907B (en) * 2017-08-16 2019-08-20 维沃移动通信有限公司 A kind of image pickup method and mobile terminal
EP3531376B1 (en) 2018-02-21 2020-09-30 Ficosa Adas, S.L.U. Calibrating a camera of a vehicle
US20190387165A1 (en) * 2018-06-07 2019-12-19 Eys3D Microelectronics, Co. Image device for generating depth images and related electronic device

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130077882A1 (en) * 2011-09-28 2013-03-28 Pelican Imaging Corporation Systems and methods for decoding light field image files
US20140225985A1 (en) * 2012-10-17 2014-08-14 DotProduct LLC Handheld portable optical scanner and method of using
US20150356358A1 (en) * 2013-02-21 2015-12-10 Sharp Kabushiki Kaisha Obstacle detection device and obstacle detection method
US20190082165A1 (en) * 2014-05-20 2019-03-14 Nextvr Inc. Methods and apparatus including or for use with one or more cameras
US20160029009A1 (en) * 2014-07-24 2016-01-28 Etron Technology, Inc. Attachable three-dimensional scan module
US20180278916A1 (en) * 2015-08-07 2018-09-27 Samsung Electronics Co., Ltd. Electronic device for generating 360-degree three-dimensional image and method therefor
US20170161561A1 (en) * 2015-10-05 2017-06-08 Pillar Vision, Inc. Systems and methods for monitoring objects at sporting events
US20180108150A1 (en) * 2016-10-17 2018-04-19 Dsi Assignments, Llc Systems and methods for in-field stereocamera calibration

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220114298A1 (en) * 2020-10-13 2022-04-14 Flyreel, Inc. Generating measurements of physical structures and environments through automated analysis of sensor data
US11699001B2 (en) * 2020-10-13 2023-07-11 Flyreel, Inc. Generating measurements of physical structures and environments through automated analysis of sensor data
US20230259667A1 (en) * 2020-10-13 2023-08-17 Flyreel, Inc. Generating measurements of physical structures and environments through automated analysis of sensor data
US11960799B2 (en) * 2020-10-13 2024-04-16 Flyreel, Inc. Generating measurements of physical structures and environments through automated analysis of sensor data

Also Published As

Publication number Publication date
US11582402B2 (en) 2023-02-14
US20190378301A1 (en) 2019-12-12
US10708520B2 (en) 2020-07-07
TWI723419B (en) 2021-04-01
TW202013956A (en) 2020-04-01
US20230231965A1 (en) 2023-07-20
TWI761684B (en) 2022-04-21
US20210160438A1 (en) 2021-05-27
US11438525B2 (en) 2022-09-06
US20190379842A1 (en) 2019-12-12
CN110581936A (en) 2019-12-17
CN110580718A (en) 2019-12-17
CN110580718B (en) 2023-08-08
TW202011731A (en) 2020-03-16

Similar Documents

Publication Publication Date Title
US11438525B2 (en) Image device for generating depth images and related electronic device
CN111052727B (en) Electronic device and control method thereof
US11049476B2 (en) Minimal-latency tracking and display for matching real and virtual worlds in head-worn displays
US7872637B2 (en) System and method for tracking a laser spot on a projected computer screen image
KR102078059B1 (en) Prism-based Eye Tracking
US9325968B2 (en) Stereo imaging using disparate imaging devices
TWI557491B (en) Combined visible and non-visible projection system
KR102214193B1 (en) Depth camera device, 3d image display system having the same and control methods thereof
US10079970B2 (en) Controlling image focus in real-time using gestures and depth sensor data
EP2824923B1 (en) Apparatus, system and method for projecting images onto predefined portions of objects
JP4707034B2 (en) Image processing method and input interface device
CN108475180B (en) Distributing video among multiple display areas
US7137707B2 (en) Projector-camera system with laser pointers
US20180146186A1 (en) Active illumination 3d imaging system
CN107439002B (en) Depth imaging
JP2004312733A (en) Device incorporating retina tracking and retina tracking system
JP2003134375A (en) Image pickup system
CN111077671A (en) Device control method and device, display device and storage medium
JP7314501B2 (en) Display control device, display control method and display control program
US20080123956A1 (en) Active environment scanning method and device
KR20190002908A (en) Curved display apparatus and operation method thereof
US10712841B2 (en) Display control device, display control system, display control method, and storage medium having stored thereon display control program
JP7412940B2 (en) Imaging device, computer program and storage medium
US20240046584A1 (en) Information processing apparatus
JP2016177121A (en) Display system, image display device, and operation device

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION