CN108347505B - Mobile terminal with 3D imaging function and image generation method - Google Patents

Mobile terminal with 3D imaging function and image generation method Download PDF

Info

Publication number
CN108347505B
CN108347505B CN201810120901.2A CN201810120901A CN108347505B CN 108347505 B CN108347505 B CN 108347505B CN 201810120901 A CN201810120901 A CN 201810120901A CN 108347505 B CN108347505 B CN 108347505B
Authority
CN
China
Prior art keywords
images
image
processing module
target object
lens
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810120901.2A
Other languages
Chinese (zh)
Other versions
CN108347505A (en
Inventor
李辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Qingyan Heshi Technology Co ltd
Original Assignee
Angrui Shanghai Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Angrui Shanghai Information Technology Co Ltd filed Critical Angrui Shanghai Information Technology Co Ltd
Priority to CN201810120901.2A priority Critical patent/CN108347505B/en
Publication of CN108347505A publication Critical patent/CN108347505A/en
Application granted granted Critical
Publication of CN108347505B publication Critical patent/CN108347505B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/02Constructional features of telephone sets
    • H04M1/0202Portable telephone sets, e.g. cordless phones, mobile phones or bar type handsets
    • H04M1/026Details of the structure or mounting of specific components
    • H04M1/0264Details of the structure or mounting of specific components for a camera module assembly

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Studio Devices (AREA)

Abstract

The invention discloses a mobile terminal with a 3D imaging function and an image generation method, wherein the mobile terminal comprises at least one 3D lens and a processing module, the 3D lens is used for shooting a target object from different shooting positions to obtain a plurality of 3D images, and the 3D images correspond to the shooting positions one by one; for any two 3D images corresponding to the adjacent shooting positions, the processing module is used for identifying feature points on the two 3D images and stitching the two 3D images in a mode of overlapping the same feature points. The mobile terminal with the 3D imaging function and the image generation method not only can acquire the 3D model, but also are convenient and quick. And the image can be clearer.

Description

Mobile terminal with 3D imaging function and image generation method
Technical Field
The invention relates to a mobile terminal with a 3D imaging function and an image generation method.
Background
The 3D camera, which is manufactured by using a 3D lens, generally has two or more image pickup lenses, and has a pitch close to the pitch of human eyes, and can capture different images of the same scene seen by similar human eyes. The holographic 3D has a disc 5 above the lens, and can view the same image in all directions through dot grating imaging or -shaped grating holographic imaging, such as being in the environment.
The first 3D camera to date the 3D revolution has all been around the hollywood heavy-pound large and major sporting events. With the advent of 3D cameras, this technology is one step closer to home users. After the camera is introduced, each memorable moment of the life, such as the first step taken by a child, a university graduation celebration and the like, can be captured by using a 3D lens in the future.
A 3D camera typically has more than two lenses. The 3D camera functions like a human brain, and can fuse two lens images together to form a 3D image. These images can be played on a 3D television, and can be viewed by viewers wearing so-called actively shuttered glasses, or directly viewed by naked-eye 3D display devices. The 3D shutter glasses can rapidly alternately open and close the lenses of the left and right glasses at a rate of 60 times per second. This means that each eye sees a slightly different picture of the same scene, so the brain can thus think that it is enjoying a single picture in 3D.
The existing 3D camera is expensive, distortion phenomenon often occurs, and the image is not clear enough.
Disclosure of Invention
The invention aims to overcome the defects that a 3D camera in the prior art is expensive, distortion phenomenon often occurs and images are not clear enough, and provides a method which can obtain a 3D model and is convenient and quick. Moreover, the image can be clearer, and the mobile terminal with the 3D imaging function and the image generation method which are simple and cheap in equipment are realized.
The invention solves the technical problems through the following technical scheme:
a mobile terminal with 3D imaging function is characterized in that the mobile terminal comprises at least one 3D lens and a processing module,
the 3D lens is used for shooting a target object from different shooting positions to obtain a plurality of 3D images, and the 3D images correspond to the shooting positions one by one;
for any two 3D images corresponding to the adjacent shooting positions, the processing module is used for identifying feature points on the two 3D images and stitching the two 3D images in a mode of overlapping the same feature points.
Preferably, for a cross section on the target object, the 3D lens is configured to sequentially capture the target object according to a preset rule to obtain the 3D images at the different capture positions corresponding to the cross section when the mobile terminal moves in a first preset manner, where the first preset manner is that the capture direction of the 3D lens is located on the cross section and rotates around the target object as a circle center;
the processing module is used for stitching the 3D image corresponding to one cross section to generate a 3D sub-model;
the processing module is also used for identifying characteristic points on the 3D submodels corresponding to the two adjacent cross sections and stitching the two 3D submodels in a mode of overlapping the same characteristic points.
Preferably, for a longitudinal section on the target object, the 3D lens is configured to sequentially capture the target object according to a preset rule to obtain 3D images at different capture positions corresponding to the longitudinal section in a second preset manner that the 3D lens moves from one end of the target object in the longitudinal direction to the other end;
the processing module is used for stitching the 3D images corresponding to the longitudinal sections to generate a 3D longitudinal sub-model;
the processing module is further used for selecting a target feature point on the 3D longitudinal sub-model, then selecting a 3D image containing the target feature point in the 3D image corresponding to the cross section, and stitching the 3D longitudinal sub-model and the 3D image containing the target feature point in a mode of overlapping the target feature points.
Preferably, the mobile terminal comprises a gyroscope and an acceleration sensor;
the preset rule is that the target object is shot frame by frame, and each frame corresponds to a 3D image; and/or the presence of a gas in the gas,
the preset rule is that the rotation angle of the mobile terminal rotating in a first preset mode is obtained through the gyroscope and the acceleration sensor, and the 3D lens is used for activating the shutter when the rotation angle reaches a preset angle.
Preferably, each 3D image comprises a pixel layer and a structural layer, and for any two 3D images corresponding to two adjacent shooting positions,
the processing module is used for identifying at least 3 peak points on the structural layers of the two 3D images;
the processing module is used for stitching the structural layers of the two 3D images in a mode of overlapping the same peak points, the peak points comprise convex points and concave points, and the number of overlapping the peak points of the two 3D images is at least 3;
the processing module is further used for attaching two pixel layers of the 3D image to the stitched 3D image.
The invention also provides an image generation method, which is characterized in that the image generation method is used for a mobile terminal, the mobile terminal comprises at least one 3D lens and a processing module, and the image generation method comprises the following steps:
the 3D lens shoots a target object from different shooting positions to obtain a plurality of 3D images, and the 3D images correspond to the shooting positions one by one;
for any two 3D images corresponding to the adjacent shooting positions, the processing module identifies feature points on the two 3D images, and stitches the two 3D images in a mode of overlapping the same feature points.
Preferably, the image generating method includes:
for a cross section on the target object, sequentially shooting the target object by the 3D lens according to a preset rule in the movement of the mobile terminal in a first preset mode so as to obtain 3D images of different shooting positions corresponding to the cross section, wherein the first preset mode is that the shooting direction of the 3D lens is located on the cross section and rotates by taking the target object as a circle center;
the processing module stitches the 3D image corresponding to one cross section to generate a 3D sub-model;
the processing module identifies characteristic points on the 3D submodels corresponding to the two adjacent cross sections, and stitches the two 3D submodels in a mode that the same characteristic points are overlapped.
Preferably, the image generating method includes:
for a longitudinal section on the target object, sequentially shooting the target object by a preset rule to obtain 3D images of different shooting positions corresponding to the longitudinal section in a second preset mode that the 3D lens moves from one longitudinal end of the target object to the other longitudinal end of the target object;
the processing module stitches the 3D images corresponding to the longitudinal sections to generate a 3D longitudinal sub-model;
the processing module selects a target characteristic point on the 3D longitudinal sub-model, then selects a 3D image containing the target characteristic point in the 3D image corresponding to the cross section, and stitches the 3D longitudinal sub-model and the 3D image containing the target characteristic point in a target characteristic point overlapping mode.
Preferably, the mobile terminal comprises a gyroscope and an acceleration sensor;
the preset rule is that the target object is shot frame by frame, and each frame corresponds to a 3D image; and/or the presence of a gas in the gas,
the preset rule is that the rotation angle of the mobile terminal rotating in a first preset mode is obtained through the gyroscope and the acceleration sensor, and the 3D lens is used for activating the shutter when the rotation angle reaches a preset angle.
Preferably, each 3D image includes a pixel layer and a structural layer, and for any two 3D images corresponding to adjacent shooting positions, the image generating method includes:
the processing module identifies at least 3 peak points on the structural layers of the two 3D images;
the processing module stitches the structural layers of the two 3D images in a mode of overlapping the same peak points, wherein the peak points comprise convex points and concave points, and the number of overlapping the peak points of the two 3D images is at least 3;
and the processing module is used for attaching two pixel layers of the 3D image to the stitched 3D image.
On the basis of the common knowledge in the field, the above preferred conditions can be combined randomly to obtain the preferred embodiments of the invention.
The positive progress effects of the invention are as follows: the mobile terminal with the 3D imaging function and the image generation method not only can acquire the 3D model, but also are convenient and quick. Moreover, the image can be clearer, and the realization equipment is simple and cheap.
Drawings
Fig. 1 is a flowchart of an image generating method according to embodiment 1 of the present invention.
Detailed Description
The invention is further illustrated by the following examples, which are not intended to limit the scope of the invention.
Example 1
The embodiment provides a mobile terminal, the mobile terminal is a mobile phone, the mobile phone comprises a 3D lens, a processing module, a gyroscope and an acceleration sensor, and the 3D lens is a front camera of the mobile phone. The 3D lens comprises an infrared beam transmitter, an infrared receiver and a camera, wherein the infrared beam transmitter and the infrared receiver are used for generating a structural layer of a 3D image, and the camera is used for generating a 3D image pixel layer.
The 3D lens is used for shooting a target object from different shooting positions to obtain a plurality of 3D images, and the 3D images correspond to the shooting positions one by one.
For any two 3D images corresponding to the adjacent shooting positions, the processing module is used for identifying feature points on the two 3D images and stitching the two 3D images in a mode of overlapping the same feature points.
In the embodiment, the 3D models of the user and the object can be acquired through the mobile phone with the 3D lens, and a complete 3D model can be acquired by splicing the 3D images acquired at different shooting positions, and the preset rule can activate the shutter once for each short time or activate the shutter once after the mobile phone is detected to move for a certain distance through the sensor.
Specifically, the mobile phone of the embodiment can implement:
for a cross section on the target object, the 3D lens is used for sequentially shooting the target object according to a preset rule to obtain 3D images of the different shooting positions corresponding to the cross section during movement of the mobile terminal in a first preset manner, where the first preset manner is that the shooting direction of the 3D lens is located on the cross section and rotates around the target object as a circle center;
the processing module is used for stitching the 3D image corresponding to one cross section to generate a 3D sub-model;
the processing module is also used for identifying characteristic points on the 3D submodels corresponding to the two adjacent cross sections and stitching the two 3D submodels in a mode of overlapping the same characteristic points.
For a target object with a higher longitudinal direction, such as a user, a mobile phone is utilized to establish a 3D sub-model layer by layer, such as a sub-model of an upper half body and a sub-model of a lower half body, and then the sub-models are spliced.
In this embodiment, the preset rule is to capture the target object frame by frame while rotating in a first preset manner, where each frame corresponds to a 3D image, and the specific manner may be to record a video, and then capture the image of each frame as the 3D image.
Further, in order to facilitate splicing and stitching of each sub-model, the mobile phone of the embodiment can implement:
for a longitudinal section on the target object, the 3D lens is used for sequentially shooting the target object according to a preset rule to acquire 3D images of different shooting positions corresponding to the longitudinal section in a second preset mode in which the mobile terminal moves, and the second preset mode is that the 3D lens moves from one longitudinal end of the target object to the other longitudinal end.
The processing module is used for stitching the 3D images corresponding to the longitudinal sections to generate a 3D longitudinal sub-model;
the processing module is further used for selecting a target feature point on the 3D longitudinal sub-model, then selecting a 3D image containing the target feature point in the 3D image corresponding to the cross section, and stitching the 3D longitudinal sub-model and the 3D image containing the target feature point in a mode of overlapping the target feature points.
Referring to fig. 1, with the mobile phone, the embodiment further provides an image generating method, including:
and step 100, for a longitudinal section on the target object, sequentially shooting the target object by the 3D lens according to a preset rule in a second preset mode of movement of the mobile terminal so as to obtain the 3D images of the different shooting positions corresponding to the longitudinal section.
In step 100, the second preset manner is that the 3D lens moves from one longitudinal end of the target object to the other longitudinal end, and the 3D images correspond to the shooting positions one by one. Taking the user himself as an example, the second preset mode is to move the shooting direction of the 3D lens from head to foot.
Step 101, the processing module stitches the 3D images corresponding to the longitudinal sections to generate a 3D longitudinal sub-model.
In step 101, for any two 3D images corresponding to two adjacent shooting positions, the processing module identifies feature points on the two 3D images, and stitches the two 3D images in a manner that the same feature points coincide, where the 3D images corresponding to the two adjacent shooting positions are all shooting positions in a second preset manner.
And step 102, for a cross section on the target object, sequentially shooting the target object by the 3D lens according to a preset rule in the movement of the mobile terminal in a first preset mode so as to obtain the 3D images of the different shooting positions corresponding to the cross section.
In step 102, the first preset mode is that the shooting direction of the 3D lens is located on the cross section and rotates around the target object, and the 3D images correspond to the shooting positions one by one.
103, the processing module selects a target feature point on the 3D longitudinal sub-model, then selects a 3D image containing the target feature point in the 3D image corresponding to the cross section, and stitches the 3D longitudinal sub-model and the 3D image containing the target feature point in a mode of overlapping the target feature points.
And step 104, stitching the 3D image corresponding to one cross section by using the 3D image stitched with the 3D longitudinal sub-model as a starting point to generate a 3D sub-model by the processing module.
In step 104, for any two 3D images corresponding to two adjacent shooting positions, the processing module identifies feature points on the two 3D images, and stitches the two 3D images in a manner that the same feature points coincide, where the 3D images corresponding to the two adjacent shooting positions are all shooting positions in a first preset manner.
After step 103, the 3D image corresponding to one cross section is already positioned by the longitudinal sub-model, and other 3D images may be sequentially stitched on the 3D image already positioned, and then the 3D longitudinal sub-model and one 3D sub-model (transverse) stitching are realized, and the stitching of the longitudinal sub-model and other transverse sub-models may be realized in the same manner.
In this embodiment, a longitudinal submodel is used as a reference, and then a transverse submodel is added, so that inferior particles are not likely to occur in model fidelity, and in an embodiment where a longitudinal submodel is not used, the transverse submodels can be sequentially spliced, and the specific method is as follows: the processing module identifies characteristic points on the 3D submodels corresponding to the two adjacent cross sections, and stitches the two 3D submodels in a mode that the same characteristic points are overlapped.
Example 2
This embodiment is substantially the same as embodiment 1 except that:
the preset rule is that the rotation angle of the mobile terminal rotating in a first preset mode is obtained through the gyroscope and the acceleration sensor, and the 3D lens is used for activating the shutter when the rotation angle reaches a preset angle.
In the mobile phone of this embodiment, each 3D image includes a pixel layer and a structural layer, and for any two 3D images corresponding to adjacent shooting positions,
the processing module is used for identifying at least 3 peak points on the structural layers of the two 3D images;
the processing module is used for stitching the structural layers of the two 3D images in a mode of overlapping the same peak points, the peak points comprise convex points and concave points, and the number of overlapping the peak points of the two 3D images is at least 3;
the processing module is further used for attaching two pixel layers of the 3D image to the stitched 3D image.
Accordingly, the image generating method of the present embodiment provides a method for performing 3D image stitching through a structural layer, including:
each 3D image comprises a pixel layer and a structural layer, and for any two 3D images corresponding to two adjacent shooting positions, the image generation method comprises the following steps:
the processing module identifies at least 3 peak points on the structural layers of the two 3D images;
the processing module stitches the structural layers of the two 3D images in a mode of overlapping the same peak points, wherein the peak points comprise convex points and concave points, and the number of overlapping the peak points of the two 3D images is at least 3;
and the processing module is used for attaching two pixel layers of the 3D image to the stitched 3D image.
While specific embodiments of the invention have been described above, it will be appreciated by those skilled in the art that these are by way of example only, and that the scope of the invention is defined by the appended claims. Various changes and modifications to these embodiments may be made by those skilled in the art without departing from the spirit and scope of the invention, and these changes and modifications are within the scope of the invention.

Claims (6)

1. A mobile terminal with 3D imaging function is characterized in that the mobile terminal comprises at least one 3D lens and a processing module, the 3D lens comprises an infrared beam transmitter, an infrared receiver and a camera,
the 3D lens is used for shooting a target object from different shooting positions to obtain a plurality of 3D images, and the 3D images correspond to the shooting positions one by one;
for any two 3D images corresponding to the adjacent shooting positions, the processing module is used for identifying feature points on the two 3D images and stitching the two 3D images in a mode of overlapping the same feature points;
for a cross section on the target object, the 3D lens is used for sequentially shooting the target object according to a preset rule to obtain 3D images of the different shooting positions corresponding to the cross section during movement of the mobile terminal in a first preset manner, where the first preset manner is that the shooting direction of the 3D lens is located on the cross section and rotates around the target object as a circle center;
the processing module is used for stitching the 3D image corresponding to one cross section to generate a 3D sub-model;
the processing module is also used for identifying characteristic points on the 3D submodels corresponding to the two adjacent cross sections and stitching the two 3D submodels in a mode of overlapping the same characteristic points;
for a longitudinal section on the target object, the 3D lens is used for sequentially shooting the target object according to a preset rule to obtain 3D images at different shooting positions corresponding to the longitudinal section in a second preset manner that the 3D lens moves from one longitudinal end of the target object to the other longitudinal end;
the processing module is used for stitching the 3D images corresponding to the longitudinal sections to generate a 3D longitudinal sub-model;
the processing module is further used for selecting a target feature point on the 3D longitudinal sub-model, then selecting a 3D image containing the target feature point in the 3D image corresponding to the cross section, and stitching the 3D longitudinal sub-model and the 3D image containing the target feature point in a mode of overlapping the target feature points.
2. The mobile terminal of claim 1, wherein the mobile terminal comprises a gyroscope and an acceleration sensor;
the preset rule is that the target object is shot frame by frame, and each frame corresponds to a 3D image; and/or the presence of a gas in the gas,
the preset rule is that the rotation angle of the mobile terminal rotating in a first preset mode is obtained through the gyroscope and the acceleration sensor, and the 3D lens is used for activating the shutter when the rotation angle reaches a preset angle.
3. The mobile terminal of claim 1, wherein each 3D image comprises a pixel layer and a structural layer, and for any two adjacent capturing positions corresponding to the 3D images,
the processing module is used for identifying at least 3 peak points on the structural layers of the two 3D images;
the processing module is used for stitching the structural layers of the two 3D images in a mode of overlapping the same peak points, the peak points comprise convex points and concave points, and the number of overlapping the peak points of the two 3D images is at least 3;
the processing module is further used for attaching two pixel layers of the 3D image to the stitched 3D image.
4. An image generation method, wherein the image generation method is applied to a mobile terminal, the mobile terminal includes at least one 3D lens and a processing module, the 3D lens includes an infrared beam transmitter, an infrared receiver and a camera, and the image generation method includes:
the 3D lens shoots a target object from different shooting positions to obtain a plurality of 3D images, and the 3D images correspond to the shooting positions one by one;
for any two 3D images corresponding to the adjacent shooting positions, the processing module identifies feature points on the two 3D images, and the two 3D images are stitched in a mode that the same feature points are overlapped;
the image generation method comprises the following steps:
for a cross section on the target object, sequentially shooting the target object by the 3D lens according to a preset rule in the movement of the mobile terminal in a first preset mode so as to obtain 3D images of different shooting positions corresponding to the cross section, wherein the first preset mode is that the shooting direction of the 3D lens is located on the cross section and rotates by taking the target object as a circle center;
the processing module stitches the 3D image corresponding to one cross section to generate a 3D sub-model;
the processing module identifies characteristic points on the 3D submodels corresponding to the two adjacent cross sections, and stitches the two 3D submodels in a mode of overlapping the same characteristic points;
the image generation method comprises the following steps:
for a longitudinal section on the target object, sequentially shooting the target object by a preset rule to obtain 3D images of different shooting positions corresponding to the longitudinal section in a second preset mode that the 3D lens moves from one longitudinal end of the target object to the other longitudinal end of the target object;
the processing module stitches the 3D images corresponding to the longitudinal sections to generate a 3D longitudinal sub-model;
the processing module selects a target characteristic point on the 3D longitudinal sub-model, then selects a 3D image containing the target characteristic point in the 3D image corresponding to the cross section, and stitches the 3D longitudinal sub-model and the 3D image containing the target characteristic point in a target characteristic point overlapping mode.
5. The image generating method as claimed in claim 4, wherein the mobile terminal comprises a gyroscope and an acceleration sensor;
the preset rule is that the target object is shot frame by frame, and each frame corresponds to a 3D image; and/or the presence of a gas in the gas,
the preset rule is that the rotation angle of the mobile terminal rotating in a first preset mode is obtained through the gyroscope and the acceleration sensor, and the 3D lens is used for activating the shutter when the rotation angle reaches a preset angle.
6. The image generation method of claim 4, wherein each 3D image comprises a pixel layer and a structural layer, and for any two adjacent capturing positions corresponding to the 3D images, the image generation method comprises:
the processing module identifies at least 3 peak points on the structural layers of the two 3D images;
the processing module stitches the structural layers of the two 3D images in a mode of overlapping the same peak points, wherein the peak points comprise convex points and concave points, and the number of overlapping the peak points of the two 3D images is at least 3;
and the processing module is used for attaching two pixel layers of the 3D image to the stitched 3D image.
CN201810120901.2A 2018-02-07 2018-02-07 Mobile terminal with 3D imaging function and image generation method Active CN108347505B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810120901.2A CN108347505B (en) 2018-02-07 2018-02-07 Mobile terminal with 3D imaging function and image generation method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810120901.2A CN108347505B (en) 2018-02-07 2018-02-07 Mobile terminal with 3D imaging function and image generation method

Publications (2)

Publication Number Publication Date
CN108347505A CN108347505A (en) 2018-07-31
CN108347505B true CN108347505B (en) 2020-11-06

Family

ID=62959073

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810120901.2A Active CN108347505B (en) 2018-02-07 2018-02-07 Mobile terminal with 3D imaging function and image generation method

Country Status (1)

Country Link
CN (1) CN108347505B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109151440B (en) * 2018-10-15 2020-06-09 盎锐(上海)信息科技有限公司 Image positioning device and method
CN109472860B (en) * 2018-11-13 2023-02-10 上海盎维信息技术有限公司 Depth map balance optimization method and device based on artificial intelligence
CN109151444B (en) * 2018-11-13 2020-09-08 盎锐(上海)信息科技有限公司 3D Intelligent Pixel enhancement Engine
CN109509262B (en) * 2018-11-13 2023-02-28 上海盎维信息技术有限公司 Intelligent enhanced modeling method and device based on artificial intelligence
CN110176034B (en) * 2019-05-27 2023-02-07 上海盎维信息技术有限公司 Positioning method and scanning terminal for VSLAM

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102208116A (en) * 2010-03-29 2011-10-05 卡西欧计算机株式会社 3D modeling apparatus and 3D modeling method
CN102917236A (en) * 2012-09-27 2013-02-06 深圳天珑无线科技有限公司 Single-camera based stereoscopic photographing method and digital photographing device
CN105241397A (en) * 2015-06-29 2016-01-13 北航温州研究院 Real-time measuring splicing method and device based on structured light
CN105306921A (en) * 2014-06-18 2016-02-03 中兴通讯股份有限公司 Three-dimensional photo shooting method based on mobile terminal and mobile terminal

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10077007B2 (en) * 2016-03-14 2018-09-18 Uber Technologies, Inc. Sidepod stereo camera system for an autonomous vehicle

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102208116A (en) * 2010-03-29 2011-10-05 卡西欧计算机株式会社 3D modeling apparatus and 3D modeling method
CN102917236A (en) * 2012-09-27 2013-02-06 深圳天珑无线科技有限公司 Single-camera based stereoscopic photographing method and digital photographing device
CN105306921A (en) * 2014-06-18 2016-02-03 中兴通讯股份有限公司 Three-dimensional photo shooting method based on mobile terminal and mobile terminal
CN105241397A (en) * 2015-06-29 2016-01-13 北航温州研究院 Real-time measuring splicing method and device based on structured light

Also Published As

Publication number Publication date
CN108347505A (en) 2018-07-31

Similar Documents

Publication Publication Date Title
CN108347505B (en) Mobile terminal with 3D imaging function and image generation method
US9927948B2 (en) Image display apparatus and image display method
JP5140210B2 (en) Imaging apparatus and image processing method
CN106507094B (en) Correct the method and device of panoramic video display view angle
CN108600729B (en) Dynamic 3D model generation device and image generation method
JP5204350B2 (en) Imaging apparatus, playback apparatus, and image processing method
US20150358539A1 (en) Mobile Virtual Reality Camera, Method, And System
JP5204349B2 (en) Imaging apparatus, playback apparatus, and image processing method
CN103370943B (en) Imaging device and formation method
US20160344999A1 (en) SYSTEMS AND METHODs FOR PRODUCING PANORAMIC AND STEREOSCOPIC VIDEOS
CA2933704A1 (en) Systems and methods for producing panoramic and stereoscopic videos
TW201351959A (en) Method of stereo 3D image synthesis and related camera
CN108391116B (en) Whole body scanning device and method based on 3D imaging technology
CN108111835B (en) Shooting device, system and method for 3D image imaging
KR20150003576A (en) Apparatus and method for generating or reproducing three-dimensional image
CN108614636A (en) A kind of 3D outdoor scenes VR production methods
CN107637067A (en) Image processing equipment and image processing method
CN108737808B (en) 3D model generation device and method
CN108513122B (en) Model adjusting method and model generating device based on 3D imaging technology
WO2012150680A1 (en) Three-dimensional image processing apparatus and three-dimensional image processing method
CN108234904B (en) Multi-video fusion method, device and system
JP5704885B2 (en) Imaging device, imaging method, and imaging control program
CN108195308B (en) 3D scanning device, system and method
CN109272453B (en) Modeling device and positioning method based on 3D camera
CN108848366B (en) Information acquisition device and method based on 3D camera

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20230727

Address after: 201703 Room 2134, Floor 2, No. 152 and 153, Lane 3938, Huqingping Road, Qingpu District, Shanghai

Patentee after: Shanghai Qingyan Heshi Technology Co.,Ltd.

Address before: 201703 No.206, building 1, no.3938 Huqingping Road, Qingpu District, Shanghai

Patentee before: UNRE (SHANGHAI) INFORMATION TECHNOLOGY Co.,Ltd.