CN108347505B - Mobile terminal with 3D imaging function and image generation method - Google Patents
Mobile terminal with 3D imaging function and image generation method Download PDFInfo
- Publication number
- CN108347505B CN108347505B CN201810120901.2A CN201810120901A CN108347505B CN 108347505 B CN108347505 B CN 108347505B CN 201810120901 A CN201810120901 A CN 201810120901A CN 108347505 B CN108347505 B CN 108347505B
- Authority
- CN
- China
- Prior art keywords
- images
- image
- processing module
- target object
- lens
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/02—Constructional features of telephone sets
- H04M1/0202—Portable telephone sets, e.g. cordless phones, mobile phones or bar type handsets
- H04M1/026—Details of the structure or mounting of specific components
- H04M1/0264—Details of the structure or mounting of specific components for a camera module assembly
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Studio Devices (AREA)
Abstract
The invention discloses a mobile terminal with a 3D imaging function and an image generation method, wherein the mobile terminal comprises at least one 3D lens and a processing module, the 3D lens is used for shooting a target object from different shooting positions to obtain a plurality of 3D images, and the 3D images correspond to the shooting positions one by one; for any two 3D images corresponding to the adjacent shooting positions, the processing module is used for identifying feature points on the two 3D images and stitching the two 3D images in a mode of overlapping the same feature points. The mobile terminal with the 3D imaging function and the image generation method not only can acquire the 3D model, but also are convenient and quick. And the image can be clearer.
Description
Technical Field
The invention relates to a mobile terminal with a 3D imaging function and an image generation method.
Background
The 3D camera, which is manufactured by using a 3D lens, generally has two or more image pickup lenses, and has a pitch close to the pitch of human eyes, and can capture different images of the same scene seen by similar human eyes. The holographic 3D has a disc 5 above the lens, and can view the same image in all directions through dot grating imaging or -shaped grating holographic imaging, such as being in the environment.
The first 3D camera to date the 3D revolution has all been around the hollywood heavy-pound large and major sporting events. With the advent of 3D cameras, this technology is one step closer to home users. After the camera is introduced, each memorable moment of the life, such as the first step taken by a child, a university graduation celebration and the like, can be captured by using a 3D lens in the future.
A 3D camera typically has more than two lenses. The 3D camera functions like a human brain, and can fuse two lens images together to form a 3D image. These images can be played on a 3D television, and can be viewed by viewers wearing so-called actively shuttered glasses, or directly viewed by naked-eye 3D display devices. The 3D shutter glasses can rapidly alternately open and close the lenses of the left and right glasses at a rate of 60 times per second. This means that each eye sees a slightly different picture of the same scene, so the brain can thus think that it is enjoying a single picture in 3D.
The existing 3D camera is expensive, distortion phenomenon often occurs, and the image is not clear enough.
Disclosure of Invention
The invention aims to overcome the defects that a 3D camera in the prior art is expensive, distortion phenomenon often occurs and images are not clear enough, and provides a method which can obtain a 3D model and is convenient and quick. Moreover, the image can be clearer, and the mobile terminal with the 3D imaging function and the image generation method which are simple and cheap in equipment are realized.
The invention solves the technical problems through the following technical scheme:
a mobile terminal with 3D imaging function is characterized in that the mobile terminal comprises at least one 3D lens and a processing module,
the 3D lens is used for shooting a target object from different shooting positions to obtain a plurality of 3D images, and the 3D images correspond to the shooting positions one by one;
for any two 3D images corresponding to the adjacent shooting positions, the processing module is used for identifying feature points on the two 3D images and stitching the two 3D images in a mode of overlapping the same feature points.
Preferably, for a cross section on the target object, the 3D lens is configured to sequentially capture the target object according to a preset rule to obtain the 3D images at the different capture positions corresponding to the cross section when the mobile terminal moves in a first preset manner, where the first preset manner is that the capture direction of the 3D lens is located on the cross section and rotates around the target object as a circle center;
the processing module is used for stitching the 3D image corresponding to one cross section to generate a 3D sub-model;
the processing module is also used for identifying characteristic points on the 3D submodels corresponding to the two adjacent cross sections and stitching the two 3D submodels in a mode of overlapping the same characteristic points.
Preferably, for a longitudinal section on the target object, the 3D lens is configured to sequentially capture the target object according to a preset rule to obtain 3D images at different capture positions corresponding to the longitudinal section in a second preset manner that the 3D lens moves from one end of the target object in the longitudinal direction to the other end;
the processing module is used for stitching the 3D images corresponding to the longitudinal sections to generate a 3D longitudinal sub-model;
the processing module is further used for selecting a target feature point on the 3D longitudinal sub-model, then selecting a 3D image containing the target feature point in the 3D image corresponding to the cross section, and stitching the 3D longitudinal sub-model and the 3D image containing the target feature point in a mode of overlapping the target feature points.
Preferably, the mobile terminal comprises a gyroscope and an acceleration sensor;
the preset rule is that the target object is shot frame by frame, and each frame corresponds to a 3D image; and/or the presence of a gas in the gas,
the preset rule is that the rotation angle of the mobile terminal rotating in a first preset mode is obtained through the gyroscope and the acceleration sensor, and the 3D lens is used for activating the shutter when the rotation angle reaches a preset angle.
Preferably, each 3D image comprises a pixel layer and a structural layer, and for any two 3D images corresponding to two adjacent shooting positions,
the processing module is used for identifying at least 3 peak points on the structural layers of the two 3D images;
the processing module is used for stitching the structural layers of the two 3D images in a mode of overlapping the same peak points, the peak points comprise convex points and concave points, and the number of overlapping the peak points of the two 3D images is at least 3;
the processing module is further used for attaching two pixel layers of the 3D image to the stitched 3D image.
The invention also provides an image generation method, which is characterized in that the image generation method is used for a mobile terminal, the mobile terminal comprises at least one 3D lens and a processing module, and the image generation method comprises the following steps:
the 3D lens shoots a target object from different shooting positions to obtain a plurality of 3D images, and the 3D images correspond to the shooting positions one by one;
for any two 3D images corresponding to the adjacent shooting positions, the processing module identifies feature points on the two 3D images, and stitches the two 3D images in a mode of overlapping the same feature points.
Preferably, the image generating method includes:
for a cross section on the target object, sequentially shooting the target object by the 3D lens according to a preset rule in the movement of the mobile terminal in a first preset mode so as to obtain 3D images of different shooting positions corresponding to the cross section, wherein the first preset mode is that the shooting direction of the 3D lens is located on the cross section and rotates by taking the target object as a circle center;
the processing module stitches the 3D image corresponding to one cross section to generate a 3D sub-model;
the processing module identifies characteristic points on the 3D submodels corresponding to the two adjacent cross sections, and stitches the two 3D submodels in a mode that the same characteristic points are overlapped.
Preferably, the image generating method includes:
for a longitudinal section on the target object, sequentially shooting the target object by a preset rule to obtain 3D images of different shooting positions corresponding to the longitudinal section in a second preset mode that the 3D lens moves from one longitudinal end of the target object to the other longitudinal end of the target object;
the processing module stitches the 3D images corresponding to the longitudinal sections to generate a 3D longitudinal sub-model;
the processing module selects a target characteristic point on the 3D longitudinal sub-model, then selects a 3D image containing the target characteristic point in the 3D image corresponding to the cross section, and stitches the 3D longitudinal sub-model and the 3D image containing the target characteristic point in a target characteristic point overlapping mode.
Preferably, the mobile terminal comprises a gyroscope and an acceleration sensor;
the preset rule is that the target object is shot frame by frame, and each frame corresponds to a 3D image; and/or the presence of a gas in the gas,
the preset rule is that the rotation angle of the mobile terminal rotating in a first preset mode is obtained through the gyroscope and the acceleration sensor, and the 3D lens is used for activating the shutter when the rotation angle reaches a preset angle.
Preferably, each 3D image includes a pixel layer and a structural layer, and for any two 3D images corresponding to adjacent shooting positions, the image generating method includes:
the processing module identifies at least 3 peak points on the structural layers of the two 3D images;
the processing module stitches the structural layers of the two 3D images in a mode of overlapping the same peak points, wherein the peak points comprise convex points and concave points, and the number of overlapping the peak points of the two 3D images is at least 3;
and the processing module is used for attaching two pixel layers of the 3D image to the stitched 3D image.
On the basis of the common knowledge in the field, the above preferred conditions can be combined randomly to obtain the preferred embodiments of the invention.
The positive progress effects of the invention are as follows: the mobile terminal with the 3D imaging function and the image generation method not only can acquire the 3D model, but also are convenient and quick. Moreover, the image can be clearer, and the realization equipment is simple and cheap.
Drawings
Fig. 1 is a flowchart of an image generating method according to embodiment 1 of the present invention.
Detailed Description
The invention is further illustrated by the following examples, which are not intended to limit the scope of the invention.
Example 1
The embodiment provides a mobile terminal, the mobile terminal is a mobile phone, the mobile phone comprises a 3D lens, a processing module, a gyroscope and an acceleration sensor, and the 3D lens is a front camera of the mobile phone. The 3D lens comprises an infrared beam transmitter, an infrared receiver and a camera, wherein the infrared beam transmitter and the infrared receiver are used for generating a structural layer of a 3D image, and the camera is used for generating a 3D image pixel layer.
The 3D lens is used for shooting a target object from different shooting positions to obtain a plurality of 3D images, and the 3D images correspond to the shooting positions one by one.
For any two 3D images corresponding to the adjacent shooting positions, the processing module is used for identifying feature points on the two 3D images and stitching the two 3D images in a mode of overlapping the same feature points.
In the embodiment, the 3D models of the user and the object can be acquired through the mobile phone with the 3D lens, and a complete 3D model can be acquired by splicing the 3D images acquired at different shooting positions, and the preset rule can activate the shutter once for each short time or activate the shutter once after the mobile phone is detected to move for a certain distance through the sensor.
Specifically, the mobile phone of the embodiment can implement:
for a cross section on the target object, the 3D lens is used for sequentially shooting the target object according to a preset rule to obtain 3D images of the different shooting positions corresponding to the cross section during movement of the mobile terminal in a first preset manner, where the first preset manner is that the shooting direction of the 3D lens is located on the cross section and rotates around the target object as a circle center;
the processing module is used for stitching the 3D image corresponding to one cross section to generate a 3D sub-model;
the processing module is also used for identifying characteristic points on the 3D submodels corresponding to the two adjacent cross sections and stitching the two 3D submodels in a mode of overlapping the same characteristic points.
For a target object with a higher longitudinal direction, such as a user, a mobile phone is utilized to establish a 3D sub-model layer by layer, such as a sub-model of an upper half body and a sub-model of a lower half body, and then the sub-models are spliced.
In this embodiment, the preset rule is to capture the target object frame by frame while rotating in a first preset manner, where each frame corresponds to a 3D image, and the specific manner may be to record a video, and then capture the image of each frame as the 3D image.
Further, in order to facilitate splicing and stitching of each sub-model, the mobile phone of the embodiment can implement:
for a longitudinal section on the target object, the 3D lens is used for sequentially shooting the target object according to a preset rule to acquire 3D images of different shooting positions corresponding to the longitudinal section in a second preset mode in which the mobile terminal moves, and the second preset mode is that the 3D lens moves from one longitudinal end of the target object to the other longitudinal end.
The processing module is used for stitching the 3D images corresponding to the longitudinal sections to generate a 3D longitudinal sub-model;
the processing module is further used for selecting a target feature point on the 3D longitudinal sub-model, then selecting a 3D image containing the target feature point in the 3D image corresponding to the cross section, and stitching the 3D longitudinal sub-model and the 3D image containing the target feature point in a mode of overlapping the target feature points.
Referring to fig. 1, with the mobile phone, the embodiment further provides an image generating method, including:
and step 100, for a longitudinal section on the target object, sequentially shooting the target object by the 3D lens according to a preset rule in a second preset mode of movement of the mobile terminal so as to obtain the 3D images of the different shooting positions corresponding to the longitudinal section.
In step 100, the second preset manner is that the 3D lens moves from one longitudinal end of the target object to the other longitudinal end, and the 3D images correspond to the shooting positions one by one. Taking the user himself as an example, the second preset mode is to move the shooting direction of the 3D lens from head to foot.
In step 101, for any two 3D images corresponding to two adjacent shooting positions, the processing module identifies feature points on the two 3D images, and stitches the two 3D images in a manner that the same feature points coincide, where the 3D images corresponding to the two adjacent shooting positions are all shooting positions in a second preset manner.
And step 102, for a cross section on the target object, sequentially shooting the target object by the 3D lens according to a preset rule in the movement of the mobile terminal in a first preset mode so as to obtain the 3D images of the different shooting positions corresponding to the cross section.
In step 102, the first preset mode is that the shooting direction of the 3D lens is located on the cross section and rotates around the target object, and the 3D images correspond to the shooting positions one by one.
103, the processing module selects a target feature point on the 3D longitudinal sub-model, then selects a 3D image containing the target feature point in the 3D image corresponding to the cross section, and stitches the 3D longitudinal sub-model and the 3D image containing the target feature point in a mode of overlapping the target feature points.
And step 104, stitching the 3D image corresponding to one cross section by using the 3D image stitched with the 3D longitudinal sub-model as a starting point to generate a 3D sub-model by the processing module.
In step 104, for any two 3D images corresponding to two adjacent shooting positions, the processing module identifies feature points on the two 3D images, and stitches the two 3D images in a manner that the same feature points coincide, where the 3D images corresponding to the two adjacent shooting positions are all shooting positions in a first preset manner.
After step 103, the 3D image corresponding to one cross section is already positioned by the longitudinal sub-model, and other 3D images may be sequentially stitched on the 3D image already positioned, and then the 3D longitudinal sub-model and one 3D sub-model (transverse) stitching are realized, and the stitching of the longitudinal sub-model and other transverse sub-models may be realized in the same manner.
In this embodiment, a longitudinal submodel is used as a reference, and then a transverse submodel is added, so that inferior particles are not likely to occur in model fidelity, and in an embodiment where a longitudinal submodel is not used, the transverse submodels can be sequentially spliced, and the specific method is as follows: the processing module identifies characteristic points on the 3D submodels corresponding to the two adjacent cross sections, and stitches the two 3D submodels in a mode that the same characteristic points are overlapped.
Example 2
This embodiment is substantially the same as embodiment 1 except that:
the preset rule is that the rotation angle of the mobile terminal rotating in a first preset mode is obtained through the gyroscope and the acceleration sensor, and the 3D lens is used for activating the shutter when the rotation angle reaches a preset angle.
In the mobile phone of this embodiment, each 3D image includes a pixel layer and a structural layer, and for any two 3D images corresponding to adjacent shooting positions,
the processing module is used for identifying at least 3 peak points on the structural layers of the two 3D images;
the processing module is used for stitching the structural layers of the two 3D images in a mode of overlapping the same peak points, the peak points comprise convex points and concave points, and the number of overlapping the peak points of the two 3D images is at least 3;
the processing module is further used for attaching two pixel layers of the 3D image to the stitched 3D image.
Accordingly, the image generating method of the present embodiment provides a method for performing 3D image stitching through a structural layer, including:
each 3D image comprises a pixel layer and a structural layer, and for any two 3D images corresponding to two adjacent shooting positions, the image generation method comprises the following steps:
the processing module identifies at least 3 peak points on the structural layers of the two 3D images;
the processing module stitches the structural layers of the two 3D images in a mode of overlapping the same peak points, wherein the peak points comprise convex points and concave points, and the number of overlapping the peak points of the two 3D images is at least 3;
and the processing module is used for attaching two pixel layers of the 3D image to the stitched 3D image.
While specific embodiments of the invention have been described above, it will be appreciated by those skilled in the art that these are by way of example only, and that the scope of the invention is defined by the appended claims. Various changes and modifications to these embodiments may be made by those skilled in the art without departing from the spirit and scope of the invention, and these changes and modifications are within the scope of the invention.
Claims (6)
1. A mobile terminal with 3D imaging function is characterized in that the mobile terminal comprises at least one 3D lens and a processing module, the 3D lens comprises an infrared beam transmitter, an infrared receiver and a camera,
the 3D lens is used for shooting a target object from different shooting positions to obtain a plurality of 3D images, and the 3D images correspond to the shooting positions one by one;
for any two 3D images corresponding to the adjacent shooting positions, the processing module is used for identifying feature points on the two 3D images and stitching the two 3D images in a mode of overlapping the same feature points;
for a cross section on the target object, the 3D lens is used for sequentially shooting the target object according to a preset rule to obtain 3D images of the different shooting positions corresponding to the cross section during movement of the mobile terminal in a first preset manner, where the first preset manner is that the shooting direction of the 3D lens is located on the cross section and rotates around the target object as a circle center;
the processing module is used for stitching the 3D image corresponding to one cross section to generate a 3D sub-model;
the processing module is also used for identifying characteristic points on the 3D submodels corresponding to the two adjacent cross sections and stitching the two 3D submodels in a mode of overlapping the same characteristic points;
for a longitudinal section on the target object, the 3D lens is used for sequentially shooting the target object according to a preset rule to obtain 3D images at different shooting positions corresponding to the longitudinal section in a second preset manner that the 3D lens moves from one longitudinal end of the target object to the other longitudinal end;
the processing module is used for stitching the 3D images corresponding to the longitudinal sections to generate a 3D longitudinal sub-model;
the processing module is further used for selecting a target feature point on the 3D longitudinal sub-model, then selecting a 3D image containing the target feature point in the 3D image corresponding to the cross section, and stitching the 3D longitudinal sub-model and the 3D image containing the target feature point in a mode of overlapping the target feature points.
2. The mobile terminal of claim 1, wherein the mobile terminal comprises a gyroscope and an acceleration sensor;
the preset rule is that the target object is shot frame by frame, and each frame corresponds to a 3D image; and/or the presence of a gas in the gas,
the preset rule is that the rotation angle of the mobile terminal rotating in a first preset mode is obtained through the gyroscope and the acceleration sensor, and the 3D lens is used for activating the shutter when the rotation angle reaches a preset angle.
3. The mobile terminal of claim 1, wherein each 3D image comprises a pixel layer and a structural layer, and for any two adjacent capturing positions corresponding to the 3D images,
the processing module is used for identifying at least 3 peak points on the structural layers of the two 3D images;
the processing module is used for stitching the structural layers of the two 3D images in a mode of overlapping the same peak points, the peak points comprise convex points and concave points, and the number of overlapping the peak points of the two 3D images is at least 3;
the processing module is further used for attaching two pixel layers of the 3D image to the stitched 3D image.
4. An image generation method, wherein the image generation method is applied to a mobile terminal, the mobile terminal includes at least one 3D lens and a processing module, the 3D lens includes an infrared beam transmitter, an infrared receiver and a camera, and the image generation method includes:
the 3D lens shoots a target object from different shooting positions to obtain a plurality of 3D images, and the 3D images correspond to the shooting positions one by one;
for any two 3D images corresponding to the adjacent shooting positions, the processing module identifies feature points on the two 3D images, and the two 3D images are stitched in a mode that the same feature points are overlapped;
the image generation method comprises the following steps:
for a cross section on the target object, sequentially shooting the target object by the 3D lens according to a preset rule in the movement of the mobile terminal in a first preset mode so as to obtain 3D images of different shooting positions corresponding to the cross section, wherein the first preset mode is that the shooting direction of the 3D lens is located on the cross section and rotates by taking the target object as a circle center;
the processing module stitches the 3D image corresponding to one cross section to generate a 3D sub-model;
the processing module identifies characteristic points on the 3D submodels corresponding to the two adjacent cross sections, and stitches the two 3D submodels in a mode of overlapping the same characteristic points;
the image generation method comprises the following steps:
for a longitudinal section on the target object, sequentially shooting the target object by a preset rule to obtain 3D images of different shooting positions corresponding to the longitudinal section in a second preset mode that the 3D lens moves from one longitudinal end of the target object to the other longitudinal end of the target object;
the processing module stitches the 3D images corresponding to the longitudinal sections to generate a 3D longitudinal sub-model;
the processing module selects a target characteristic point on the 3D longitudinal sub-model, then selects a 3D image containing the target characteristic point in the 3D image corresponding to the cross section, and stitches the 3D longitudinal sub-model and the 3D image containing the target characteristic point in a target characteristic point overlapping mode.
5. The image generating method as claimed in claim 4, wherein the mobile terminal comprises a gyroscope and an acceleration sensor;
the preset rule is that the target object is shot frame by frame, and each frame corresponds to a 3D image; and/or the presence of a gas in the gas,
the preset rule is that the rotation angle of the mobile terminal rotating in a first preset mode is obtained through the gyroscope and the acceleration sensor, and the 3D lens is used for activating the shutter when the rotation angle reaches a preset angle.
6. The image generation method of claim 4, wherein each 3D image comprises a pixel layer and a structural layer, and for any two adjacent capturing positions corresponding to the 3D images, the image generation method comprises:
the processing module identifies at least 3 peak points on the structural layers of the two 3D images;
the processing module stitches the structural layers of the two 3D images in a mode of overlapping the same peak points, wherein the peak points comprise convex points and concave points, and the number of overlapping the peak points of the two 3D images is at least 3;
and the processing module is used for attaching two pixel layers of the 3D image to the stitched 3D image.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810120901.2A CN108347505B (en) | 2018-02-07 | 2018-02-07 | Mobile terminal with 3D imaging function and image generation method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810120901.2A CN108347505B (en) | 2018-02-07 | 2018-02-07 | Mobile terminal with 3D imaging function and image generation method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN108347505A CN108347505A (en) | 2018-07-31 |
CN108347505B true CN108347505B (en) | 2020-11-06 |
Family
ID=62959073
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810120901.2A Active CN108347505B (en) | 2018-02-07 | 2018-02-07 | Mobile terminal with 3D imaging function and image generation method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108347505B (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109151440B (en) * | 2018-10-15 | 2020-06-09 | 盎锐(上海)信息科技有限公司 | Image positioning device and method |
CN109472860B (en) * | 2018-11-13 | 2023-02-10 | 上海盎维信息技术有限公司 | Depth map balance optimization method and device based on artificial intelligence |
CN109151444B (en) * | 2018-11-13 | 2020-09-08 | 盎锐(上海)信息科技有限公司 | 3D Intelligent Pixel enhancement Engine |
CN109509262B (en) * | 2018-11-13 | 2023-02-28 | 上海盎维信息技术有限公司 | Intelligent enhanced modeling method and device based on artificial intelligence |
CN110176034B (en) * | 2019-05-27 | 2023-02-07 | 上海盎维信息技术有限公司 | Positioning method and scanning terminal for VSLAM |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102208116A (en) * | 2010-03-29 | 2011-10-05 | 卡西欧计算机株式会社 | 3D modeling apparatus and 3D modeling method |
CN102917236A (en) * | 2012-09-27 | 2013-02-06 | 深圳天珑无线科技有限公司 | Single-camera based stereoscopic photographing method and digital photographing device |
CN105241397A (en) * | 2015-06-29 | 2016-01-13 | 北航温州研究院 | Real-time measuring splicing method and device based on structured light |
CN105306921A (en) * | 2014-06-18 | 2016-02-03 | 中兴通讯股份有限公司 | Three-dimensional photo shooting method based on mobile terminal and mobile terminal |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10077007B2 (en) * | 2016-03-14 | 2018-09-18 | Uber Technologies, Inc. | Sidepod stereo camera system for an autonomous vehicle |
-
2018
- 2018-02-07 CN CN201810120901.2A patent/CN108347505B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102208116A (en) * | 2010-03-29 | 2011-10-05 | 卡西欧计算机株式会社 | 3D modeling apparatus and 3D modeling method |
CN102917236A (en) * | 2012-09-27 | 2013-02-06 | 深圳天珑无线科技有限公司 | Single-camera based stereoscopic photographing method and digital photographing device |
CN105306921A (en) * | 2014-06-18 | 2016-02-03 | 中兴通讯股份有限公司 | Three-dimensional photo shooting method based on mobile terminal and mobile terminal |
CN105241397A (en) * | 2015-06-29 | 2016-01-13 | 北航温州研究院 | Real-time measuring splicing method and device based on structured light |
Also Published As
Publication number | Publication date |
---|---|
CN108347505A (en) | 2018-07-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108347505B (en) | Mobile terminal with 3D imaging function and image generation method | |
US9927948B2 (en) | Image display apparatus and image display method | |
JP5140210B2 (en) | Imaging apparatus and image processing method | |
CN106507094B (en) | Correct the method and device of panoramic video display view angle | |
CN108600729B (en) | Dynamic 3D model generation device and image generation method | |
JP5204350B2 (en) | Imaging apparatus, playback apparatus, and image processing method | |
US20150358539A1 (en) | Mobile Virtual Reality Camera, Method, And System | |
JP5204349B2 (en) | Imaging apparatus, playback apparatus, and image processing method | |
CN103370943B (en) | Imaging device and formation method | |
US20160344999A1 (en) | SYSTEMS AND METHODs FOR PRODUCING PANORAMIC AND STEREOSCOPIC VIDEOS | |
CA2933704A1 (en) | Systems and methods for producing panoramic and stereoscopic videos | |
TW201351959A (en) | Method of stereo 3D image synthesis and related camera | |
CN108391116B (en) | Whole body scanning device and method based on 3D imaging technology | |
CN108111835B (en) | Shooting device, system and method for 3D image imaging | |
KR20150003576A (en) | Apparatus and method for generating or reproducing three-dimensional image | |
CN108614636A (en) | A kind of 3D outdoor scenes VR production methods | |
CN107637067A (en) | Image processing equipment and image processing method | |
CN108737808B (en) | 3D model generation device and method | |
CN108513122B (en) | Model adjusting method and model generating device based on 3D imaging technology | |
WO2012150680A1 (en) | Three-dimensional image processing apparatus and three-dimensional image processing method | |
CN108234904B (en) | Multi-video fusion method, device and system | |
JP5704885B2 (en) | Imaging device, imaging method, and imaging control program | |
CN108195308B (en) | 3D scanning device, system and method | |
CN109272453B (en) | Modeling device and positioning method based on 3D camera | |
CN108848366B (en) | Information acquisition device and method based on 3D camera |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
TR01 | Transfer of patent right | ||
TR01 | Transfer of patent right |
Effective date of registration: 20230727 Address after: 201703 Room 2134, Floor 2, No. 152 and 153, Lane 3938, Huqingping Road, Qingpu District, Shanghai Patentee after: Shanghai Qingyan Heshi Technology Co.,Ltd. Address before: 201703 No.206, building 1, no.3938 Huqingping Road, Qingpu District, Shanghai Patentee before: UNRE (SHANGHAI) INFORMATION TECHNOLOGY Co.,Ltd. |