CN100594519C - Method for real-time generating reinforced reality surroundings by spherical surface panoramic camera - Google Patents

Method for real-time generating reinforced reality surroundings by spherical surface panoramic camera Download PDF

Info

Publication number
CN100594519C
CN100594519C CN200810101290A CN200810101290A CN100594519C CN 100594519 C CN100594519 C CN 100594519C CN 200810101290 A CN200810101290 A CN 200810101290A CN 200810101290 A CN200810101290 A CN 200810101290A CN 100594519 C CN100594519 C CN 100594519C
Authority
CN
China
Prior art keywords
model
real
illumination
light source
augmented reality
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN200810101290A
Other languages
Chinese (zh)
Other versions
CN101246600A (en
Inventor
吴威
赵旭
张淑军
周忠
李艳丽
赵沁平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beihang University
Original Assignee
Beihang University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beihang University filed Critical Beihang University
Priority to CN200810101290A priority Critical patent/CN100594519C/en
Publication of CN101246600A publication Critical patent/CN101246600A/en
Application granted granted Critical
Publication of CN100594519C publication Critical patent/CN100594519C/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Processing Or Creating Images (AREA)

Abstract

The present invention provides a method of generating augmented reality surroundings illumination model by a spherical panoramic camera real time, including: (1) placing the spherical panoramic camerain the augmented reality surroundings, placing in the viewport of the virtual worlds, real-time collecting panoramic picture; (2) processing the texture map having optical property aimed at the realworld object after being collected panoramic picture of the step 1; (3) calculating collected panoramagram light intensity value aimed at the virtual world object after being collected the panoramic picture of the step 1, processing illumination model pattern matching, generating virtual world illumination model by the calculation; (4) calculating the illumination model of the augmented reality site under the condition of interactive local impact of the consideration reconstruction model and the virtual object model. The invention real-time generates illumination model in the complex photoenvironment, having advantage of support augmented reality site real time interactive.

Description

Method with real-time generating reinforced reality surroundings by spherical surface panoramic camera
Technical field
The invention belongs to the augmented reality field, specifically a kind of computing machine photo realism graphic method of generation in real time that is applied to.
Background technology
Augmented reality system is with virtual world stack or is blended in man-machine interactive system on the real world.Because the object object of systemic effect comprises real-world objects, require the generation sense of reality geometric model that system can be real-time.Illumination model is most important to the generation of photo realism graphic.
Setting up illumination model, mainly is the relevant law according to optical physics, adopts the physical process of computer simulation occurring in nature optical illumination.According to whether considering the light source reflective characteristic of other object outward, illumination model is divided into local illumination model and global illumination model.Wherein, the effect of surrounding environment to object ignored by local illumination model, only considers the radiation response of light source to body surface.The global illumination model is then considered the influence of surrounding environment to the scenery surface.In general, local illumination model is a kind of ideal situation, and gained result and real world have than big difference, are not suitable for using in the augmented reality field, uses the global illumination model to generate photo realism graphic usually.
Light source is the core of illumination model.Can be summed up as pointolite, line source, area source and body light source according to its geometric configuration light source.Wherein, pointolite calculates the simplest, and is the basis that other a few class light sources calculate.
On the basis that light source is determined, set up illumination model.The method of setting up illumination model at present roughly has three classes: a kind of is the mathematical simulation method; A kind of is method according to three-dimensional symbol thing light and shade backstepping illumination model; And a kind of illumination model generation method based on image.
The method that mathematical simulation generates illumination model is the most original illumination model generation method, and extensively is used.Common method comprises Lambet illumination model generation method, Phong illumination model generation method, Witted illumination model generation method, ray tracing method and radiosity method etc. at present.But the method also is not suitable for the complicated and requirement that generate in real time of light source model under the augmented reality environment.Augmented reality system requires physical presence light source in the true environment is registered one by one, uses this method complexity, and workload is big, and limitation is arranged.Show and whenever change an environment, need re-register, but often a lot of owing to influencing system's illumination factor, in fact can't register fully.
Proposition in article " augmented reality system illumination model set up research " such as Zhou Ya a kind of according to registered images in the light and shade situation of mark, utilize the optical illumination Model Calculation method of computer graphics to carry out the method that backstepping is set up illumination model.More effectively solved augmented reality system illumination factor complicated problems, but this method depends on the mark of registered images, and when viewpoint changes, can cause the illumination error, can not satisfy lighting requirement the object of any position and material.
Wang Jun etc. propose a kind of illumination model generation method based on image in article " based on the research and the realization of the illumination model of image ", the any object that is about to the perception of vision institute is regarded light source as, utilize scene around the camera record, and be converted into radiation pattern, panorama sketch and flash ranging figure successively, thereby realize illumination scene.But this method only is fit to the foundation of static scene illumination model to be recovered, and the object of real world can not dynamically move, and does not satisfy the real-time requirement of augmented reality system to man-machine interaction.
Generally speaking, existing scene illumination model generating method exists the problem that the calculation of complex workload is big, be not suitable for calculating in real time under the complex illumination environment, can't supporting the augmented reality system real time human-machine interaction.
Summary of the invention
The technical matters that the present invention solves is: overcome the existing illumination model method that generates and exist calculation of complex, shortcoming that can not real-time generating reinforced reality scene illumination model, a kind of method of using the panoramic camera real-time generating reinforced reality surroundings is provided, supports the augmented reality system real time human-machine interaction.
The present invention proposes a kind of method of using the panoramic camera real-time generating reinforced reality surroundings, may further comprise the steps:
(1) spherical surface panoramic camera is placed the augmented reality environment, the position places the observation viewpoint of virtual world, in real time capturing panoramic view image;
(2) behind step (1) capturing panoramic view image,, its three-dimensional reconstruction model is had the texture mapping of light attribute at real-world objects;
(3) behind step (1) capturing panoramic view image, at virtual world object, calculate the panorama sketch light intensity value that collects, carry out the illumination model pattern match, generate the virtual world illumination model by calculating;
(4) take all factors into consideration under the mutual local influence situation of reconstruction model and virtual objects model, calculate the illumination model of augmented reality scene.
The invention has the beneficial effects as follows:
(1) the present invention overcomes the shortcoming that prior art is used the static collection of camera scene image information calculations illumination model, can realize real-time generating reinforced reality surroundings, for augmented reality environment Real-time and Dynamic provides the illumination support alternately.
(2) the inventive method is divided into virtual objects and real-world objects various computing mode with the calculating of illumination model, and is comparatively simple with respect to existing illumination model mathematic calculation.Prior art is often supposed augmented reality virtual world illumination model parameter, require the iteration adjustment according to vision then, and the present invention carries out reverse computational analysis from the panoramic picture of real scene, generate virtual illumination model on this basis, improved the degree true to nature that the augmented reality virtual world is played up.
(3) in the present invention, when carrying out the light source mode coupling, remove pointolite pattern, lamp pattern etc., definition illumination mode that can be artificial to adapt to different complicated scene environment, has extendability preferably.
Description of drawings
Fig. 1 is panoramic camera of the present invention and augmented reality scene environment graph of a relation;
Fig. 2 is that panorama sketch real-time generating reinforced reality of the present invention is looked after the model method process flow diagram;
Fig. 3 is a panorama sketch coordinate world coordinate system coordinate conversion synoptic diagram of the present invention.
Embodiment
Step 1 as shown in Figure 1, the augmented reality environment scene is a cubic space, people or other object are finished with the virtual objects of augmented reality in cubic space alternately.Place virtual world to observe the position of viewpoint a panoramic camera, carry out the panoramic video collection.
Step 2 provides as shown in Figure 2 by analyzing the process flow diagram that panorama sketch generates illumination model.The method that the present invention proposes comprises six steps, is respectively: the texture mapping of panoramic picture collection, three-dimensional reconstruction model, calculating panorama sketch light intensity value, light source mode are mated, are calculated light source position and input virtual objects model parameter, calculate illumination model.
The collection of step 2.1 panoramic picture.Use places the spherical surface panoramic camera of virtual scene viewpoint to gather panoramic video.
Step 2.2 contains the texture mapping of illumination information as on the content basis as described in the step 2.1 to the three-dimensional reconstruction model of real world in the augmented reality scene.Augmented reality virtual world object model is divided into two classes, and a class is the virtual objects model, and another kind of is the three-dimensional reconstruction model of real world.Determine its illumination feature for the employing of three-dimensional reconstruction model from the mode of panoramic video stream intercepting texture image and pinup picture.At first the real object in the augmented reality scene is carried out background segment, get foreground image, and the Pixel Information in the panorama sketch is mapped to virtual world reconstruction model surface.
Common virtual world coordinate system is the 3 D stereo coordinate system, but the panoramic picture of panoramic camera collection is based on spherical coordinate, need be in the virtual world coordinate system with the coordinate conversion of panoramic picture.
As shown in Figure 3, panoramic camera and scene space relation according to Fig. 1 provide panoramic picture coordinate conversion formula.If the coordinate of panoramic picture be (u, v), solid space length of side length is d, the formula that is transformed into coordinate under the world coordinate system (x ', y ', z ') is as follows:
Figure C20081010129000061
x ′ = x y ′ = y + d z ′ = z + d
Step 2.3 is calculated and is collected the panorama sketch light intensity value as on the content basis as described in the step 2.1.If the panorama map reference is that (u, v), the rgb value of corresponding pixel points is respectively: N R, N G, N B, use RGB conversion HIS formula, calculate the light intensity value of corresponding point, formula is:
N I=N R+N G+N B
Step 2.4 is carried out the light source mode coupling as on the content basis as described in the step 2.3.So-called light source mode, the i.e. number of light source and position distribution rule.At first the panorama plot of light intensity is filtered, and remove noise spot, obtain the panorama sketch light source mode according to certain light intensity threshold values (such as N=220).If have only a pointolite, judge that promptly environment is the single-point light source mode; If the some spots light source is arranged, and the comparatively dispersion that distributes each other, judge that then environment is the multiple spot light source mode; If a large amount of pointolites is arranged, and be evenly distributed, judge that then environment is the lamp pattern; In addition, for special photoenvironment, can manually formulate its light source mode.
Step 2.5 is calculated the position of light source in virtual world, coupling back as on the content basis as described in the step 2.4, and input virtual objects model parameter.For the virtual objects model in the augmented reality virtual world, adopt Phong illumination model computing method.Promptly at first calculate the coordinate position of light source at virtual world according to step 2, and the parameter of input virtual objects model, comprise the diffuse-reflection factor of object, the diffuse-reflection factor of object surface and the specularity factor of object to surround lighting.Substitution Phong illumination model computing formula:
I = k a I a + Σ i = 1 M f i ( d ) I ni [ k d ( N · L i ) + k s ( N · H i ) n ]
Wherein, M is the total number of pointolite in the scene; I is illumination surface point (x, the light intensity of y) locating; I aBe incident surround lighting light intensity, i.e. the average intensity value of the whole pixels of panorama sketch; I Ni, f i(d) be incident light light intensity and the intensity of light source decay factor that i pointolite sends; k aBe the diffuse-reflection factor of model surface to surround lighting; k dIt is the diffuse-reflection factor of model surface; k sIt is the specularity factor of model surface; N is the DE Specular Lighting index; L iBe i pointolite transmit direction vector of unit length; N is that (x y) locates the surface unit normal vector to point; H iBe the unit normal vector that incident light is reflexed to the ideal mirror of observer's direction, V is observer's sight line vector of unit length.
Step 2.6 is calculated the virtual world illumination model as on the content basis as described in step 2.2 and the step 2.5.Consider the interaction of three-dimensional reconstruction model and virtual objects model, adopt ray tracing method, the light reflection case when calculating the blocking mutually of reconstruction model and virtual objects model, computing formula is as follows:
R=L-2N(N·L)
Wherein, R is the reflection ray vector of unit length; L is the incident ray direction vector; N is the surface unit normal vector, and R, L, and the N coplane realizes that by ray tracing method scene blanking and shade generate, and finally finishes the calculating of the illumination model of whole augmented reality virtual world, computing formula is:
I ( x , y ) = k a Σ j = 1 T ( N R + N G + N B ) j T +
Σ i = 1 M f i ( d ) ( N R + N G + N B ) n i ( V · [ L i - 2 N ( N · L i ) ] ) [ k d ( N · L i ) + k s ( N · L i + V 2 ) n ]
Wherein, I (x, y) expression point (x, y) light intensity value of calculating gained; k aBe the diffuse-reflection factor of model surface to surround lighting; k dIt is the diffuse-reflection factor of model surface; k sIt is the specularity factor of model surface; N is the DE Specular Lighting index; f i(d) be the intensity of light source decay factor that i pointolite sends light; N R, N G, N BThe rgb value of expression picture element; L iBe i pointolite transmit direction vector of unit length; N is that (x y) locates the surface unit normal vector to point; V is observer's sight line vector of unit length; T is all pixel summations; M is the illumination summation of counting, and the method by aforesaid illumination model pattern match obtains.
It should be noted that at last; the above only is a preferred implementation of the present invention; should be understood that; for those skilled in the art; utilize under the principle prerequisite of real-time generating reinforced reality surroundings by spherical surface panoramic camera not breaking away from the present invention; can also make some improvement or be equal to replacement, these improvement and be equal to replacement and also should be considered as protection scope of the present invention.

Claims (5)

1, a kind of method with real-time generating reinforced reality surroundings by spherical surface panoramic camera is characterized in that comprising:
(1) spherical surface panoramic camera is placed the augmented reality environment, the position places the observation viewpoint of virtual world, in real time capturing panoramic view image;
(2) behind step (1) capturing panoramic view image, at real-world objects, the texture mapping that its three-dimensional reconstruction model is had the light attribute, the method of the described texture mapping that the three-dimensional reconstruction model is had a light attribute is as follows: at first the real object in the augmented reality scene is carried out background segment, extract foreground image, and be the virtual world coordinate system the coordinate conversion of panoramic picture;
(3) behind step (1) capturing panoramic view image, at virtual world object, calculate the panorama sketch light intensity value that collects, carry out the illumination model pattern match, generate the virtual world illumination model by calculating;
(4) take all factors into consideration under the mutual local influence situation of reconstruction model and virtual objects model, calculate the illumination model of augmented reality scene, the method for the illumination model of described calculating augmented reality scene is:
(4.1) adopt ray tracing method, the light reflection case when calculating the blocking mutually of reconstruction model and virtual objects model, computing formula is as follows:
R=L-2N(N·L)
Wherein, R is that reflection ray vector of unit length, L are that incident ray direction vector, N are the surface unit normal vectors, and R, L, N coplane;
(4.2) realize that by ray tracing method scene blanking and shade generate, finally finish the calculating of the illumination model of whole augmented reality virtual world, computing formula is:
I ( x , y ) = k a Σ j = 1 T ( N R + N G + N B ) j T +
Σ i = 1 m f i ( d ) ( N R + N G + N B ) n i ( V · [ L i - 2 N ( N · L i ) ] ) [ k d ( N · L i ) + k s ( N · L i + V 2 ) 2 ]
Wherein, I (x, y) expression point (x, y) light intensity value of calculating gained; k aBe the diffuse-reflection factor of model surface to surround lighting; k bIt is the diffuse-reflection factor of model surface; k sIt is the specularity factor of model surface; N is the DE Specular Lighting index; f i(d) be the intensity of light source decay factor that i pointolite sends light; N R, N G, N BThe rgb value of expression picture element; L iBe i pointolite transmit direction vector of unit length; N is that (x y) locates the surface unit normal vector to point; V is observer's sight line vector of unit length; T is all pixel summations; M is the illumination summation of counting, and the method by the illumination model pattern match obtains.
2, the method with real-time generating reinforced reality surroundings by spherical surface panoramic camera according to claim 1, it is characterized in that: described step (2) is that the formula of virtual world coordinate system is with the coordinate conversion of panoramic picture: the coordinate of establishing panoramic picture is (u, v), solid space length of side length is d, be transformed into coordinate under the world coordinate system (x ', y ', z ') formula is as follows:
Figure C2008101012900003C1
3, the method with real-time generating reinforced reality surroundings by spherical surface panoramic camera according to claim 1, it is characterized in that: the panorama sketch light intensity value is calculated as follows in the described step (3):
If the coordinate of panorama sketch is that (u, v), the red R of corresponding pixel points, green G, blue B value are respectively: N R, N G, N B, use RGB conversion HIS formula, calculate the light intensity value of corresponding point, formula is:
N I=N R+I G+N B
4, the method with real-time generating reinforced reality surroundings by spherical surface panoramic camera according to claim 1, it is characterized in that: the method for carrying out the illumination model pattern match in the described step (3) is:
(3.1) at first the panorama plot of light intensity is filtered, and remove noise spot, obtain the panorama sketch light source mode according to certain light intensity threshold values;
(3.2) light source mode is divided into single-point light source mode, multiple spot light source mode, lamp pattern and self-defined light source mode,, judges that promptly environment is the single-point light source mode if promptly have only a pointolite; If the some spots light source is arranged, and the comparatively dispersion that distributes each other, judge that then environment is the multiple spot light source mode; If a large amount of pointolites is arranged, and be evenly distributed, judge that then environment is the lamp pattern; For special photoenvironment, can manually formulate its light source mode.
5, the method with real-time generating reinforced reality surroundings by spherical surface panoramic camera according to claim 1 is characterized in that: calculating the formula that generates the virtual world illumination model in the described step (3) is:
I = k a I a + Σ i = 1 M f i ( d ) I ni [ k d ( N · L i ) + k s ( N · H i ) n ]
Wherein: M is the total number of pointolite in the scene; I is illumination surface point (x, the light intensity of y) locating; I aBe incident surround lighting light intensity, i.e. the average intensity value of the whole pixels of panorama sketch; I Ni, f i(d) be incident light light intensity and the intensity of light source decay factor that i pointolite sends; k aBe the diffuse-reflection factor of model surface to surround lighting; k dIt is the diffuse-reflection factor of model surface; k sIt is the specularity factor of model surface; N is the DE Specular Lighting index; L iBe i pointolite transmit direction vector of unit length; N is that (x y) locates the surface unit normal vector to point; H iBe the unit normal vector that incident light is reflexed to the ideal mirror of observer's direction,
Figure C2008101012900003C5
V is observer's sight line vector of unit length.
CN200810101290A 2008-03-03 2008-03-03 Method for real-time generating reinforced reality surroundings by spherical surface panoramic camera Expired - Fee Related CN100594519C (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN200810101290A CN100594519C (en) 2008-03-03 2008-03-03 Method for real-time generating reinforced reality surroundings by spherical surface panoramic camera

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN200810101290A CN100594519C (en) 2008-03-03 2008-03-03 Method for real-time generating reinforced reality surroundings by spherical surface panoramic camera

Publications (2)

Publication Number Publication Date
CN101246600A CN101246600A (en) 2008-08-20
CN100594519C true CN100594519C (en) 2010-03-17

Family

ID=39947036

Family Applications (1)

Application Number Title Priority Date Filing Date
CN200810101290A Expired - Fee Related CN100594519C (en) 2008-03-03 2008-03-03 Method for real-time generating reinforced reality surroundings by spherical surface panoramic camera

Country Status (1)

Country Link
CN (1) CN100594519C (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102436663A (en) * 2010-08-12 2012-05-02 株式会社泛泰 User equipment, server, and method for selectively filtering augmented reality
US9449428B2 (en) 2009-12-21 2016-09-20 Thomson Licensing Method for generating an environment map

Families Citing this family (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA2774257C (en) * 2008-09-30 2021-04-27 Universite De Montreal Method and device for assessing, training and improving perceptual-cognitive abilities of individuals
FR2965652A1 (en) * 2010-09-30 2012-04-06 Thomson Licensing METHOD FOR ESTIMATING LIGHT QUANTITY RECEIVED IN ONE POINT OF A VIRTUAL ENVIRONMENT
EP2650843A4 (en) * 2010-12-09 2018-03-28 Samsung Electronics Co., Ltd. Image processor, lighting processor and method therefor
CN102236912A (en) * 2011-07-08 2011-11-09 清华大学 Three-dimensional reconstruction method and device of moving target under variable illumination condition
CN102538761B (en) * 2012-01-09 2014-09-03 刘进 Photography measurement method for spherical panoramic camera
CN103814382B (en) * 2012-09-14 2016-10-05 华为技术有限公司 The augmented reality processing method and processing device of mobile terminal
US9524585B2 (en) 2012-11-05 2016-12-20 Microsoft Technology Licensing, Llc Constructing augmented reality environment with pre-computed lighting
CN103108452B (en) * 2013-01-10 2015-01-21 北京航空航天大学 Scene illumination reappearing method driven by dynamic light field data
CN103489002B (en) * 2013-09-27 2017-03-29 广州中国科学院软件应用技术研究所 A kind of augmented reality method and system
CN104580920B (en) * 2013-10-21 2018-03-13 华为技术有限公司 The method and user terminal of a kind of imaging
CN103761763B (en) * 2013-12-18 2017-01-04 微软技术许可有限责任公司 For the method using precalculated illumination to build augmented reality environment
CN104268939B (en) * 2014-09-28 2017-02-08 国家电网公司 Transformer substation virtual-reality management system based on three-dimensional panoramic view and implementation method of transformer substation virtual-reality management system based on three-dimensional panoramic view
CN104331929B (en) * 2014-10-29 2018-02-02 深圳先进技术研究院 Scene of a crime restoring method based on video map and augmented reality
CN104392481B (en) * 2014-11-25 2017-12-05 无锡梵天信息技术股份有限公司 A kind of method and device that high light reflectivity definition is controlled using textures
US9779512B2 (en) * 2015-01-29 2017-10-03 Microsoft Technology Licensing, Llc Automatic generation of virtual materials from real-world materials
CN105138763A (en) * 2015-08-19 2015-12-09 中山大学 Method for real scene and reality information superposition in augmented reality
CN105447906B (en) * 2015-11-12 2018-03-13 浙江大学 The method that weight illumination render is carried out based on image and model calculating illumination parameter
CN105488840B (en) * 2015-11-26 2019-04-23 联想(北京)有限公司 A kind of information processing method and electronic equipment
CN105844695B (en) * 2016-03-18 2017-05-24 山东大学 Illumination modeling method based on real material measurement data
CN106056550B (en) * 2016-05-25 2019-02-22 广东工业大学 A kind of rendering method and device based on high dynamic range images
CN106327438B (en) * 2016-08-12 2019-02-26 武汉秀宝软件有限公司 A kind of pair of bloom and the augmented reality method and Crawl mat application for repeating texture elimination
CN106840389A (en) * 2016-12-30 2017-06-13 歌尔科技有限公司 Light source estimating and measuring method and device, intelligent electronic device based on multiple balls
CN106658148B (en) * 2017-01-16 2020-04-10 深圳创维-Rgb电子有限公司 VR playing method, VR playing device and VR playing system
CN106940897A (en) * 2017-03-02 2017-07-11 苏州蜗牛数字科技股份有限公司 A kind of method that real shadow is intervened in AR scenes
CN106981087A (en) * 2017-04-05 2017-07-25 杭州乐见科技有限公司 Lighting effect rendering intent and device
CN109427089B (en) * 2017-08-25 2023-04-28 微软技术许可有限责任公司 Mixed reality object presentation based on ambient lighting conditions
CN107749076B (en) * 2017-11-01 2021-04-20 太平洋未来科技(深圳)有限公司 Method and device for generating real illumination in augmented reality scene
CN108509887A (en) * 2018-03-26 2018-09-07 深圳超多维科技有限公司 A kind of acquisition ambient lighting information approach, device and electronic equipment
CN109883414B (en) * 2019-03-20 2021-08-27 百度在线网络技术(北京)有限公司 Vehicle navigation method and device, electronic equipment and storage medium
CN110033423B (en) * 2019-04-16 2020-08-28 北京字节跳动网络技术有限公司 Method and apparatus for processing image
CN111866489A (en) * 2019-04-29 2020-10-30 浙江开奇科技有限公司 Method for realizing immersive panoramic teaching
CN110471061A (en) * 2019-07-16 2019-11-19 青岛擎鹰信息科技有限责任公司 A kind of emulation mode and its system for realizing airborne synthetic aperture radar imaging
CN111145333B (en) * 2019-12-11 2022-08-12 江苏艾佳家居用品有限公司 Indoor scene illumination layout method
CN111127624A (en) * 2019-12-27 2020-05-08 珠海金山网络游戏科技有限公司 Illumination rendering method and device based on AR scene
CN111932641B (en) 2020-09-27 2021-05-14 北京达佳互联信息技术有限公司 Image processing method and device, electronic equipment and storage medium
CN112562051B (en) * 2020-11-30 2023-06-27 腾讯科技(深圳)有限公司 Virtual object display method, device, equipment and storage medium
CN114979457B (en) * 2021-02-26 2023-04-07 华为技术有限公司 Image processing method and related device
CN113743380B (en) * 2021-11-03 2022-02-15 江苏博子岛智能产业技术研究院有限公司 Active tracking method based on video image dynamic monitoring
CN116664752B (en) * 2023-08-01 2023-10-17 南京维赛客网络科技有限公司 Method, system and storage medium for realizing panoramic display based on patterned illumination

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
基于图像的光照模型的研究与实现. 王骏.计算机工程与设计,第26卷第1期. 2005
基于图像的光照模型的研究与实现. 王骏.计算机工程与设计,第26卷第1期. 2005 *
增强现实***光照模型建立研究. 周雅.中国图象图形学报,第9卷第8期. 2004
增强现实***光照模型建立研究. 周雅.中国图象图形学报,第9卷第8期. 2004 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9449428B2 (en) 2009-12-21 2016-09-20 Thomson Licensing Method for generating an environment map
CN102436663A (en) * 2010-08-12 2012-05-02 株式会社泛泰 User equipment, server, and method for selectively filtering augmented reality

Also Published As

Publication number Publication date
CN101246600A (en) 2008-08-20

Similar Documents

Publication Publication Date Title
CN100594519C (en) Method for real-time generating reinforced reality surroundings by spherical surface panoramic camera
CN106803267B (en) Kinect-based indoor scene three-dimensional reconstruction method
CN110517355B (en) Ambient composition for illuminating mixed reality objects
JP6425780B1 (en) Image processing system, image processing apparatus, image processing method and program
CN103226830B (en) The Auto-matching bearing calibration of video texture projection in three-dimensional virtual reality fusion environment
CN110148204B (en) Method and system for representing virtual objects in a view of a real environment
CN103400409B (en) A kind of coverage 3D method for visualizing based on photographic head attitude Fast estimation
CN107341853B (en) Virtual-real fusion method and system for super-large virtual scene and dynamic screen shooting
JP4804256B2 (en) Information processing method
CN102096941B (en) Consistent lighting method under falsehood-reality fused environment
CN104330074B (en) Intelligent surveying and mapping platform and realizing method thereof
Sawhney et al. Video flashlights: real time rendering of multiple videos for immersive model visualization
JP2006053694A (en) Space simulator, space simulation method, space simulation program and recording medium
CN109314753A (en) Medial view is generated using light stream
CN109754461A (en) Image processing method and related product
Maurus et al. Realistic heatmap visualization for interactive analysis of 3D gaze data
CN106940897A (en) A kind of method that real shadow is intervened in AR scenes
CN113436559B (en) Sand table dynamic landscape real-time display system and display method
CN106683163B (en) Imaging method and system for video monitoring
Okura et al. Mixed-reality world exploration using image-based rendering
CN108629828B (en) Scene rendering transition method in the moving process of three-dimensional large scene
Du et al. Video fields: fusing multiple surveillance videos into a dynamic virtual environment
CN107562185B (en) Light field display system based on head-mounted VR equipment and implementation method
CN111145341A (en) Single light source-based virtual-real fusion illumination consistency drawing method
Li et al. Augmented reality and virtual reality

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
C17 Cessation of patent right
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20100317

Termination date: 20130303