JPWO2020165076A5 - - Google Patents

Download PDF

Info

Publication number
JPWO2020165076A5
JPWO2020165076A5 JP2021541447A JP2021541447A JPWO2020165076A5 JP WO2020165076 A5 JPWO2020165076 A5 JP WO2020165076A5 JP 2021541447 A JP2021541447 A JP 2021541447A JP 2021541447 A JP2021541447 A JP 2021541447A JP WO2020165076 A5 JPWO2020165076 A5 JP WO2020165076A5
Authority
JP
Japan
Prior art keywords
light intensity
image
value
depth
determining
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
JP2021541447A
Other languages
Japanese (ja)
Other versions
JP7460641B2 (en
JP2022520525A (en
Publication date
Priority claimed from EP19157331.0A external-priority patent/EP3696773A1/en
Application filed filed Critical
Publication of JP2022520525A publication Critical patent/JP2022520525A/en
Publication of JPWO2020165076A5 publication Critical patent/JPWO2020165076A5/ja
Application granted granted Critical
Publication of JP7460641B2 publication Critical patent/JP7460641B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Claims (14)

光強度画像を生成する方法であって、
第1視点からのシーンを表す第1テクスチャマップおよび第1メッシュ受信するステップと、
第2視点からの前記シーンを表す第2テクスチャマップおよび第2メッシュ受信するステップと、
前記第1テクスチャマップ、前記第1メッシュ、前記第2テクスチャマップおよび前記第2メッシュに応じて、第3視点からの前記シーンを表す前記光強度画像を決定するステップと、を有し、
前記光強度画像を決定する前記ステップは、
前記光強度画像内の対応する第1画像位置および前記第1メッシュの頂点のための第1画像位置奥行き値を決定し、
前記光強度画像内の対応する第2画像位置および前記第2メッシュの頂点のための第2画像位置奥行き値を決定し、
前記光強度画像内の複数の位置のうちの各第1位置について、
前記第1テクスチャマップおよび前記第1画像位置に基づく第1光強度値、ならびに、前記第1画像位置および前記第1画像位置奥行き値に基づく第1奥行き値を決定し、
前記第2テクスチャマップおよび前記第2画像位置に基づく第2光強度値、ならびに、前記第2画像位置および前記第2画像位置奥行き値に基づく第2奥行き値を決定し、
少なくとも前記第1光強度値および前記第2光強度値の重み付け組み合わせとして前記第1位置における前記光強度画像の光強度値を決定し、
第1奥行き値が増加するほど前記第2光強度値に対する前記第1光強度値の重みが増加前記重みが、前記第1奥行き値と前記第2奥行き値との間の差分測度の大きさに依存する、方法。
A method of generating a light intensity image, comprising:
receiving a first texture map and a first mesh representing the scene from a first viewpoint;
receiving a second texture map and a second mesh representing the scene from a second viewpoint;
determining the light intensity image representing the scene from a third viewpoint according to the first texture map, the first mesh, the second texture map and the second mesh;
The step of determining the light intensity image comprises:
determining first image position depth values for corresponding first image positions in said light intensity image and vertices of said first mesh;
determining second image location depth values for corresponding second image locations in said light intensity image and vertices of said second mesh;
For each first position of a plurality of positions within the light intensity image,
determining a first light intensity value based on the first texture map and the first image location and a first depth value based on the first image location and the first image location depth value;
determining a second light intensity value based on the second texture map and the second image location and a second depth value based on the second image location and the second image location depth value;
determining a light intensity value of the light intensity image at the first location as a weighted combination of at least the first light intensity value and the second light intensity value;
As the first depth value increases, the weight of the first light intensity value with respect to the second light intensity value increases, and the weight increases the difference measure between the first depth value and the second depth value. Depends on how.
前記重みが、前記第1奥行き値と前記第2奥行き値との間の差の符号に依存する、請求項1に記載の方法。 2. Method according to claim 1 , wherein said weight depends on the sign of the difference between said first depth value and said second depth value. 前記第1奥行き値および前記第2奥行き値のうち小さい方の奥行き値に対する前記第1奥行き値および前記第2奥行き値のうち大きい方の奥行き値に対する重みが、前記差分測度の大きさが増加するほど、増加する、請求項
に記載の方法。
The weight of the larger one of the first depth value and the second depth value with respect to the smaller one of the first depth value and the second depth value increases as the magnitude of the difference measure increases. Claims that increase as much as
1. The method according to 1.
前記光強度画像の前記光強度値の決定が、前記第1光強度値および前記第2光強度値のうち奥行き値が大きい方の値に前記光強度値を設定することを含む、請求項1からのいずれか一項に記載の方法。 2. The determining of the light intensity value of the light intensity image comprises setting the light intensity value to one of the first light intensity value and the second light intensity value that has a greater depth value. 4. The method of any one of 3 to 4. 前記光強度画像内の対応する第1画像位置および前記第1メッシュの頂点に対する第1画像位置奥行き値の決定が頂点シェーダによって実行される、請求項1から請求項のいずれか一項に記載の方法。 5. A vertex shader according to any one of claims 1 to 4 , wherein determination of first image position depth values for corresponding first image positions in said light intensity image and vertices of said first mesh is performed by a vertex shader. the method of. 前記第1光強度値を決定するステップが第1フラグメントシェーダによって実行される、請求項1から請求項のいずれか一項に記載の方法。 6. The method of any one of claims 1 to 5 , wherein determining the first light intensity value is performed by a first fragment shader. 前記光強度画像の前記光強度値を決定するステップが第2フラグメントシェーダによって実行される、請求項1からのいずれか一項に記載の方法。 7. The method of any one of claims 1 to 6 , wherein determining the light intensity values of the light intensity image is performed by a second fragment shader. 前記第2フラグメントシェーダが、
少なくとも1つの光強度成分値および1つの透明成分値を含む多成分値を有する第2シェーダテクスチャマップを受信し、前記透明成分値に応じて複数の前記テクスチャマップからの光強度成分値および組み合わせることにより出力光強度値を生成するように構成され、前記方法がさらに、前記第1光強度値を有する多成分値の透明成分値を第1奥行き値に設定するステップを有する、請求項に記載の方法。
The second fragment shader
Receiving a second shader texture map having multi-component values including at least one light intensity component value and one transparency component value, and combining light intensity component values from a plurality of said texture maps according to said transparency component value. 8. The method of claim 7 , wherein the method further comprises setting a transparency component value of a multi-component value having the first light intensity value to a first depth value. the method of.
前記重み付け組み合わせが、第3メッシュおよび第3テクスチャマップから生成される第3光強度値をさらに含む、請求項1からのいずれか一項に記載の方法。 9. A method according to any preceding claim, wherein said weighted combination further comprises a third light intensity value generated from a third mesh and a third texture map. 前記第2光強度値に対する前記第1光強度値の前記重みが、前記第1光強度値の周りの第1近傍における第1光強度値および前記第2光強度値の周りの第2近傍における第2光強度値に依存する、請求項1からのいずれか一項に記載の方法。 The weight of the first light intensity value relative to the second light intensity value is a first light intensity value in a first neighborhood around the first light intensity value and a second light intensity value in a second neighborhood around the second light intensity value. 10. A method according to any one of claims 1 to 9 , depending on the second light intensity value. 前記重みが、前記第1視点と前記第3視点との間の差にさらに依存する、請求項1から10のいずれか一項に記載の方法。 11. A method according to any one of claims 1 to 10 , wherein said weight further depends on the difference between said first viewpoint and said third viewpoint. 前記第2光強度値に対する前記第1光強度値の前記重みが、前記第1奥行きマップにおける前記第1メッシュ位置における奥行きを決定するために使用される奥行き推定の信頼性を示す信頼性指標にさらに依存し、前記第2光強度値に対する前記第1光強度値の前記重みが、前記奥行き推定の信頼性が増加することを示す前記信頼性指標に対して増加する、請求項1から11のいずれか一項に記載の方法。 wherein the weight of the first light intensity value relative to the second light intensity value is a reliability metric that indicates the reliability of a depth estimate used to determine depth at the first mesh location in the first depth map. 12. The method of claims 1 to 11 , further dependent, wherein the weight of the first light intensity value with respect to the second light intensity value increases with respect to the reliability index indicating that the reliability of the depth estimation increases. A method according to any one of paragraphs. 光強度画像を生成するための装置であって、
第1視点からのシーンを表す第1テクスチャマップと第1メッシュとを受信する第1受信機と、
第2視点からの前記シーンを表す第2テクスチャマップと第2メッシュとを受信する第2受信機と、
前記第1テクスチャマップ、前記第1メッシュ、前記第2テクスチャマップおよび前記第2メッシュに応じて、前記第3視点からの前記シーンを表す前記光強度画像を決定する画像生成器と、
を有し、
前記画像生成器は、
前記光強度画像における対応する第1画像位置および前記第1メッシュの頂点のための第1画像位置奥行き値を決定するための第1ビュー変換器と、
前記光強度画像における対応する第2画像位置および前記第2メッシュの頂点のための第2画像位置奥行き値を決定するための第2ビュー変換器と、
前記光強度画像における複数の位置のうちの各第1位置のために、前記第1テクスチャマップおよび前記第1画像位置に基づく第1光強度値、ならびに、前記第1画像位置および前記第1画像位置奥行き値に基づく第1奥行き値を決定するための第1シェーダと、
前記光強度画像における複数の位置のうちの各第1位置のために、前記第2テクスチャマップおよび前記第2画像位置に基づく第2光強度値、ならびに、前記第2画像位置および前記第2画像位置奥行き値に基づく第2奥行き値を決定するための第2シェーダと、
前記光強度画像における複数の位置のうちの各第1位置のために、少なくとも前記第1光強度値と前記第2光強度値との重み付け組み合わせとして当該第1位置における前記光強度画像のための光強度値を決定する結合器と、
を有し、第1奥行き値が増加するほど、前記第2光強度値に対する前記第1光強度値の重みが増加前記重みが、前記第1奥行き値と前記第2奥行き値との間の差分測度の大きさに依存する、装置。
An apparatus for generating a light intensity image, comprising:
a first receiver that receives a first texture map and a first mesh representing a scene from a first viewpoint;
a second receiver for receiving a second texture map and a second mesh representing the scene from a second viewpoint;
an image generator for determining the light intensity image representing the scene from the third viewpoint according to the first texture map, the first mesh, the second texture map and the second mesh;
has
The image generator comprises:
a first view transformer for determining first image position depth values for corresponding first image positions in said light intensity image and vertices of said first mesh;
a second view transformer for determining second image location depth values for corresponding second image locations in said light intensity image and vertices of said second mesh;
a first light intensity value based on the first texture map and the first image position, and the first image position and the first image, for each first position of a plurality of positions in the light intensity image; a first shader for determining a first depth value based on the position depth value;
a second light intensity value based on the second texture map and the second image position, and the second image position and the second image, for each first position of a plurality of positions in the light intensity image; a second shader for determining a second depth value based on the positional depth value;
for each first position of a plurality of positions in the light intensity image as a weighted combination of at least the first light intensity value and the second light intensity value for the light intensity image at that first position; a coupler for determining light intensity values;
and the weight of the first light intensity value with respect to the second light intensity value increases as the first depth value increases, and the weight is between the first depth value and the second depth value A device that depends on the magnitude of the difference measure of .
コンピュータにより実行されて、当該コンピュータに請求項1から12のいずれか一項に記載の方法を実行させるコンピュータプログラム。 A computer program, being executed by a computer, causing said computer to perform the method according to any one of claims 1 to 12 .
JP2021541447A 2019-02-15 2020-02-10 Apparatus and method for generating a light intensity image - Patents.com Active JP7460641B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
EP19157331.0 2019-02-15
EP19157331.0A EP3696773A1 (en) 2019-02-15 2019-02-15 Apparatus and method for generating a light intensity image
PCT/EP2020/053281 WO2020165076A1 (en) 2019-02-15 2020-02-10 Apparatus and method for generating a light intensity image

Publications (3)

Publication Number Publication Date
JP2022520525A JP2022520525A (en) 2022-03-31
JPWO2020165076A5 true JPWO2020165076A5 (en) 2023-01-30
JP7460641B2 JP7460641B2 (en) 2024-04-02

Family

ID=65443692

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2021541447A Active JP7460641B2 (en) 2019-02-15 2020-02-10 Apparatus and method for generating a light intensity image - Patents.com

Country Status (9)

Country Link
US (1) US11783527B2 (en)
EP (2) EP3696773A1 (en)
JP (1) JP7460641B2 (en)
KR (1) KR20210129685A (en)
CN (1) CN113424231A (en)
BR (1) BR112021015772A2 (en)
CA (1) CA3129658A1 (en)
TW (1) TWI831919B (en)
WO (1) WO2020165076A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3792876A1 (en) * 2019-09-13 2021-03-17 FRAUNHOFER-GESELLSCHAFT zur Förderung der angewandten Forschung e.V. Apparatus, method and computer program for rendering a visual scene
US11928787B2 (en) * 2020-07-29 2024-03-12 Intel Corporation Deep novel view synthesis from unstructured input

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7068272B1 (en) * 2000-05-31 2006-06-27 Nvidia Corporation System, method and article of manufacture for Z-value and stencil culling prior to rendering in a computer graphics processing pipeline
JP4187188B2 (en) * 2002-04-26 2008-11-26 株式会社バンダイナムコゲームス Image generation system, program, and information storage medium
US7142209B2 (en) 2004-08-03 2006-11-28 Microsoft Corporation Real-time rendering system and process for interactive viewpoint video that was generated using overlapping images of a scene captured from viewpoints forming a grid
WO2009013682A2 (en) * 2007-07-26 2009-01-29 Koninklijke Philips Electronics N.V. Method and apparatus for depth-related information propagation
US9648346B2 (en) * 2009-06-25 2017-05-09 Microsoft Technology Licensing, Llc Multi-view video compression and streaming based on viewpoints of remote viewer
JP2011028428A (en) * 2009-07-23 2011-02-10 Kddi Corp Method and apparatus for high-speed interpolation of ray information
US8698832B1 (en) * 2010-05-26 2014-04-15 Disney Enterprises, Inc. Perceptual detail and acutance enhancement for digital images
US8847968B2 (en) * 2011-07-12 2014-09-30 Qualcomm Incorporated Displaying static images
WO2013133648A1 (en) * 2012-03-07 2013-09-12 엘지전자 주식회사 Method and device for processing video signal
AU2014218390B2 (en) * 2014-08-27 2017-01-12 Canon Kabushiki Kaisha Method, system and apparatus for forming a high resolution depth map
US9613587B2 (en) * 2015-01-20 2017-04-04 Snaptrack, Inc. Apparatus and method for adaptive image rendering based on ambient light levels
US10325402B1 (en) * 2015-07-17 2019-06-18 A9.Com, Inc. View-dependent texture blending in 3-D rendering
EP3552183B1 (en) 2016-12-06 2021-09-15 Koninklijke Philips N.V. Apparatus and method for generating a light intensity image
EP3358844A1 (en) * 2017-02-07 2018-08-08 Koninklijke Philips N.V. Method and apparatus for processing an image property map
US10482618B2 (en) * 2017-08-21 2019-11-19 Fotonation Limited Systems and methods for hybrid depth regularization
US11528461B2 (en) * 2018-11-16 2022-12-13 Electronics And Telecommunications Research Institute Method and apparatus for generating virtual viewpoint image

Similar Documents

Publication Publication Date Title
US9916684B2 (en) Generating portable three-dimensional print-preview renderings of three-dimensional models
JP7045721B2 (en) Identification device, identification method, and program to identify the position of the wall surface in the tunnel shown in the photographic image.
US10593096B2 (en) Graphics processing employing cube map texturing
KR100920225B1 (en) Method and apparatus for accuracy measuring of?3d graphical model by using image
US9892550B2 (en) Photorealistic rendering of scenes with dynamic content
JP2006190308A (en) Depth image-based modeling method and apparatus
US20130063472A1 (en) Customized image filters
JP2019536174A (en) Apparatus and method for generating light intensity images
KR20120064641A (en) Image processing apparatus, apparatus and method for lighting processing
CN114119853A (en) Image rendering method, device, equipment and medium
TWI439961B (en) Conversion algorithm for voids generated after converting 2D images
CN112365572B (en) Rendering method based on surface subdivision and related products thereof
KR101585998B1 (en) Image processing apparatus and method
JPWO2020165076A5 (en)
CN112509108A (en) GPU-based vertex ambient light shading generation method and image rendering method
JP5432571B2 (en) Image processing apparatus and method
Marrugo et al. Fourier transform profilometry in LabVIEW
CN110378948B (en) 3D model reconstruction method and device and electronic equipment
JP6393153B2 (en) Program, recording medium, luminance calculation device, and luminance calculation method
KR101227155B1 (en) Graphic image processing apparatus and method for realtime transforming low resolution image into high resolution image
CN110458935A (en) A kind of section Real-time Generation of 3D model
JP2009244229A (en) Three-dimensional image processing method, three-dimensional image processing device, and three-dimensional image processing program
KR20190067070A (en) System and method for projection mapping in virtual space
Thorne Healing Cracks in Cyberspace: towards best practice
RU2021127206A (en) Device and method for forming a light-intensive image