JPWO2020165076A5 - - Google Patents
Download PDFInfo
- Publication number
- JPWO2020165076A5 JPWO2020165076A5 JP2021541447A JP2021541447A JPWO2020165076A5 JP WO2020165076 A5 JPWO2020165076 A5 JP WO2020165076A5 JP 2021541447 A JP2021541447 A JP 2021541447A JP 2021541447 A JP2021541447 A JP 2021541447A JP WO2020165076 A5 JPWO2020165076 A5 JP WO2020165076A5
- Authority
- JP
- Japan
- Prior art keywords
- light intensity
- image
- value
- depth
- determining
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 claims 16
- 239000012634 fragment Substances 0.000 claims 3
- 238000004590 computer program Methods 0.000 claims 1
- 230000001419 dependent effect Effects 0.000 claims 1
Claims (14)
第1視点からのシーンを表す第1テクスチャマップおよび第1メッシュ受信するステップと、
第2視点からの前記シーンを表す第2テクスチャマップおよび第2メッシュ受信するステップと、
前記第1テクスチャマップ、前記第1メッシュ、前記第2テクスチャマップおよび前記第2メッシュに応じて、第3視点からの前記シーンを表す前記光強度画像を決定するステップと、を有し、
前記光強度画像を決定する前記ステップは、
前記光強度画像内の対応する第1画像位置および前記第1メッシュの頂点のための第1画像位置奥行き値を決定し、
前記光強度画像内の対応する第2画像位置および前記第2メッシュの頂点のための第2画像位置奥行き値を決定し、
前記光強度画像内の複数の位置のうちの各第1位置について、
前記第1テクスチャマップおよび前記第1画像位置に基づく第1光強度値、ならびに、前記第1画像位置および前記第1画像位置奥行き値に基づく第1奥行き値を決定し、
前記第2テクスチャマップおよび前記第2画像位置に基づく第2光強度値、ならびに、前記第2画像位置および前記第2画像位置奥行き値に基づく第2奥行き値を決定し、
少なくとも前記第1光強度値および前記第2光強度値の重み付け組み合わせとして前記第1位置における前記光強度画像の光強度値を決定し、
第1奥行き値が増加するほど前記第2光強度値に対する前記第1光強度値の重みが増加し、前記重みが、前記第1奥行き値と前記第2奥行き値との間の差分測度の大きさに依存する、方法。 A method of generating a light intensity image, comprising:
receiving a first texture map and a first mesh representing the scene from a first viewpoint;
receiving a second texture map and a second mesh representing the scene from a second viewpoint;
determining the light intensity image representing the scene from a third viewpoint according to the first texture map, the first mesh, the second texture map and the second mesh;
The step of determining the light intensity image comprises:
determining first image position depth values for corresponding first image positions in said light intensity image and vertices of said first mesh;
determining second image location depth values for corresponding second image locations in said light intensity image and vertices of said second mesh;
For each first position of a plurality of positions within the light intensity image,
determining a first light intensity value based on the first texture map and the first image location and a first depth value based on the first image location and the first image location depth value;
determining a second light intensity value based on the second texture map and the second image location and a second depth value based on the second image location and the second image location depth value;
determining a light intensity value of the light intensity image at the first location as a weighted combination of at least the first light intensity value and the second light intensity value;
As the first depth value increases, the weight of the first light intensity value with respect to the second light intensity value increases, and the weight increases the difference measure between the first depth value and the second depth value. Depends on how.
1に記載の方法。 The weight of the larger one of the first depth value and the second depth value with respect to the smaller one of the first depth value and the second depth value increases as the magnitude of the difference measure increases. Claims that increase as much as
1. The method according to 1.
少なくとも1つの光強度成分値および1つの透明成分値を含む多成分値を有する第2シェーダテクスチャマップを受信し、前記透明成分値に応じて複数の前記テクスチャマップからの光強度成分値および組み合わせることにより出力光強度値を生成するように構成され、前記方法がさらに、前記第1光強度値を有する多成分値の透明成分値を第1奥行き値に設定するステップを有する、請求項7に記載の方法。 The second fragment shader
Receiving a second shader texture map having multi-component values including at least one light intensity component value and one transparency component value, and combining light intensity component values from a plurality of said texture maps according to said transparency component value. 8. The method of claim 7 , wherein the method further comprises setting a transparency component value of a multi-component value having the first light intensity value to a first depth value. the method of.
第1視点からのシーンを表す第1テクスチャマップと第1メッシュとを受信する第1受信機と、
第2視点からの前記シーンを表す第2テクスチャマップと第2メッシュとを受信する第2受信機と、
前記第1テクスチャマップ、前記第1メッシュ、前記第2テクスチャマップおよび前記第2メッシュに応じて、前記第3視点からの前記シーンを表す前記光強度画像を決定する画像生成器と、
を有し、
前記画像生成器は、
前記光強度画像における対応する第1画像位置および前記第1メッシュの頂点のための第1画像位置奥行き値を決定するための第1ビュー変換器と、
前記光強度画像における対応する第2画像位置および前記第2メッシュの頂点のための第2画像位置奥行き値を決定するための第2ビュー変換器と、
前記光強度画像における複数の位置のうちの各第1位置のために、前記第1テクスチャマップおよび前記第1画像位置に基づく第1光強度値、ならびに、前記第1画像位置および前記第1画像位置奥行き値に基づく第1奥行き値を決定するための第1シェーダと、
前記光強度画像における複数の位置のうちの各第1位置のために、前記第2テクスチャマップおよび前記第2画像位置に基づく第2光強度値、ならびに、前記第2画像位置および前記第2画像位置奥行き値に基づく第2奥行き値を決定するための第2シェーダと、
前記光強度画像における複数の位置のうちの各第1位置のために、少なくとも前記第1光強度値と前記第2光強度値との重み付け組み合わせとして当該第1位置における前記光強度画像のための光強度値を決定する結合器と、
を有し、第1奥行き値が増加するほど、前記第2光強度値に対する前記第1光強度値の重みが増加し、前記重みが、前記第1奥行き値と前記第2奥行き値との間の差分測度の大きさに依存する、装置。 An apparatus for generating a light intensity image, comprising:
a first receiver that receives a first texture map and a first mesh representing a scene from a first viewpoint;
a second receiver for receiving a second texture map and a second mesh representing the scene from a second viewpoint;
an image generator for determining the light intensity image representing the scene from the third viewpoint according to the first texture map, the first mesh, the second texture map and the second mesh;
has
The image generator comprises:
a first view transformer for determining first image position depth values for corresponding first image positions in said light intensity image and vertices of said first mesh;
a second view transformer for determining second image location depth values for corresponding second image locations in said light intensity image and vertices of said second mesh;
a first light intensity value based on the first texture map and the first image position, and the first image position and the first image, for each first position of a plurality of positions in the light intensity image; a first shader for determining a first depth value based on the position depth value;
a second light intensity value based on the second texture map and the second image position, and the second image position and the second image, for each first position of a plurality of positions in the light intensity image; a second shader for determining a second depth value based on the positional depth value;
for each first position of a plurality of positions in the light intensity image as a weighted combination of at least the first light intensity value and the second light intensity value for the light intensity image at that first position; a coupler for determining light intensity values;
and the weight of the first light intensity value with respect to the second light intensity value increases as the first depth value increases, and the weight is between the first depth value and the second depth value A device that depends on the magnitude of the difference measure of .
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP19157331.0 | 2019-02-15 | ||
EP19157331.0A EP3696773A1 (en) | 2019-02-15 | 2019-02-15 | Apparatus and method for generating a light intensity image |
PCT/EP2020/053281 WO2020165076A1 (en) | 2019-02-15 | 2020-02-10 | Apparatus and method for generating a light intensity image |
Publications (3)
Publication Number | Publication Date |
---|---|
JP2022520525A JP2022520525A (en) | 2022-03-31 |
JPWO2020165076A5 true JPWO2020165076A5 (en) | 2023-01-30 |
JP7460641B2 JP7460641B2 (en) | 2024-04-02 |
Family
ID=65443692
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
JP2021541447A Active JP7460641B2 (en) | 2019-02-15 | 2020-02-10 | Apparatus and method for generating a light intensity image - Patents.com |
Country Status (9)
Country | Link |
---|---|
US (1) | US11783527B2 (en) |
EP (2) | EP3696773A1 (en) |
JP (1) | JP7460641B2 (en) |
KR (1) | KR20210129685A (en) |
CN (1) | CN113424231A (en) |
BR (1) | BR112021015772A2 (en) |
CA (1) | CA3129658A1 (en) |
TW (1) | TWI831919B (en) |
WO (1) | WO2020165076A1 (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3792876A1 (en) * | 2019-09-13 | 2021-03-17 | FRAUNHOFER-GESELLSCHAFT zur Förderung der angewandten Forschung e.V. | Apparatus, method and computer program for rendering a visual scene |
US11928787B2 (en) * | 2020-07-29 | 2024-03-12 | Intel Corporation | Deep novel view synthesis from unstructured input |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7068272B1 (en) * | 2000-05-31 | 2006-06-27 | Nvidia Corporation | System, method and article of manufacture for Z-value and stencil culling prior to rendering in a computer graphics processing pipeline |
JP4187188B2 (en) * | 2002-04-26 | 2008-11-26 | 株式会社バンダイナムコゲームス | Image generation system, program, and information storage medium |
US7142209B2 (en) | 2004-08-03 | 2006-11-28 | Microsoft Corporation | Real-time rendering system and process for interactive viewpoint video that was generated using overlapping images of a scene captured from viewpoints forming a grid |
WO2009013682A2 (en) * | 2007-07-26 | 2009-01-29 | Koninklijke Philips Electronics N.V. | Method and apparatus for depth-related information propagation |
US9648346B2 (en) * | 2009-06-25 | 2017-05-09 | Microsoft Technology Licensing, Llc | Multi-view video compression and streaming based on viewpoints of remote viewer |
JP2011028428A (en) * | 2009-07-23 | 2011-02-10 | Kddi Corp | Method and apparatus for high-speed interpolation of ray information |
US8698832B1 (en) * | 2010-05-26 | 2014-04-15 | Disney Enterprises, Inc. | Perceptual detail and acutance enhancement for digital images |
US8847968B2 (en) * | 2011-07-12 | 2014-09-30 | Qualcomm Incorporated | Displaying static images |
WO2013133648A1 (en) * | 2012-03-07 | 2013-09-12 | 엘지전자 주식회사 | Method and device for processing video signal |
AU2014218390B2 (en) * | 2014-08-27 | 2017-01-12 | Canon Kabushiki Kaisha | Method, system and apparatus for forming a high resolution depth map |
US9613587B2 (en) * | 2015-01-20 | 2017-04-04 | Snaptrack, Inc. | Apparatus and method for adaptive image rendering based on ambient light levels |
US10325402B1 (en) * | 2015-07-17 | 2019-06-18 | A9.Com, Inc. | View-dependent texture blending in 3-D rendering |
EP3552183B1 (en) | 2016-12-06 | 2021-09-15 | Koninklijke Philips N.V. | Apparatus and method for generating a light intensity image |
EP3358844A1 (en) * | 2017-02-07 | 2018-08-08 | Koninklijke Philips N.V. | Method and apparatus for processing an image property map |
US10482618B2 (en) * | 2017-08-21 | 2019-11-19 | Fotonation Limited | Systems and methods for hybrid depth regularization |
US11528461B2 (en) * | 2018-11-16 | 2022-12-13 | Electronics And Telecommunications Research Institute | Method and apparatus for generating virtual viewpoint image |
-
2019
- 2019-02-15 EP EP19157331.0A patent/EP3696773A1/en not_active Withdrawn
-
2020
- 2020-02-10 EP EP20703050.3A patent/EP3924941A1/en active Pending
- 2020-02-10 US US17/430,345 patent/US11783527B2/en active Active
- 2020-02-10 JP JP2021541447A patent/JP7460641B2/en active Active
- 2020-02-10 KR KR1020217029614A patent/KR20210129685A/en unknown
- 2020-02-10 CA CA3129658A patent/CA3129658A1/en active Pending
- 2020-02-10 BR BR112021015772-7A patent/BR112021015772A2/en unknown
- 2020-02-10 CN CN202080014424.7A patent/CN113424231A/en active Pending
- 2020-02-10 WO PCT/EP2020/053281 patent/WO2020165076A1/en unknown
- 2020-02-14 TW TW109104730A patent/TWI831919B/en active
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9916684B2 (en) | Generating portable three-dimensional print-preview renderings of three-dimensional models | |
JP7045721B2 (en) | Identification device, identification method, and program to identify the position of the wall surface in the tunnel shown in the photographic image. | |
US10593096B2 (en) | Graphics processing employing cube map texturing | |
KR100920225B1 (en) | Method and apparatus for accuracy measuring of?3d graphical model by using image | |
US9892550B2 (en) | Photorealistic rendering of scenes with dynamic content | |
JP2006190308A (en) | Depth image-based modeling method and apparatus | |
US20130063472A1 (en) | Customized image filters | |
JP2019536174A (en) | Apparatus and method for generating light intensity images | |
KR20120064641A (en) | Image processing apparatus, apparatus and method for lighting processing | |
CN114119853A (en) | Image rendering method, device, equipment and medium | |
TWI439961B (en) | Conversion algorithm for voids generated after converting 2D images | |
CN112365572B (en) | Rendering method based on surface subdivision and related products thereof | |
KR101585998B1 (en) | Image processing apparatus and method | |
JPWO2020165076A5 (en) | ||
CN112509108A (en) | GPU-based vertex ambient light shading generation method and image rendering method | |
JP5432571B2 (en) | Image processing apparatus and method | |
Marrugo et al. | Fourier transform profilometry in LabVIEW | |
CN110378948B (en) | 3D model reconstruction method and device and electronic equipment | |
JP6393153B2 (en) | Program, recording medium, luminance calculation device, and luminance calculation method | |
KR101227155B1 (en) | Graphic image processing apparatus and method for realtime transforming low resolution image into high resolution image | |
CN110458935A (en) | A kind of section Real-time Generation of 3D model | |
JP2009244229A (en) | Three-dimensional image processing method, three-dimensional image processing device, and three-dimensional image processing program | |
KR20190067070A (en) | System and method for projection mapping in virtual space | |
Thorne | Healing Cracks in Cyberspace: towards best practice | |
RU2021127206A (en) | Device and method for forming a light-intensive image |