WO2013023325A1 - 基于图像运动信息的2d转3d方法 - Google Patents
基于图像运动信息的2d转3d方法 Download PDFInfo
- Publication number
- WO2013023325A1 WO2013023325A1 PCT/CN2011/001377 CN2011001377W WO2013023325A1 WO 2013023325 A1 WO2013023325 A1 WO 2013023325A1 CN 2011001377 W CN2011001377 W CN 2011001377W WO 2013023325 A1 WO2013023325 A1 WO 2013023325A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- depth
- value
- pixel
- input
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 38
- 239000013598 vector Substances 0.000 claims description 10
- 239000010432 diamond Substances 0.000 claims description 5
- 238000009825 accumulation Methods 0.000 claims description 4
- 238000010845 search algorithm Methods 0.000 claims description 4
- 229910003460 diamond Inorganic materials 0.000 claims description 3
- 230000000694 effects Effects 0.000 abstract description 2
- 230000000007 visual effect Effects 0.000 description 6
- 238000006243 chemical reaction Methods 0.000 description 3
- 235000002566 Capsicum Nutrition 0.000 description 1
- 239000006002 Pepper Substances 0.000 description 1
- 235000016761 Piper aduncum Nutrition 0.000 description 1
- 235000017804 Piper guineense Nutrition 0.000 description 1
- 244000203593 Piper nigrum Species 0.000 description 1
- 235000008184 Piper nigrum Nutrition 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000001737 promoting effect Effects 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 150000003839 salts Chemical class 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000016776 visual perception Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/579—Depth or shape recovery from multiple images from motion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/223—Analysis of motion using block-matching
- G06T7/238—Analysis of motion using block-matching using non-full search, e.g. three-step search
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/261—Image signal generators with monoscopic-to-stereoscopic image conversion
- H04N13/264—Image signal generators with monoscopic-to-stereoscopic image conversion using the relative movement of objects in two video frames or fields
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20021—Dividing image into blocks, subimages or windows
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2213/00—Details of stereoscopic systems
- H04N2213/003—Aspects relating to the "2D+depth" image format
Definitions
- the present invention relates to the field of 2D to 3D technology, and in particular, to a 2D to 3D method based on image motion information.
- 3D TV Three Dimensions
- 3D applications are becoming more and more popular in people's lives, but despite the constant 3D filming, 3D sources are still unable to meet the current market needs.
- Automatic conversion of 2D (Dimensions, 2D) sources to 3D becomes a new market need.
- the conversion between 2D and 3D is to generate a second view video based on 2D view content.
- the process includes two aspects of processing: one for depth estimation to obtain depth map (image); the other is based on depth map Depth Image Based Rendering (DIBR).
- the depth map stores depth information in 8-bit grayscale values (0 grayscale represents the farthest value, and 255 grayscale represents the most recent value).
- many algorithms have appeared in the field of 2D to 3D, and the more common ones are Based on the motion estimation based 2D to 3D algorithm, the method obtains the depth map of the input image by motion estimation.
- the existing motion estimation based 2D to 3D algorithm is obtained.
- the depth map is sparse, and the object can not be distinguished by different objects, which affects the image quality obtained by DIBR. Therefore, the popularization of the method is limited.
- the technical problem to be solved by the present invention is: How to improve the quality of the image generated by the 2D to 3D method based on image motion information.
- the present invention provides a 2D to 3D method based on motion estimation, the method comprising the steps of:
- the motion estimation method is used to obtain the depth value of each pixel of the 2D image
- step S4 The left eye image and the right eye image of step S4 are combined and output to obtain a 3D image.
- step S1 further includes:
- S1.2 calculates the depth value of each pixel according to the motion vector obtained in step S1.1.
- the method of motion estimation is a diamond search algorithm.
- step S2 further includes:
- I ( x, y ) is the luminance value of the pixel at the (x, y ) position, and its value ranges from [0, 255]; SCALE is the scaling factor of the luminance value; width is the input 2D image The width value, height is the height value of the input 2D image; DEPTH SCALE is the depth value scaling factor,
- step S2.1 further includes:
- step S2.12 is performed
- D(x,y)' min(D(xl,y)'+
- SCALE 0.1.
- DEPTH-SCALE 120.
- step S3 further includes:
- x1 and xr are the positions of the corresponding input 2D image xc positions in the left eye image and the right eye image
- f is the focal length of the eye
- tx is the distance between the eyes
- Z is the distance of the pixel point from the human eye.
- Dzero is the position of the zero plane, which has a value range of [0, 255];
- Dzero 255.
- the obtained depth map is continuous and dense, and the quality of the reconstructed image and the 3D visual effect are improved.
- Figure 1/2 is a flow chart of a 2D to 3D method based on image motion information according to an embodiment of the present invention
- Figure 2/2 is a schematic diagram of a visual model of a dual camera.
- the 2D to 3D method based on image motion information includes the following steps: 51. Obtaining a depth value of each pixel of the input 2D image based on a motion estimation method;
- step S4 The left eye image and the right eye image of step S4 are combined and output to obtain a 3D image.
- step S1 further includes:
- S1.1 Calculates the motion vector of each pixel based on the motion estimation method.
- the motion estimation method adopts the diamond search algorithm, first searches for large diamonds, then searches for small diamonds, and finally moves to integer pixel precision.
- Vector of course, other search algorithms are equally applicable here, not as a limitation on the method of the invention;
- S1.2 calculates the depth value of each pixel according to the motion vector obtained in step S1.1.
- y is the row of the pixel
- X is the column of the pixel
- D ( x, y ) is the depth value of the pixel at the unknown (x, y )
- MV X and MV y are the horizontal and vertical movements of the pixel, respectively.
- the input 2D image can be input before the motion search in step S1.1 is performed.
- Performing a denoising process which is well known to those skilled in the art, does not perform motion vector discontinuity due to motion search. If the directly calculated depth map is sparse, the actual depth map should be dense. Therefore, the present invention accumulates the depth values calculated by the motion vectors based on the luminance information of each pixel.
- step S2 further includes:
- D(x,y)' min(D(xl,y) , +
- D(x,y)' min(D(xl,y)'+
- step S2.14 returns to step S2.ll if y ⁇ height, otherwise, outputs D(x, y) ' obtained in step S2.12 or S2.13.
- height is the height value of the input 2D image;
- the horizontal direction should keep the continuity of the depth value as much as possible, and avoid the excessive noise caused by the motion search. Therefore, the present invention does not use the horizontal gradient value for the scaling motion to obtain the depth value.
- step S3 left The eye image is taken as an example, that is, the left eye image is reconstructed based on DIBR according to the depth map obtained in step S2 in step S3.
- Equation (9) and (10) are the geometric relationships of the same pixel points in the corresponding Cl, Cr, and Cc in Fig. 2.
- the xl or xr values of the corresponding input 2D image xc are calculated according to formulas (9) and (10). Then, the pixel value at (xc, y) is copied to the corresponding (xl, y) or (xr, y) (copy to (xl, y) in this embodiment).
- step S3 further includes:
- x1 and xr are the positions of the corresponding input 2D image xc positions in the left eye image and the right eye image
- f is the focal length of the eye
- tx is the distance between the eyes
- Z is the distance of the pixel point from the human eye.
- Dzero is the position of the zero plane, which has a value range of [0, 255];
- the horizontal direction of the input 2D image is first scaled to improve the pixel precision during projection.
- the image is stretched four times in the horizontal direction, and the X value of the 1/4 pixel accuracy corresponding to x1 per line is calculated based on the above human visual relationship. If the value of X corresponding to xl exceeds the image range, the pixel value of the xl position is obtained according to the interpolation; if multiple xl correspond to the same x, then D(x, y) is taken as the largest xl, and the other xl position values are passed. Interpolation is obtained; if X corresponding to xl is unique, the pixel value at position xl is the pixel value of the X position of the input 2D image.
- the image reconstructed image obtained by the 2D-to-3D method based on the image motion information of the present invention has high quality and good 3D visual effect, and is important for promoting the automatic conversion of the 2D source to the 3D market.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Signal Processing (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
Claims
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP11870997.1A EP2629531A4 (en) | 2011-08-18 | 2011-08-18 | METHOD FOR CONVERTING 2D IMAGES IN 3D BASED ON MOTION IMAGE INFORMATION |
CN201180028889.9A CN103053165B (zh) | 2011-08-18 | 2011-08-18 | 基于图像运动信息的2d转3d方法 |
PCT/CN2011/001377 WO2013023325A1 (zh) | 2011-08-18 | 2011-08-18 | 基于图像运动信息的2d转3d方法 |
US13/818,101 US20130235155A1 (en) | 2011-08-18 | 2011-08-18 | Method of converting 2d into 3d based on image motion information |
JP2013540213A JP2014504468A (ja) | 2011-08-18 | 2011-08-18 | 画像運動情報に基づく2dから3dへの変換方法 |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2011/001377 WO2013023325A1 (zh) | 2011-08-18 | 2011-08-18 | 基于图像运动信息的2d转3d方法 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2013023325A1 true WO2013023325A1 (zh) | 2013-02-21 |
Family
ID=47714669
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2011/001377 WO2013023325A1 (zh) | 2011-08-18 | 2011-08-18 | 基于图像运动信息的2d转3d方法 |
Country Status (5)
Country | Link |
---|---|
US (1) | US20130235155A1 (zh) |
EP (1) | EP2629531A4 (zh) |
JP (1) | JP2014504468A (zh) |
CN (1) | CN103053165B (zh) |
WO (1) | WO2013023325A1 (zh) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103533329A (zh) * | 2013-10-09 | 2014-01-22 | 上海大学 | 一种2d转3d的视频自动评估方法 |
CN103826032A (zh) * | 2013-11-05 | 2014-05-28 | 四川长虹电器股份有限公司 | 深度图后期处理方法 |
CN104243950A (zh) * | 2013-06-06 | 2014-12-24 | 索尼公司 | 用于将2维内容实时转换为3维内容的方法和设备 |
CN109274951A (zh) * | 2017-07-13 | 2019-01-25 | 富泰华工业(深圳)有限公司 | 深度计算方法及其装置 |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9483836B2 (en) * | 2011-02-28 | 2016-11-01 | Sony Corporation | Method and apparatus for real-time conversion of 2-dimensional content to 3-dimensional content |
KR20130033125A (ko) * | 2011-09-26 | 2013-04-03 | 삼성전자주식회사 | 컨텐츠변환장치 및 컨텐츠변환방법 |
CN104113745A (zh) | 2013-04-17 | 2014-10-22 | 咏传电子科技(上海)有限公司 | 显示装置及其影像显示方法 |
CN105989326B (zh) * | 2015-01-29 | 2020-03-03 | 北京三星通信技术研究有限公司 | 人眼三维位置信息的确定方法和装置 |
CN111369612B (zh) * | 2018-12-25 | 2023-11-24 | 北京欣奕华科技有限公司 | 一种三维点云图像生成方法及设备 |
TWI784428B (zh) * | 2021-03-03 | 2022-11-21 | 宏碁股份有限公司 | 立體影像產生方法與使用該方法的電子裝置 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001103513A (ja) * | 1999-09-27 | 2001-04-13 | Sanyo Electric Co Ltd | 2次元映像を3次元映像に変換する方法 |
CN101271578A (zh) * | 2008-04-10 | 2008-09-24 | 清华大学 | 一种平面视频转立体视频技术中的深度序列生成方法 |
CN101631256A (zh) * | 2009-08-13 | 2010-01-20 | 浙江大学 | 用于三维电视***中2d视频到3d视频的转换方法 |
EP2268047A2 (en) * | 2009-06-16 | 2010-12-29 | Samsung Electronics Co., Ltd. | Conversion device and method converting a two dimensional image to a three dimensional image |
CN102075780A (zh) * | 2011-02-25 | 2011-05-25 | 福建华映显示科技有限公司 | 立体影像产生装置及其方法 |
US20110141237A1 (en) * | 2009-12-15 | 2011-06-16 | Himax Technologies Limited | Depth map generation for a video conversion system |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH08331607A (ja) * | 1995-03-29 | 1996-12-13 | Sanyo Electric Co Ltd | 三次元表示画像生成方法 |
JP2001016609A (ja) * | 1999-06-05 | 2001-01-19 | Soft Foo Deii:Kk | Mpegデータを用いた立体映像生成装置及び方法 |
EP1719079B1 (en) * | 2004-02-17 | 2008-01-02 | Koninklijke Philips Electronics N.V. | Creating a depth map |
JP2007300169A (ja) * | 2006-04-27 | 2007-11-15 | Toshiba Corp | 動きベクトル検出器 |
US9418433B2 (en) * | 2007-07-03 | 2016-08-16 | Koninklijke Philips N.V. | Computing a depth map |
KR20100002032A (ko) * | 2008-06-24 | 2010-01-06 | 삼성전자주식회사 | 영상 생성 방법, 영상 처리 방법, 및 그 장치 |
KR101468267B1 (ko) * | 2008-10-02 | 2014-12-15 | 프라운호퍼-게젤샤프트 추르 푀르데룽 데어 안제반텐 포르슝 에 파우 | 중간 뷰 합성 및 멀티-뷰 데이터 신호 추출 |
KR20100040236A (ko) * | 2008-10-09 | 2010-04-19 | 삼성전자주식회사 | 시각적 관심에 기반한 2차원 영상의 3차원 영상 변환기 및 변환 방법 |
JP5428454B2 (ja) * | 2009-03-30 | 2014-02-26 | 凸版印刷株式会社 | 画像生成方法 |
TWI564843B (zh) * | 2009-07-01 | 2017-01-01 | 晨星半導體股份有限公司 | 移動估測方法及移動估測裝置 |
-
2011
- 2011-08-18 US US13/818,101 patent/US20130235155A1/en not_active Abandoned
- 2011-08-18 WO PCT/CN2011/001377 patent/WO2013023325A1/zh active Application Filing
- 2011-08-18 JP JP2013540213A patent/JP2014504468A/ja active Pending
- 2011-08-18 EP EP11870997.1A patent/EP2629531A4/en not_active Withdrawn
- 2011-08-18 CN CN201180028889.9A patent/CN103053165B/zh active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001103513A (ja) * | 1999-09-27 | 2001-04-13 | Sanyo Electric Co Ltd | 2次元映像を3次元映像に変換する方法 |
CN101271578A (zh) * | 2008-04-10 | 2008-09-24 | 清华大学 | 一种平面视频转立体视频技术中的深度序列生成方法 |
EP2268047A2 (en) * | 2009-06-16 | 2010-12-29 | Samsung Electronics Co., Ltd. | Conversion device and method converting a two dimensional image to a three dimensional image |
CN101631256A (zh) * | 2009-08-13 | 2010-01-20 | 浙江大学 | 用于三维电视***中2d视频到3d视频的转换方法 |
US20110141237A1 (en) * | 2009-12-15 | 2011-06-16 | Himax Technologies Limited | Depth map generation for a video conversion system |
CN102075780A (zh) * | 2011-02-25 | 2011-05-25 | 福建华映显示科技有限公司 | 立体影像产生装置及其方法 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2629531A4 * |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104243950A (zh) * | 2013-06-06 | 2014-12-24 | 索尼公司 | 用于将2维内容实时转换为3维内容的方法和设备 |
CN104243950B (zh) * | 2013-06-06 | 2016-08-24 | 索尼公司 | 用于将2维内容实时转换为3维内容的方法和设备 |
CN103533329A (zh) * | 2013-10-09 | 2014-01-22 | 上海大学 | 一种2d转3d的视频自动评估方法 |
CN103826032A (zh) * | 2013-11-05 | 2014-05-28 | 四川长虹电器股份有限公司 | 深度图后期处理方法 |
CN103826032B (zh) * | 2013-11-05 | 2017-03-15 | 四川长虹电器股份有限公司 | 深度图后期处理方法 |
CN109274951A (zh) * | 2017-07-13 | 2019-01-25 | 富泰华工业(深圳)有限公司 | 深度计算方法及其装置 |
CN109274951B (zh) * | 2017-07-13 | 2020-11-10 | 富泰华工业(深圳)有限公司 | 深度计算方法及其装置 |
Also Published As
Publication number | Publication date |
---|---|
EP2629531A1 (en) | 2013-08-21 |
US20130235155A1 (en) | 2013-09-12 |
JP2014504468A (ja) | 2014-02-20 |
CN103053165B (zh) | 2015-02-11 |
CN103053165A (zh) | 2013-04-17 |
EP2629531A4 (en) | 2015-01-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2013023325A1 (zh) | 基于图像运动信息的2d转3d方法 | |
Martin-Brualla et al. | Lookingood: Enhancing performance capture with real-time neural re-rendering | |
Liu et al. | Joint geodesic upsampling of depth images | |
CN111194550B (zh) | 处理3d视频内容 | |
CN111105432B (zh) | 基于深度学习的无监督端到端的驾驶环境感知方法 | |
US20110304708A1 (en) | System and method of generating stereo-view and multi-view images for rendering perception of depth of stereoscopic image | |
CN109963048B (zh) | 降噪方法、降噪装置及降噪电路*** | |
JP2021535466A (ja) | シーンの色及び深度の情報を再構成するための方法及びシステム | |
US9973694B1 (en) | Image stitching to form a three dimensional panoramic image | |
WO2011138472A1 (es) | Método de generación de mapas de profundidad para conversión de imágenes animadas 2d en 3d | |
Berdnikov et al. | Real-time depth map occlusion filling and scene background restoration for projected-pattern-based depth cameras | |
WO2018032841A1 (zh) | 绘制三维图像的方法及其设备、*** | |
JP5893166B2 (ja) | 3dモデル・モーフィングのための方法および装置 | |
Yang et al. | Global auto-regressive depth recovery via iterative non-local filtering | |
Yuan et al. | Temporal upsampling of depth maps using a hybrid camera | |
TWI608447B (zh) | 立體影像深度圖產生裝置及方法 | |
WO2016078162A1 (zh) | 一种基于hevc的3d视频快速编码方法 | |
CN102819662B (zh) | 一种视频流体高度的计算方法 | |
CN107610070B (zh) | 基于三摄像采集的自由立体匹配方法 | |
Bae et al. | 5d light field synthesis from a monocular video | |
CN103198464B (zh) | 一种基于单个参考视频的人脸视频光影迁移生成方法 | |
Kowalczuk et al. | Real-time temporal stereo matching using iterative adaptive support weights | |
TWI439119B (zh) | 視訊影像的深度資訊處理系統與其相關處理方法 | |
US9336596B2 (en) | System and methods to efficiently measure stereoscopic disparity and to generate related mismatch images | |
EP2966867A1 (en) | Methods and devices for encoding and decoding a sequence of frames representing a 3D scene, and corresponding computer program products and computer-readable medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201180028889.9 Country of ref document: CN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13818101 Country of ref document: US |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11870997 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2011870997 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2013540213 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |