CN102333230A - Method for improving quality of synthetized virtual views in three-dimensional video system - Google Patents

Method for improving quality of synthetized virtual views in three-dimensional video system Download PDF

Info

Publication number
CN102333230A
CN102333230A CN201110281837A CN201110281837A CN102333230A CN 102333230 A CN102333230 A CN 102333230A CN 201110281837 A CN201110281837 A CN 201110281837A CN 201110281837 A CN201110281837 A CN 201110281837A CN 102333230 A CN102333230 A CN 102333230A
Authority
CN
China
Prior art keywords
video
rec
virtual view
video system
texture
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201110281837A
Other languages
Chinese (zh)
Inventor
元辉
刘琚
孙建德
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shandong University
Original Assignee
Shandong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shandong University filed Critical Shandong University
Priority to CN201110281837A priority Critical patent/CN102333230A/en
Publication of CN102333230A publication Critical patent/CN102333230A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
  • Image Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The invention discloses a method for improving the quality of synthetized virtual views in a three-dimensional video system. In the method, uncompressed video and deeply compressed video and deeply synthetized virtual views are respectively adopted at server sides of the three-dimensional video system, and a wiener filter coefficient is obtained through computation and is sent to a user terminal of the three-dimensional video system; at the user terminal of the three-dimensional video system, wiener filtering is carried out on the synthetized virtual views by the wiener filter coefficient; and further the quality of the synthetized virtual views is improved. According to the invention, the influence of compression distortion of video information and deep information in the three-dimensional video system on the quality of the synthetized virtual views is considered, so the quality of the synthetized virtual views is improved, the frame of the three-dimensional video system is perfected, and the method can be applied to the application of three-position videos.

Description

Synthetic virtual view method for quality in a kind of raising 3 D video system
Technical field
The present invention relates to synthetic virtual view method for quality in a kind of raising 3 D video system, belong to the virtual view synthesis technical field in the 3 D stereo video coding standard.
Background technology
The 3 D stereo video is meant that as main video applications in future the user can enjoy real 3 D stereo video content through 3 D stereo video display device.The correlation technique of 3 D video, such as, the technology such as demonstration of the collection of 3 D stereo video, 3 D stereo video coding, 3 D stereo video are paid close attention to widely.In order to promote the standardization of 3 D stereo video technology, 2002, (Motion Picture Experts Group MPEG) proposed any viewpoint TV (Free View Television, notion FTV) in Motion Picture Experts Group.It can provide vividly real, interactively 3 D stereo audiovisual system.The user can watch the 3 D stereo video of this angle from different angles, makes the user have and incorporates the sense of reality in the video scene.FTV can be widely used in fields such as broadcast communication, amusement, education, medical treatment and video monitoring.In order to make the user can watch 3 D stereo video at any angle, FTV system service end uses the video camera array of having demarcated to obtain the video on certain viewpoint.And, utilize corrected video information to generate the virtual view of virtual view through the virtual view synthetic technology to the video correction on the different points of view.MPEG suggestion is at present specifically used based on the degree of depth-image (Depth-Image Based Rendering, virtual view synthetic technology DIBR).Depth information is generally represented through depth map.The main process that virtual view synthesizes is following:
1). confirm to want the relative position of virtual view in video camera array.
2). confirm to be used for the texture video of synthetic virtual view.
3). confirm step 2) the corresponding depth map of texture video.
4). according to step 2) with 3) and in texture video and depth map, adopt the DIBR technology, synthetic virtual view.
The standardization effort of FTV is divided into two stages to carry out.Phase I is 2006 to 2008 the expansion scheme-MVC H.264/AVC (Multi-View Video Coding) that formulated by JVT (Joint Video Team, joint video code sets).MVC can encode to many viewpoints texture video.But to finally realize the function of FTV system, also must encode depth information.The standardization formulation work of FTV has at present got into second stage, i.e. 3DVC (Three Dimensional Video Coding).3DVC mainly pays close attention to the expression and the coding of depth information, and the combined coding of texture video and depth information.Among the 3DVC, depth information is represented through depth map.
The leading indicator of weighing the 3DVC performance is the quality of synthetic virtual view, and the encoder bit rate of texture video, depth map.The quality of virtual view:
Usually (Peak Signal-to-Noise Ratio PSNR) weighs the quality of video to adopt Y-PSNR.The computing formula of PSNR is shown below,
PSNR = 10 × log ( 255 2 MSE )
MSE representes the mean square error between original view and the synthetic virtual view in the following formula, is used for weighing the distortion of virtual view, and the coding distortion of the coding distortion of texture video, depth map.
In practical application, the view of virtual view is non-existent, does not also promptly have original view.But,, at first adopt the existing texture video of un-encoded and the corresponding synthetic virtual view V of depth map thereof therefore for weighing the performance of 3DVC because 3DVC mainly pays close attention to coding efficiency Orig, the depth map that adopts the texture video of the reconstruction after process is encoded and the back of encoding to rebuild then synthesizes virtual view V Rec, at last through calculating V RecWith V OrigBetween MSE, and then obtain PSNR, to weigh the performance of 3DVC.
As shown in Figure 1, in the 3D video system, the virtual view that the user watches is that video, the depth information through decoding and rebuilding is synthetic through the DIBR technology.As shown in Figure 2, the pixel m in the left side camera can project on the object point M in the 3D world coordinate system through following formula,
(X,Y,Z) T=RgA -1(x,y,1) TgZ+t,
In the following formula, (X .Y, Z) TBe the coordinate of M, Z is the depth value of object point M, (x, y, 1) TBe the homogeneous coordinates of pixel m, use usually
Figure BDA0000093161410000021
Expression, A and R are the inside and outside parameter matrixs of video camera, t=(t x, t y, t z) TBe the coordinate of camera lens center in the 3D world coordinate system.Object point M in the 3D world coordinate system can project on the pixel m ' in the camera plane of right side through following formula,
sg(x′,y′,1) T=A′gR′ -1g[(X,Y,Z) T-t′],
Wherein, (x ', y ', 1) TBe the homogeneous coordinates of m ', use usually
Figure BDA0000093161410000022
Expression, s is a scalar factor, A ', R ', t ' they are respectively the coordinate of optical center in the 3D world coordinate system of inner parameter matrix, external parameter matrix and the right side video camera of right side video camera.3D camera acquisition system for calibration accuracy; A ' is identical with A; R ' is identical with R; Only there is the difference on the horizontal coordinate between t ' and the t; Thus, the relation between and
Figure BDA0000093161410000024
can be represented with following formula
Figure BDA0000093161410000025
That is,
s·(x′,y′,1) T=Z·(x,y,1) T+A·R -1·(t x-t x′,t y-t y′,t z-t z′) T.
Thus, the pixel m in the left side camera just can be corresponding with the right side video camera in pixel m ', the basic principle of DIBR technology that Here it is.
In addition, for fear of block the cavitation that phenomenon produces in virtual view, the synthetic virtual view of the video content of two video cameras was as shown in Figure 3 about existing virtual view synthetic technology often adopted.Among Fig. 3, ω LAnd ω RBe two weight coefficients, satisfy ω R+ ω L=1.
Existing virtual view synthetic technology has been considered and has been blocked linearity, improved the quality of synthetic virtual view.But the 3D video system by shown in Figure 1 can find out that the coding distortion of video, the degree of depth also has very big influence to the quality of synthetic virtual view.Existing virtual view composition algorithm is not considered the influence of the coding distortion of video, the degree of depth to synthetic virtual view quality.
Summary of the invention
Consider not that to what existing virtual view synthetic method existed coding distortion is to synthesizing the problem of virtual view quality influence; The present invention is according to the influence to synthetic virtual view quality of the compression artefacts of video, the degree of depth, proposes synthetic virtual view method for quality in a kind of raising 3 D video system.
Synthetic virtual view method for quality is in the 3 D video system in the raising 3 D video of the present invention system, adopts Weiner filter that synthetic virtual view is carried out filtering, improves the quality of synthetic virtual view, specifically may further comprise the steps:
(1) adopt depth map (Depth) and texture video (Texture) to synthesize virtual view V without overcompression Orig
(2) at the coding side of 3 D video system depth map and texture video are encoded, and obtain the depth map Depth_rec of local decode reconstruction and the texture video Texture_rec of reconstruction;
(3) according to Depth_rec and the synthetic virtual view V of Texture_rec Rec
(4) with V OrigAs primary signal, V RecAs the signal behind the primary signal interpolation noise, and calculate V RecThe auto-correlation function matrix and the V of signal OrigAnd V RecBetween the cross-correlation function matrix, obtain wiener filter coefficients through finding the solution Wei Na-Hough Equation for Calculating;
(5) coefficient with Weiner filter is sent to the terminal of 3 D video system, the terminal that depth map after will encoding simultaneously and texture video send to the 3 D video system through encoding;
(6), at first depth map and texture video are decoded, and adopt based on the synthetic virtual view V of the virtual view synthetic technology of the degree of depth-image at the terminal of 3 D video system Rec_terminal, and wiener filter coefficients decoded;
(7) utilize wiener filter coefficients that decoding obtains to V Rec_terminalCarry out filtering, obtain final virtual view.
The present invention has considered the influence of the compression artefacts of video, the degree of depth to synthetic virtual view quality, adopts Weiner filter to improve the quality of the synthetic virtual view of 3 D video system terminal.
Description of drawings
Fig. 1 is the system block diagram of 3 D video;
Fig. 2 is a DIBR technology sketch map;
Fig. 3 is existing virtual view synthetic technology sketch map;
Fig. 4 adopts 3 D video system block diagram of the present invention.
Fig. 5 is the experimental result sketch map to 3 D video sequence B ookarrival.
Fig. 6 is the experimental result sketch map to 3 D video sequence dancer.
Fig. 7 is the experimental result sketch map to 3 D video sequence Lovebird1.
Embodiment
As shown in Figure 4, synthetic virtual view method for quality specifically comprises the steps: in the raising 3 D video of the present invention system
(1) adopt depth map (Depth) and texture video (Texture) to synthesize virtual view (V without overcompression Orig);
(2) at the coding side of 3 D video system depth map and texture video are encoded, and obtain the depth map Depth_rec of local decode reconstruction and the texture video Texture_rec of reconstruction;
(3) according to Depth_rec and Texture_rec at synthetic virtual view V Rec
(4) with V OrigAs primary signal, V RecAs the signal behind the primary signal interpolation noise, and calculate V RecThe auto-correlation function matrix and the V of signal OrigAnd V RecBetween the cross-correlation function matrix, obtain wiener filter coefficients through finding the solution following Wei Na-Hough Equation for Calculating,
Figure BDA0000093161410000041
Wherein,
Figure BDA0000093161410000042
Be V RecThe auto-correlation function matrix,
Figure BDA0000093161410000043
Be V OrigAnd V RecBetween the cross-correlation function matrix, C is the wiener filter coefficients matrix;
(5) coefficient with Weiner filter is sent to the terminal of 3 D video system, the terminal that depth map after will encoding simultaneously and texture video send to the 3 D video system through encoding;
(6), at first depth map and texture video are decoded, and adopt based on the synthetic virtual view V of the virtual view synthetic technology of the degree of depth-image at the terminal of 3 D video system Rec_terminal, and wiener filter coefficients decoded;
(7) utilize wiener filter coefficients that decoding obtains to V Rec_terminalCarry out filtering, obtain final virtual view.
Effect of the present invention can further specify through experiment.
Experiment test under the condition of different quantization parameters coding, the quality of the synthetic virtual view of a plurality of 3 D video sequences.Fig. 5 has compared in the 3 D video system, adopts the present invention and does not adopt rate distortion curve of the present invention.Wherein Fig. 5 is the experimental result to 3 D video sequence B ookarrival, and Fig. 6 is the experimental result to 3 D video sequence dancer, and Fig. 7 is the experimental result to 3 D video sequence Lovebird1.Visible by Fig. 5, Fig. 6 and Fig. 7, behind employing the present invention, under the identical condition of 3 D video system coding code check, the objective quality of synthetic virtual view is higher, explains that the present invention has improved the code efficiency of depth map.As far as 3 D video sequence B ookarrival, the objective quality of synthetic virtual view on average increases 0.531dB; As far as 3 D video sequence dancer, the objective quality of synthetic virtual view on average increases 0.217dB; As far as 3 D video sequence Lovebird1, the objective quality of synthetic virtual view on average increases 0.473dB.

Claims (1)

1. one kind is improved synthetic virtual view method for quality in the 3 D video system, it is characterized in that:
In the 3 D video system, adopt Weiner filter that synthetic virtual view is carried out filtering, improve the quality of synthetic virtual view, specifically may further comprise the steps:
(1) adopt depth map and texture video to synthesize virtual view V without overcompression Orig
(2) at the coding side of 3 D video system depth map and texture video are encoded, and obtain the depth map Depth_rec of local decode reconstruction and the texture video Texture_rec of reconstruction;
(3) according to Depth_rec and the synthetic virtual view V of Texture_rec Rec
(4) with V OrigAs primary signal, V RecAs the signal behind the primary signal interpolation noise, and calculate V RecThe auto-correlation function matrix and the V of signal OrigAnd V RecBetween the cross-correlation function matrix, obtain wiener filter coefficients through finding the solution Wei Na-Hough Equation for Calculating;
(5) coefficient with Weiner filter is sent to the terminal of 3 D video system, the terminal that depth map after will encoding simultaneously and texture video send to the 3 D video system through encoding;
(6), at first depth map and texture video are decoded, and adopt based on the synthetic virtual view V of the virtual view synthetic technology of the degree of depth-image at the terminal of 3 D video system Rec_terminal, and wiener filter coefficients decoded;
(7) utilize wiener filter coefficients that decoding obtains to V Rec_terminalCarry out filtering, obtain final virtual view.
CN201110281837A 2011-09-21 2011-09-21 Method for improving quality of synthetized virtual views in three-dimensional video system Pending CN102333230A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201110281837A CN102333230A (en) 2011-09-21 2011-09-21 Method for improving quality of synthetized virtual views in three-dimensional video system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201110281837A CN102333230A (en) 2011-09-21 2011-09-21 Method for improving quality of synthetized virtual views in three-dimensional video system

Publications (1)

Publication Number Publication Date
CN102333230A true CN102333230A (en) 2012-01-25

Family

ID=45484812

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201110281837A Pending CN102333230A (en) 2011-09-21 2011-09-21 Method for improving quality of synthetized virtual views in three-dimensional video system

Country Status (1)

Country Link
CN (1) CN102333230A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013185336A1 (en) * 2012-06-15 2013-12-19 Intel Corporation Adaptive filtering for scalable video coding
CN103873867A (en) * 2014-03-31 2014-06-18 清华大学深圳研究生院 Free viewpoint video depth map distortion prediction method and free viewpoint video depth map coding method
CN106507132A (en) * 2016-12-07 2017-03-15 清华大学深圳研究生院 A kind of cloud discharging method synthesized based on mobile terminal virtual view and system
CN108769646A (en) * 2018-05-06 2018-11-06 Oppo广东移动通信有限公司 Three-dimensional video communication method and system, electronic device, server and readable storage medium storing program for executing
CN111405264A (en) * 2020-01-20 2020-07-10 杭州电子科技大学 3D video comfort level improving method based on depth adjustment

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101937578A (en) * 2010-09-08 2011-01-05 宁波大学 Method for drawing virtual view color image

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101937578A (en) * 2010-09-08 2011-01-05 宁波大学 Method for drawing virtual view color image

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013185336A1 (en) * 2012-06-15 2013-12-19 Intel Corporation Adaptive filtering for scalable video coding
US10979703B2 (en) 2012-06-15 2021-04-13 Intel Corporation Adaptive filtering for scalable video coding
CN103873867A (en) * 2014-03-31 2014-06-18 清华大学深圳研究生院 Free viewpoint video depth map distortion prediction method and free viewpoint video depth map coding method
CN103873867B (en) * 2014-03-31 2017-01-25 清华大学深圳研究生院 Free viewpoint video depth map distortion prediction method and free viewpoint video depth map coding method
CN106507132A (en) * 2016-12-07 2017-03-15 清华大学深圳研究生院 A kind of cloud discharging method synthesized based on mobile terminal virtual view and system
CN106507132B (en) * 2016-12-07 2019-03-08 清华大学深圳研究生院 A kind of cloud discharging method and system based on the synthesis of mobile terminal virtual view
CN108769646A (en) * 2018-05-06 2018-11-06 Oppo广东移动通信有限公司 Three-dimensional video communication method and system, electronic device, server and readable storage medium storing program for executing
CN108769646B (en) * 2018-05-06 2020-02-18 Oppo广东移动通信有限公司 Three-dimensional video communication method and system, electronic device, server and readable storage medium
US10785468B2 (en) 2018-05-06 2020-09-22 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Communication methods and systems, electronic devices, servers, and readable storage media
US11595635B2 (en) 2018-05-06 2023-02-28 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Communication methods and systems, electronic devices, servers, and readable storage media
CN111405264A (en) * 2020-01-20 2020-07-10 杭州电子科技大学 3D video comfort level improving method based on depth adjustment
CN111405264B (en) * 2020-01-20 2022-04-12 杭州电子科技大学 3D video comfort level improving method based on depth adjustment

Similar Documents

Publication Publication Date Title
US11412233B2 (en) Methods for MR-DIBR disparity map merging and disparity threshold determination
US11051039B2 (en) Methods for full parallax light field compression
Merkle et al. Multi-view video plus depth representation and coding
JP5763184B2 (en) Calculation of parallax for 3D images
WO2018050725A1 (en) A method and a device for reconstructing a point cloud representative of a scene using light-field data
CN106131531B (en) Method for processing video frequency and device
CN101835056B (en) Allocation method for optimal code rates of texture video and depth map based on models
US9635345B2 (en) Method and system for acquisition, representation, compression, and transmission of three-dimensional data
CN102685532B (en) Coding method for free view point four-dimensional space video coding system
US20080205791A1 (en) Methods and systems for use in 3d video generation, storage and compression
CN102742282B (en) It is block-based to interlock
Graziosi et al. Depth assisted compression of full parallax light fields
US10827161B2 (en) Depth codec for 3D-video recording and streaming applications
CN102333230A (en) Method for improving quality of synthetized virtual views in three-dimensional video system
US20150249839A1 (en) Picture encoding method, picture decoding method, picture encoding apparatus, picture decoding apparatus, picture encoding program, picture decoding program, and recording media
Morvan et al. System architecture for free-viewpoint video and 3D-TV
CN103561267A (en) 3D video coding transmission method based on motion information and depth information
EP3301926A1 (en) A method and a device for reconstructing a point cloud representative of a scene using light-field data
KR102505130B1 (en) A method and a device for encoding a signal representative of a light-field content
CN102158710B (en) Depth view encoding rate distortion judgment method for virtual view quality
CN102595166B (en) Lagrange factor calculation method applied for depth image encoding
CN103139554A (en) Method and device for optimizing three-dimensional video frequency distortion
CN104427323B (en) Three dimensional image processing method based on the degree of depth
EP2839437B1 (en) View synthesis using low resolution depth maps
CN116325721A (en) Method for encoding and decoding multiview video

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20120125

RJ01 Rejection of invention patent application after publication