WO2015192314A1 - A simplified method for depth based block partitioning - Google Patents

A simplified method for depth based block partitioning Download PDF

Info

Publication number
WO2015192314A1
WO2015192314A1 PCT/CN2014/080062 CN2014080062W WO2015192314A1 WO 2015192314 A1 WO2015192314 A1 WO 2015192314A1 CN 2014080062 W CN2014080062 W CN 2014080062W WO 2015192314 A1 WO2015192314 A1 WO 2015192314A1
Authority
WO
WIPO (PCT)
Prior art keywords
block
sub
kxm
size
mask
Prior art date
Application number
PCT/CN2014/080062
Other languages
French (fr)
Inventor
Xianguo Zhang
Kai Zhang
Jicheng An
Han HUANG
Original Assignee
Mediatek Singapore Pte. Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mediatek Singapore Pte. Ltd. filed Critical Mediatek Singapore Pte. Ltd.
Priority to PCT/CN2014/080062 priority Critical patent/WO2015192314A1/en
Publication of WO2015192314A1 publication Critical patent/WO2015192314A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/86Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/119Adaptive subdivision aspects, e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding

Definitions

  • the invention relates generally to Multi-view video coding and Three-Dimensional (3D) video coding.
  • the present invention relates to simplified methods for depth-based block partitioning in 3D video coding.
  • 3D video coding is developed for encoding or decoding video data of multiple views simultaneously captured by several cameras. Since all cameras capture the same scene for both the input texture videos and depth videos, the depth information can be utilized to improve the motion compensation efficiency of texture videos. Especially, the corresponding depth block of the texture block can represent the pixel level object segmentation, so it is reasonable to realize pixel-level segment based motion compensation by utilizing the depth information. Therefore, a depth-based block partitioning (DBBP) is adopted for texture video coding in the current 3D- HEVC.
  • DBBP depth-based block partitioning
  • the current depth-based block partitioning comprises steps of virtual depth derivation, block partitioning, block segmentation, predicted block combination and post-filtering.
  • Virtual depth derivation utilizes the disparity vector from neighboring blocks ( BDV) to address a virtual block of depth from the depth picture in the dependent depth view;
  • One partition among 2NxN, Nx2N, 2NxnU, 2NxnD, nLx2N and nRx2N is selected by the block partitioning to derive two predictive motion vectors (PMVs), which are utilized for the compensation procedure for the to-be-divided two segments;
  • Block segmentation process divides the current texture block into two segments according to segment mask generated from the virtual depth block;
  • the predicted block combination utilizes the PMVs and motion vector differences (MVDs) from block partitioning and the derived segment mask to fetch the predicted pixels, and then completes the combination by two-time 2Nx2N predictions with their own decoded motion vectors.
  • MDV motion vector differences
  • a mean value is firstly calculated by averaging each pixel of the virtual depth block.
  • the left-up corner pixel is larger than the mean value
  • all positions in the mask where the corresponding pixels in virtual depth are larger than the mean value are marked with ' ⁇ ', others are marked with T; Otherwise, all positions in the mask where the corresponding pixels in virtual depth block are smaller or equal than the mean value are marked with ' ⁇ ', others are marked with T.
  • the basic rule is, the 1 st 2Nx2N predicted block's pixels with "0" in the segment mask and 2 nd 2Nx2N predicted block's pixels with "1" in the segment mask are combined as the predicted pixels of the decoded CU.
  • a blending operation in the final post-filtering step is utilized to average the two predicted values at the same position in the two predicted 2Nx2N blocks.
  • the mask generation process requires pixel-level mean-value calculation and pixel-by-pixel comparison.
  • Fig. 1 is a diagram illustrating the mask based motion compensation steps of depth-based block partitioning (DBBP).
  • DBBP depth-based block partitioning
  • Fig. 2 is a diagram illustrating the modified mask based steps of depth-based block partitioning (DBBP). DETAILED DESCRIPTION
  • the mean value for the virtual depth block is equal to the average results of all pixels at (x, y) positions of all the kxm-size sub blocks, where x and y are respectively smaller than k and m.
  • the size of the segment mask is unchanged, and the samples of each kxm-size sub-block in the segment mask are the same and equal to the comparison between one specified (x,y)-position pixel in the corresponding kxm-size sub-block in the virtual depth block and the mean value.
  • An example is shown is Fig. 2, where k and m are equal to 2, x and y equal to 0.
  • the size of the segment mask is changed, and the samples at (x, y) position in the segment mask is equal to the comparison between one specified (kxx+i,mxy+j)- position pixel value in the virtual depth block and the mean value, where i and j should be respectively smaller than k and m. For example, k and m are equal to 2, i and j are equal to 0.
  • the size of the segment mask is unchanged, and the process of combing each kxm-size sub-block of two predicted 2Nx2N block is only according to the value at (x,y)-position in the corresponding kxm-size sub-block in the segment mask. For example, k and m are equal to 2, x and y are equal to 0.
  • the size of the segment mask is changed, and the process of combing each kxm-size sub-block at (x,y) position of two predicted 2Nx2N block is according to the pixel at (x/k, y/m) position of the segment mask.
  • k and m are equal to 2.
  • the process of filtering each pixel at (x,y) position of the kxm-size sub-block of combined block of the two predicted 2Nx2N block is related to the values x and y.
  • k and m are equal to 2.
  • any combinations of the first to sixth embodiments are included.
  • an embodiment of the present invention can be a circuit integrated into a video compression chip or program codes integrated into video compression software to perform the processing described herein.
  • An embodiment of the present invention may also be program codes to be executed on a Digital Signal Processor (DSP) to perform the processing described herein.
  • DSP Digital Signal Processor
  • the invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA).
  • processors can be configured to perform particular tasks according to the invention, by executing machine-readable software code or firmware code that defines the particular methods embodied by the invention.
  • the software code or firmware codes may be developed in different programming languages and different format or style.
  • the software code may also be compiled for different target platform.
  • different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Methods of depth-based block partitioning (DBBP) for multi-view video coding and 3D video coding are disclosed. Several methods are proposed to reduce the calculation or accessing of the samples in the segment mask by performing the DBBP process sub-block by sub-block.

Description

A SIMPLIFIED METHOD FOR DEPTH BASED BLOCK
PARTITIONING
FIELD OF INVENTION
The invention relates generally to Multi-view video coding and Three-Dimensional (3D) video coding. In particular, the present invention relates to simplified methods for depth-based block partitioning in 3D video coding.
BACKGROUND OF THE INVENTION
3D video coding is developed for encoding or decoding video data of multiple views simultaneously captured by several cameras. Since all cameras capture the same scene for both the input texture videos and depth videos, the depth information can be utilized to improve the motion compensation efficiency of texture videos. Especially, the corresponding depth block of the texture block can represent the pixel level object segmentation, so it is reasonable to realize pixel-level segment based motion compensation by utilizing the depth information. Therefore, a depth-based block partitioning (DBBP) is adopted for texture video coding in the current 3D- HEVC.
The current depth-based block partitioning comprises steps of virtual depth derivation, block partitioning, block segmentation, predicted block combination and post-filtering. (1) Virtual depth derivation utilizes the disparity vector from neighboring blocks ( BDV) to address a virtual block of depth from the depth picture in the dependent depth view; (2) One partition among 2NxN, Nx2N, 2NxnU, 2NxnD, nLx2N and nRx2N is selected by the block partitioning to derive two predictive motion vectors (PMVs), which are utilized for the compensation procedure for the to-be-divided two segments; (3) Block segmentation process divides the current texture block into two segments according to segment mask generated from the virtual depth block; (4) Whereas the predicted block combination utilizes the PMVs and motion vector differences (MVDs) from block partitioning and the derived segment mask to fetch the predicted pixels, and then completes the combination by two-time 2Nx2N predictions with their own decoded motion vectors. Thus the combination is completed through two 2Nx2N compensations under the control of the calculated segment mask. (5) Finally, a post filtering process is utilized to smooth the internal pixel boundaries(where neighboring mask values are different) of the combined block.
From the description above, it can be found that segment mask generation, the predicted block combination and post filtering are all based on the mask, as shown in Fig. 1. Therefore, improved mask generation is the key point for DBBP.
In the process of mask generation, a mean value is firstly calculated by averaging each pixel of the virtual depth block. In the following, if the left-up corner pixel is larger than the mean value, all positions in the mask where the corresponding pixels in virtual depth are larger than the mean value are marked with 'Ο', others are marked with T; Otherwise, all positions in the mask where the corresponding pixels in virtual depth block are smaller or equal than the mean value are marked with 'Ο', others are marked with T.
In the mask based predicted block combination, the basic rule is, the 1st 2Nx2N predicted block's pixels with "0" in the segment mask and 2nd 2Nx2N predicted block's pixels with "1" in the segment mask are combined as the predicted pixels of the decoded CU. In addition, for border positions where 1 and 0 are neighboring in the segment mask, a blending operation in the final post-filtering step is utilized to average the two predicted values at the same position in the two predicted 2Nx2N blocks.
However, complexity problems exist in the following two aspects. (1) The mask generation process requires pixel-level mean-value calculation and pixel-by-pixel comparison.
(2) The combination of the two 2Nx2N predicted block must be operated pixel by pixel.
Thereby, a simplified method with computational complexity reduction for the mask based motion prediction will be welcome for DBBP process.
SUMMARY OF THE INVENTION
In this invention, it is proposed to reduce the calculation or access of the samples in the segment mask by performing the DBBP process of mask generation, block combination and post filtering sub-block by sub-block.
Other aspects and features of the invention will become apparent to those with ordinary skill in the art upon review of the following descriptions of specific embodiments.
BRIEF DESCRIPTION OF THE DRAWINGS
The invention can be more fully understood by reading the subsequent detailed description and examples with references made to the accompanying drawings, wherein:
Fig. 1 is a diagram illustrating the mask based motion compensation steps of depth-based block partitioning (DBBP).
Fig. 2 is a diagram illustrating the modified mask based steps of depth-based block partitioning (DBBP). DETAILED DESCRIPTION
The following description is of the best-contemplated mode of carrying out the invention. This description is made for the purpose of illustrating the general principles of the invention and should not be taken in a limiting sense. The scope of the invention is best determined by reference to the appended claims.
There are many embodiments proposed in this invention to simplify depth-based block partitioning (DBBP).
In a first embodiment, the mean value for the virtual depth block is equal to the average results of all pixels at (x, y) positions of all the kxm-size sub blocks, where x and y are respectively smaller than k and m.
In a second embodiment, the size of the segment mask is unchanged, and the samples of each kxm-size sub-block in the segment mask are the same and equal to the comparison between one specified (x,y)-position pixel in the corresponding kxm-size sub-block in the virtual depth block and the mean value. An example is shown is Fig. 2, where k and m are equal to 2, x and y equal to 0.
In a third embodiment, the size of the segment mask is changed, and the samples at (x, y) position in the segment mask is equal to the comparison between one specified (kxx+i,mxy+j)- position pixel value in the virtual depth block and the mean value, where i and j should be respectively smaller than k and m. For example, k and m are equal to 2, i and j are equal to 0.
In a fourth embodiment, the size of the segment mask is unchanged, and the process of combing each kxm-size sub-block of two predicted 2Nx2N block is only according to the value at (x,y)-position in the corresponding kxm-size sub-block in the segment mask. For example, k and m are equal to 2, x and y are equal to 0.
In a fifth embodiment, the size of the segment mask is changed, and the process of combing each kxm-size sub-block at (x,y) position of two predicted 2Nx2N block is according to the pixel at (x/k, y/m) position of the segment mask. For example, k and m are equal to 2.
In a sixth embodiment, the process of filtering each pixel at (x,y) position of the kxm-size sub-block of combined block of the two predicted 2Nx2N block is related to the values x and y. For example, k and m are equal to 2.
In a seventh embodiment, any combinations of the first to sixth embodiments are included.
The proposed method described above can be used in a video encoder as well as in a video decoder. Embodiments of the method according to the present invention as described above may be implemented in various hardware, software codes, or a combination of both. For example, an embodiment of the present invention can be a circuit integrated into a video compression chip or program codes integrated into video compression software to perform the processing described herein. An embodiment of the present invention may also be program codes to be executed on a Digital Signal Processor (DSP) to perform the processing described herein. The invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA). These processors can be configured to perform particular tasks according to the invention, by executing machine-readable software code or firmware code that defines the particular methods embodied by the invention. The software code or firmware codes may be developed in different programming languages and different format or style. The software code may also be compiled for different target platform. However, different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.
The invention may be embodied in other specific forms without departing from its spirit or essential characteristics. The described examples are to be considered in all respects only as illustrative and not restrictive. To the contrary, it is intended to cover various modifications and similar arrangements (as would be apparent to those skilled in the art). Therefore, the scope of the appended claims should be accorded the broadest interpretation so as to encompass all such modifications and similar arrangements.

Claims

1. A method of depth-based block partition (DBBP) for multi-view video coding or 3D video coding comprising, a mean value calculation of a virtual depth, a mask generation, a predicted block combination or a post filtering process considers pixel values in each kxm sub- block of the virtual depth block are the same.
2. The method of as claimed in claim 1, wherein all pixels in each kxm sub-block of a current block utilize the same value in a segment mask for the predicted block combination.
3. The method of as claimed in claim 1, wherein the mean value for the virtual depth block is equal to average results of all pixels at (x, y) positions of all the kxm-size sub blocks, where x and y are respectively smaller than k and m.
4. The method of as claimed in claim 2, wherein the size of the segment mask is unchanged, and samples of each kxm-size sub-block in the segment mask are the same and equal to the comparison between one specified (x,y)-position pixel in the corresponding kxm- size sub-block in the virtual depth block and the mean value, x and y are must smaller than k and m, and typically equal to 0.
5. The method of as claimed in claim 4, wherein k and m are the same.
6. The method of as claimed in claim 2, wherein the size of the segment mask is changed, and samples at (x, y) position in the segment mask is equal to the comparison between one specified (kxx+i,mxy+j) position pixel value and the mean value, where i and j should be respectively smaller than k and m.
7. The method of as claimed in claim 2, wherein the size of the segment mask is unchanged, and the process of combing each kxm-size sub-block of two predicted 2Nx2N block is only according to the value at (x,y)-position in the corresponding kxm-size sub-block in the segment mask, wherein x and y are smaller than k and m.
8. The method of as claimed in claim 2, wherein the size of the segment mask is changed, and the process of combing each kxm-size sub-block at (x,y) position of two predicted 2Nx2N block is according to the pixel at (x/k, y/m) position of the segment mask.
9. The method of as claimed in claim 2, wherein the process of filtering each pixel at (x,y) position of the kxm-size sub-block of the combined block of the two predicted 2Nx2N block is related to the values x and y.
10. The method of as claimed in claim 1, wherein the mask generation, the predicted block combination or the post filtering process works sub-block by sub-block, wherein the sub-block sizes are the same or different with each other.
PCT/CN2014/080062 2014-06-17 2014-06-17 A simplified method for depth based block partitioning WO2015192314A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2014/080062 WO2015192314A1 (en) 2014-06-17 2014-06-17 A simplified method for depth based block partitioning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2014/080062 WO2015192314A1 (en) 2014-06-17 2014-06-17 A simplified method for depth based block partitioning

Publications (1)

Publication Number Publication Date
WO2015192314A1 true WO2015192314A1 (en) 2015-12-23

Family

ID=54934670

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2014/080062 WO2015192314A1 (en) 2014-06-17 2014-06-17 A simplified method for depth based block partitioning

Country Status (1)

Country Link
WO (1) WO2015192314A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111107368A (en) * 2018-10-26 2020-05-05 北京字节跳动网络技术有限公司 Fast method for split tree decision

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140085416A1 (en) * 2011-06-15 2014-03-27 Mediatek Inc. Method and apparatus of texture image compress in 3d video coding
WO2014053095A1 (en) * 2012-10-03 2014-04-10 Mediatek Inc. Method and apparatus for inter-component motion prediction in three-dimensional video coding
US20140098882A1 (en) * 2012-10-04 2014-04-10 Qualcomm Incorporated Inter-view predicted motion vector for 3d video

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140085416A1 (en) * 2011-06-15 2014-03-27 Mediatek Inc. Method and apparatus of texture image compress in 3d video coding
WO2014053095A1 (en) * 2012-10-03 2014-04-10 Mediatek Inc. Method and apparatus for inter-component motion prediction in three-dimensional video coding
US20140098882A1 (en) * 2012-10-04 2014-04-10 Qualcomm Incorporated Inter-view predicted motion vector for 3d video

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
FABIAN, JAGER.: "Depth-based Block Partitioning for 3D Video Coding", 2013 PICTURE CODING SYMPOSIUM, 11 December 2013 (2013-12-11), pages 410 - 413, XP032566956 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111107368A (en) * 2018-10-26 2020-05-05 北京字节跳动网络技术有限公司 Fast method for split tree decision
CN111107368B (en) * 2018-10-26 2024-05-14 北京字节跳动网络技术有限公司 Fast method for segmentation tree decision

Similar Documents

Publication Publication Date Title
JP6633694B2 (en) Multi-view signal codec
US11234002B2 (en) Method and apparatus for encoding and decoding a texture block using depth based block partitioning
JP6501808B2 (en) Efficient partition coding with high degree of freedom
US10212411B2 (en) Methods of depth based block partitioning
EP3167613B1 (en) Method of motion vector derivation for video coding
JP6472877B2 (en) Method for 3D or multi-view video encoding including view synthesis prediction
KR101706309B1 (en) Method and apparatus of inter-view candidate derivation for three-dimensional video coding
WO2019143841A1 (en) Affine motion compensation in video coding
WO2015062002A1 (en) Methods for sub-pu level prediction
JP2019068455A (en) Effective prediction using partition coding
WO2015003383A1 (en) Methods for inter-view motion prediction
WO2015109598A1 (en) Methods for motion parameter hole filling
TWI688266B (en) Method and apparatus for intra prediction fusion in image and video coding
WO2015192706A1 (en) Method of coding for depth based block partitioning mode in three-dimensional or multi-view video coding
WO2015135175A1 (en) Simplified depth based block partitioning method
WO2020058958A1 (en) Construction for motion candidates list
WO2015006922A1 (en) Methods for residual prediction
WO2015192314A1 (en) A simplified method for depth based block partitioning
CN104853216B (en) Block dividing method based on depth and electronic device
WO2015127581A1 (en) Methods for switching off the intra prediction filters
WO2015113245A1 (en) Methods for merging candidates list construction
WO2014106327A1 (en) Method and apparatus for inter-view residual prediction in multiview video coding
WO2015103747A1 (en) Motion parameter hole filling
WO2015196364A1 (en) Methods for inter-view advanced residual prediction
WO2015139206A1 (en) Methods for 3d video coding

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14894928

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14894928

Country of ref document: EP

Kind code of ref document: A1