CN115066895A - Method and apparatus for encoding/decoding image by using palette mode, and recording medium - Google Patents

Method and apparatus for encoding/decoding image by using palette mode, and recording medium Download PDF

Info

Publication number
CN115066895A
CN115066895A CN202080095770.2A CN202080095770A CN115066895A CN 115066895 A CN115066895 A CN 115066895A CN 202080095770 A CN202080095770 A CN 202080095770A CN 115066895 A CN115066895 A CN 115066895A
Authority
CN
China
Prior art keywords
block
target block
information
mode
prediction
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202080095770.2A
Other languages
Chinese (zh)
Inventor
林雄
方健
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Electronics and Telecommunications Research Institute ETRI
Original Assignee
Electronics and Telecommunications Research Institute ETRI
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Electronics and Telecommunications Research Institute ETRI filed Critical Electronics and Telecommunications Research Institute ETRI
Priority claimed from PCT/KR2020/017794 external-priority patent/WO2021112651A1/en
Publication of CN115066895A publication Critical patent/CN115066895A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/11Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

A method and apparatus for encoding/decoding an image by using a palette mode, and a recording medium are disclosed. Prediction using the palette mode for a block to be predicted may be performed. Various types of prediction may be performed for a block to be predicted, and a plurality of pieces of information to be signaled from an encoding apparatus to a decoding apparatus may be changed according to the type of prediction to be performed for the block. The determination of the signaled pieces of information may be optimized for when the palette mode is to be used for the block, so that encoding and decoding efficiency of the image may be improved.

Description

Method and apparatus for encoding/decoding image by using palette mode, and recording medium
Technical Field
The present disclosure generally relates to a method, apparatus, and storage medium for image encoding/decoding. More particularly, the present disclosure relates to a method, apparatus, and storage medium for image encoding/decoding using a palette mode.
The present application claims the benefit of korean patent application No. 10-2019-.
Background
With the continuous development of the information and communication industry, broadcasting services supporting High Definition (HD) resolution have been popularized throughout the world. Through this popularity, a large number of users have become accustomed to high resolution and high definition images and/or videos.
In order to meet the demand of users for high definition, a large number of mechanisms have accelerated the development of next-generation imaging devices. In addition to high definition TV (hdtv) and Full High Definition (FHD) TV, user interest in UHD TV has also increased, where the resolution of UHD TV is more than four times the resolution of Full High Definition (FHD) TV. With the increase of interest thereof, image encoding/decoding techniques for images with higher resolution and higher definition are now required.
As an image compression technique, there are various techniques (such as an inter-prediction technique, an intra-prediction technique, a transform, a quantization technique, and an entropy coding technique).
The inter prediction technique is a technique for predicting values of pixels included in a current picture using a picture before the current picture and/or a picture after the current picture. The intra prediction technique is a technique for predicting values of pixels included in a current picture using information on the pixels in the current picture. The transform and quantization techniques may be techniques for compressing the energy of the residual image. Entropy coding techniques are techniques for assigning short codewords to frequently occurring values and long codewords to less frequently occurring values.
By using these image compression techniques, data on an image can be efficiently compressed, transmitted, and stored.
Disclosure of Invention
Technical problem
Embodiments are directed to an apparatus, method, and storage medium using a palette mode.
Embodiments are directed to an apparatus, method, and storage medium for efficiently processing an application of a palette mode.
Technical scheme
According to an aspect, there is provided an image encoding method including: determining a prediction mode for a chroma component of a target block; and performing encoding on the chrominance components of the target block according to the prediction mode.
The step of performing encoding may be configured to: determining whether information regarding the prediction mode is to be encoded based on whether a palette mode is used for a chroma component of the target block when a partition structure of a luma component of the target block is the same as a partition structure of a chroma component of the target block.
The information regarding the prediction mode may include information indicating whether block-Based Delta Pulse Code Modulation (BDPCM) is applied to a chrominance component of the target block.
The information on the prediction mode may include information indicating whether cross-component linear model CCLM intra prediction is used for a chroma component of the target block.
The information on the prediction mode may include information indicating an intra prediction mode for chroma sampling of the target block.
The step of performing encoding may be configured to: encoding information on the prediction mode if a palette mode is not used for a chroma component of the target block when a partition structure of a luma component of the target block is the same as a partition structure of a chroma component of the target block.
The step of performing encoding may be configured to: when the partition structure of the luminance component of the target block is the same as the partition structure of the chrominance component of the target block, if a palette mode is used for the chrominance component of the target block, information on the prediction mode is not encoded.
According to another aspect, there is provided a storage medium storing a bitstream generated by an image encoding method.
According to another aspect, there is provided an image decoding method including: determining a prediction mode for a chroma component of a target block; and performing decoding on a chrominance component of the target block according to the prediction mode.
The step of determining the prediction mode may be configured to: determining whether information regarding the prediction mode is to be decoded based on whether a palette mode is used for a chroma component of the target block when a partition structure of a luma component of the target block is the same as a partition structure of a chroma component of the target block.
The information regarding the prediction mode may include information indicating whether block-Based Delta Pulse Code Modulation (BDPCM) is applied to a chrominance component of the target block.
The information on the prediction mode may include information indicating whether cross-component linear model (CCLM) intra prediction is used for a chroma component of the target block.
The information on the prediction mode may include information indicating an intra prediction mode for chroma sampling of the target block.
The step of determining the prediction mode may be configured to: decoding information on the prediction mode if a palette mode is not used for a chroma component of the target block when a partition structure of a luma component of the target block is the same as a partition structure of a chroma component of the target block.
The step of determining the prediction mode may be configured to: when the partition structure of the luminance component of the target block is the same as the partition structure of the chrominance component of the target block, if a palette mode is used for the chrominance component of the target block, information on the prediction mode is not decoded.
According to yet another aspect, there is provided a computer-readable storage medium storing a bitstream, wherein the bitstream includes encoding information regarding a chroma component of a target block, decoding is performed on the chroma component of the target block using the encoding information, a prediction mode for the chroma component of the target block is determined, and encoding is performed on the chroma component of the target block according to the prediction mode.
The determination of the prediction mode may be configured to: determining whether information regarding the prediction mode is to be encoded based on whether a palette mode is used for a chroma component of the target block when a partition structure of a luma component of the target block is the same as a partition structure of a chroma component of the target block.
The information on the prediction mode may include information indicating an intra prediction mode for chroma sampling of the target block.
The determination of the prediction mode may be configured to: encoding information on the prediction mode if a palette mode is not used for a chroma component of the target block when a partition structure of a luma component of the target block is the same as a partition structure of a chroma component of the target block.
The determination of the prediction mode may be configured to: when the partition structure of the luminance component of the target block is the same as the partition structure of the chrominance component of the target block, if a palette mode is used for the chrominance component of the target block, information on the prediction mode is not encoded.
Advantageous effects
An apparatus, method, and storage medium using a palette mode are provided.
An apparatus, method, and storage medium for efficiently processing application of a palette mode are provided.
Drawings
Fig. 1 is a block diagram showing a configuration of an embodiment of an encoding apparatus to which the present disclosure is applied;
fig. 2 is a block diagram showing a configuration of an embodiment of a decoding apparatus to which the present disclosure is applied;
fig. 3 is a diagram schematically showing a partition structure of an image when the image is encoded and decoded;
fig. 4 is a diagram illustrating a form of a prediction unit that a coding unit can include;
fig. 5 is a diagram showing a form of a transform unit that can be included in an encoding unit;
FIG. 6 illustrates partitioning of blocks according to an example;
FIG. 7 is a diagram for explaining an embodiment of an intra prediction process;
fig. 8 is a diagram illustrating reference samples used in an intra prediction process;
FIG. 9 is a diagram for explaining an embodiment of an inter prediction process;
FIG. 10 illustrates spatial candidates according to an embodiment;
fig. 11 illustrates an order of adding motion information of spatial candidates to a merge list according to an embodiment;
FIG. 12 illustrates a transform and quantization process according to an example;
FIG. 13 illustrates a diagonal scan according to an example;
FIG. 14 shows a horizontal scan according to an example;
FIG. 15 shows a vertical scan according to an example;
fig. 16 is a configuration diagram of an encoding device according to an embodiment;
fig. 17 is a configuration diagram of a decoding apparatus according to an embodiment;
fig. 18 shows a first syntax for a target block, in accordance with an embodiment;
fig. 19 shows a second syntax for a target block, in accordance with an embodiment;
fig. 20 shows a modified second syntax for a target block, in accordance with an embodiment;
fig. 21 is a flowchart of an image encoding method according to an embodiment; and
fig. 22 is a flowchart of an image decoding method according to an embodiment.
Detailed Description
The present invention may be variously modified and may have various embodiments, and specific embodiments will be described in detail below with reference to the accompanying drawings. It should be understood, however, that these examples are not intended to limit the invention to the particular forms disclosed, but to include all changes, equivalents, and modifications that are within the spirit and scope of the invention.
The following exemplary embodiments will be described in detail with reference to the accompanying drawings showing specific embodiments. These embodiments are described so that those of ordinary skill in the art to which this disclosure pertains will be readily able to practice them. It should be noted that the various embodiments are distinct from one another, but are not necessarily mutually exclusive. For example, particular shapes, structures, and characteristics described herein may be implemented as one embodiment without departing from the spirit and scope of other embodiments associated with the other embodiments. Further, it is to be understood that the location or arrangement of individual components within each disclosed embodiment can be modified without departing from the spirit and scope of the embodiments. Therefore, the appended detailed description is not intended to limit the scope of the disclosure, and the scope of exemplary embodiments is defined only by the appended claims and equivalents thereof, as they are properly described.
In the drawings, like numerals are used to designate the same or similar functions in various respects. The shapes, sizes, and the like of components in the drawings may be exaggerated for clarity of the description.
Terms such as "first" and "second" may be used to describe various components, but the components are not limited by the terms. The terms are only used to distinguish one component from another component. For example, a first component may be termed a second component without departing from the scope of the present description. Similarly, the second component may be referred to as the first component. The term "and/or" may include a combination of multiple related descriptive items or any of multiple related descriptive items.
It will be understood that when an element is referred to as being "connected" or "coupled" to another element, the two elements may be directly connected or coupled to each other or intervening elements may be present between the two elements. On the other hand, it will be understood that when components are referred to as being "directly connected or coupled", there are no intervening components between the two components.
The components described in the embodiments are independently illustrated to indicate different feature functions, but this does not mean that each component is formed of a separate piece of hardware or software. That is, a plurality of components are individually arranged and included for convenience of description. For example, at least two of the plurality of components may be integrated into a single component. Instead, one component may be divided into a plurality of components. Embodiments in which a plurality of components are integrated or embodiments in which some components are separated are included in the scope of the present specification as long as they do not depart from the essence of the present specification.
Furthermore, in exemplary embodiments, the expression that a component "includes" a specific component means that another component may be included within the scope of practical or technical spirit of the exemplary embodiments, but does not exclude the presence of components other than the specific component.
The terminology used in the description is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. Singular references include plural references unless the context specifically indicates the contrary. In this specification, it is to be understood that terms such as "including" or "having" are only intended to indicate that there are features, numbers, steps, operations, components, parts, or combinations thereof, and are not intended to preclude the possibility that one or more other features, numbers, steps, operations, components, parts, or combinations thereof will be present or added. That is, in the present invention, the expression that a component is described as "including" a specific component means that another component may be included in the scope of the practice of the present invention or the technical spirit of the present invention, but does not exclude the presence of components other than the specific component.
Some components of the present invention are not essential components for performing essential functions but may be optional components only for improving performance. An embodiment may be implemented using only the necessary components to implement the essence of the embodiment. For example, a structure including only necessary components (not including only optional components for improving performance) is also included in the scope of the embodiments.
The embodiments will be described in detail below with reference to the accompanying drawings so that those skilled in the art to which the embodiments pertain can easily implement the embodiments. In the following description of the embodiments, a detailed description of known functions or configurations incorporated herein will be omitted. In addition, the same reference numerals are used to designate the same components throughout the drawings, and repeated description of the same components will be omitted.
Hereinafter, "image" may represent a single picture constituting a video, or may represent the video itself. For example, "encoding and/or decoding of an image" may mean "encoding and/or decoding of a video", and may also mean "encoding and/or decoding of any one of a plurality of images constituting a video".
Hereinafter, the terms "video" and "moving picture" may be used to have the same meaning and may be used interchangeably with each other.
Hereinafter, the target image may be an encoding target image that is a target to be encoded and/or a decoding target image that is a target to be decoded. Further, the target image may be an input image input to the encoding apparatus or an input image input to the decoding apparatus. Also, the target image may be a current image, i.e., a target that is currently to be encoded and/or decoded. For example, the terms "target image" and "current image" may be used to have the same meaning and may be used interchangeably with each other.
Hereinafter, the terms "image", "picture", "frame", and "screen" may be used to have the same meaning and may be used interchangeably with each other.
Hereinafter, the target block may be an encoding target block (i.e., a target to be encoded) and/or a decoding target block (i.e., a target to be decoded). Furthermore, the target block may be a current block, i.e., a target that is currently to be encoded and/or decoded. Here, the terms "target block" and "current block" may be used to have the same meaning and may be used interchangeably with each other. The current block may represent an encoding target block that is an encoding target during encoding and/or a decoding target block that is a decoding target during decoding. Further, the current block may be at least one of an encoding block, a prediction block, a residual block, and a transform block.
Hereinafter, the terms "block" and "unit" may be used to have the same meaning and may be used interchangeably with each other. Alternatively, a "block" may represent a particular unit.
Hereinafter, the terms "region" and "fragment" are used interchangeably with each other.
In the following embodiments, particular information, data, flags, indices, elements, and attributes may have their respective values. A value of "0" corresponding to each of the information, data, flags, indices, elements, and attributes may indicate false, logical false, or a first predefined value. In other words, the values "0", false, logical false, and the first predefined value may be used interchangeably with each other. A value of "1" corresponding to each of the information, data, flags, indices, elements, and attributes may indicate true, logically true, or a second predefined value. In other words, the values "1", true, logically true, and second predefined values may be used interchangeably with each other.
When a variable such as i or j is used to indicate a row, column, or index, the value i may be an integer 0 or greater than 0, or may be an integer 1 or greater than 1. In other words, in embodiments, each of the rows, columns, and indices may count from 0, or may count from 1.
In embodiments, the term "one or more" or the term "at least one" may mean the term "a plurality". The term "one or more" or the term "at least one" may be used interchangeably with "plurality".
Next, terms to be used in the embodiments will be described.
An encoder: the encoder represents an apparatus for performing encoding. That is, the encoder may represent an encoding apparatus.
A decoder: the decoder represents means for performing decoding. That is, the decoder may represent a decoding apparatus.
A unit: the unit may represent a unit of image encoding and decoding. The terms "unit" and "block" may be used to have the same meaning and may be used interchangeably with each other.
The cell may be an M × N array of samples. Each of M and N may be a positive integer. The cells may generally represent a two-dimensional form of an array of samples.
During the encoding and decoding of an image, a "unit" may be a region generated by partitioning an image. In other words, a "cell" may be a region designated in one image. A single image may be partitioned into multiple cells. Alternatively, one image may be partitioned into sub-parts, and a unit may represent each partitioned sub-part when encoding or decoding is performed on the partitioned sub-parts.
During the encoding and decoding of the image, a predefined processing may be performed on each unit according to the type of unit.
Unit types may be classified into macro-units, Coding Units (CUs), Prediction Units (PUs), residual units, Transform Units (TUs), etc., according to function. Alternatively, the unit may represent a block, a macroblock, a coding tree unit, a coding tree block, a coding unit, a coding block, a prediction unit, a prediction block, a residual unit, a residual block, a transform unit, a transform block, and the like according to functions. For example, a target unit that is a target of encoding and/or decoding may be at least one of a CU, a PU, a residual unit, and a TU.
The term "unit" may denote information including a luminance (luma) component block, a chrominance (chroma) component block corresponding to the luminance component block, and syntax elements for the respective blocks, such that the unit is designated to be distinguished from the blocks.
The size and shape of the cells can be implemented differently. Further, the cells may have any of a variety of sizes and shapes. Specifically, the shape of the cell may include not only a square but also a geometric shape (such as a rectangle, a trapezoid, a triangle, and a pentagon) that can be represented in two dimensions (2D).
Further, the unit information may include one or more of a type of the unit, a size of the unit, a depth of the unit, an encoding order of the unit, a decoding order of the unit, and the like. For example, the type of the unit may indicate one of a CU, a PU, a residual unit, and a TU.
A unit may be partitioned into sub-units, each sub-unit having a size smaller than the size of the associated unit.
Depth: depth may represent the degree to which a cell is partitioned. Further, the depth of a cell may indicate a level at which the corresponding cell exists when the cell is represented by a tree structure.
The unit partition information may comprise a depth indicating a depth of the unit. The depth may indicate the number of times a cell is partitioned and/or the extent to which the cell is partitioned.
In the tree structure, the depth of the root node can be considered to be the smallest and the depth of the leaf nodes the largest. The root node may be the highest (top) node. The leaf node may be the lowest node.
A single unit may be hierarchically partitioned into a plurality of sub-units, while the single unit has tree structure based depth information. In other words, a unit and a child unit generated by partitioning the unit may correspond to a node and a child node of the node, respectively. Each partitioned sub-unit may have a unit depth. Since the depth indicates the number of times the unit is partitioned and/or the degree to which the unit is partitioned, the partition information of the sub-unit may include information on the size of the sub-unit.
In the tree structure, the top node may correspond to the initial node before partitioning. The top node may be referred to as the "root node". Further, the root node may have a minimum depth value. Here, the depth of the top node may be level "0".
A node with a depth of level "1" may represent a cell generated when an initial cell is partitioned once. A node with a depth of level "2" may represent a cell generated when an initial cell is partitioned twice.
A leaf node with a depth of level "n" may represent a cell generated when an initial cell is partitioned n times.
A leaf node may be the bottom node that cannot be partitioned further. The depth of a leaf node may be a maximum level. For example, the predefined value for the maximum level may be 3.
the-QT depth may represent the depth for a quad-partition. BT depth may represent depth for a bipartite partition. The TT depth may represent a depth for a tri-partition.
-sampling points: the samples may be elementary units that constitute a block. Available from 0 to 2 according to the bit depth (Bd) Bd -a value of 1 to represent a sample point.
The samples may be pixels or pixel values.
In the following, the terms "pixel" and "sample" may be used with the same meaning and may be used interchangeably with each other.
Code Tree Unit (CTU): a CTU may be composed of a single luma component (Y) coding tree block and two chroma component (i.e., Cb, Cr) coding tree blocks associated with the luma component coding tree block. In addition, the CTU may represent information including the above-described blocks and syntax elements for each block.
-each Coding Tree Unit (CTU) may be partitioned using one or more partitioning methods, such as Quadtree (QT), Binary Tree (BT) and Ternary Tree (TT), in order to configure sub-units, such as coding units, prediction units and transform units. The quadtree may represent a quadtree. Further, each coding tree unit may be partitioned using a multi-type tree (MTT) using one or more partitioning methods.
"CTU" may be used as a term designating a pixel block as a processing unit in an image decoding and encoding process, such as in the case of partitioning an input image.
Coded Tree Block (CTB): "CTB" may be used as a term designating any one of a Y coding tree block, a Cb coding tree block, and a Cr coding tree block.
Adjacent blocks: the neighboring blocks (or neighboring blocks) may represent blocks adjacent to the target block. The neighboring blocks may represent reconstructed neighboring blocks.
Hereinafter, the terms "adjacent block" and "adjacent block" may be used to have the same meaning and may be used interchangeably with each other.
The neighboring blocks may represent reconstructed neighboring blocks.
Spatially adjacent blocks: the spatially neighboring block may be a block spatially adjacent to the target block. The neighboring blocks may include spatially neighboring blocks.
The target block and the spatially neighboring blocks may be comprised in the target picture.
Spatially neighboring blocks may represent blocks whose boundaries are in contact with the target block or blocks which are located within a predetermined distance from the target block.
The spatially neighboring blocks may represent blocks adjacent to the vertex of the target block. Here, the blocks adjacent to the vertex of the target block may represent blocks vertically adjacent to an adjacent block horizontally adjacent to the target block or blocks horizontally adjacent to an adjacent block vertically adjacent to the target block.
Temporal neighboring blocks: the temporally adjacent block may be a block temporally adjacent to the target block. The neighboring blocks may include temporally neighboring blocks.
The temporally adjacent blocks may comprise co-located blocks (col blocks).
A col block may be a block in a previously reconstructed co-located picture (col picture). The location of the col block in the col picture may correspond to the location of the target block in the target picture. Alternatively, the location of the col block in the col picture may be equal to the location of the target block in the target picture. The col picture may be a picture included in the reference picture list.
The temporally neighboring blocks may be blocks temporally adjacent to spatially neighboring blocks of the target block.
Prediction mode: the prediction mode may be information indicating a mode in which encoding and/or decoding is performed for intra prediction or a mode in which encoding and/or decoding is performed for inter prediction.
A prediction unit: the prediction unit may be a basic unit for prediction such as inter prediction, intra prediction, inter compensation, intra compensation, and motion compensation.
A single prediction unit may be divided into multiple partitions or sub-prediction units of smaller size. The plurality of partitions may also be basic units in performing prediction or compensation. The partition generated by dividing the prediction unit may also be the prediction unit.
Prediction unit partitioning: the prediction unit partition may be a shape into which the prediction unit is divided.
Reconstructed neighboring cells: the reconstructed neighboring unit may be a unit that is neighboring the target unit and has been decoded and reconstructed.
The reconstructed neighboring cells may be cells spatially adjacent to the target cell or temporally adjacent to the target cell.
The reconstructed spatially neighboring units may be units comprised in the target picture that have been reconstructed by encoding and/or decoding.
The reconstructed temporal neighboring cells may be cells comprised in the reference image that have been reconstructed by encoding and/or decoding. The position of the reconstructed temporally neighboring cell in the reference image may be the same as the position of the target cell in the target picture or may correspond to the position of the target cell in the target picture. Further, the reconstructed temporal neighboring cell may be a block neighboring the corresponding block in the reference image. Here, the position of the corresponding block in the reference image may correspond to the position of the target block in the target image. Here, the fact that the positions of the blocks correspond to each other may mean that the positions of the blocks are the same as each other, may mean that one block is included in another block, or may mean that one block occupies a specific position in another block.
And (3) sub-picture: a picture may be divided into one or more sub-pictures. A sprite may be composed of one or more parallel block rows and one or more parallel block columns.
A sprite may be a region in a picture that has a square or rectangular (i.e., non-square, rectangular) shape. Further, a sprite may include one or more CTUs.
A single sprite may comprise one or more parallel blocks, one or more tiles (swick) and/or one or more stripes.
Parallel block: a parallel block may be a region in a picture having a square or rectangular (i.e., non-square, rectangular) shape.
A parallel block may comprise one or more CTUs.
A parallel block may be partitioned into one or more partitions.
Partitioning: a block may represent one or more rows of CTUs in a parallel block.
A parallel block may be partitioned into one or more partitions. Each partition may include one or more rows of CTUs.
Parallel blocks that are not partitioned into two parts may also represent partitions.
Strip: a stripe may comprise one or more parallel blocks in a picture. Optionally, a stripe may comprise one or more partitions of parallel blocks.
Parameter set: the parameter set may correspond to header information in an internal structure of the bitstream.
The parameter set may comprise at least one of a Video Parameter Set (VPS), a Sequence Parameter Set (SPS), a Picture Parameter Set (PPS), an Adaptation Parameter Set (APS), a Decoding Parameter Set (DPS), etc.
The information signaled by each parameter set may be applied to a picture referring to the corresponding parameter set. For example, information in the VPS can be applied to pictures that reference the VPS. Information in the SPS may be applied to pictures that reference the SPS. Information in the PPS may be applied to pictures that reference the PPS.
Each parameter set may refer to a higher parameter set. For example, a PPS may reference an SPS. SPS may refer to VPS.
Further, the parameter set may include a parallel block group, slice header information, and parallel block header information. The parallel block group may be a group including a plurality of parallel blocks. Further, the meaning of "parallel block group" may be the same as that of "stripe".
And (3) rate distortion optimization: an encoding device may use rate-distortion optimization to provide high encoding efficiency by utilizing a combination of: a size of a Coding Unit (CU), a prediction mode, a size of a Prediction Unit (PU), motion information, and a size of a Transform Unit (TU).
The rate-distortion optimization scheme may calculate the rate-distortion cost of each combination to select the optimal combination from the combinations. The rate-distortion cost may be calculated using the equation "D + λ R". In general, the combination that minimizes the rate-distortion cost may be selected as the optimal combination under the rate-distortion optimization scheme.
D may represent distortion. D may be the average of the squares of the differences between the original transform coefficients and the reconstructed transform coefficients in the transform unit (i.e., the mean square error).
-R may represent the rate, which may represent the bit rate using the relevant context information.
- λ represents the lagrange multiplier. R may include not only coding parameter information such as a prediction mode, motion information, and a coding block flag, but also bits generated as a result of coding transform coefficients.
The coding device may perform processes such as inter-and/or intra-prediction, transformation, quantization, entropy coding, inverse quantization (dequantization) and/or inverse transformation in order to calculate the exact D and R. These processes can add significant complexity to the encoding device.
Bit stream: the bitstream may represent a stream of bits including encoded image information.
And (3) analysis: parsing may be a decision on the value of a syntax element made by performing entropy decoding on the bitstream. Alternatively, the term "parsing" may denote such entropy decoding itself.
Symbol: the symbol may be at least one of a syntax element, an encoding parameter, and a transform coefficient of the encoding target unit and/or the decoding target unit. Further, the symbol may be a target of entropy encoding or a result of entropy decoding.
Reference picture: the reference picture may be an image that is unit-referenced in order to perform inter prediction or motion compensation. Alternatively, the reference picture may be an image including a reference unit that is referred to by the target unit in order to perform inter prediction or motion compensation.
Hereinafter, the terms "reference picture" and "reference image" may be used to have the same meaning and may be used interchangeably with each other.
Reference picture list: the reference picture list may be a list including one or more reference pictures used for inter prediction or motion compensation.
The types of the reference picture list may include a combination List (LC), a list 0(L0), a list 1(L1), a list 2(L2), a list 3(L3), and the like.
For inter prediction, one or more reference picture lists may be used.
Inter prediction indicator: the inter prediction indicator may indicate an inter prediction direction for the target unit. The inter prediction may be one of unidirectional prediction and bidirectional prediction. Alternatively, the inter prediction indicator may represent the number of reference pictures used to generate the prediction unit of the target unit. Alternatively, the inter prediction indicator may represent the number of prediction blocks used for inter prediction or motion compensation of the target unit.
Prediction list utilization flag: the prediction list utilization flag may indicate whether at least one reference picture in a particular reference picture list is used to generate a prediction unit.
-the prediction list utilization flag may be used to derive the inter prediction indicator. Instead, the prediction list utilization flag may be derived using the inter prediction indicator. For example, a case where the prediction list indicates "0" (as a first value) with the flag may indicate that, for the target unit, the reference picture in the reference picture list is not used to generate the prediction block. The case where the prediction list indicates "1" (as a second value) with the flag may indicate that, for the target unit, the prediction unit is generated using the reference picture list.
Reference picture index: the reference picture index may be an index indicating a specific reference picture in the reference picture list.
Picture Order Count (POC): the POC value of a picture may represent an order in which the corresponding picture is displayed.
Motion Vector (MV): the motion vector may be a 2D vector for inter prediction or motion compensation. The motion vector may represent an offset between the target image and the reference image.
For example, may be represented by a symbol such as (mv) x ,mv y ) Represents the MV. mv x Can indicate the horizontal component, mv y A vertical component may be indicated.
The search range is as follows: the search range may be a 2D region in which a search for MVs is performed during inter prediction. For example, the size of the search range may be M × N. M and N may be positive integers, respectively.
Motion vector candidates: the motion vector candidate may be a block that is a prediction candidate when the motion vector is predicted or a motion vector of a block that is a prediction candidate.
The motion vector candidate may be comprised in a motion vector candidate list.
Motion vector candidate list: the motion vector candidate list may be a list configured using one or more motion vector candidates.
Motion vector candidate index: the motion vector candidate index may be an indicator for indicating a motion vector candidate in the motion vector candidate list. Alternatively, the motion vector candidate index may be an index of a motion vector predictor.
Motion information: the motion information may be information including at least one of a reference picture list, a reference picture, a motion vector candidate index, a merge candidate, and a merge index, and a motion vector, a reference picture index, and an inter prediction indicator.
Merging the candidate lists: the merge candidate list may be a list using one or more merge candidate configurations.
Merging candidates: the merge candidate may be a spatial merge candidate, a temporal merge candidate, a combined bi-predictive merge candidate, a history-based candidate, a candidate based on the average of the two candidates, a zero merge candidate, etc. The merge candidate may include an inter prediction indicator and may include motion information such as prediction type information, a reference picture index for each list, a motion vector, a prediction list utilization flag, and an inter prediction indicator.
Merging indexes: the merge index may be an indicator for indicating a merge candidate in the merge candidate list.
The merging index may indicate a reconstruction unit used for deriving the merging candidate among reconstruction units spatially neighboring the target unit and reconstruction units temporally neighboring the target unit.
The merge index may indicate at least one of pieces of motion information of the merge candidates.
A transformation unit: the transform unit may be a basic unit of residual signal encoding and/or residual signal decoding, such as transform, inverse transform, quantization, inverse quantization, transform coefficient encoding, and transform coefficient decoding. A single transform unit may be partitioned into a plurality of sub-transform units having smaller sizes. Here, the transform may include one or more of a primary transform and a secondary transform, and the inverse transform may include one or more of a primary inverse transform and a secondary inverse transform.
Zooming: scaling may refer to the process of multiplying a factor by a transform coefficient level.
-as a result of scaling the transform coefficient level, transform coefficients may be generated. Scaling may also be referred to as "inverse quantization".
Quantization Parameter (QP): the quantization parameter may be a value used to generate a transform coefficient level for a transform coefficient in quantization. Alternatively, the quantization parameter may also be a value used to generate a transform coefficient by scaling the transform coefficient level in inverse quantization. Alternatively, the quantization parameter may be a value mapped to a quantization step.
Delta (Delta) quantization parameter: the delta quantization parameter may represent a difference between the quantization parameter of the target unit and the predicted quantization parameter.
Scanning: scanning may represent a method of arranging the order of coefficients in a cell, block, or matrix. For example, a method for arranging a 2D array in the form of a one-dimensional (1D) array may be referred to as "scanning". Alternatively, the method for arranging the 1D array in the form of a 2D array may also be referred to as "scanning" or "inverse scanning".
Transform coefficients: the transform coefficient may be a coefficient value generated when the encoding apparatus performs the transform. Alternatively, the transform coefficient may be a coefficient value generated when the decoding apparatus performs at least one of entropy decoding and inverse quantization.
Quantized levels generated by applying quantization to the transform coefficients or the residual signal or quantized transform coefficient levels may also be included in the meaning of the term "transform coefficients".
Level of quantization: the level of quantization may be a value generated when the encoding apparatus performs quantization on the transform coefficient or the residual signal. Alternatively, the quantized level may be a value that is a target of inverse quantization when the decoding apparatus performs inverse quantization.
The quantized transform coefficient levels as a result of the transform and quantization may also be included in the meaning of quantized levels.
Non-zero transform coefficients: the non-zero transform coefficient may be a transform coefficient having a value other than 0, or may be a transform coefficient level having a value other than 0. Alternatively, the non-zero transform coefficient may be a transform coefficient whose value is not 0 in magnitude, or may be a transform coefficient level whose value is not 0 in magnitude.
Quantization matrix: the quantization matrix may be a matrix used in a quantization process or an inverse quantization process in order to improve subjective image quality or objective image quality of an image. The quantization matrix may also be referred to as a "scaling list".
Quantization matrix coefficients: the quantization matrix coefficient may be each element in the quantization matrix. The quantized matrix coefficients may also be referred to as "matrix coefficients".
A default matrix: the default matrix may be a quantization matrix predefined by the encoding device and the decoding device.
Non-default matrix: the non-default matrix may be a quantization matrix that is not predefined by the encoding device and the decoding device. The non-default matrix may represent a quantization matrix signaled by a user from an encoding device to a decoding device.
Most Probable Mode (MPM): the MPM may represent an intra prediction mode in which a high probability is used for intra prediction for the target block.
The encoding apparatus and the decoding apparatus may determine one or more MPMs based on the encoding parameters related to the target block and the attributes of the entity related to the target block.
The encoding device and the decoding device may determine the one or more MPMs based on an intra prediction mode of the reference block. The reference block may include a plurality of reference blocks. The plurality of reference blocks may include a spatially neighboring block adjacent to the left side of the target block and a spatially neighboring block adjacent to the upper side of the target block. In other words, one or more different MPMs may be determined according to which intra prediction modes have been used for the reference block.
One or more MPMs may be determined in the same way in both the encoding device and the decoding device. That is, the encoding apparatus and the decoding apparatus may share the same MPM list including one or more MPMs.
List of MPMs: the MPM list may be a list including one or more MPMs. The number of one or more MPMs in the MPM list may be predefined.
MPM indicator: the MPM indicator may indicate an MPM to be used for intra prediction for the target block among one or more MPMs in the MPM list. For example, the MPM indicator may be an index for an MPM list.
Since the MPM list is determined in the same manner in both the encoding device and the decoding device, it may not be necessary to transmit the MPM list itself from the encoding device to the decoding device.
The MPM indicator may be signaled from the encoding device to the decoding device. Since the MPM indicator is signaled, the decoding apparatus may determine an MPM to be used for intra prediction for the target block among MPMs in the MPM list.
MPM usage indicator: the MPM usage indicator may indicate whether an MPM usage mode is to be used for prediction for the target block. The MPM use mode may be a mode that determines an MPM to be used for intra prediction for the target block using the MPM list.
The MPM usage indicator may be signaled from the encoding device to the decoding device.
Signaling: "signaling" may mean that information is sent from an encoding device to a decoding device. Alternatively, "signaling" may mean that information is included in a bitstream or a recording medium. The information signaled by the encoding device may be used by the decoding device.
The encoding device may generate the encoded information by performing an encoding of the information to be signaled. The encoded information may be transmitted from the encoding device to the decoding device. The decoding apparatus may obtain the information by decoding the transmitted encoded information. Here, the encoding may be entropy encoding, and the decoding may be entropy decoding.
And (3) statistical value: variables, coding parameters, constants, etc. may have calculable values. The statistical value may be a value generated by performing calculation (operation) on a value specifying a target. For example, the statistical value may indicate one or more of an average, a weighted sum, a minimum, a maximum, a mode, a median, and an interpolation of values of a particular variable, a particular encoding parameter, a particular constant, and the like.
Fig. 1 is a block diagram showing a configuration of an embodiment of an encoding apparatus to which the present disclosure is applied.
The encoding device 100 may be an encoder, a video encoding device, or an image encoding device. A video may comprise one or more images (pictures). The encoding apparatus 100 may sequentially encode one or more images of a video.
Referring to fig. 1, the encoding apparatus 100 includes an inter prediction unit 110, an intra prediction unit 120, a switch 115, a subtractor 125, a transform unit 130, a quantization unit 140, an entropy encoding unit 150, an inverse quantization (inverse quantization) unit 160, an inverse transform unit 170, an adder 175, a filter unit 180, and a reference picture buffer 190.
The encoding apparatus 100 may perform encoding on the target image using an intra mode and/or an inter mode. In other words, the prediction mode of the target block may be one of an intra mode and an inter mode.
Hereinafter, the terms "intra mode", "intra prediction mode", "intra mode", and "intra prediction mode" may be used to have the same meaning and may be used interchangeably with each other.
Hereinafter, the terms "inter mode", "inter prediction mode", "inter mode", and "inter prediction mode" may be used to have the same meaning and may be used interchangeably with each other.
Hereinafter, the term "image" may indicate only a partial image, or may indicate a block. Further, the processing of an "image" may indicate sequential processing of a plurality of blocks.
Further, the encoding apparatus 100 may generate a bitstream including encoded information by encoding the target image, and may output and store the generated bitstream. The generated bitstream may be stored in a computer-readable storage medium and may be streamed over a wired and/or wireless transmission medium.
When the intra mode is used as the prediction mode, the switch 115 may switch to the intra mode. When the inter mode is used as the prediction mode, the switch 115 may switch to the inter mode.
The encoding apparatus 100 may generate a prediction block of a target block. Also, after the prediction block has been generated, the encoding apparatus 100 may encode a residual block for the target block using a residual between the target block and the prediction block.
When the prediction mode is the intra mode, the intra prediction unit 120 may use pixels of a previously encoded/decoded neighboring block adjacent to the target block as reference samples. The intra prediction unit 120 may perform spatial prediction on the target block using the reference sample points, and may generate prediction sample points for the target block via the spatial prediction. The prediction samples may represent samples in a prediction block.
The inter prediction unit 110 may include a motion prediction unit and a motion compensation unit.
When the prediction mode is the inter mode, the motion prediction unit may search for a region that best matches the target block in the reference image in the motion prediction process, and may derive a motion vector for the target block and the found region based on the found region. Here, the motion prediction unit may use the search range as a target region for the search.
The reference image may be stored in the reference picture buffer 190. More specifically, when encoding and/or decoding of a reference image has been processed, the encoded and/or decoded reference image may be stored in the reference picture buffer 190.
The reference picture buffer 190 may be a Decoded Picture Buffer (DPB) since decoded pictures are stored.
The motion compensation unit may generate the prediction block for the target block by performing motion compensation using the motion vector. Here, the motion vector may be a two-dimensional (2D) vector for inter prediction. Further, the motion vector may indicate an offset between the target image and the reference image.
When the motion vector has a value other than an integer, the motion prediction unit and the motion compensation unit may generate the prediction block by applying an interpolation filter to a partial region of the reference image. To perform inter prediction or motion compensation, it may be determined which one of a skip mode, a merge mode, an Advanced Motion Vector Prediction (AMVP) mode, and a current picture reference mode corresponds to a method for predicting and compensating for motion of a PU included in a CU based on the CU, and the inter prediction or motion compensation may be performed according to the mode.
The subtractor 125 may generate a residual block, wherein the residual block is a difference between the target block and the prediction block. The residual block may also be referred to as a "residual signal".
The residual signal may be the difference between the original signal and the predicted signal. Alternatively, the residual signal may be a signal generated by transforming or quantizing the difference between the original signal and the prediction signal or a signal generated by transforming and quantizing the difference. The residual block may be a residual signal for a block unit.
The transform unit 130 may generate a transform coefficient by transforming the residual block, and may output the generated transform coefficient. Here, the transform coefficient may be a coefficient value generated by transforming the residual block.
The transform unit 130 may use one of a plurality of predefined transform methods when performing the transform.
The plurality of predefined transform methods may include Discrete Cosine Transform (DCT), Discrete Sine Transform (DST), Karhunen-Loeve transform (KLT), and the like.
The transform method for transforming the residual block may be determined according to at least one of the encoding parameters for the target block and/or the neighboring blocks. For example, the transform method may be determined based on at least one of an inter prediction mode for the PU, an intra prediction mode for the PU, a size of the TU, and a shape of the TU. Alternatively, transformation information indicating a transformation method may be signaled from the encoding apparatus 100 to the decoding apparatus 200.
When the transform skip mode is used, the transform unit 130 may omit an operation of transforming the residual block.
By performing quantization on the transform coefficients, quantized transform coefficient levels or quantized levels may be generated. Hereinafter, in the embodiment, each of the quantized transform coefficient level and the quantized level may also be referred to as a "transform coefficient".
The quantization unit 140 may generate quantized transform coefficient levels (i.e., quantized levels or quantized coefficients) by quantizing the transform coefficients according to a quantization parameter. The quantization unit 140 may output the generated quantized transform coefficient levels. In this case, the quantization unit 140 may quantize the transform coefficient using a quantization matrix.
The entropy encoding unit 150 may generate a bitstream by performing probability distribution-based entropy encoding based on the values calculated by the quantization unit 140 and/or the encoding parameter values calculated in the encoding process. The entropy encoding unit 150 may output the generated bitstream.
The entropy encoding unit 150 may perform entropy encoding on information regarding pixels of the image and information required to decode the image. For example, information required for decoding an image may include syntax elements and the like.
When entropy coding is applied, fewer bits may be allocated to more frequently occurring symbols and more bits may be allocated to less frequently occurring symbols. Since the symbols are represented by this allocation, the size of the bit string for the target symbol to be encoded can be reduced. Accordingly, the compression performance of video encoding can be improved by entropy encoding.
Also, in order to perform entropy encoding, the entropy encoding unit 150 may use an encoding method such as exponential golomb, Context Adaptive Variable Length Coding (CAVLC), or Context Adaptive Binary Arithmetic Coding (CABAC). For example, entropy encoding unit 150 may perform entropy encoding using a variable length coding/code (VLC) table. For example, the entropy encoding unit 150 may derive a binarization method for the target symbol. Furthermore, entropy encoding unit 150 may derive a probability model for the target symbol/bin. The entropy encoding unit 150 may perform arithmetic encoding using the derived binarization method, probability model, and context model.
The entropy encoding unit 150 may transform the coefficients in the form of 2D blocks into the form of 1D vectors by a transform coefficient scanning method so as to encode the quantized transform coefficient levels.
The encoding parameter may be information required for encoding and/or decoding. The encoding parameter may include information encoded by the encoding apparatus 100 and transmitted from the encoding apparatus 100 to the decoding apparatus, and may also include information that may be derived in an encoding or decoding process. For example, the information sent to the decoding device may include syntax elements.
The encoding parameters may include not only information (or flags or indexes) such as syntax elements encoded by the encoding apparatus and signaled to the decoding apparatus by the encoding apparatus, but also information derived in the encoding or decoding process. In addition, the encoding parameter may include information required to encode or decode the image. For example, the encoding parameters may include at least one of the following, a combination of the following, or statistics: size of unit/block, shape/form of unit/block, depth of unit/block, partition information of unit/block, partition structure of unit/block, information indicating whether unit/block is partitioned in a quad-tree structure, information indicating whether unit/block is partitioned in a binary-tree structure, partition direction (horizontal direction or vertical direction) of a binary-tree structure, partition form (symmetric partition or asymmetric partition) of a binary-tree structure, information indicating whether unit/block is partitioned in a tri-tree structure, partition direction (horizontal direction or vertical direction) of a tri-tree structure, partition form (symmetric partition or asymmetric partition, etc.) of a tri-tree structure, information indicating whether unit/block is partitioned in a multi-type tree structure, combination and direction (horizontal direction or vertical direction, etc.) of partitions of a multi-type tree structure, Partition form of partitions of multi-type tree structure (symmetric partition or asymmetric partition, etc.), partition tree of multi-type tree form (binary tree or ternary tree), prediction type (intra prediction or inter prediction), intra prediction mode/direction, intra luma prediction mode/direction, intra chroma prediction mode/direction, intra partition information, inter partition information, coding block partition flag, prediction block partition flag, transform block partition flag, reference sample point filtering method, reference sample point filter tap, reference sample point filter coefficient, prediction block filtering method, prediction block filter tap, prediction block filter coefficient, prediction block boundary filtering method, prediction block boundary filter tap, prediction block boundary filter coefficient, inter prediction mode, motion information, motion vector difference, reference picture index, prediction mode, motion vector, motion information, motion vector, reference picture index, and/mode, Inter prediction direction, inter prediction indicator, prediction list utilization flag, reference picture list, reference picture, POC, motion vector predictor, motion vector prediction index, motion vector prediction candidate, motion vector candidate list, information indicating whether merge mode is used, merge index, merge candidate list, information indicating whether skip mode is used, type of interpolation filter, tap of interpolation filter, filter coefficient of interpolation filter, size of motion vector, accuracy of motion vector representation, transform type, transform size, information indicating whether first transform is used, information indicating whether additional (second) transform is used, first transform selection information (or first transform index), second transform selection information (or second transform index), information indicating presence or absence of residual signal, motion vector prediction index, motion vector prediction candidate, motion vector prediction index, and motion vector prediction index, A coding block pattern, a coding block flag, a quantization parameter, a residual quantization parameter, a quantization matrix, information on an in-loop filter, information indicating whether an in-loop filter is applied, a coefficient of an in-loop filter, a tap of an in-loop filter, a shape/form of an in-loop filter, information indicating whether a deblocking filter is applied, a coefficient of a deblocking filter, a tap of a deblocking filter, a deblocking filter strength, a shape/form of a deblocking filter, information indicating whether an adaptive sample offset is applied, a value of an adaptive sample offset, a class of an adaptive sample offset, a type of an adaptive sample offset, information indicating whether an adaptive loop filter is applied, a coefficient of an adaptive loop filter, a tap of an adaptive loop filter, a shape/form of an adaptive loop filter, a binarization/inverse binarization method, a computer program, and a computer-readable storage medium, Context model, context model deciding method, context model updating method, information indicating whether normal mode is executed or not, information indicating whether bypass (bypass) mode is executed or not, significant coefficient flag, last significant coefficient flag, coding flag of coefficient group, position of last significant coefficient, information indicating whether value of coefficient is greater than 1, information indicating whether value of coefficient is greater than 2, information indicating whether value of coefficient is greater than 3, residual coefficient value information, sign information, reconstructed luma sample, reconstructed chroma sample, context binary, bypass binary, residual luma sample, residual chroma sample, transform coefficient, luma transform coefficient, chroma transform coefficient, quantized level, luma quantized level, chroma quantized level, transform coefficient level scanning method, size of motion vector search region on decoding apparatus side, and method of decoding apparatus, A shape/form of a motion vector search region on a decoding apparatus side, a number of motion vector searches on the decoding apparatus side, a size of a CTU, a minimum block size, a maximum block depth, a minimum block depth, an image display/output order, slice identification information, a slice type, slice partition information, parallel block group identification information, a parallel block group type, parallel block group partition information, parallel block identification information, a parallel block type, parallel block partition information, a picture type, a bit depth, an input sample bit depth, a reconstructed sample bit depth, a residual sample bit depth, a transform coefficient bit depth, a quantized level bit depth, information on a luminance signal, information on a chrominance signal, a color space of a target block, and a color space of a residual block. In addition, the above-described encoding parameter-related information may also be included in the encoding parameter. Information for calculating and/or deriving the above-described encoding parameters may also be included in the encoding parameters. Information calculated or derived using the above-described encoding parameters may also be included in the encoding parameters.
The prediction scheme may represent one of an intra prediction mode and an inter prediction mode.
The first transform selection information may indicate a first transform applied to the target block.
The second transform selection information may indicate a second transform applied to the target block.
The residual signal may represent the difference between the original signal and the predicted signal. Alternatively, the residual signal may be a signal generated by transforming a difference between the original signal and the prediction signal. Alternatively, the residual signal may be a signal generated by transforming and quantizing the difference between the original signal and the prediction signal. The residual block may be a residual signal for the block.
Here, signaling the information may indicate that the encoding apparatus 100 includes entropy-encoded information generated by performing entropy encoding on the flag or the index in the bitstream, and may indicate that the decoding apparatus 200 acquires the information by performing entropy decoding on the entropy-encoded information extracted from the bitstream. Here, the information may include a flag, an index, and the like.
The signal may represent information to be signaled. Hereinafter, information on the image and the block may be referred to as a "signal". Also, hereinafter, the terms "information" and "signal" may be used to have the same meaning and may be used interchangeably with each other. For example, the specific signal may be a signal representing a specific block. The original signal may be a signal representing the target block. The prediction signal may be a signal representing a prediction block. The residual signal may be a signal representing a residual block.
The bitstream may include information based on a specific syntax. The encoding apparatus 100 may generate a bitstream including information according to a specific syntax. The decoding apparatus 200 may acquire information from the bitstream according to a specific syntax.
Since the encoding apparatus 100 performs encoding via inter prediction, the encoded target image can be used as a reference image for another image to be subsequently processed. Accordingly, the encoding apparatus 100 may reconstruct or decode the encoded target image and store the reconstructed or decoded image as a reference image in the reference picture buffer 190. For decoding, inverse quantization and inverse transformation of the encoded target image may be performed.
The quantized levels may be inverse quantized by the inverse quantization unit 160 and inverse transformed by the inverse transformation unit 170. The inverse quantization unit 160 may generate inverse quantized coefficients by performing an inverse transform on the quantized levels. The inverse transform unit 170 may generate the inverse quantized and inverse transformed coefficients by performing an inverse transform on the inverse quantized coefficients.
The inverse quantized and inverse transformed coefficients may be added to the prediction block by adder 175. The inverse quantized and inverse transformed coefficients and the prediction block are added, and then a reconstructed block may be generated. Here, the inverse quantized and/or inverse transformed coefficients may represent coefficients on which one or more of inverse quantization and inverse transformation are performed, and may also represent a reconstructed residual block. Here, the reconstructed block may represent a restored block or a decoded block.
The reconstructed block may be filtered by the filter unit 180. Filter unit 180 may apply one or more of a deblocking filter, a Sample Adaptive Offset (SAO) filter, an Adaptive Loop Filter (ALF), and a non-local filter (NLF) to the reconstructed samples, reconstructed blocks, or reconstructed pictures. The filter unit 180 may also be referred to as an "in-loop filter".
The deblocking filter may remove block distortion occurring at boundaries between blocks in the reconstructed picture. In order to determine whether to apply the deblocking filter, it may be decided to be included in the block and include the number of columns or lines of pixels on which it is determined whether to apply the deblocking filter to the target block.
When the deblocking filter is applied to the target block, the applied filter may be different according to the strength of the deblocking filtering required. In other words, among different filters, a filter decided in consideration of the strength of the deblocking filtering may be applied to the target block. When the deblocking filter is applied to the target block, a filter corresponding to any one of a long tap filter, a strong filter, a weak filter, and a gaussian filter may be applied to the target block according to the strength of the required deblocking filter.
Further, when vertical filtering and horizontal filtering are performed on the target block, the horizontal filtering and the vertical filtering may be performed in parallel.
The SAO may add appropriate offsets to the pixel values to compensate for the coding error. The SAO may perform a correction on the image to which the deblocking is applied on a pixel basis, wherein the correction uses an offset of a difference between the original image and the image to which the deblocking is applied. In order to perform offset correction for an image, a method for dividing pixels included in the image into a certain number of regions, determining a region to which an offset is to be applied among the divided regions, and applying the offset to the determined region may be used, and a method for applying an offset in consideration of edge information of each pixel may also be used.
ALF may perform filtering based on values obtained by comparing a reconstructed image with an original image. After pixels included in an image have been divided into a predetermined number of groups, a filter to be applied to each group may be determined, and filtering may be performed differently for the respective groups. Information about whether to apply the adaptive loop filter may be signaled for each CU. Such information may be signaled for a luminance signal. The shape and filter coefficients of the ALF to be applied to each block may be different for each block. Alternatively, ALF having a fixed form may be applied to a block regardless of the characteristics of the block.
The non-local filter may perform filtering based on a reconstructed block similar to the target block. A region similar to the target block may be selected from the reconstructed picture, and filtering of the target block may be performed using statistical properties of the selected similar region. Information about whether to apply a non-local filter may be signaled for a Coding Unit (CU). Further, the shape and filter coefficients of the non-local filter to be applied to a block may be different according to the block.
The reconstructed block or the reconstructed image filtered by the filter unit 180 may be stored as a reference picture in the reference picture buffer 190. The reconstructed block filtered by the filter unit 180 may be a portion of a reference picture. In other words, the reference picture may be a reconstructed picture composed of the reconstructed block filtered by the filter unit 180. The stored reference pictures can then be used for inter prediction or motion compensation.
Fig. 2 is a block diagram showing a configuration of an embodiment of a decoding apparatus to which the present disclosure is applied.
The decoding apparatus 200 may be a decoder, a video decoding apparatus, or an image decoding apparatus.
Referring to fig. 2, the decoding apparatus 200 may include an entropy decoding unit 210, an inverse quantization (inverse quantization) unit 220, an inverse transformation unit 230, an intra prediction unit 240, an inter prediction unit 250, a switch 245, an adder 255, a filter unit 260, and a reference picture buffer 270.
The decoding apparatus 200 may receive the bitstream output from the encoding apparatus 100. The decoding apparatus 200 may receive a bitstream stored in a computer-readable storage medium and may receive a bitstream transmitted through a wired/wireless transmission medium stream.
The decoding apparatus 200 may perform decoding on the bitstream in an intra mode and/or an inter mode. Further, the decoding apparatus 200 may generate a reconstructed image or a decoded image via decoding, and may output the reconstructed image or the decoded image.
For example, an operation of switching to an intra mode or an inter mode based on a prediction mode for decoding may be performed by the switch 245. When the prediction mode used for decoding is intra mode, switch 245 may be operated to switch to intra mode. When the prediction mode for decoding is an inter mode, the switch 245 may be operated to switch to the inter mode.
The decoding apparatus 200 may acquire a reconstructed residual block by decoding an input bitstream and may generate a prediction block. When the reconstructed residual block and the prediction block are acquired, the decoding apparatus 200 may generate a reconstructed block, which is a target to be decoded, by adding the reconstructed residual block to the prediction block.
The entropy decoding unit 210 may generate symbols by performing entropy decoding on the bitstream based on the probability distribution of the bitstream. The generated symbols may comprise symbols in the form of quantized transform coefficient levels (i.e. quantized levels or quantized coefficients). Here, the entropy decoding method may be similar to the entropy encoding method described above. That is, the entropy decoding method may be the inverse process of the entropy encoding method described above.
The entropy decoding unit 210 may change coefficients having a one-dimensional (1D) vector form into a 2D block shape by a transform coefficient scanning method in order to decode quantized transform coefficient levels.
For example, the coefficients of a block may be changed to a 2D block shape by scanning the block coefficients using an upper right diagonal scan. Alternatively, which one of the upper right diagonal scan, the vertical scan, and the horizontal scan is to be used may be determined according to the size of the corresponding block and/or the intra prediction mode.
The quantized coefficients may be inverse quantized by the inverse quantization unit 220. The inverse quantization unit 220 may generate inverse quantized coefficients by performing inverse quantization on the quantized coefficients. Also, the inverse quantized coefficients may be inverse transformed by the inverse transform unit 230. The inverse transform unit 230 may generate a reconstructed residual block by performing an inverse transform on the inversely quantized coefficients. As a result of inverse quantization and inverse transformation performed on the quantized coefficients, a reconstructed residual block may be generated. Here, when generating the reconstructed residual block, the inverse quantization unit 220 may apply a quantization matrix to the quantized coefficients.
When using the intra mode, the intra prediction unit 240 may generate a prediction block by performing spatial prediction on a target block, wherein the spatial prediction uses pixel values of previously decoded neighboring blocks adjacent to the target block.
The inter prediction unit 250 may include a motion compensation unit. Alternatively, the inter prediction unit 250 may be designated as a "motion compensation unit".
When the inter mode is used, the motion compensation unit may generate the prediction block by performing motion compensation for the target block, wherein the motion compensation uses the reference image stored in the reference picture buffer 270 and the motion vector.
The motion compensation unit may apply an interpolation filter to a partial region of the reference image when the motion vector has a value other than an integer, and may generate the prediction block using the reference image to which the interpolation filter is applied. To perform motion compensation, the motion compensation unit may determine, based on the CU, which one of a skip mode, a merge mode, an Advanced Motion Vector Prediction (AMVP) mode, and a current picture reference mode corresponds to a motion compensation method for a PU included in the CU, and may perform motion compensation according to the determined mode.
The reconstructed residual block and the prediction block may be added to each other by an adder 255. The adder 255 may generate a reconstructed block by adding the reconstructed residual block and the predicted block.
The reconstructed block may be filtered by the filter unit 260. Filter unit 260 may apply at least one of a deblocking filter, SAO filter, ALF, and NLF to the reconstructed block or the reconstructed image. The reconstructed image may be a picture that includes the reconstructed block.
The filter unit may output a reconstructed image.
The reconstructed image and/or reconstructed block filtered by the filter unit 260 may be stored as a reference picture in the reference picture buffer 270. The reconstructed block filtered by the filter unit 260 may be a portion of a reference picture. In other words, the reference picture may be an image composed of the reconstructed block filtered by the filter unit 260. The stored reference pictures can then be used for inter prediction or motion compensation.
Fig. 3 is a diagram schematically showing a partition structure of an image when the image is encoded and decoded.
Fig. 3 may schematically illustrate an example in which a single cell is partitioned into a plurality of sub-cells.
In order to efficiently partition an image, a Coding Unit (CU) may be used in encoding and decoding. The term "unit" may be used to collectively specify 1) a block comprising image samples and 2) syntax elements. For example, "partition of a unit" may represent "partition of a block corresponding to the unit".
A CU can be used as a basic unit for image encoding/decoding. A CU can be used as a unit to which one mode selected from an intra mode and an inter mode is applied in image encoding/decoding. In other words, in image encoding/decoding, it may be determined which one of an intra mode and an inter mode is to be applied to each CU.
Also, a CU may be a basic unit that predicts, transforms, quantizes, inversely transforms, inversely quantizes, and encodes/decodes transform coefficients.
Referring to fig. 3, a picture 300 may be sequentially partitioned into units corresponding to maximum coding units (LCUs), and a partition structure may be determined for each LCU. Here, the LCU may be used to have the same meaning as a Coding Tree Unit (CTU).
Partitioning a unit may mean partitioning a block corresponding to the unit. The block partition information may include depth information regarding a depth of the unit. The depth information may indicate a number of times the unit is partitioned and/or a degree to which the unit is partitioned. A single unit may be hierarchically partitioned into a plurality of sub-units while the single unit has depth information based on a tree structure.
Each partitioned sub-unit may have depth information. The depth information may be information indicating a size of the CU. Depth information may be stored for each CU.
Each CU may have depth information. When a CU is partitioned, the depth of the CU generated from the partition may be increased by 1 from the depth of the partitioned CU.
The partition structure may represent the distribution of Coding Units (CUs) in the LCU 310 for efficient encoding of the image. Such a distribution may be determined according to whether a single CU is to be partitioned into multiple CUs. The number of CUs generated by partitioning may be a positive integer of 2 or more, including 2, 3, 4, 8, 16, etc.
According to the number of CUs generated by performing partitioning, the horizontal size and the vertical size of each CU generated by performing partitioning may be smaller than those of the CUs before being partitioned. For example, the horizontal and vertical dimensions of each CU generated by partitioning may be half the horizontal and vertical dimensions of the CU before partitioning.
Each partitioned CU may be recursively partitioned into four CUs in the same manner. At least one of a horizontal size and a vertical size of each partitioned CU may be reduced via recursive partitioning compared to at least one of a horizontal size and a vertical size of a CU before being partitioned.
Partitioning of CUs may be performed recursively until a predefined depth or a predefined size.
For example, the depth of a CU may have a value ranging from 0 to 3. The size of a CU may range from a size of 64 × 64 to a size of 8 × 8, depending on the depth of the CU.
For example, the depth of the LCU 310 may be 0 and the depth of the minimum coding unit (SCU) may be a predefined maximum depth. Here, as described above, the LCU may be a CU having a maximum coding unit size, and the SCU may be a CU having a minimum coding unit size.
Partitioning may begin at LCU 310, and the depth of a CU may increase by 1 each time the horizontal and/or vertical dimensions of the CU are reduced by partitioning.
For example, for each depth, a CU that is not partitioned may have a size of 2N × 2N. Further, in the case where CUs are partitioned, CUs of a size of 2N × 2N may be partitioned into four CUs each of a size of N × N. The value of N may be halved each time the depth is increased by 1.
Referring to fig. 3, an LCU having a depth of 0 may have 64 × 64 pixels or 64 × 64 blocks. 0 may be a minimum depth. An SCU of depth 3 may have 8 × 8 pixels or 8 × 8 blocks. 3 may be the maximum depth. Here, a CU having 64 × 64 blocks as an LCU may be represented by depth 0. A CU with 32 x 32 blocks may be represented with depth 1. A CU with 16 x 16 blocks may be represented with depth 2. A CU with 8 x 8 blocks as SCU can be represented by depth 3.
The information on whether the corresponding CU is partitioned or not may be represented by partition information of the CU. The partition information may be 1-bit information. All CUs except the SCU may include partition information. For example, the value of the partition information of the CU that is not partitioned may be the first value. The value of the partition information of the partitioned CU may be the second value. When the partition information indicates whether the CU is partitioned, the first value may be "0" and the second value may be "1".
For example, when a single CU is partitioned into four CUs, the horizontal and vertical sizes of each of the four CUs generated by partitioning may be half the horizontal and vertical sizes of the CU before being partitioned. When a CU having a size of 32 × 32 is partitioned into four CUs, the size of each of the partitioned four CUs may be 16 × 16. When a single CU is partitioned into four CUs, the CUs may be considered to have been partitioned in a quadtree structure. In other words, the quadtree partition may be considered to have been applied to the CU.
For example, when a single CU is partitioned into two CUs, the horizontal size or the vertical size of each of the two CUs generated by partitioning may be half the horizontal size or the vertical size of the CU before being partitioned. When a CU having a size of 32 × 32 is vertically partitioned into two CUs, the size of each of the partitioned two CUs may be 16 × 32. When a CU having a size of 32 × 32 is horizontally partitioned into two CUs, the size of each of the partitioned two CUs may be 32 × 16. When a single CU is partitioned into two CUs, the CUs may be considered to have been partitioned in a binary tree structure. In other words, the binary tree partition may be considered to have been applied to the CU.
For example, when a single CU is partitioned (or divided) into three CUs, the original CU before being partitioned is partitioned such that its horizontal or vertical size is 1: 2: the ratio of 1 is divided, thus enabling generation of three sub-CUs. For example, when a CU of size 16 × 32 is horizontally partitioned into three sub-CUs, the three sub-CUs generated by the partitioning may have sizes of 16 × 8, 16 × 16, and 16 × 8, respectively, in the direction from top to bottom. For example, when a CU of size 32 × 32 is vertically partitioned into three sub-CUs, the three sub-CUs generated by the partitioning may have sizes of 8 × 32, 16 × 32, and 8 × 32, respectively, in the left-to-right direction. When a single CU is partitioned into three CUs, the CUs may be considered to be partitioned in a ternary tree. In other words, a ternary tree partition may be considered to have been applied to a CU.
Both quad tree and binary tree partitioning are applied to LCU 310 of fig. 3.
In the encoding apparatus 100, a Coding Tree Unit (CTU) having a size of 64 × 64 may be partitioned into a plurality of smaller CUs by a recursive quadtree structure. A single CU may be partitioned into four CUs having the same size. Each CU may be recursively partitioned and may have a quadtree structure.
By recursive partitioning of CUs, the optimal partitioning method that incurs the smallest rate-distortion cost can be selected.
The Coding Tree Unit (CTU)320 in fig. 3 is an example of a CTU to which a quad tree partition, a binary tree partition, and a ternary tree partition are all applied.
As described above, in order to partition the CTU, at least one of the quadtree partition, the binary tree partition, and the ternary tree partition may be applied to the CTU. Partitions may be applied based on a particular priority.
For example, quadtree partitioning may be preferentially applied to CTUs. CUs that cannot be further partitioned in a quadtree fashion may correspond to leaf nodes of the quadtree. CUs corresponding to leaf nodes of a quadtree may be root nodes of a binary tree and/or a ternary tree. That is, CUs corresponding to leaf nodes of a quadtree may be partitioned in binary or ternary tree form, or may not be further partitioned. In this case, each CU generated by applying binary tree partitioning or ternary tree partitioning to CUs corresponding to leaf nodes of the quadtree is prevented from being partitioned again by the quadtree, thereby efficiently performing partitioning of blocks and/or signaling of block partition information.
The partition of the CU corresponding to each node of the quadtree may be signaled using the four-partition information. The four-partition information having a first value (e.g., "1") may indicate that the corresponding CU is partitioned in a quadtree form. The four-partition information having a second value (e.g., "0") may indicate that the corresponding CU is not partitioned in a quadtree form. The quad-partition information may be a flag having a specific length (e.g., 1 bit).
There may be no priority between the binary tree partition and the ternary tree partition. That is, CUs corresponding to leaf nodes of a quadtree may be partitioned in a binary tree form or a ternary tree form. Furthermore, CUs generated by binary tree partitioning or ternary tree partitioning may or may not be further partitioned in binary tree form or ternary tree form.
Partitions that are executed when there is no priority between a binary tree partition and a ternary tree partition may be referred to as "multi-type tree partitions". That is, a CU corresponding to a leaf node of a quadtree may be a root node of a multi-type tree. The partition of the CU corresponding to each node of the multi-type tree may be signaled using at least one of information indicating whether the CU is partitioned by the multi-type tree, partition direction information, and partition tree information. For the partition of the CU corresponding to each node of the multi-type tree, information indicating whether or not the partition by the multi-type tree is performed, partition direction information, and partition tree information may be sequentially signaled.
For example, the information indicating whether a CU is partitioned in a multi-type tree and has a first value (e.g., "1") may indicate that the corresponding CU is partitioned in a multi-type tree form. The information indicating whether the CU is partitioned by the multi-type tree and has a second value (e.g., "0") may indicate that the corresponding CU is not partitioned in the multi-type tree form.
When a CU corresponding to each node of the multi-type tree is partitioned in the multi-type tree form, the corresponding CU may further include partition direction information.
The partition direction information may indicate a partition direction of the multi-type tree partition. The partition direction information having a first value (e.g., "1") may indicate that the corresponding CU is partitioned in the vertical direction. The partition direction information having the second value (e.g., "0") may indicate that the corresponding CU is partitioned in the horizontal direction.
When a CU corresponding to each node of the multi-type tree is partitioned in the multi-type tree form, the corresponding CU may further include partition tree information. The partition tree information may indicate a tree that is used for multi-type tree partitioning.
For example, partition tree information having a first value (e.g., "1") may indicate that the corresponding CU is partitioned in binary tree form. The partition tree information having the second value (e.g., "0") may indicate that the corresponding CU is partitioned in a ternary tree form.
Here, each of the above-described information indicating whether partitioning by the multi-type tree is performed, the partition tree information, and the partition direction information may be a flag having a specific length (e.g., 1 bit).
At least one of the above-described four partition information, information indicating whether or not partitioning is performed per multi-type tree, partition direction information, and partition tree information may be entropy-encoded and/or entropy-decoded. To perform entropy encoding/decoding of such information, information of neighboring CUs adjacent to the target CU may be used.
For example, it may be considered that there is a high probability that the partition form (i.e., partition/non-partition, partition tree, and/or partition direction) of the left-side CU and/or the upper CU and the partition form of the target CU may be similar to each other. Thus, based on the information of neighboring CUs, context information for entropy encoding and/or entropy decoding of the information of the target CU may be derived. Here, the information of the neighboring CU may include at least one of: 1) four partition information of neighboring CUs, 2) information indicating whether the neighboring CUs are partitioned by a multi-type tree, 3) partition direction information of the neighboring CUs, and 4) partition tree information of the neighboring CUs.
In another embodiment of binary tree partitioning and ternary tree partitioning, binary tree partitioning may be performed preferentially. That is, binary tree partitioning may be applied first, and then CUs corresponding to leaf nodes of the binary tree may be set as root nodes of the ternary tree. In this case, the quadtree partitioning or the binary tree partitioning may not be performed on CUs corresponding to the nodes of the ternary tree.
CUs that are not further partitioned by quadtree partitioning, binary tree partitioning, and/or ternary tree partitioning may be units of coding, prediction, and/or transformation. That is, a CU may not be further partitioned for prediction and/or transform. Accordingly, a partition structure for partitioning a CU into Prediction Units (PUs)/or Transform Units (TUs), partition information thereof, and the like may not exist in a bitstream.
However, when the size of a CU, which is a unit of partitioning, is larger than the size of the largest transform block, the CU may be recursively partitioned until the size of the CU becomes smaller than or equal to the size of the largest transform block. For example, when the size of a CU is 64 × 64 and the size of the largest transform block is 32 × 32, the CU may be partitioned into four 32 × 32 blocks in order to perform the transform. For example, when the size of a CU is 32 × 64 and the size of the largest transform block is 32 × 32, the CU may be partitioned into two 32 × 32 blocks.
In this case, the information indicating whether a CU is partitioned for transformation may not be separately signaled. Without signaling, it may be determined whether a CU is partitioned via a comparison between the horizontal size (and/or vertical size) of the CU and the horizontal size (and/or vertical size) of the largest transform block. For example, a CU may be vertically halved when the horizontal size of the CU is larger than the horizontal size of the largest transform block. Furthermore, when the vertical size of a CU is larger than the vertical size of the largest transform block, the CU may be horizontally halved.
The information on the maximum size and/or the minimum size of the CU and the information on the maximum size and/or the minimum size of the transform block may be signaled or determined at a level higher than the level of the CU. For example, the higher level may be a sequence level, a picture level, a parallel block group level, or a stripe level. For example, the minimum size of a CU may be set to 4 × 4. For example, the maximum size of the transform block may be set to 64 × 64. For example, the maximum size of the transform block may be set to 4 × 4.
Information about a minimum size of a CU corresponding to a leaf node of the quadtree (i.e., the minimum size of the quadtree) and/or information about a maximum depth of a path from a root node of the multi-type tree to the leaf node (i.e., the maximum depth of the multi-type tree) may be signaled or determined at a level higher than that of the CU. For example, the higher level may be a sequence level, a picture level, a stripe level, a parallel block group level, or a parallel block level. Information regarding a minimum size of the quadtree and/or information regarding a maximum depth of the multi-type tree may be separately signaled or determined at each of the intra-stripe level and the inter-stripe level.
Information about the difference between the size of the CTU and the maximum size of the transform block may be signaled or determined at a level higher than the level of the CU. For example, the higher level may be a sequence level, a picture level, a slice level, a parallel block group level, or a parallel block level. Information about the maximum size of the CU corresponding to each node of the binary tree (i.e., the maximum size of the binary tree) may be determined based on the size of the CTU and the information of the difference. The maximum size of the CU corresponding to each node of the ternary tree (i.e., the maximum size of the ternary tree) may have different values according to the type of the strip. For example, the maximum size of the ternary tree at the intra-stripe level may be 32 × 32. For example, the maximum size of the tri-ary tree at the inter-band level may be 128 × 128. For example, the minimum size of the CU corresponding to each node of the binary tree (i.e., the minimum size of the binary tree) and/or the minimum size of the CU corresponding to each node of the ternary tree (i.e., the minimum size of the ternary tree) may be set to the minimum size of the CU.
In another example, the maximum size of the binary tree and/or the maximum size of the ternary tree may be signaled or determined at the slice level. Further, a minimum size of the binary tree and/or a minimum size of the ternary tree may be signaled or determined at the slice level.
Based on the various block sizes and depths described above, the four-partition information, information indicating whether partitioning by the multi-type tree is performed, partition tree information, and/or partition direction information may or may not be present in the bitstream.
For example, when the size of the CU is not greater than the minimum size of the quadtree, the CU may not include the four-partition information, and the four-partition information of the CU may be inferred to be a second value.
For example, when the size (horizontal size and vertical size) of a CU corresponding to each node of the multi-type tree is larger than the maximum size (horizontal size and vertical size) of the binary tree and/or the maximum size (horizontal size and vertical size) of the ternary tree, the CU may not be partitioned in the binary tree form and/or the ternary tree form. By this determination, the information indicating whether partitioning is performed per multi-type tree may not be signaled, but may be inferred as a second value.
Alternatively, a CU may not be partitioned in binary tree form and/or ternary tree form when the size (horizontal size and vertical size) of the CU corresponding to each node of the multi-type tree is equal to the minimum size (horizontal size and vertical size) of the binary tree, or when the size (horizontal size and vertical size) of the CU is equal to twice the minimum size (horizontal size and vertical size) of the ternary tree. By this determination, the information indicating whether partitioning is performed per multi-type tree may not be signaled, but may be inferred as the second value. The reason for this is that when a CU is partitioned in binary tree form and/or ternary tree form, a CU smaller than the minimum size of the binary tree and/or the minimum size of the ternary tree is generated.
Alternatively, binary or ternary tree partitioning may be restricted based on the size of the virtual pipeline data unit (i.e., the size of the pipeline buffer). For example, binary or ternary tree partitioning may be limited when a CU is partitioned into sub-CUs that do not fit the size of the pipeline buffer by binary or ternary tree partitioning. The size of the pipeline buffer may be equal to the maximum size of the transform block (e.g., 64 x 64).
For example, when the size of the pipeline buffer is 64 × 64, the following partitions may be restricted.
Ternary tree partitioning for nxm CU (where N and/or M are 128)
Horizontal binary tree partitioning for 128 × N CUs (where N < ═ 64)
Vertical binary tree partitioning for nx128 CU (where N < ═ 64)
Alternatively, a CU may not be partitioned in binary and/or ternary tree form when the depth of the CU corresponding to each node of the multi-type tree is equal to the maximum depth of the multi-type tree. By this determination, information indicating whether partitioning is performed per multi-type tree may be signaled but may be inferred as a second value.
Alternatively, the information indicating whether partitioning per multi-type tree is performed may be signaled only when at least one of the vertical binary tree partition, the horizontal binary tree partition, the vertical ternary tree partition, and the horizontal ternary tree partition is possible for a CU corresponding to each node of the multi-type tree. Otherwise, the CU may not be partitioned in binary tree form and/or ternary tree form. By this determination, information indicating whether partitioning is performed per multi-type tree may not be signaled, but may be inferred as a second value.
Alternatively, for a CU corresponding to each node of the multi-type tree, partition direction information may be signaled only when both vertical and horizontal binary tree partitions are feasible or only when both vertical and horizontal ternary tree partitions are feasible. Otherwise, partition direction information may be not signaled, but may be inferred as a value indicating the direction in which the CU may be partitioned.
Alternatively, for a CU corresponding to each node of the multi-type tree, partition tree information may be signaled only when both vertical binary tree partitioning and vertical ternary tree partitioning are feasible, or only when both horizontal binary tree partitioning and horizontal ternary tree partitioning are feasible. Otherwise, partition tree information may be not signaled, but may be inferred as a value indicating a tree applicable to the partitions of the CU.
Fig. 4 is a diagram illustrating a form of a prediction unit that a coding unit can include.
Among CUs partitioned from the LCU, CUs that are no longer partitioned may be divided into one or more Prediction Units (PUs). This division is also referred to as "partitioning".
A PU may be the basic unit for prediction. The PU may be encoded and decoded in any one of a skip mode, an inter mode, and an intra mode. The PU may be partitioned into various shapes according to various modes. For example, the target block described above with reference to fig. 1 and the target block described above with reference to fig. 2 may both be PUs.
A CU may not be partitioned into PUs. When a CU is not divided into PUs, the size of the CU and the size of the PU may be equal to each other.
In skip mode, there may be no partition in a CU. In the skip mode, the 2N × 2N mode 410 may be supported without partitioning, wherein the size of the PU and the size of the CU are the same as each other in the 2N × 2N mode 410.
In inter mode, there may be 8 types of partition shapes in a CU. For example, in the inter mode, a 2N × 2N mode 410, a 2N × N mode 415, an N × 2N mode 420, an N × N mode 425, a 2N × nU mode 430, a 2N × nD mode 435, an nL × 2N mode 440, and an nR × 2N mode 445 may be supported.
In intra mode, a 2N × 2N mode 410 and an N × N mode 425 may be supported.
In the 2 nx 2N mode 410, PUs of size 2 nx 2N may be encoded. A PU of size 2N × 2N may represent a PU of the same size as the CU. For example, a PU of size 2N × 2N may have a size 64 × 64, 32 × 32, 16 × 16, or 8 × 8.
In the nxn mode 425, PUs of size nxn may be encoded.
For example, in intra prediction, when the size of a PU is 8 × 8, four partitioned PUs may be encoded. The size of each partitioned PU may be 4 x 4.
When a PU is encoded in intra mode, the PU may be encoded using any one of a plurality of intra prediction modes. For example, High Efficiency Video Coding (HEVC) techniques may provide 35 intra prediction modes, a PU may be encoded in any one of the 35 intra prediction modes.
Which of the 2N × 2N mode 410 and the N × N mode 425 is to be used to encode the PU may be determined based on the rate-distortion cost.
The encoding apparatus 100 may perform an encoding operation on PUs having a size of 2N × 2N. Here, the encoding operation may be an operation of encoding the PU in each of a plurality of intra prediction modes that can be used by the encoding apparatus 100. Through the encoding operation, the optimal intra prediction mode for a PU of size 2N × 2N may be derived. The optimal intra prediction mode may be an intra prediction mode in which a minimum rate-distortion cost occurs when a PU having a size of 2N × 2N is encoded, among a plurality of intra prediction modes that can be used by the encoding apparatus 100.
Further, the encoding apparatus 100 may sequentially perform an encoding operation on the respective PUs obtained by performing the N × N partitioning. Here, the encoding operation may be an operation of encoding the PU in each of a plurality of intra prediction modes that can be used by the encoding apparatus 100. Through the encoding operation, the optimal intra prediction mode for a PU of size N × N may be derived. The optimal intra prediction mode may be an intra prediction mode in which a minimum rate-distortion cost occurs when a PU of size N × N is encoded, among a plurality of intra prediction modes that can be used by the encoding apparatus 100.
The encoding apparatus 100 may determine which one of a PU of size 2N × 2N and a PU of size N × N is to be encoded based on a comparison between a rate distortion cost of the PU of size 2N × 2N and a rate distortion cost of the PU of size N × N.
A single CU may be partitioned into one or more PUs, and a PU may be partitioned into multiple PUs.
For example, when a single PU is partitioned into four PUs, the horizontal and vertical dimensions of each of the four PUs generated by the partitioning may be half the horizontal and vertical dimensions of the PU before being partitioned. When a PU of size 32 x 32 is partitioned into four PUs, the size of each of the four partitioned PUs may be 16 x 16. When a single PU is partitioned into four PUs, the PUs may be considered to have been partitioned in a quad-tree structure.
For example, when a single PU is partitioned into two PUs, the horizontal or vertical size of each of the two PUs generated by the partitioning may be half the horizontal or vertical size of the PU before being partitioned. When a PU of size 32 x 32 is vertically partitioned into two PUs, the size of each of the two partitioned PUs may be 16 x 32. When a PU of size 32 x 32 is horizontally partitioned into two PUs, the size of each of the two partitioned PUs may be 32 x 16. When a single PU is partitioned into two PUs, the PUs may be considered to have been partitioned in a binary tree structure.
Fig. 5 is a diagram illustrating a form of a transform unit that can be included in an encoding unit.
A Transform Unit (TU) may be a basic unit used in a CU for processes such as transform, quantization, inverse transform, inverse quantization, entropy coding, and entropy decoding.
The TU may have a square shape or a rectangular shape. The shape of a TU may be determined based on the size and/or shape of the CU.
Among CUs partitioned from the LCU, CUs that are no longer partitioned into CUs may be partitioned into one or more TUs. Here, the partition structure of the TU may be a quad-tree structure. For example, as shown in fig. 5, a single CU 510 may be partitioned one or more times according to a quadtree structure. With such partitioning, a single CU 510 may be composed of TUs having various sizes.
A CU may be considered to be recursively divided when a single CU is divided two or more times. By the division, a single CU may be composed of Transform Units (TUs) having various sizes.
Alternatively, a single CU may be divided into one or more TUs based on the number of vertical and/or horizontal lines dividing the CU.
A CU may be divided into symmetric TUs or asymmetric TUs. For the division into asymmetric TUs, information regarding the size and/or shape of each TU may be signaled from the encoding apparatus 100 to the decoding apparatus 200. Alternatively, the size and/or shape of each TU may be derived from information on the size and/or shape of the CU.
A CU may not be divided into TUs. When a CU is not divided into TUs, the size of the CU and the size of the TU may be equal to each other.
A single CU may be partitioned into one or more TUs, and a TU may be partitioned into multiple TUs.
For example, when a single TU is partitioned into four TUs, the horizontal size and the vertical size of each of the four TUs generated by the partitioning may be half of the horizontal size and the vertical size of the TU before being partitioned. When a TU having a size of 32 × 32 is partitioned into four TUs, the size of each of the four partitioned TUs may be 16 × 16. When a single TU is partitioned into four TUs, the TUs may be considered to have been partitioned in a quadtree structure.
For example, when a single TU is partitioned into two TUs, the horizontal size or the vertical size of each of the two TUs generated by the partitioning may be half of the horizontal size or the vertical size of the TU before being partitioned. When a TU of a size of 32 × 32 is vertically partitioned into two TUs, each of the two partitioned TUs may be of a size of 16 × 32. When a TU having a size of 32 × 32 is horizontally partitioned into two TUs, the size of each of the two partitioned TUs may be 32 × 16. When a single TU is partitioned into two TUs, the TUs may be considered to have been partitioned in a binary tree structure.
A CU may be partitioned in a different manner than shown in fig. 5.
For example, a single CU may be divided into three CUs. The horizontal sizes or vertical sizes of the three CUs generated by the division may be 1/4, 1/2, and 1/4, respectively, of the horizontal size or vertical size of the original CU before being divided.
For example, when a CU having a size of 32 × 32 is vertically divided into three CUs, the sizes of the three CUs generated by the division may be 8 × 32, 16 × 32, and 8 × 32, respectively. In this way, when a single CU is divided into three CUs, the CU can be considered to be divided in a form of a ternary tree.
One of exemplary division forms (i.e., quadtree division, binary tree division, and ternary tree division) may be applied to the division of the CU, and a variety of division schemes may be combined and used together for the division of the CU. Here, a case where a plurality of division schemes are combined and used together may be referred to as "composite tree-like division".
Fig. 6 illustrates partitioning of blocks according to an example.
In the video encoding and/or decoding process, as shown in fig. 6, the target block may be divided. For example, the target block may be a CU.
For the division of the target block, an indicator indicating division information may be signaled from the encoding apparatus 100 to the decoding apparatus 200. The partition information may be information indicating how the target block is partitioned.
The partition information may be one or more of a partition flag (hereinafter, referred to as "split _ flag"), a quad-binary flag (hereinafter, referred to as "QB _ flag"), a quad-tree flag (hereinafter, referred to as "quadtree _ flag"), a binary tree flag (hereinafter, referred to as "binary _ flag"), and a binary type flag (hereinafter, referred to as "Btype _ flag").
The "split _ flag" may be a flag indicating whether the block is divided. For example, a split _ flag value of 1 may indicate that the corresponding block is divided. A split _ flag value of 0 may indicate that the corresponding block is not divided.
"QB _ flag" may be a flag indicating which of the quad tree form and the binary tree form corresponds to the shape in which the block is divided. For example, a QB _ flag value of 0 may indicate that the block is divided in a quad tree form. A QB _ flag value of 1 may indicate that the block is divided in a binary tree. Alternatively, a QB _ flag value of 0 may indicate that the block is divided in a binary tree form. A QB _ flag value of 1 may indicate that the block is divided in a quad tree form.
"quadtree _ flag" may be a flag indicating whether a block is divided in a quad-tree form. For example, a quadtree _ flag value of 1 may indicate that the block is divided in a quadtree form. A quadtree _ flag value of 0 may indicate that the block is not partitioned in a quadtree form.
"binarytree _ flag" may be a flag indicating whether a block is divided in a binary tree form. For example, a binarytree _ flag value of 1 may indicate that the block is divided in a binary tree form. A binarytree _ flag value of 0 may indicate that the block is not divided in a binary tree form.
"Btype _ flag" may be a flag indicating which one of the vertical division and the horizontal division corresponds to the division direction when the block is divided in the binary tree form. For example, a Btype _ flag value of 0 may indicate that the block is divided in the horizontal direction. A Btype _ flag value of 1 may indicate that the block is divided in the vertical direction. Alternatively, a Btype _ flag value of 0 may indicate that the block is divided in the vertical direction. A Btype _ flag value of 1 may indicate that the block is divided in the horizontal direction.
For example, the partition information of the block in fig. 6 may be derived by signaling at least one of quadtree _ flag, binytree _ flag, and Btype _ flag, as shown in table 1 below.
TABLE 1
Figure BDA0003783682360000361
For example, the partition information of the block in fig. 6 may be derived by signaling at least one of split _ flag, QB _ flag, and Btype _ flag, as shown in table 2 below.
TABLE 2
Figure BDA0003783682360000371
The partitioning method may be limited to only a quad tree or a binary tree depending on the size and/or shape of the block. When this restriction is applied, the split _ flag may be a flag indicating whether the block is divided in a quad tree form or a flag indicating whether the block is divided in a binary tree form. The size and shape of the block may be derived from the depth information of the block, and the depth information may be signaled from the encoding apparatus 100 to the decoding apparatus 200.
When the size of the block falls within a certain range, division in the form of only a quad tree is possible. For example, the specific range may be defined by at least one of a maximum block size and a minimum block size that can be divided only in a quad-tree form.
Information indicating the maximum block size and the minimum block size that can be divided only in the form of a quadtree may be signaled from the encoding apparatus 100 to the decoding apparatus 200 through a bitstream. Further, this information may be signaled for at least one of units such as video, sequences, pictures, parameters, parallel block groups, and stripes (or slices).
Alternatively, the maximum block size and/or the minimum block size may be a fixed size predefined by the encoding apparatus 100 and the decoding apparatus 200. For example, when the size of the block is larger than 64 × 64 and smaller than 256 × 256, only the division in the form of a quad tree is possible. In this case, the split _ flag may be a flag indicating whether to perform partitioning in the form of a quadtree.
When the size of the block is larger than the maximum size of the transform block, only partitioning in the form of a quadtree is possible. Here, the sub-blocks generated by the partitioning may be at least one of the CU and the TU.
In this case, split _ flag may be a flag indicating whether partitioning in the form of a quad-tree is performed.
When the size of the block falls within a specific range, division in only a binary tree form or a ternary tree form is possible. For example, the specific range may be defined by at least one of a maximum block size and a minimum block size that can be divided only in a binary tree form or a ternary tree form.
Information indicating the maximum block size and/or the minimum block size that can be divided only in a binary tree form or in a ternary tree form may be signaled from the encoding apparatus 100 to the decoding apparatus 200 through a bitstream. Further, this information may be signaled for at least one of the units such as sequence, picture, and slice (or slice).
Alternatively, the maximum block size and/or the minimum block size may be a fixed size predefined by the encoding apparatus 100 and the decoding apparatus 200. For example, when the size of the block is larger than 8 × 8 and smaller than 16 × 16, only division in a binary tree form is possible. In this case, the split _ flag may be a flag indicating whether to perform partitioning in a binary tree form or a ternary tree form.
The above description of partitioning in a quadtree form can be equally applied to a binary tree form and/or a ternary tree form.
The partitioning of a block may be limited by previous partitions. For example, when a block is partitioned in a specific binary tree form and a plurality of sub-blocks are generated from the partition, each sub-block may be additionally partitioned only in a specific tree form. Here, the specific tree form may be at least one of a binary tree form, a ternary tree form, and a quaternary tree form.
The indicator may not be signaled when the horizontal size or the vertical size of the partition block is a size that cannot be further divided.
Fig. 7 is a diagram for explaining an embodiment of an intra prediction process.
The arrows extending radially from the center of the graph in fig. 7 indicate the prediction directions of the directional intra prediction modes. Further, numbers appearing near the arrows indicate examples of mode values assigned to the intra prediction mode or the prediction direction of the intra prediction mode.
In fig. 7, the number "0" may represent a planar mode as a non-directional intra prediction mode. The number "1" may represent a DC mode as a non-directional intra prediction mode.
Intra-coding and/or decoding may be performed using reference samples of neighboring units of the target block. The neighboring blocks may be reconstructed neighboring blocks. The reference samples may represent neighboring samples.
For example, intra-coding and/or decoding may be performed using values of reference samples included in the reconstructed neighboring blocks or encoding parameters of the reconstructed neighboring blocks.
The encoding apparatus 100 and/or the decoding apparatus 200 may generate the prediction block by performing intra prediction on the target block based on the information on the sampling points in the target image. When the intra prediction is performed, the encoding apparatus 100 and/or the decoding apparatus 200 may generate a prediction block for the target block by performing the intra prediction based on the information on the sampling points in the target image. When the intra prediction is performed, the encoding apparatus 100 and/or the decoding apparatus 200 may perform directional prediction and/or non-directional prediction based on the at least one reconstructed reference sample.
The prediction block may be a block generated as a result of performing intra prediction. The prediction block may correspond to at least one of a CU, a PU, and a TU.
The units of the prediction block may have a size corresponding to at least one of the CU, the PU, and the TU. The prediction block may have a square shape with a size of 2N × 2N or N × N. The size N × N may include sizes 4 × 4, 8 × 8, 16 × 16, 32 × 32, 64 × 64, and so on.
Alternatively, the prediction block may be a square block having a size of 2 × 2, 4 × 4, 8 × 8, 16 × 16, 32 × 32, 64 × 64, or the like or a rectangular block having a size of 2 × 8, 4 × 8, 2 × 16, 4 × 16, 8 × 16, or the like.
The intra prediction may be performed in consideration of an intra prediction mode for the target block. The number of intra prediction modes that the target block may have may be a predefined fixed value, and may be a value differently determined according to the properties of the prediction block. For example, the properties of the prediction block may include the size of the prediction block, the type of prediction block, and the like. Furthermore, the properties of the prediction block may indicate the coding parameters used for the prediction block.
For example, the number of intra prediction modes may be fixed to N regardless of the size of the prediction block. Alternatively, the number of intra prediction modes may be, for example, 3, 5, 9, 17, 34, 35, 36, 65, 67, or 95.
The intra prediction mode may be a non-directional mode or a directional mode.
For example, the intra prediction modes may include two non-directional modes and 65 directional modes corresponding to numbers 0 to 66 shown in fig. 7.
For example, in the case of using a specific intra prediction method, the intra prediction modes may include two non-directional modes corresponding to numbers-14 to 80 shown in fig. 7 and 93 directional modes.
The two non-directional modes may include a DC mode and a planar mode.
The directional mode may be a prediction mode having a specific direction or a specific angle. The directional mode may also be referred to as an "angular mode".
The intra prediction mode may be represented by at least one of a mode number, a mode value, a mode angle, and a mode direction. In other words, the terms "a (mode) number of an intra prediction mode", "a (mode) value of an intra prediction mode", "a (mode) angle of an intra prediction mode", and "a (mode) direction of an intra prediction mode" may be used to have the same meaning, and may be used interchangeably with each other.
The number of intra prediction modes may be M. The value of M may be 1 or greater. In other words, the number of intra prediction modes may be M, where M includes the number of non-directional modes and the number of directional modes.
The number of intra prediction modes may be fixed to M regardless of the size and/or color components of the block. For example, the number of intra prediction modes may be fixed to any one of 35 and 67 regardless of the size of the block.
Alternatively, the number of intra prediction modes may be different according to the shape, size, and/or type of color component of the block.
For example, in fig. 7, the directional prediction mode as shown by the dotted line may be applied only to prediction for non-square blocks.
For example, the larger the size of the block, the larger the number of intra prediction modes. Alternatively, the larger the size of the block, the smaller the number of intra prediction modes. When the size of the block is 4 × 4 or 8 × 8, the number of intra prediction modes may be 67. When the size of the block is 16 × 16, the number of intra prediction modes may be 35. When the size of the block is 32 × 32, the number of intra prediction modes may be 19. When the size of the block is 64 × 64, the number of intra prediction modes may be 7.
For example, the number of intra prediction modes may be different according to whether a color component is a luminance signal or a chrominance signal. Alternatively, the number of intra prediction modes corresponding to the luminance component block may be greater than the number of intra prediction modes corresponding to the chrominance component block.
For example, in the vertical mode having a mode value of 50, prediction may be performed in the vertical direction based on the pixel value of the reference sampling point. For example, in the horizontal mode with the mode value of 18, prediction may be performed in the horizontal direction based on the pixel value of the reference sampling point.
Even in a directional mode other than the above-described modes, the encoding apparatus 100 and the decoding apparatus 200 may perform intra prediction on a target unit using reference samples according to an angle corresponding to the directional mode.
The intra prediction mode located at the right side with respect to the vertical mode may be referred to as a "vertical-right mode". The intra prediction mode located below the horizontal mode may be referred to as a "horizontal-below mode". For example, in fig. 7, the intra prediction mode having one of the mode values 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, and 66 may be a vertical-right mode. The intra prediction mode having a mode value of one of 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, and 17 may be a horizontal-lower mode.
The non-directional mode may include a DC mode and a planar mode. For example, the value of the DC mode may be 1. The value of the planar mode may be 0.
The directional pattern may include an angular pattern. Among the plurality of intra prediction modes, the remaining modes except for the DC mode and the planar mode may be directional modes.
When the intra prediction mode is the DC mode, the prediction block may be generated based on an average value of pixel values of the plurality of reference pixels. For example, values of pixels of the prediction block may be determined based on an average of pixel values of a plurality of reference pixels.
The number of intra prediction modes and the mode values of the respective intra prediction modes described above are merely exemplary. The number of intra prediction modes described above and the mode values of the respective intra prediction modes may be defined differently according to embodiments, implementations, and/or requirements.
In order to perform intra prediction on the target block, a step of checking whether or not a sample included in the reconstructed neighboring block can be used as a reference sample of the target block may be performed. When there are samples that cannot be used as reference samples of the target block among samples in the neighboring block, a value generated via interpolation and/or duplication using at least one sample value among samples included in the reconstructed neighboring block may replace sample values of samples that cannot be used as reference samples. When a value generated via replication and/or interpolation replaces a sample value of an existing sample, the sample may be used as a reference sample for the target block.
When intra prediction is used, a filter may be applied to at least one of the reference samples and the prediction samples based on at least one of the size of the target block and the intra prediction mode.
The type of the filter to be applied to at least one of the reference samples and the prediction samples may be different according to at least one of an intra prediction mode of the target block, a size of the target block, and a shape of the target block. The type of filter may be classified according to one or more of the length of the filter tap, the value of the filter coefficient, and the filter strength. The length of the filter taps may represent the number of filter taps. Further, the number of filter taps may represent the length of the filter.
When the intra prediction mode is the planar mode, the sample value of the prediction target block may be generated using a weighted sum of the upper reference sample of the target block, the left reference sample of the target block, the upper right reference sample of the target block, and the lower left reference sample of the target block according to the position of the prediction target sample in the prediction block when generating the prediction block of the target block.
When the intra prediction mode is the DC mode, an average value of the reference samples above the target block and the reference samples to the left of the target block may be used in generating the prediction block of the target block. Further, filtering using the value of the reference sampling point may be performed on a specific row or a specific column in the target block. The particular row may be one or more upper rows adjacent to the reference sample point. The particular column may be one or more left-hand columns adjacent to the reference sample point.
When the intra prediction mode is a directional mode, the prediction block may be generated using the upper reference sample, the left reference sample, the upper right reference sample, and/or the lower left reference sample of the target block.
To generate the predicted samples described above, real-based interpolation may be performed.
The intra prediction mode of the target block may be predicted from intra prediction modes of neighboring blocks adjacent to the target block, and information for prediction may be entropy-encoded/entropy-decoded.
For example, when the intra prediction modes of the target block and the neighboring block are identical to each other, the intra prediction modes of the target block and the neighboring block may be signaled to be identical using a predefined flag.
For example, an indicator indicating the same intra prediction mode as that of the target block among intra prediction modes of a plurality of neighboring blocks may be signaled.
When the intra prediction modes of the target block and the neighboring block are different from each other, information regarding the intra prediction mode of the target block may be encoded and/or decoded using entropy encoding and/or entropy decoding.
Fig. 8 is a diagram illustrating reference samples used in an intra prediction process.
The reconstructed reference samples for intra prediction of the target block may include a lower left reference sample, a left reference sample, an upper right reference sample, and an upper right reference sample.
For example, the left reference sample point may represent a reconstructed reference pixel adjacent to the left side of the target block. The upper reference sample point may represent a reconstructed reference pixel adjacent to the top of the target block. The upper left reference sample point may represent a reconstructed reference pixel located at the upper left corner of the target block. The lower-left reference sampling point may represent a reference sampling point located below a left side sampling point line composed of the left reference sampling points among sampling points located on the same line as the left side sampling point line. The upper right reference sampling point may represent a reference sampling point located at the right side of an upper sampling point line composed of the upper reference sampling points among sampling points located on the same line as the upper sampling point line.
When the size of the target block is N × N, the numbers of the lower-left reference samples, the upper reference samples, and the upper-right reference samples may all be N.
By performing intra prediction on the target block, a prediction block may be generated. The process of generating the prediction block may include determining values of pixels in the prediction block. The sizes of the target block and the prediction block may be the same.
The reference sampling point used for intra prediction of the target block may be changed according to the intra prediction mode of the target block. The direction of the intra prediction mode may represent a dependency between the reference samples and the pixels of the prediction block. For example, a value specifying a reference sample may be used as a value of one or more specified pixels in the prediction block. In this case, the specified reference samples and the one or more specified pixels in the prediction block may be samples and pixels located on a straight line along a direction of the intra prediction mode. In other words, the value of the specified reference sample point may be copied as the value of the pixel located in the direction opposite to the direction of the intra prediction mode. Alternatively, the value of a pixel in the prediction block may be a value of a reference sample point located in the direction of the intra prediction mode with respect to the position of the pixel.
In an example, when the intra prediction mode of the target block is a vertical mode, the above-reference samples may be used for intra prediction. When the intra prediction mode is a vertical mode, the value of a pixel in the prediction block may be a value of a reference sample point vertically located above the position of the pixel. Therefore, the upper reference samples adjacent to the top of the target block may be used for intra prediction. In addition, the values of pixels in a row of the prediction block may be the same as those of the pixels of the upper reference sample point.
In an example, when the intra prediction mode of the target block is a horizontal mode, the left reference sample may be used for intra prediction. When the intra prediction mode is a horizontal mode, the value of a pixel in the prediction block may be a value of a reference sample point horizontally located to the left of the position of the pixel. Therefore, the left reference samples adjacent to the left side of the target block may be used for intra prediction. Furthermore, the values of pixels in a column of the prediction block may be the same as the values of pixels of the left reference sample point.
In an example, when a mode value of an intra prediction mode of the current block is 34, at least some of the left reference samples, the upper-left corner reference samples, and at least some of the upper reference samples may be used for intra prediction. When the mode value of the intra prediction mode is 18, the value of a pixel in the prediction block may be a value of a reference sample point located diagonally at an upper left corner of the pixel.
Further, in the case of an intra prediction mode in which the mode value is a value ranging from 52 to 66, at least a portion of the upper-right reference samples may be used for intra prediction.
Further, in the case of an intra prediction mode in which the mode value is a value ranging from 2 to 17, at least a part of the lower left reference sample may be used for intra prediction.
Further, in the case of an intra prediction mode in which the mode value is a value ranging from 19 to 49, the upper left reference sample may be used for intra prediction.
The number of reference samples used to determine the pixel value of one pixel in the prediction block may be 1 or 2 or more.
As described above, the pixel values of the pixels in the prediction block may be determined according to the positions of the pixels and the positions of the reference samples indicated by the direction of the intra prediction mode. When the position of the pixel and the position of the reference sample point indicated by the direction of the intra prediction mode are integer positions, the value of one reference sample point indicated by the integer position may be used to determine the pixel value of the pixel in the prediction block.
When the position of the pixel and the position of the reference sample point indicated by the direction of the intra prediction mode are not integer positions, an interpolated reference sample point based on two reference sample points closest to the position of the reference sample point may be generated. The values of the interpolated reference samples may be used to determine pixel values for pixels in the prediction block. In other words, when the position of the pixel in the prediction block and the position of the reference sample point indicated by the direction of the intra prediction mode indicate a position between two reference sample points, an interpolation based on the values of the two sample points may be generated.
The prediction block generated via prediction may be different from the original target block. In other words, there may be a prediction error, which is a difference between the target block and the prediction block, and there may also be a prediction error between pixels of the target block and pixels of the prediction block.
Hereinafter, the terms "difference", "error" and "residual" may be used to have the same meaning and may be used interchangeably with each other.
For example, in the case of directional intra prediction, the longer the distance between the pixels of the predicted block and the reference sample, the larger the prediction error that may occur. Such a prediction error may cause discontinuity between the generated prediction block and the neighboring block.
To reduce the prediction error, a filtering operation for the prediction block may be used. The filtering operation may be configured to adaptively apply a filter to a region in the prediction block that is considered to have a large prediction error. For example, a region considered to have a large prediction error may be a boundary of a prediction block. In addition, regions that are considered to have a large prediction error in a prediction block may be different according to an intra prediction mode, and characteristics of a filter may also be different according to the intra prediction mode.
As shown in fig. 8, for intra prediction of a target block, at least one of reference line 0 to reference line 3 may be used. Each reference line may indicate a reference sample line. When the number of reference lines is smaller, a reference sample line closer to the target block may be indicated.
The samples in segment a and segment F may be obtained by padding using the samples in segment B and segment E that are closest to the target block, rather than from reconstructed neighboring blocks.
Index information indicating a reference sample line to be used for intra prediction of a target block may be signaled. The index information may indicate a reference sample line of the plurality of reference sample lines to be used for intra prediction of the target block. For example, the index information may have a value corresponding to any one of 0 to 3.
When the upper boundary of the target block is the boundary of the CTU, only the reference sample line 0 may be available. Therefore, in this case, the index information may not be signaled. When an additional reference sample line other than the reference sample line 0 is used, filtering of a prediction block, which will be described later, may not be performed.
In the case of inter-color intra prediction, a prediction block of a target block of a second color component may be generated based on a corresponding reconstructed block of a first color component.
For example, the first color component may be a luminance component and the second color component may be a chrominance component.
To perform inter-color intra prediction, parameters of a linear model between the first color component and the second color component may be derived based on the template.
The template may include reference samples (upper reference samples) above the target block and/or reference samples (left reference samples) to the left of the target block, and may include upper reference samples and/or left reference samples of a reconstructed block of the first color component corresponding to the reference samples.
For example, the following values may be used to derive the parameters of the linear model: 1) a value of a sample point of a first color component having a maximum value among sample points in the template, 2) a value of a sample point of a second color component corresponding to a sample point of the first color component, 3) a value of a sample point of a first color component having a minimum value among sample points in the template, and 4) a value of a sample point of a second color component corresponding to a sample point of the first color component.
When the parameters of the linear model are derived, the prediction block of the target block may be generated by applying the corresponding reconstructed block to the linear model.
According to the image format, subsampling may be performed on samples adjacent to the reconstructed block of the first color component and the corresponding reconstructed block of the first color component. For example, when one sampling point of the second color component corresponds to four sampling points of the first color component, one corresponding sampling point may be calculated by performing sub-sampling on the four sampling points of the first color component. When performing sub-sampling, derivation of parameters of the linear model and inter-color intra prediction may be performed based on the sub-sampled corresponding sampling points.
Information regarding whether to perform inter-color intra prediction and/or the range of templates may be signaled in the intra prediction mode.
The target block may be partitioned into two or four sub-blocks in the horizontal direction and/or the vertical direction.
The sub-blocks generated by the partitioning may be sequentially reconstructed. That is, when intra prediction is performed on each sub block, sub prediction blocks of the sub blocks may be generated. Further, when inverse quantization (inverse quantization) and/or inverse transformation is performed on each sub-block, a sub-residual block for the corresponding sub-block may be generated. The reconstructed sub-block may be generated by adding the sub-prediction block to the sub-residual block. The reconstructed sub-block may be used as a reference sample point for intra prediction of a sub-block having a next priority.
A sub-block may be a block that includes a certain number (e.g., 16) or more samples. For example, when the target block is an 8 × 4 block or a 4 × 8 block, the target block may be partitioned into two sub-blocks. Further, when the target block is a 4 × 4 block, the target block cannot be partitioned into sub-blocks. When the target block has another size, the target block may be partitioned into four sub-blocks.
Information on whether to perform intra prediction based on the sub-blocks and/or information on a partition direction (horizontal direction or vertical direction) may be signaled.
Such sub-block based intra prediction may be limited such that it is only performed when the reference sample line 0 is used. When the sub-block-based intra prediction is performed, filtering of a prediction block, which will be described below, may not be performed.
The final prediction block may be generated by performing filtering on the prediction block generated through intra prediction.
The filtering may be performed by applying a specific weight to the filtering target samples, the left reference samples, the upper reference samples, and/or the upper left reference samples, which are targets to be filtered.
The weight for filtering and/or the reference samples (e.g., the range of the reference samples, the location of the reference samples, etc.) may be determined based on at least one of the block size, the intra prediction mode, and the location of the filtering target samples in the prediction block.
For example, the filtering may be performed only in a specific intra prediction mode (e.g., DC mode, planar mode, vertical mode, horizontal mode, diagonal mode, and/or adjacent diagonal mode).
The adjacent diagonal patterns may be patterns having numbers obtained by adding k to the numbers of the diagonal patterns, and may be patterns having numbers obtained by subtracting k from the numbers of the diagonal patterns. In other words, the number of the adjacent diagonal patterns may be the sum of the number of the diagonal patterns and k, or may be the difference between the number of the diagonal patterns and k. For example, k may be a positive integer of 8 or less.
The intra prediction mode of the target block may be derived using intra prediction modes of neighboring blocks existing near the target block, and such derived intra prediction modes may be entropy-encoded and/or entropy-decoded.
For example, when the intra prediction mode of the target block is the same as the intra prediction modes of the neighboring blocks, information indicating that the intra prediction mode of the target block is the same as the intra prediction modes of the neighboring blocks may be signaled using the specific flag information.
Also, for example, indicator information of neighboring blocks of which intra prediction modes are the same as the intra prediction mode of the target block among the intra prediction modes of the plurality of neighboring blocks may be signaled.
For example, when the intra prediction mode of the target block is different from the intra prediction modes of the neighboring blocks, entropy encoding and/or entropy decoding may be performed on information regarding the intra prediction mode of the target block by performing entropy encoding and/or entropy decoding based on the intra prediction modes of the neighboring blocks.
Fig. 9 is a diagram for explaining an embodiment of an inter prediction process.
The rectangle shown in fig. 9 may represent an image (or picture). In addition, in fig. 9, an arrow may indicate a prediction direction. The arrow pointing from the first picture to the second picture indicates that the second picture refers to the first picture. That is, each image may be encoded and/or decoded according to a prediction direction.
Images can be classified into an intra picture (I picture), a mono-predictive picture or a predictive coded picture (P picture), and a bi-predictive picture or a bi-predictive coded picture (B picture) according to coding types. Each picture may be encoded and/or decoded according to its coding type.
When the target image that is the target to be encoded is an I picture, the target image can be encoded using data contained in the image itself without performing inter prediction with reference to other images. For example, an I picture may be encoded via intra prediction only.
When the target image is a P picture, the target image may be encoded via inter prediction using a reference picture existing in one direction. Here, the one direction may be a forward direction or a backward direction.
When the target image is a B picture, the image may be encoded via inter prediction using reference pictures existing in both directions, or may be encoded via inter prediction using reference pictures existing in one of a forward direction and a backward direction. Here, the two directions may be a forward direction and a backward direction.
P-pictures and B-pictures encoded and/or decoded using reference pictures may be considered images using inter prediction.
Hereinafter, inter prediction in the inter mode according to the embodiment will be described in detail.
Inter prediction or motion compensation may be performed using the reference picture and the motion information.
In the inter mode, the encoding apparatus 100 may perform inter prediction and/or motion compensation on the target block. The decoding apparatus 200 may perform inter prediction and/or motion compensation corresponding to the inter prediction and/or motion compensation performed by the encoding apparatus 100 on the target block.
The motion information of the target block may be separately derived by the encoding apparatus 100 and the decoding apparatus 200 during inter prediction. The motion information may be derived using motion information of reconstructed neighboring blocks, motion information of a col block, and/or motion information of blocks adjacent to the col block.
For example, the encoding apparatus 100 or the decoding apparatus 200 may perform prediction and/or motion compensation by using motion information of a spatial candidate and/or a temporal candidate as motion information of a target block. The target blocks may represent PUs and/or PU partitions.
The spatial candidate may be a reconstructed block spatially adjacent to the target block.
The temporal candidate may be a reconstructed block corresponding to the target block in a previously reconstructed co-located picture (col picture).
In the inter prediction, the encoding apparatus 100 and the decoding apparatus 200 may improve encoding efficiency and decoding efficiency by using motion information of spatial candidates and/or temporal candidates. The motion information of the spatial candidates may be referred to as "spatial motion information". The motion information of the temporal candidates may be referred to as "temporal motion information".
Next, the motion information of the spatial candidate may be the motion information of the PU including the spatial candidate. The motion information of the temporal candidate may be the motion information of the PU including the temporal candidate. The motion information of the candidate block may be motion information of a PU that includes the candidate block.
Inter prediction may be performed using a reference picture.
The reference picture may be at least one of a picture preceding the target picture and a picture following the target picture. The reference picture may be an image used for prediction of the target block.
In inter prediction, a region in a reference picture may be specified using a reference picture index (or refIdx) indicating the reference picture, a motion vector to be described later, or the like. Here, the area specified in the reference picture may indicate a reference block.
Inter prediction may select a reference picture, and may also select a reference block corresponding to the target block from the reference picture. Further, inter prediction may generate a prediction block for a target block using the selected reference block.
The motion information may be derived by each of the encoding apparatus 100 and the decoding apparatus 200 during inter prediction.
The spatial candidates may be 1) blocks that exist in the target picture that 2) have been previously reconstructed via encoding and/or decoding and 3) are adjacent to the target block or located at corners of the target block. Here, the "block located at a corner of the target block" may be a block vertically adjacent to an adjacent block horizontally adjacent to the target block, or a block horizontally adjacent to an adjacent block vertically adjacent to the target block. Further, "a block located at a corner of the target block" may have the same meaning as "a block adjacent to the corner of the target block". The meaning of "a block located at a corner of a target block" may be included in the meaning of "a block adjacent to the target block".
For example, the spatial candidate may be a reconstructed block located to the left of the target block, a reconstructed block located above the target block, a reconstructed block located in the lower left corner of the target block, a reconstructed block located in the upper right corner of the target block, or a reconstructed block located in the upper left corner of the target block.
Each of the encoding apparatus 100 and the decoding apparatus 200 can identify a block existing in a position spatially corresponding to a target block in a col picture. The position of the target block in the target picture and the position of the identified block in the col picture may correspond to each other.
Each of the encoding apparatus 100 and the decoding apparatus 200 may determine, as a time candidate, a col block existing at a predefined correlation position with respect to the identified block. The predefined relative location may be a location that exists inside and/or outside the identified block.
For example, the col blocks may include a first col block and a second col block. When the coordinates of the identified block are (xP, yP) and the size of the identified block is represented by (nPSW, nPSH), the first col block may be a block located at coordinates (xP + nPSW, yP + nPSH). The second col block may be a block located at coordinates (xP + (nPSW > >1), yP + (nPSH > > 1)). When the first col block is not available, the second col block may be selectively used.
The motion vector of the target block may be determined based on the motion vector of the col block. Each of the encoding apparatus 100 and the decoding apparatus 200 may scale the motion vector of the col block. The scaled motion vector of the col block can be used as the motion vector of the target block. Further, the motion vector of the motion information of the temporal candidate stored in the list may be a scaled motion vector.
The ratio of the motion vector of the target block relative to the motion vector of the col block may be the same as the ratio of the first temporal distance relative to the second temporal distance. The first temporal distance may be a distance between the reference picture and a target picture of the target block. The second temporal distance may be a distance between the reference picture and a col picture of the col block.
The scheme for deriving the motion information may vary according to the inter prediction mode of the target block. For example, as an inter prediction mode applied to inter prediction, there may be an Advanced Motion Vector Predictor (AMVP) mode, a merge mode, a skip mode, a merge mode with a motion vector difference, a sub-block merge mode, a triangle partition mode, an inter-intra combined prediction mode, an affine inter mode, a current picture reference mode, and the like. The merge mode may also be referred to as a "motion merge mode". Each mode will be described in detail below.
1) AMVP mode
When using the AMVP mode, the encoding apparatus 100 may search for similar blocks in a neighboring area of a target block. The encoding apparatus 100 may acquire a prediction block by performing prediction on a target block using motion information of the found similar block. The encoding apparatus 100 may encode a residual block that is a difference between the target block and the prediction block.
1-1) creating a list of predicted motion vector candidates
When the AMVP mode is used as the prediction mode, each of the encoding apparatus 100 and the decoding apparatus 200 may create a list of predicted motion vector candidates using a motion vector of a spatial candidate, a motion vector of a temporal candidate, and a zero vector. The predicted motion vector candidate list may include one or more predicted motion vector candidates. At least one of a motion vector of the spatial candidate, a motion vector of the temporal candidate, and a zero vector may be determined and used as the prediction motion vector candidate.
Hereinafter, the terms "prediction motion vector (candidate)" and "motion vector (candidate)" may be used to have the same meaning and may be used interchangeably with each other.
Hereinafter, the terms "prediction motion vector candidate" and "AMVP candidate" may be used to have the same meaning and may be used interchangeably with each other.
Hereinafter, the terms "prediction motion vector candidate list" and "AMVP candidate list" may be used to have the same meaning and may be used interchangeably with each other.
The spatial candidates may comprise reconstructed spatially neighboring blocks. In other words, the motion vectors of the reconstructed neighboring blocks may be referred to as "spatial prediction motion vector candidates".
The temporal candidates may include a col block and blocks adjacent to the col block. In other words, a motion vector of a col block or a motion vector of a block adjacent to the col block may be referred to as a "temporal prediction motion vector candidate".
The zero vector may be a (0,0) motion vector.
The predicted motion vector candidate may be a motion vector predictor for predicting a motion vector. Further, in the encoding apparatus 100, each predicted motion vector candidate may be an initial search position for a motion vector.
1-2) searching for motion vector using list of predicted motion vector candidates
The encoding apparatus 100 may determine a motion vector to be used for encoding the target block within the search range using the list of predicted motion vector candidates. Further, the encoding apparatus 100 may determine a predicted motion vector candidate to be used as the predicted motion vector of the target block among the predicted motion vector candidates existing in the predicted motion vector candidate list.
The motion vector to be used for encoding the target block may be a motion vector that can be encoded at a minimum cost.
In addition, the encoding apparatus 100 may determine whether to encode the target block using the AMVP mode.
1-3) Transmission of inter-frame prediction information
The encoding apparatus 100 may generate a bitstream including inter prediction information required for inter prediction. The decoding apparatus 200 may perform inter prediction on the target block using inter prediction information of the bitstream.
The inter prediction information may include 1) mode information indicating whether the AMVP mode is used, 2) a prediction motion vector index, 3) a Motion Vector Difference (MVD), 4) a reference direction, and 5) a reference picture index.
Hereinafter, the terms "prediction motion vector index" and "AMVP index" may be used to have the same meaning and may be used interchangeably with each other.
In addition, the inter prediction information may include a residual signal.
When the mode information indicates that the AMVP mode is used, the decoding apparatus 200 may acquire a prediction motion vector index, an MVD, a reference direction, and a reference picture index from the bitstream through entropy decoding.
The prediction motion vector index may indicate a prediction motion vector candidate to be used for predicting the target block among prediction motion vector candidates included in the prediction motion vector candidate list.
1-4) inter prediction in AMVP mode using inter prediction information
The decoding apparatus 200 may derive a predicted motion vector candidate using the predicted motion vector candidate list, and may determine motion information of the target block based on the derived predicted motion vector candidate.
The decoding apparatus 200 may determine a motion vector candidate for the target block among the predicted motion vector candidates included in the predicted motion vector candidate list using the predicted motion vector index. The decoding apparatus 200 may select a predicted motion vector candidate indicated by the predicted motion vector index as the predicted motion vector of the target block from among the predicted motion vector candidates included in the predicted motion vector candidate list.
The encoding apparatus 100 may generate an entropy-encoded prediction motion vector index by applying entropy encoding to the prediction motion vector index, and may generate a bitstream including the entropy-encoded prediction motion vector index. The entropy-encoded prediction motion vector index may be signaled from the encoding apparatus 100 to the decoding apparatus 200 through a bitstream. The decoding apparatus 200 may extract an entropy-encoded prediction motion vector index from a bitstream, and may acquire the prediction motion vector index by applying entropy decoding to the entropy-encoded prediction motion vector index.
The motion vector that is actually to be used for inter prediction of the target block may not match the predicted motion vector. To indicate the difference between the motion vector that will actually be used for inter-predicting the target block and the predicted motion vector, MVD may be used. The encoding apparatus 100 may derive a prediction motion vector similar to a motion vector that will actually be used for inter-prediction of the target block in order to use an MVD as small as possible.
The MVD may be the difference between the motion vector of the target block and the predicted motion vector. The encoding apparatus 100 may calculate an MVD and may generate an entropy-encoded MVD by applying entropy encoding to the MVD. The encoding apparatus 100 may generate a bitstream including the entropy-encoded MVDs.
The MVD may be transmitted from the encoding apparatus 100 to the decoding apparatus 200 through a bitstream. The decoding apparatus 200 may extract entropy-encoded MVDs from the bitstream and may acquire the MVDs by applying entropy decoding to the entropy-encoded MVDs.
The decoding apparatus 200 may derive a motion vector of the target block by summing the MVD and the prediction motion vector. In other words, the motion vector of the target block derived by the decoding apparatus 200 may be the sum of the MVD and the motion vector candidate.
Also, the encoding apparatus 100 may generate entropy-encoded MVD resolution information by applying entropy encoding to the calculated MVD resolution information, and may generate a bitstream including the entropy-encoded MVD resolution information. The decoding apparatus 200 may extract entropy-encoded MVD resolution information from a bitstream, and may acquire the MVD resolution information by applying entropy decoding to the entropy-encoded MVD resolution information. The decoding apparatus 200 may adjust the resolution of the MVD using the MVD resolution information.
In addition, the encoding apparatus 100 may calculate the MVD based on an affine model. The decoding apparatus 200 may derive an affine control motion vector of the target block through the sum of the MVD and the affine control motion vector candidate, and may derive a motion vector of the sub-block using the affine control motion vector.
The reference direction may indicate a list of reference pictures to be used for predicting the target block. For example, the reference direction may indicate one of the reference picture list L0 and the reference picture list L1.
The reference direction indicates only a reference picture list to be used for prediction of the target block, and may not mean that the direction of the reference picture is limited to a forward direction or a backward direction. In other words, each of the reference picture list L0 and the reference picture list L1 may include pictures in the forward direction and/or the backward direction.
The reference direction being unidirectional may mean that a single reference picture list is used. The reference direction being bi-directional may mean that two reference picture lists are used. In other words, the reference direction may indicate one of the following: the case of using only the reference picture list L0, the case of using only the reference picture list L1, and the case of using two reference picture lists.
The reference picture index may indicate a reference picture for the prediction target block among reference pictures existing in the reference picture list. The encoding apparatus 100 may generate an entropy-encoded reference picture index by applying entropy encoding to the reference picture index, and may generate a bitstream including the entropy-encoded reference picture index. The entropy-encoded reference picture index may be signaled from the encoding apparatus 100 to the decoding apparatus 200 through a bitstream. The decoding apparatus 200 may extract an entropy-encoded reference picture index from a bitstream, and may acquire the reference picture index by applying entropy decoding to the entropy-encoded reference picture index.
When two reference picture lists are used for prediction of a target block, a single reference picture index and a single motion vector may be used for each of the reference picture lists. Further, when two reference picture lists are used for predicting the target block, two prediction blocks may be specified for the target block. For example, an average or a weighted sum of two prediction blocks for a target block may be used to generate a (final) prediction block for the target block.
The motion vector of the target block may be derived by predicting a motion vector index, an MVD, a reference direction, and a reference picture index.
The decoding apparatus 200 may generate a prediction block for the target block based on the derived motion vector and the reference picture index. For example, the prediction block may be a reference block indicated by a derived motion vector in a reference picture indicated by a reference picture index.
Since the prediction motion vector index and the MVD are encoded while the motion vector of the target block itself is not encoded, the number of bits transmitted from the encoding apparatus 100 to the decoding apparatus 200 may be reduced and the encoding efficiency may be improved.
For the target block, motion information of the reconstructed neighboring blocks may be used. In a specific inter prediction mode, the encoding apparatus 100 may not encode actual motion information of the target block separately. The motion information of the target block is not encoded, but additional information that enables the motion information of the target block to be derived using the reconstructed motion information of the neighboring blocks may be encoded. Since the additional information is encoded, the number of bits transmitted to the decoding apparatus 200 may be reduced and encoding efficiency may be improved.
For example, as an inter prediction mode in which motion information of a target block is not directly encoded, a skip mode and/or a merge mode may exist. Here, each of the encoding apparatus 100 and the decoding apparatus 200 may use an identifier and/or an index indicating a unit of which motion information is to be used as motion information of the target unit among the reconstructed neighboring units.
2) Merge mode
As a scheme for deriving motion information of a target block, there is merging. The term "merging" may mean merging motion of multiple blocks. "merging" may mean that motion information of one block is also applied to other blocks. In other words, the merge mode may be a mode in which motion information of the target block is derived from motion information of neighboring blocks.
When the merge mode is used, the encoding apparatus 100 may predict motion information of the target block using motion information of the spatial candidate and/or motion information of the temporal candidate. The spatial candidates may include reconstructed spatially neighboring blocks that are spatially adjacent to the target block. The spatially neighboring blocks may include a left neighboring block and an upper neighboring block. The temporal candidates may include col blocks. The terms "spatial candidate" and "spatial merge candidate" may be used to have the same meaning and may be used interchangeably with each other. The terms "time candidate" and "time merge candidate" may be used to have the same meaning and may be used interchangeably with each other.
The encoding apparatus 100 may acquire a prediction block via prediction. The encoding apparatus 100 may encode a residual block that is a difference between the target block and the prediction block.
2-1) creating a merge candidate list
When the merge mode is used, each of the encoding apparatus 100 and the decoding apparatus 200 may create a merge candidate list using motion information of spatial candidates and/or motion information of temporal candidates. The motion information may include 1) a motion vector, 2) a reference picture index, and 3) a reference direction. The reference direction may be unidirectional or bidirectional. The reference direction may represent an inter prediction indicator.
The merge candidate list may include merge candidates. The merge candidate may be motion information. In other words, the merge candidate list may be a list storing a plurality of pieces of motion information.
The merge candidate may be motion information of a plurality of temporal candidates and/or spatial candidates. In other words, the merge candidate list may include motion information of temporal candidates and/or spatial candidates, and the like.
Further, the merge candidate list may include a new merge candidate generated by combining merge candidates already existing in the merge candidate list. In other words, the merge candidate list may include new motion information generated by combining a plurality of pieces of motion information previously existing in the merge candidate list.
Further, the merge candidate list may include history-based merge candidates. The history-based merge candidate may be motion information of a block that is encoded and/or decoded before the target block.
Further, the merge candidate list may include a merge candidate based on an average of the two merge candidates.
The merging candidate may be a specific mode of deriving inter prediction information. The merge candidate may be information indicating a specific mode of deriving inter prediction information. Inter prediction information for the target block may be derived from a particular mode indicated by the merge candidate. Further, the particular mode may include a process of deriving a series of inter prediction information. This particular mode may be an inter prediction information derivation mode or a motion information derivation mode.
The inter prediction information of the target block may be derived according to a mode indicated by a merge candidate selected among merge candidates in the merge candidate list by a merge index.
For example, the motion information derivation mode in the merge candidate list may be at least one of the following modes: 1) a motion information derivation mode for sub-block units and 2) an affine motion information derivation mode.
In addition, the merge candidate list may include motion information of a zero vector. The zero vector may also be referred to as a "zero merge candidate".
In other words, the pieces of motion information in the merge candidate list may be at least one of: 1) motion information of a spatial candidate, 2) motion information of a temporal candidate, 3) motion information generated by combining pieces of motion information previously existing in the merge candidate list, and 4) a zero vector.
The motion information may include 1) a motion vector, 2) a reference picture index, and 3) a reference direction. The reference direction may also be referred to as an "inter prediction indicator". The reference direction may be unidirectional or bidirectional. The unidirectional reference direction may indicate L0 prediction or L1 prediction.
The merge candidate list may be created before performing prediction in merge mode.
The number of merge candidates in the merge candidate list may be predefined. Each of the encoding apparatus 100 and the decoding apparatus 200 may add the merge candidates to the merge candidate list according to a predefined scheme and a predefined priority such that the merge candidate list has a predefined number of merge candidates. The merge candidate list of the encoding apparatus 100 and the merge candidate list of the decoding apparatus 200 may be made identical to each other using a predefined scheme and a predefined priority.
Merging may be applied on a CU or PU basis. When the merging is performed on a CU or PU basis, the encoding apparatus 100 may transmit a bitstream including predefined information to the decoding apparatus 200. For example, the predefined information may include 1) information indicating whether to perform merging for respective block partitions, and 2) information on a block on which merging is to be performed among blocks that are spatial candidates and/or temporal candidates for a target block.
2-2) searching for motion vector using merge candidate list
The encoding apparatus 100 may determine a merge candidate to be used for encoding the target block. For example, the encoding apparatus 100 may perform prediction on the target block using the merge candidate in the merge candidate list, and may generate a residual block for the merge candidate. The encoding apparatus 100 may encode the target block using a merging candidate that generates the minimum cost in the encoding of the prediction and residual blocks.
In addition, the encoding apparatus 100 may determine whether to encode the target block using the merge mode.
2-3) Transmission of inter-frame prediction information
The encoding apparatus 100 may generate a bitstream including inter prediction information required for inter prediction. The encoding apparatus 100 may generate entropy-encoded inter prediction information by performing entropy encoding on the inter prediction information, and may transmit a bitstream including the entropy-encoded inter prediction information to the decoding apparatus 200. The entropy-encoded inter prediction information may be signaled to the decoding apparatus 200 through a bitstream by the encoding apparatus 100. The decoding apparatus 200 may extract entropy-encoded inter prediction information from a bitstream, and may acquire the inter prediction information by applying entropy decoding to the entropy-encoded inter prediction information.
The decoding apparatus 200 may perform inter prediction on the target block using inter prediction information of the bitstream.
The inter prediction information may include 1) mode information indicating whether a merge mode is used, 2) a merge index, and 3) correction information.
In addition, the inter prediction information may include a residual signal.
The decoding apparatus 200 may acquire the merge index from the bitstream only when the mode information indicates that the merge mode is used.
The mode information may be a merge flag. The unit of the mode information may be a block. The information on the block may include mode information, and the mode information may indicate whether a merge mode is applied to the block.
The merge index may indicate a merge candidate to be used for prediction of the target block among merge candidates included in the merge candidate list. Alternatively, the merge index may indicate a block to be merged with the target block among neighboring blocks spatially or temporally adjacent to the target block.
The encoding apparatus 100 may select a merge candidate having the highest encoding performance among merge candidates included in the merge candidate list, and may set a value of the merge index to indicate the selected merge candidate.
The correction information may be information for correcting a motion vector. The encoding apparatus 100 may generate correction information. The decoding apparatus 200 may correct the motion vector of the merge candidate selected by the merge index based on the correction information.
The correction information may include at least one of information indicating whether correction is to be performed, correction direction information, and correction size information. The prediction mode for correcting the motion vector based on the signaled correction information may be referred to as a "merge mode with motion vector difference".
2-4) inter prediction of merge mode using inter prediction information
The decoding apparatus 200 may perform prediction on the target block using the merge candidate indicated by the merge index among the merge candidates included in the merge candidate list.
The motion vector of the target block may be specified by the motion vector of the merging candidate indicated by the merging index, the reference picture index, and the reference direction.
3) Skip mode
The skip mode may be a mode in which motion information of a spatial candidate or motion information of a temporal candidate is applied to the target block without change. Also, the skip mode may be a mode that does not use a residual signal. In other words, when the skip mode is used, the reconstructed block may be the same as the predicted block.
The difference between the merge mode and the skip mode is whether a residual signal is sent or used. That is, the skip mode may be similar to the merge mode except that no residual signal is sent or used.
When the skip mode is used, the encoding apparatus 100 may transmit information on a block whose motion information is to be used as motion information of a target block among blocks that are spatial candidates or temporal candidates to the decoding apparatus 200 through a bitstream. The encoding apparatus 100 may generate entropy-encoded information by performing entropy encoding on the information, and may signal the entropy-encoded information to the decoding apparatus 200 through a bitstream. The decoding apparatus 200 may extract entropy-encoded information from a bitstream and may acquire the information by applying entropy decoding to the entropy-encoded information.
Also, when the skip mode is used, the encoding apparatus 100 may not send other syntax information (such as MVD) to the decoding apparatus 200. For example, when the skip mode is used, the encoding apparatus 100 may not signal syntax elements related to at least one of an MVD, a coded block flag, and a transform coefficient level to the decoding apparatus 200.
3-1) creating a merge candidate list
The skip mode may also use a merge candidate list. In other words, the merge candidate list may be used in both the merge mode and the skip mode. In this regard, the merge candidate list may also be referred to as a "skip candidate list" or a "merge/skip candidate list".
Alternatively, the skip mode may use an additional candidate list different from the candidate list of the merge mode. In this case, in the following description, the merge candidate list and the merge candidate may be replaced with the skip candidate list and the skip candidate, respectively.
The merge candidate list may be created before performing prediction in skip mode.
3-2) searching for motion vector using merge candidate list
The encoding apparatus 100 may determine a merging candidate to be used for encoding the target block. For example, the encoding apparatus 100 may perform prediction on the target block using the merge candidate in the merge candidate list. The encoding apparatus 100 may encode the target block using the merge candidate that generates the smallest cost in the prediction.
In addition, the encoding apparatus 100 may determine whether to encode the target block using the skip mode.
3-3) Transmission of inter-frame prediction information
The encoding apparatus 100 may generate a bitstream including inter prediction information required for inter prediction. The decoding apparatus 200 may perform inter prediction on the target block using inter prediction information of the bitstream.
The inter prediction information may include 1) mode information indicating whether a skip mode is used and 2) a skip index.
The skip index may be the same as the merge index described above.
When the skip mode is used, the target block may be encoded without using a residual signal. The inter prediction information may not include a residual signal. Alternatively, the bitstream may not include a residual signal.
The decoding apparatus 200 may acquire the skip index from the bitstream only when the mode information indicates that the skip mode is used. As described above, the merge index and the skip index may be identical to each other. The decoding apparatus 200 may acquire the skip index from the bitstream only when the mode information indicates that the merge mode or the skip mode is used.
The skip index may indicate a merge candidate to be used for predicting the target block among merge candidates included in the merge candidate list.
3-4) inter prediction in skip mode using inter prediction information
The decoding apparatus 200 may perform prediction on the target block using a merge candidate indicated by the skip index among merge candidates included in the merge candidate list.
The motion vector of the target block may be specified by the motion vector of the merging candidate indicated by the skip index, the reference picture index, and the reference direction.
4) Current picture reference mode
The current picture reference mode may represent a prediction mode: the prediction mode uses a previously reconstructed region in a target picture to which the target block belongs.
A motion vector for specifying a previously reconstructed region may be used. The reference picture index of the target block may be used to determine whether the target block has been encoded in the current picture reference mode.
A flag or index indicating whether the target block is a block encoded in the current picture reference mode may be signaled by the encoding apparatus 100 to the decoding apparatus 200. Alternatively, whether the target block is a block encoded in the current picture reference mode may be inferred by the reference picture index of the target block.
When the target block is encoded in the current picture reference mode, the current picture may exist at a fixed position or an arbitrary position in the reference picture list for the target block.
For example, the fixed position may be a position where the value of the reference picture index is 0 or the last position.
When the target picture exists at an arbitrary position in the reference picture list, an additional reference picture index indicating such an arbitrary position may be signaled by the encoding apparatus 100 to the decoding apparatus 200.
5) Subblock merge mode
The sub-block merging mode may be a mode in which motion information is derived from sub-blocks of the CU.
When the subblock merging mode is applied, a subblock merging candidate list may be generated using motion information of a co-located subblock (col-sub-block) of a target subblock (i.e., a subblock-based temporal merging candidate) in a reference image and/or an affine control point motion vector merging candidate.
6) Triangle partition mode
In the triangle partition mode, the target block may be partitioned in a diagonal direction, and a child target block generated by the partitioning may be generated. For each sub-target block, motion information for the corresponding sub-target block may be derived, and the derived motion information may be used to derive a prediction sample for each sub-target block. The predicted samples of the target block may be derived by a weighted sum of the predicted samples of the sub-target blocks generated via partitioning.
7) Combined inter-intra prediction mode
The combined inter-intra prediction mode may be a mode in which a predicted sample of the target block is derived using a weighted sum of predicted samples generated via inter prediction and predicted samples generated via intra prediction.
In the above-described mode, the decoding apparatus 200 may autonomously correct the derived motion information. For example, the decoding apparatus 200 may search for motion information having a minimum Sum of Absolute Differences (SAD) in a specific region based on a reference block indicated by the derived motion information, and may derive the found motion information as corrected motion information.
In the above-described mode, the decoding apparatus 200 may compensate for prediction samples derived through inter-prediction using optical flow.
In the AMVP mode, the merge mode, the skip mode, and the like described above, among pieces of motion information in the list, motion information to be used for prediction of a target block may be specified using index information of the list.
In order to improve encoding efficiency, the encoding apparatus 100 may signal only an index of an element that generates the smallest cost in inter prediction of the target block among elements in the list. The encoding apparatus 100 may encode the index and may signal the encoded index.
Therefore, it is necessary to be able to derive the above-described lists (i.e., the predictive motion vector candidate list and the merge candidate list) based on the same data using the same scheme by the encoding apparatus 100 and the decoding apparatus 200. Here, the same data may include a reconstructed picture and a reconstructed block. Further, in order to specify an element using an index, the order of the elements in the list must be fixed.
Fig. 10 illustrates spatial candidates according to an embodiment.
In fig. 10, the positions of the spatial candidates are shown.
The large block at the center of the graph may represent the target block. Five small blocks may represent spatial candidates.
The coordinates of the target block may be (xP, yP), and the size of the target block may be represented by (nPSW, nPSH).
Spatial candidate A 0 May be a block adjacent to the lower left corner of the target block. A. the 0 May be a block occupying a pixel located at the coordinates (xP-1, yP + nPSH + 1).
Spatial candidate A 1 May be the block adjacent to the left side of the target block. A. the 1 May be the lowermost block among blocks adjacent to the left side of the target block. Alternatively, A 1 May be with A 0 Top adjacent block of (a). A. the 1 May be a block occupying pixels located at coordinates (xP-1, yP + nPSH).
Spatial candidate B 0 May be a block adjacent to the upper right corner of the target block. B 0 May be a block occupying a pixel located at the coordinates (xP + nPSW +1, yP-1).
Spatial candidate B 1 May be a block adjacent to the top of the target block. B 1 May be the rightmost block among blocks adjacent to the top of the target block. Alternatively, B 1 May be with B 0 Left adjacent block. B is 1 May be a block occupying a pixel located at the coordinates (xP + nPSW, yP-1).
Spatial candidate B 2 May be a block adjacent to the upper left corner of the target block. B is 2 May be a block occupying a pixel located at the coordinates (xP-1, yP-1).
Determination of availability of spatial and temporal candidates
In order to include motion information of a spatial candidate or motion information of a temporal candidate in a list, it must be determined whether motion information of a spatial candidate or motion information of a temporal candidate is available.
Hereinafter, the candidate block may include a spatial candidate and a temporal candidate.
For example, the determination may be performed by sequentially applying the following steps 1) to 4).
Step 1) when a PU including a candidate block is located outside the boundary of the picture, the availability of the candidate block may be set to "false". The expression "availability is set to false" may have the same meaning as "set to unavailable".
Step 2) when a PU including a candidate block is located outside the boundary of a slice, the availability of the candidate block may be set to "false". When the target block and the candidate block are located in different stripes, the availability of the candidate block may be set to "false".
Step 3) when the PU including the candidate block is outside the boundary of the parallel block, the availability of the candidate block may be set to "false". When the target block and the candidate block are located in different parallel blocks, the availability of the candidate block may be set to "false".
Step 4) when the prediction mode of the PU including the candidate block is an intra prediction mode, the availability of the candidate block may be set to "false". The availability of a candidate block may be set to "false" when a PU that includes the candidate block does not use inter prediction.
Fig. 11 illustrates an order of adding motion information of spatial candidates to a merge list according to an embodiment.
As shown in fig. 11, a may be used when pieces of motion information of spatial candidates are added to the merge list 1 、B 1 、B 0 、A 0 And B 2 The order of (a). That is, can be according to A 1 、B 1 、B 0 、A 0 And B 2 The order of (a) adds pieces of motion information of the available spatial candidates to the merge list.
Method for deriving merge lists in merge mode and skip mode
As described above, the maximum number of merging candidates in the merge list may be set. The maximum number of settings may be indicated by "N". The set number may be transmitted from the encoding apparatus 100 to the decoding apparatus 200. The head of the strip may comprise N. In other words, the maximum number of merging candidates in the merging list for the target block of the slice may be set by the slice header. For example, the value of N may be substantially 5.
Pieces of motion information (i.e., merging candidates) may be added to the merge list in the order of the following steps 1) to 4).
Step 1)Among the spatial candidates, the available spatial candidates may be added to the merge list. The pieces of motion information of the available spatial candidates may be added to the merge list in the order shown in fig. 10. Here, when the motion information of the available spatial candidate overlaps with other motion information already existing in the merge list, the motion information of the available spatial candidate may not be added to the merge list. The operation of checking whether the corresponding motion information overlaps with other motion information present in the list may be simply referred to as "overlap check".
The maximum number of pieces of motion information added may be N.
Step 2)When the number of pieces of motion information in the merge list is less than N and a temporal candidate is available, the motion information of the temporal candidate may be added to the merge list. Here, when the motion information of the available temporal candidate overlaps with other motion information already existing in the merge list, the motion information of the available temporal candidate may not be added to the merge list.
Step 3)When the number of pieces of motion information in the merge list is less than N and the type of the target slice is "B", combined motion information generated by combining bi-prediction (bi-prediction) may be added to the merge list.
The target stripe may be a stripe that includes the target block.
The combined motion information may be a combination of the L0 motion information and the L1 motion information. The L0 motion information may be motion information referring only to the reference picture list L0. The L1 motion information may be motion information referring only to the reference picture list L1.
In the merge list, there may be one or more pieces of L0 motion information. Further, in the merge list, there may be one or more pieces of L1 motion information.
The combined motion information may include one or more pieces of combined motion information. When generating the combined motion information, L0 motion information and L1 motion information to be used for the step of generating the combined motion information among the one or more pieces of L0 motion information and the one or more pieces of L1 motion information may be defined in advance. One or more pieces of combined motion information may be generated in a predefined order via combined bi-prediction using a pair of different pieces of motion information in the merge list. One piece of motion information of the pair of different motion information may be L0 motion information, and the other piece of motion information of the pair of different motion information may be L1 motion information.
For example, the combined motion information added with the highest priority may be a combination of L0 motion information having a merge index of 0 and L1 motion information having a merge index of 1. When the motion information having the merge index 0 is not the L0 motion information or when the motion information having the merge index 1 is not the L1 motion information, the combined motion information may be neither generated nor added. Next, the combined motion information added with the next priority may be a combination of L0 motion information having a merge index of 1 and L1 motion information having a merge index of 0. The detailed combinations that follow may conform to other combinations in the video encoding/decoding field.
Here, when the combined motion information overlaps with other motion information already existing in the merge list, the combined motion information may not be added to the merge list.
Step 4)When the number of pieces of motion information in the merge list is less than N, the motion information of the zero vector may be added to the merge list.
The zero vector motion information may be motion information in which the motion vector is a zero vector.
The number of pieces of zero vector motion information may be one or more. The reference picture indices of one or more pieces of zero vector motion information may be different from each other. For example, the value of the reference picture index of the first zero vector motion information may be 0. The reference picture index of the second zero vector motion information may have a value of 1.
The number of pieces of zero vector motion information may be the same as the number of reference pictures in the reference picture list.
The reference direction of the zero vector motion information may be bi-directional. Both motion vectors may be zero vectors. The number of pieces of zero vector motion information may be the smaller one of the number of reference pictures in the reference picture list L0 and the number of reference pictures in the reference picture list L1. Alternatively, when the number of reference pictures in the reference picture list L0 and the number of reference pictures in the reference picture list L1 are different from each other, the reference direction, which is unidirectional, may be used for the reference picture index that can be applied to only a single reference picture list.
The encoding apparatus 100 and/or the decoding apparatus 200 may then add zero vector motion information to the merge list while changing the reference picture index.
Zero vector motion information may not be added to the merge list when it overlaps with other motion information already present in the merge list.
The order of the above-described steps 1) to 4) is merely exemplary, and may be changed. Furthermore, some of the above steps may be omitted according to predefined conditions.
Method for deriving predicted motion vector candidate list in AMVP mode
The maximum number of predicted motion vector candidates in the predicted motion vector candidate list may be predefined. A predefined maximum number may be indicated by N. For example, the predefined maximum number may be 2.
The pieces of motion information (i.e., predicted motion vector candidates) may be added to the predicted motion vector candidate list in the order of step 1) to step 3) below.
Step 1)An available spatial candidate among the spatial candidates may be added to the predicted motion vector candidate list. The spatial candidates may include a first spatial candidate and a second spatial candidate.
The first spatial candidate may be a 0 、A 1 Zoomed A 0 And scaled A 1 One of them. The second spatial candidate may be B 0 、B 1 、B 2 Scaled B 0 Scaled B 1 And scaled B 2 One of them.
The plurality of pieces of motion information of the available spatial candidates may be added to the prediction motion vector candidate list in the order of the first spatial candidate and the second spatial candidate. In this case, when the motion information of the available spatial candidate overlaps with other motion information already existing in the predicted motion vector candidate list, the motion information of the available spatial candidate may not be added to the predicted motion vector candidate list. In other words, when the value of N is 2, if the motion information of the second spatial candidate is the same as the motion information of the first spatial candidate, the motion information of the second spatial candidate may not be added to the predicted motion vector candidate list.
The maximum number of pieces of motion information added may be N.
Step 2)When the number of pieces of motion information in the predicted motion vector candidate list is less than N and a temporal candidate is available, the motion information of the temporal candidate may be added to the predicted motion vector candidate list. In this case, when the motion information of the available temporal candidate overlaps with other motion information already existing in the predicted motion vector candidate list, the motion information of the available temporal candidate may not be added to the predicted motion vector candidate list.
Step 3)When the number of pieces of motion information in the predicted motion vector candidate list is less than N, zero vector motion information may be added to the predicted motion vector candidate list.
The zero vector motion information may include one or more pieces of zero vector motion information. The reference picture indices of the one or more pieces of zero vector motion information may be different from each other.
The encoding apparatus 100 and/or the decoding apparatus 200 may sequentially add pieces of zero vector motion information to the predicted motion vector candidate list while changing the reference picture index.
When the zero vector motion information overlaps with other motion information already existing in the predicted motion vector candidate list, the zero vector motion information may not be added to the predicted motion vector candidate list.
The description of zero vector motion information made above in connection with the merge list is also applicable to zero vector motion information. A repetitive description thereof will be omitted.
The order of step 1) to step 3) described above is merely exemplary and may be changed. Furthermore, some of the steps may be omitted according to predefined conditions.
Fig. 12 illustrates a transform and quantization process according to an example.
As shown in fig. 12, the quantized level may be generated by performing a transform and/or quantization process on the residual signal.
A residual signal may be generated as a difference between the original block and the prediction block. Here, the prediction block may be a block generated via intra prediction or inter prediction.
The residual signal may be transformed into a signal in the frequency domain by a transform process that is part of a quantization process.
The transform kernels used for the transform may include various DCT kernels, such as Discrete Cosine Transform (DCT) type 2(DCT-II) and Discrete Sine Transform (DST) kernels.
These transform kernels may perform separable transforms or two-dimensional (2D) inseparable transforms on the residual signal. The separable transform may be a transform indicating that a one-dimensional (1D) transform is performed on the residual signal in each of a horizontal direction and a vertical direction.
The DCT type and the DST type adaptively used for the 1D transform may include DCT-V, DCT-VIII, DST-I, and DST-VII in addition to DCT-II, as shown in each of Table 3 and Table 4 below.
TABLE 3
Figure BDA0003783682360000631
Figure BDA0003783682360000641
TABLE 4
Transformation set Transformation candidates
0 DST-VII,DCT-VIII,DST-I
1 DST-VII,DST-I,DCT-VIII
2 DST-VII,DCT-V,DST-I
As shown in tables 3 and 4, when a DCT type or a DST type to be used for transformation is derived, a transformation set may be used. Each transform set may include a plurality of transform candidates. Each transform candidate may be a DCT type or a DST type.
Table 5 below shows an example of a transform set to be applied to the horizontal direction and a transform set to be applied to the vertical direction according to the intra prediction mode.
TABLE 5
Intra prediction mode 0 1 2 3 4 5 6 7 8 9
Vertical direction transformation set 2 1 0 1 0 1 0 1 0 1
Horizontal direction transformation set 2 1 0 1 0 1 0 1 0 1
Intra prediction mode 10 11 12 13 14 15 16 17 18 19
Vertical direction transformation set 0 1 0 1 0 0 0 0 0 0
Horizontal direction transformation set 0 1 0 1 2 2 2 2 2 2
Intra prediction mode 20 21 22 23 24 25 26 27 28 29
Vertical direction transformation set 0 0 0 1 0 1 0 1 0 1
Horizontal direction transformation set 2 2 2 1 0 1 0 1 0 1
Intra prediction mode 30 31 32 33 34 35 36 37 38 39
Vertical direction transformation set 0 1 0 1 0 1 0 1 0 1
Horizontal direction transformation set 0 1 0 1 0 1 0 1 0 1
Intra prediction mode 40 41 42 43 44 45 46 47 48 49
Vertical direction transformation set 0 1 0 1 0 1 2 2 2 2
Horizontal direction transformation set 0 1 0 1 0 1 0 0 0 0
Intra prediction mode 50 51 52 53 54 55 56 57 58 59
Vertical direction transformation set 2 2 2 2 2 1 0 1 0 1
Horizontal direction transformation set 0 0 0 0 0 1 0 1 0 1
Intra prediction mode 60 61 62 63 64 65 66
Vertical direction transformation set 0 1 0 1 0 1 0
Horizontal direction transformation set 0 1 0 1 0 1 0
In table 5, the numbers of the vertical transform set and the horizontal transform set to be applied to the horizontal direction of the residual signal according to the intra prediction mode of the target block are shown.
As illustrated in table 5, a transform set to be applied to the horizontal direction and the vertical direction may be predefined according to the intra prediction mode of the target block. The encoding apparatus 100 may perform transformation and inverse transformation on the residual signal using the transformation included in the transformation set corresponding to the intra prediction mode of the target block. Further, the decoding apparatus 200 may perform inverse transformation on the residual signal using the transformation included in the transformation set corresponding to the intra prediction mode of the target block.
In the transform and inverse transform, as illustrated in table 3, table 4, and table 5, a transform set to be applied to a residual signal may be determined and may not be signaled. The transformation indication information may be signaled from the encoding apparatus 100 to the decoding apparatus 200. The transformation indication information may be information indicating which one of a plurality of transformation candidates included in a transformation set to be applied to the residual signal is used.
For example, when the size of the target block is 64 × 64 or less, transform sets each having three transforms may be configured according to the intra prediction mode. The optimal transformation method may be selected from a total of nine multi-transformation methods resulting from a combination of three transformations in the horizontal direction and three transformations in the vertical direction. By such an optimal transformation method, a residual signal may be encoded and/or decoded, and thus encoding efficiency may be improved.
Here, the information indicating which one of the plurality of transforms belonging to each transform set has been used for at least one of the vertical transform and the horizontal transform may be entropy-encoded and/or entropy-decoded. Here, truncated unary binarization may be used to encode and/or decode such information.
As described above, a method using various transforms may be applied to a residual signal generated via intra prediction or inter prediction.
The transform may include at least one of a first transform and a secondary transform. The transform coefficient may be generated by performing a first transform on the residual signal, and the secondary transform coefficient may be generated by performing a secondary transform on the transform coefficient.
The first transformation may be referred to as the "primary transformation". Further, the first transformation may also be referred to as an "adaptive multi-transformation (AMT) scheme". As described above, the AMT may represent applying different transforms to respective 1D directions (i.e., vertical and horizontal directions).
The secondary transform may be a transform for increasing the energy concentration of transform coefficients generated by the first transform. Similar to the first transform, the secondary transform may be a separable transform or a non-separable transform. Such an inseparable transform may be an inseparable secondary transform (NSST).
The first transformation may be performed using at least one of a predefined plurality of transformation methods. For example, the predefined plurality of transform methods may include Discrete Cosine Transform (DCT), Discrete Sine Transform (DST), Karhunen-Loeve transform (KLT), and the like.
Further, the first transform may be a transform having various types according to a kernel function defining a Discrete Cosine Transform (DCT) or a Discrete Sine Transform (DST).
For example, the transformation type may be determined based on at least one of: 1) a prediction mode (e.g., one of intra prediction and inter prediction) of the target block, 2) a size of the target block, 3) a shape of the target block, 3) an intra prediction mode of the target block, 4) a component (e.g., one of a luminance component and a chrominance component) of the target block, 5) a partition type (e.g., one of a Quadtree (QT), a Binary Tree (BT), and a Ternary Tree (TT) applied to the target block.
For example, the first transform may include transforms such as DCT-2, DCT-5, DCT-7, DST-1, DST-8, and DCT-8 according to the transform kernel presented in Table 6 below. In table 6 below, various transform types and transform kernels for Multiple Transform Selection (MTS) are illustrated.
MTS may refer to the selection of a combination of one or more DCT and/or DST kernels to transform the residual signal in the horizontal and/or vertical directions.
TABLE 6
Figure BDA0003783682360000661
In Table 6, i and j may be integer values equal to or greater than 0 and less than or equal to N-1.
A secondary transform may be performed on transform coefficients generated by performing the first transform.
As in the first transformation, a set of transformations may also be defined in the secondary transformation. The method for deriving and/or determining the above-described set of transforms may be applied not only to the first transform but also to the secondary transform.
The first transform and the secondary transform may be determined for a particular target.
For example, the first transform and the secondary transform may be applied to signal components corresponding to one or more of a luminance (luma) component and a chrominance (chroma) component. Whether to apply the first transform and/or the secondary transform may be determined according to at least one of encoding parameters for the target block and/or the neighboring blocks. For example, whether to apply the first transform and/or the secondary transform may be determined according to the size and/or shape of the target block.
In the encoding apparatus 100 and the decoding apparatus 200, transformation information indicating a transformation method to be used for a target may be derived by using the designation information.
For example, the transformation information may include transformation indices to be used for the primary transformation and/or the secondary transformation. Optionally, the transformation information may indicate that the primary transformation and/or the secondary transformation is not used.
For example, when the primary transform and the secondary transform are targeted to a target block, a transform method to be applied to the primary transform and/or the secondary transform indicated by the transform information may be determined according to at least one of encoding parameters for the target block and/or blocks adjacent to the target block.
Alternatively, transformation information indicating a transformation method for a specific object may be signaled from the encoding apparatus 100 to the decoding apparatus 200.
For example, whether to use the primary transform, the index indicating the primary transform, whether to use the secondary transform, and the index indicating the secondary transform may be derived as transform information by the decoding apparatus 200 for a single CU. Alternatively, for a single CU, transform information indicating the following may be signaled: whether to use a primary transformation, an index indicating a primary transformation, whether to use a secondary transformation, and an index indicating a secondary transformation.
The quantized transform coefficients (i.e., quantized levels) may be generated by performing quantization on a result generated by performing the first transform and/or the secondary transform or performing quantization on the residual signal.
Fig. 13 illustrates a diagonal scan according to an example.
Fig. 14 shows a horizontal scan according to an example.
Fig. 15 illustrates a vertical scan according to an example.
The quantized transform coefficients may be scanned via at least one of a (top right) diagonal scan, a vertical scan, and a horizontal scan according to at least one of an intra prediction mode, a block size, and a block shape. The block may be a Transform Unit (TU).
Each scan may be initiated at a particular starting point and may be terminated at a particular ending point.
For example, the quantized transform coefficients may be changed into a 1D vector form by scanning the coefficients of the block using the diagonal scan of fig. 13. Alternatively, the horizontal scan of fig. 14 or the vertical scan of fig. 15 may be used according to the size of the block and/or the intra prediction mode, instead of using the diagonal scan.
The vertical scanning may be an operation of scanning the 2D block type coefficients in the column direction. The horizontal scanning may be an operation of scanning the 2D block type coefficients in a row direction.
In other words, which one of the diagonal scan, the vertical scan, and the horizontal scan is to be used may be determined according to the size of the block and/or the inter prediction mode.
As shown in fig. 13, 14, and 15, the quantized transform coefficients may be scanned in a diagonal direction, a horizontal direction, or a vertical direction.
The quantized transform coefficients may be represented by block shapes. Each block may include a plurality of sub-blocks. Each sub-block may be defined according to a minimum block size or a minimum block shape.
In the scanning, a scanning order according to the type or direction of the scanning may be first applied to the subblocks. Further, a scanning order according to the direction of scanning may be applied to the quantized transform coefficients in each sub-block.
For example, as shown in fig. 13, 14, and 15, when the size of the target block is 8 × 8, the quantized transform coefficient may be generated by the first transform, the secondary transform, and the quantization of the residual signal of the target block. Thus, one of three types of scanning orders may be applied to four 4 × 4 sub-blocks, and the quantized transform coefficients may also be scanned for each 4 × 4 sub-block according to the scanning order.
The encoding apparatus 100 may generate entropy-encoded quantized transform coefficients by performing entropy encoding on the scanned quantized transform coefficients, and may generate a bitstream including the entropy-encoded quantized transform coefficients.
The decoding apparatus 200 may extract entropy-encoded quantized transform coefficients from a bitstream, and may generate the quantized transform coefficients by performing entropy decoding on the entropy-encoded quantized transform coefficients. The quantized transform coefficients may be arranged in the form of 2D blocks via inverse scanning. Here, as a method of the inverse scanning, at least one of the upper right diagonal scanning, the vertical scanning, and the horizontal scanning may be performed.
In the decoding apparatus 200, inverse quantization may be performed on the quantized transform coefficients. The secondary inverse transform may be performed on a result generated by performing inverse quantization according to whether the secondary inverse transform is performed. Further, the first inverse transform may be performed on a result generated by performing the secondary inverse transform according to whether the first inverse transform is to be performed. The reconstructed residual signal may be generated by performing a first inverse transform on a result generated by performing the secondary inverse transform.
For the luminance component reconstructed via intra prediction or inter prediction, inverse mapping with dynamic range may be performed before loop filtering.
The dynamic range may be divided into 16 equal segments and the mapping function of the respective segments may be signaled. Such mapping functions may be signaled at the stripe level or parallel block group level.
An inverse mapping function for performing inverse mapping may be derived based on the mapping function.
Loop filtering, storage of reference pictures, and motion compensation may be performed in the inverse mapped region.
The prediction block generated via inter prediction may be transformed to a mapping region by mapping using a mapping function, and the transformed prediction block may be used to generate a reconstructed block. However, since the intra prediction is performed in the mapping region, the prediction block generated via the intra prediction may be used to generate the reconstructed block without the need for mapping and/or inverse mapping.
For example, when the target block is a residual block of the chrominance component, the residual block may be transformed to the inverse mapping region by scaling the chrominance component of the mapping region.
Whether scaling is available may be signaled at the stripe level or the parallel block group level.
For example, scaling may only be applied to the case where the mapping is available for the luma component and the partitions of the chroma component follow the same tree structure.
Scaling may be performed based on an average of values of samples in a luma prediction block corresponding to a chroma prediction block. Here, when the target block uses inter prediction, the luma prediction block may represent a mapped luma prediction block.
The values required for scaling may be derived by referring to a look-up table using the index of the slice to which the average of the sample values of the luma prediction block belongs.
The residual block may be transformed to the inverse mapping region by scaling the residual block using the finally derived value. Thereafter, for the block of the chrominance component, reconstruction, intra prediction, inter prediction, loop filtering, and storage of a reference picture may be performed in the inverse mapping region.
For example, information indicating whether mapping and/or inverse mapping of the luminance component and the chrominance component is available may be signaled by the sequence parameter set.
A prediction block for the target block may be generated based on the block vector. The block vector may indicate a displacement between the target block and the reference block. The reference block may be a block in the target image.
In this way, a prediction mode in which a prediction block is generated by referring to a target image may be referred to as an "Intra Block Copy (IBC) mode".
The IBC mode may be applied to a CU having a specific size. For example, the IBC mode may be applied to an mxn CU. Here, M and N may be less than or equal to 64.
The IBC mode may include a skip mode, a merge mode, an AMVP mode, and the like. In the case of the skip mode or the merge mode, the merge candidate list may be configured and the merge index may be signaled, and thus a single merge candidate may be specified among merge candidates existing in the merge candidate list. The block vector of the specified merging candidate may be used as the block vector of the target block.
In the case of AMVP mode, a differential block vector may be signaled. Furthermore, the prediction block vector may be derived from a left neighboring block and an upper neighboring block of the target block. Further, an index indicating which neighboring block is to be used may be signaled.
The prediction block in the IBC mode may be included in the target CTU or the left CTU, and may be limited to a block within the previous reconstruction region. For example, the value of the block vector may be restricted such that the prediction block of the target block is located in a specific region. The specific region may be a region defined by three 64 × 64 blocks that are encoded and/or decoded before a 64 × 64 block including the target block. Limiting the value of the block vector in this manner, memory consumption and device complexity caused by implementation of the IBC mode can thus be reduced.
Fig. 16 is a configuration diagram of an encoding apparatus according to an embodiment.
The encoding apparatus 1600 may correspond to the encoding apparatus 100 described above.
The encoding apparatus 1600 may include a processing unit 1610, a memory 1630, a User Interface (UI) input device 1650, a UI output device 1660, and a storage 1640 that communicate with each other over a bus 1690. The encoding device 1600 may also include a communication unit 1620 connected to the network 1699.
The processing unit 1610 may be a Central Processing Unit (CPU) or semiconductor device for executing processing instructions stored in the memory 1630 or the storage 1640. The processing unit 1610 may be at least one hardware processor.
The processing unit 1610 may generate and process a signal, data, or information input to the encoding apparatus 1600, output from the encoding apparatus 1600, or used in the encoding apparatus 1600, and may perform checking, comparison, determination, or the like related to the signal, data, or information. In other words, in embodiments, the generation and processing of data or information, as well as the inspection, comparison, and determination related to the data or information, may be performed by the processing unit 1610.
The processing unit 1610 may include an inter prediction unit 110, an intra prediction unit 120, a switch 115, a subtractor 125, a transform unit 130, a quantization unit 140, an entropy encoding unit 150, an inverse quantization unit 160, an inverse transform unit 170, an adder 175, a filter unit 180, and a reference picture buffer 190.
At least some of the inter prediction unit 110, the intra prediction unit 120, the switch 115, the subtractor 125, the transform unit 130, the quantization unit 140, the entropy encoding unit 150, the inverse quantization unit 160, the inverse transform unit 170, the adder 175, the filter unit 180, and the reference picture buffer 190 may be program modules and may communicate with an external device or system. The program modules may be included in the encoding device 1600 in the form of an operating system, application program modules, or other program modules.
The program modules may be physically stored in various types of well-known storage devices. Additionally, at least some of the program modules may also be stored in remote memory storage devices that are capable of communicating with the encoding apparatus 1600.
Program modules may include, but are not limited to, routines, subroutines, programs, objects, components, and data structures for performing functions or operations in accordance with the embodiments or for implementing abstract data types in accordance with the embodiments.
The program modules may be implemented using instructions or code executed by at least one processor of the encoding apparatus 1600.
The processing unit 1610 may execute instructions or code in the inter-prediction unit 110, the intra-prediction unit 120, the switch 115, the subtractor 125, the transform unit 130, the quantization unit 140, the entropy encoding unit 150, the inverse quantization unit 160, the inverse transform unit 170, the adder 175, the filter unit 180, and the reference picture buffer 190.
The memory unit may represent the memory 1630 and/or the memory 1640. Each of memory 1630 and storage 1640 may be any of various types of volatile or non-volatile storage media. For example, the memory 1630 may include at least one of Read Only Memory (ROM)1631 and Random Access Memory (RAM) 1632.
The storage unit may store data or information for the operation of the encoding device 1600. In an embodiment, data or information of the encoding apparatus 1600 may be stored in a storage unit.
For example, the storage unit may store pictures, blocks, lists, motion information, inter prediction information, bitstreams, and the like.
The encoding device 1600 may be implemented in a computer system including a computer-readable storage medium.
The storage medium may store at least one module required for the operation of the encoding apparatus 1600. Memory 1630 may store at least one module and may be configured to cause the at least one module to be executed by processing unit 1610.
Functions related to communication of data or information of the encoding apparatus 1600 may be performed by the communication unit 1620.
For example, the communication unit 1620 may transmit the bit stream to the decoding apparatus 1700 to be described later.
Fig. 17 is a configuration diagram of a decoding apparatus according to an embodiment.
The decoding apparatus 1700 may correspond to the decoding apparatus 200 described above.
The decoding apparatus 1700 may include a processing unit 1710, a memory 1730, a User Interface (UI) input device 1750, a UI output device 1760, and a storage 1740 that communicate with each other through a bus 1790. The decoding apparatus 1700 may further include a communication unit 1720 connected to a network 1799.
The processing unit 1710 may be a Central Processing Unit (CPU) or a semiconductor device for executing processing instructions stored in the memory 1730 or the storage 1740. The processing unit 1710 can be at least one hardware processor.
The processing unit 1710 may generate and process a signal, data, or information input to the decoding apparatus 1700, output from the decoding apparatus 1700, or used in the decoding apparatus 1700, and may perform checking, comparison, determination, or the like related to the signal, data, or information. In other words, in embodiments, the generation and processing of data or information, as well as the examination, comparison, and determination of data or information related thereto, may be performed by the processing unit 1710.
The processing unit 1710 may include the entropy decoding unit 210, the inverse quantization unit 220, the inverse transform unit 230, the intra prediction unit 240, the inter prediction unit 250, the switch 245, the adder 255, the filter unit 260, and the reference picture buffer 270.
At least some of the entropy decoding unit 210, the inverse quantization unit 220, the inverse transform unit 230, the intra prediction unit 240, the inter prediction unit 250, the adder 255, the switch 245, the filter unit 260, and the reference picture buffer 270 of the decoding apparatus 200 may be program modules and may communicate with an external device or system. The program modules may be included in the decoding apparatus 1700 in the form of an operating system, application program modules, or other program modules.
Program modules may be physically stored in various types of well-known memory devices. Furthermore, at least some of the program modules may also be stored in a remote memory storage device that is capable of communicating with the decoding apparatus 1700.
Program modules may include, but are not limited to, routines, subroutines, programs, objects, components, and data structures for performing functions or operations in accordance with the embodiments or for implementing abstract data types in accordance with the embodiments.
The program modules may be implemented using instructions or code executed by at least one processor of the decoding apparatus 1700.
Processing unit 1710 may execute instructions or code in entropy decoding unit 210, inverse quantization unit 220, inverse transform unit 230, intra prediction unit 240, inter prediction unit 250, switch 245, adder 255, filter unit 260, and reference picture buffer 270.
The memory unit may represent memory 1730 and/or memory 1740. Memory 1730 and storage 1740 can each be any of a variety of types of volatile or non-volatile storage media. For example, memory 1730 may include at least one of ROM 1731 and RAM 1732.
The storage unit may store data or information for the operation of the decoding apparatus 1700. In an embodiment, data or information of the decoding apparatus 1700 may be stored in a storage unit.
For example, the storage unit may store pictures, blocks, lists, motion information, inter prediction information, bitstreams, and the like.
The decoding apparatus 1700 may be implemented in a computer system including a computer-readable storage medium.
The storage medium may store at least one module required for the operation of the decoding apparatus 1700. The memory 1730 may store at least one module and may be configured to cause the at least one module to be executed by the processing unit 1710.
Functions related to communication of data or information of the decoding apparatus 1700 can be performed by the communication unit 1720.
For example, the communication unit 1720 may receive a bitstream from the encoding device 1600.
Coding using palette mode
As described above, the prediction modes for the target block may include an intra mode, an inter mode, and an Intra Block Copy (IBC) mode.
The prediction mode for the target block may also include a palette mode. The target block may be encoded/decoded based on a prediction method selected for the target block among an intra mode, an inter mode, an IBC mode, and a palette mode.
Hereinafter, the terms "intra mode" and "intra coding mode" may be used to have the same meaning and may be used interchangeably with each other.
Hereinafter, the terms "inter mode" and "inter coding mode" may be used to have the same meaning and may be used interchangeably with each other.
Hereinafter, the terms "IBC mode" and "IBC coding mode" may be used to have the same meaning and may be used interchangeably with each other.
Hereinafter, the terms "palette mode" and "palette coding mode" may be used to have the same meaning, and may be used interchangeably with each other.
A palette to be utilized when using the palette mode may form a set of a particular number of representative color values. The particular number of representative color values of the palette may be used to represent a pixel value for each pixel in the target CU.
Pixels having a same color value as a representative color value of the palette or pixels having a value closer to the representative color value of the palette may be represented by values of indices of the palette instead of the original pixel values.
In other words, a value of a particular palette index may correspond to a particular representative color value. The palette index values may correspond to respective representative color values of the palette.
Further, among pixel values of pixels in the target block, color values that do not correspond to the representative color values of the palette may be represented by a separate scheme including escape (esc) symbols and quantized color components. The color values may be used to signal information including escape (esc) symbols and quantized color components that do not correspond to a color value of a palette's representative color values.
Image coding using syntax for target block
Fig. 18 shows a first syntax for a target block according to an embodiment.
One line of the code shown in fig. 18 may correspond to one step of the encoding method and one step of the decoding method.
Hereinafter, the code lines of fig. 18 will be described as individual steps from the viewpoint of encoding.
Steps 1810, 1815, 1820, 1825, 1830, 1835, 1840, 1845, 1850, 1855, 1860, and 1865 of fig. 18 may be performed by the processing unit 1610 of the encoding apparatus 1600.
The target block may be a Coding Unit (CU).
At step 1810, encoding of the target block may be initiated.
x0 may be the x coordinate of the upper left portion of the target block.
y0 may be the y coordinate of the upper left portion of the target block.
cbWidth may be the width of the target block.
cbHeight may be the height of the target block.
The cqtDepth may be the depth of the currently encoded quadtree.
the treeType may be information indicating whether a Coding Tree Unit (CTU) is a single tree or a dual tree used for partitioning the CTU.
Further, when a dual tree is used for the CTU, treeType may be information indicating whether a luminance component or a chrominance component is currently being processed.
the value of treeType may be one of SINGLE _ TREE, DUAL _ TREE _ LUMA, and DUAL _ TREE _ CHROMA.
SINGLE TREE may indicate that a SINGLE TREE is used for the CTU. SINGLE _ TREE may represent that the partition structure of the luminance component of the CTU (i.e., the partition structure of the luminance component of the target block) is the same as the partition structure of the chrominance component of the target block.
DUAL TREE LUMA may indicate that a DUAL TREE is used for the CTU and that the LUMA component of the CTU is currently being processed.
The DUAL _ TREE _ CHROMA may indicate that the CHROMA component of the CTU is currently being processed.
The modeType may be information indicating the following for a CU in the CTU: 1) whether all of the intra prediction mode, the IBC mode, the palette mode, and the inter prediction mode can be used, 2) whether only the intra prediction mode, the IBC mode, and the palette mode can be used, and 3) whether only the inter mode can be used.
Hereinafter, the terms "CTU" and "coding tree node" may be used to have the same meaning and may be used interchangeably with each other.
The value of modeType may be one of MODE _ TYPE _ ALL, MODE _ TYPE _ INTRA, and MODE _ TYPE _ INTER.
For a CU in the CTU, MODE _ TYPE _ ALL may indicate that an intra prediction MODE, IBC MODE, palette MODE, and inter prediction MODE may ALL be used.
For a CU in the CTU, MODE _ TYPE _ INTRA may indicate that INTRA prediction MODE, IBC MODE, and palette MODE may be used.
For a CU in the CTU, MODE _ TYPE _ INTER may indicate that only INTER MODE may be used.
In step 1815, it may be determined whether the palette information pred _ mode _ plt _ flag is to be written to the bitstream.
pred _ mode _ plt _ flag may be information indicating whether the target block is a palette block. The palette blocks may be blocks that are encoded using a palette mode.
If it is determined that pred _ mode _ plt _ flag is to be written to the bitstream, step 1820 may be performed.
If it is determined that pred _ mode _ plt _ flag is not to be written to the bitstream, step 1820 may be performed.
Whether pred _ mode _ plt _ flag is to be written to the bitstream can be determined based on the following equation 1.
[ formula 1]
CuPredMode[chType][x0][y0]==MODE_INTRA&&sps_palette_enabled_flag&&cbWidth<=64&&cbHeight<=64&&cu_skip_flag[x0][y0]==0&&modeType!=MODE_TYPE_INTER
In equation 1 AND subsequent equations, "& &" represents a logical AND operation. "|" represents a logical OR operation.
"! "denotes a logical NOT operation.
The chType may be information indicating whether a value of treeType, which will be described later, is DUAL _ TREE _ CHROMA.
CuPredMode [ chType ] [ x0] [ y0] may be information indicating a prediction mode of a target block.
The value of CuPredMode [ chType ] [ x0] [ y0] may be one of MODE _ INTRA, MODE _ INTER, MODE _ PLT and MODE _ IBC.
MODE INTRA may indicate that the prediction MODE for the target block is an INTRA MODE.
MODE INTER may indicate that the prediction MODE for the target block is INTER MODE.
MODE _ PLT may indicate that the prediction MODE for the target block is a palette MODE.
MODE _ IBC may indicate that the prediction MODE for the target block is an IBC MODE.
The SPS _ palette _ enabled _ flag may be information indicating whether or not the palette mode is enabled for SPS.
cu _ skip _ flag may be information indicating whether a skip mode is applied to the target block.
When the value of equation 1 is "0", it may be determined that pred _ mode _ plt _ flag is not written to the bitstream, and step 1825 may be performed. In an embodiment, the result or information value "0" of equation may be replaced with the first value or with "false". Hereinafter, a repetitive explanation will be omitted.
When the value of equation 1 is "1", it may be determined that pred _ mode _ plt _ flag is written to the bitstream, and step 1820 may be performed. In an embodiment, the result or information value "1" of the equation may be replaced with a second value or with "true". Hereinafter, a repetitive explanation will be omitted.
Formula 1 above may be replaced with formula 2 below.
[ formula 2]
CuPredMode[chType][x0][y0]==MODE_INTRA&&sps_palette_enabled_flag&&cbWidth<=64&&cbHeight<=64&&cu_skip_flag[x0][y0]==0&&modeType!=MODE_TYPE_INTER&&((cbWidth*cbHeight)>(treeType!=DUAL_TREE_CHROMA16:16*SubWidthC*SubHeightC))&&(modeType!=MODE_TYPE_INTRA||treeType!=DUAL_TREE_CHROMA))
The SubWidthC may be information indicating the number of samples to be sampled in the width direction of the chrominance component. For example, the width of the chrominance component of the target block may be calculated using the following equation 3.
[ formula 3]
cbWidth/SubWidthC
The subheight c may be information indicating the number of samples to be sampled in the height direction of the chrominance component. For example, the height of the chrominance component of the target block may be calculated using equation 4 below.
[ formula 4]
cbHeight/SubHeightC
In step 1820, pred _ mode _ plt _ flag may be written to the bitstream.
When step 1820 is performed, step 1825 may then be performed.
At step 1825, it may be determined whether step 1830 is to be performed.
If it is determined that step 1830 is to be performed, then step 1830 may be performed.
If it is determined that step 1830 is not to be performed, step 1910, which will be described later with reference to FIG. 19, may be performed.
Whether step 1830 is to be performed may be determined based on equation 5 below.
[ formula 5]
CuPredMode[chType][x0][y0]==MODE_INTRA||CuPredMode[chType][x0][y0]==MODE_PLT
When the value of equation 5 is "0", step 1830 may not be performed and step 1910, which will be described later with reference to fig. 19, may be performed.
When the value of equation 5 is "1," step 1830 may be performed.
At step 1830, it may be determined whether step 1835 is to be performed.
If it is determined that step 1835 is to be performed, step 1835 may be performed.
If it is determined that step 1835 is not to be performed, step 1910, which will be described later with reference to FIG. 19, may be performed.
Whether step 1835 is to be performed may be determined based on equation 6 below.
[ formula 6]
treeType==SINGLE_TREE||treeType==DUAL_TREE_LUMA
When the value of equation 6 is "0", step 1835 may not be performed and step 1910, which will be described later with reference to fig. 19, may be performed.
When the value of equation 6 is "1," step 1835 may be performed.
At step 1835, it may be determined whether palette coding is to be performed on the target block.
Palette coding may be an operation of performing coding on a target block using a palette mode.
If it is determined that palette coding is to be performed on the target block, step 1840 may be performed.
If it is determined that palette coding will not be performed on the target block, steps 1850 and 1860 may be performed by step 1845.
Whether palette coding is to be performed on the target block may be determined based on equation 7 below.
[ formula 7]
pred_mode_plt_flag
When the value of equation 7 is "0," it may be determined that palette coding is to be performed on the target block, and step 1840 may be performed.
When the value of equation 7 is "1", palette coding may not be performed on the target block, and steps 1850 and 1860 may be performed through step 1845.
At step 1840, palette coding may be performed on the target block.
palette _ coding may specify a palette coding process for the target block.
In step 1850, it may be determined whether intra _ sub _ partitions _ mode _ flag, which is intra sub partition mode information, is to be written into the bitstream.
The intra _ sub _ modes _ flag may be information indicating an intra sub partition mode for the target block.
A value of intra _ subportions _ mode _ flag equal to "1" may specify that the current intra-coded unit is partitioned into numintrasubportions [ x0] [ y0] transform block sub-partitions.
A value of intra _ sub partitions _ mode _ flag equal to "0" may specify that the current intra coding unit is not partitioned into transform block sub-partitions.
If it is determined that intra _ sub _ modes _ flag is to be written to the bitstream, step 1855 may be performed.
If it is determined that intra _ sub _ modes _ flag is not to be written to the bitstream, step 1860 may be performed.
Whether the intra _ sub _ modes _ flag is to be written to the bitstream may be determined based on the following equation 8.
[ formula 8]
sps_isp_enabled_flag&&intra_luma_ref_idx==0&&(cbWidth<=MaxTbSizeY&&cbHeight<=MaxTbSizeY)&&(cbWidth*cbHeight>MinTbSizeY*MinTbSizeY)&&!cu_act_enabled_flag
The sps _ isp _ enabled _ flag may be information indicating whether intra prediction with sub-partitions is enabled for a Coded Layer Video Sequence (CLVS).
A value of sps _ isp _ enabled _ flag equal to "1" may specify that intra prediction with sub-partitions is enabled for CLVS.
A value of sps _ isp _ enabled _ flag equal to "0" may specify that intra prediction with sub-partitions is disabled for CLVS.
intra _ luma _ ref _ idx may be index information specifying a reference sample line to be used for intra prediction.
MaxTbSizeY may be the maximum transform size (maximum size of transform unit) in the luminance component.
MinTbSizeY may be the minimum transform size (minimum size of a transform unit) in the luminance component.
cu _ act _ enabled _ flag may be information indicating whether color space conversion is used for a decoding residual of the target block. This may be information indicating whether color space conversion is enabled.
A value of cu act enabled flag equal to "1" may specify that the decoding residual of the target block is applied in case of using color space conversion.
A value of cu act enabled flag equal to "0" may specify that the decoding residual of the target block is applied without using color space conversion.
When the value of equation 8 is "0", it may be determined that intra _ sub _ modes _ flag is not written to the bitstream, and step 1860 may be performed.
When the value of equation 8 is "1", it may be determined that intra _ sub _ modes _ flag is to be written to the bitstream, and step 1855 may be performed.
In step 1855, the intra _ sub _ modes _ flag may be written to the bitstream.
When step 1855 is performed, step 1860 may then be performed.
In step 1860, it may be determined whether intra sub-partition division information intra _ sub partitions _ split _ flag is to be written into the bitstream.
The intra _ sub _ partitions _ split _ flag may be information specifying an intra sub-partition division type for the target block.
When the value of intra _ subpartitions _ mode _ flag is "0", the value of intrasubpartitionsplittype may be 0.
When the value of intra _ subpartitions _ mode _ flag is "1", the value of intrasubpartitionspatitype may be determined to be represented by the following equation 9.
[ formula 9]
IntraSubPartitionsSplitType=intra_subpartitions_mode_flag+intra_subpartitions_split_flag
A value of intrasubportionssplittype equal to "0" may specify that the target block is not partitioned.
A value of intrasubpartitionsplit type equal to "1" may specify that the target block is divided in the horizontal direction.
A value of intrasubpartitionsplit type equal to "2" may specify that the target block is divided in the vertical direction.
If it is determined that intra _ sub _ partitions _ split _ flag is to be written to the bitstream, step 1865 may be performed.
If it is determined that intra _ sub _ partitions _ split _ flag will not be written to the bitstream, step 1910, which will be described later with reference to fig. 19, may be performed.
Whether the intra _ sub _ modes _ flag is to be written to the bitstream may be determined based on the following equation 10.
[ formula 10]
intra_subpartitions_mode_flag==1
When the value of equation 10 is "0", it may be determined that intra _ sub _ partitions _ split _ flag will not be written to the bitstream, and step 1910, which will be described later with reference to fig. 19, may be performed.
When the value of equation 10 is "1", it may be determined that intra _ sub _ splits _ flag is to be written to the bitstream, and step 1865 may be performed.
In step 1865, the intra _ sub _ partitions _ split _ flag may be written to the bitstream.
When step 1865 is performed, step 1910, which will be described later with reference to fig. 19, may be performed.
Fig. 19 shows a second syntax for a target block according to an embodiment.
One line of the code shown in fig. 19 may correspond to one step of the encoding method and one step of the decoding method.
Hereinafter, the code lines of fig. 19 will be described as individual steps from the perspective of encoding.
Steps 1910, 1915, 1920, 1925, 1930, 1935, 1940, 1945, 1950, 1955, 1960, 1965, 1970, 1975, 1980, and 1985 of fig. 19 may be performed by the processing unit 1610 of the encoding apparatus 1600.
The target block may be a CU.
The second syntax, which will be described with reference to fig. 19, may be performed after the first syntax described above with reference to fig. 18.
At step 1910, it may be determined whether step 1915 is to be performed.
If it is determined that step 1915 is to be performed, step 1915 may be performed.
If it is determined that step 1915 is not to be performed, the process may be terminated, or alternatively, other steps not shown may be performed.
Based on equation 11 below, it may be determined whether step 1915 is to be performed.
[ formula 11]
(treeType==SINGLE_TREE||treeType==DUAL_TREE_CHROMA)&&ChromaArrayType!=0
The ChromaArrayType may be information that specifies the chroma format of the target block.
When the value of equation 11 is "0", step 1815 may not be performed and the process may be terminated, or alternatively, other steps not shown may be performed.
When the value of equation 11 is "1," step 1915 may be performed.
Formula 11 above may be replaced with formula 12 below.
[ formula 12]
(treeType==SINGLE_TREE||treeType==DUAL_TREE_CHROMA)&&sps_chroma_format_idc!=0
The sps _ chroma _ format _ idc may be information indicating a chroma format of the target block, a value of subwidtc, and a value of subwight c.
When the value of sps _ chroma _ format _ idc is "0", the chroma format may be monochrome, the value of subwidtc may be "1", and the value of subwight c may be "1".
When the value of sps _ chroma _ format _ idc is "1", the chroma format may be "4: 2: 0", the value of subwidtc may be "2", and the value of subwight c may be "2".
When the value of sps _ chroma _ format _ idc is "2", the chroma format may be "4: 2: 2", the value of SubWidthC may be "2", and the value of subwight c may be "1".
When the value of sps _ chroma _ format _ idc is "3", the chroma format may be "4: 4: 4", the value of SubWidthC may be "1", and the value of subwight c may be "1".
When the value of equation 12 is "0", step 1815 may not be performed, and the process may be terminated, or alternatively, other steps not shown may be performed.
When the value of equation 12 is "1," step 1915 may be performed.
At step 1915, it may be determined whether palette coding is to be performed on the target block.
Palette coding may be an operation of performing coding on a target block using a palette mode.
If it is determined that palette coding is to be performed on the target block, step 1920 may be performed.
If it is determined that palette coding is not to be performed on the target block, step 1930 may be performed by step 1925.
Whether palette coding is to be performed on the target block may be determined based on equation 13 below.
[ formula 13]
pred_mode_plt_flag&&treeType==DUAL_TREE_CHROMA
When the value of equation 13 is "0," it may be determined that palette coding is to be performed on the target block, and step 1920 may be performed.
When the value of equation 13 is "1", palette coding may not be performed on the target block, and step 1930 may be performed through step 1925.
At step 1920, palette coding may be performed on the target block.
palette _ coding may specify a palette coding process for a target block.
When step 1920 is performed, the process may be terminated, or other steps not shown may alternatively be performed.
At step 1930, it may be determined whether step 1935 is to be performed.
If it is determined that step 1935 is to be performed, step 1935 may be performed.
If it is determined that step 1935 is not to be performed, the process can be terminated, or other steps not shown can alternatively be performed.
Whether step 1935 is to be performed may be determined based on equation 14 below.
[ formula 14]
!cu_act_enabled_flag
When the value of equation 14 is "0," step 1935 may not be performed and the process may be terminated, or alternatively, other steps not shown may be performed.
When the value of equation 14 is "1," step 1935 may be performed.
In step 1935, information to be written to the bitstream can be selected from the first information and the second information.
Steps 1940 and 1945 may be performed when the first information is selected as information to be written to the bitstream.
When the second information is selected as information to be written to the bitstream, steps 1960 and 19070 may be performed by 1955.
The first information may include at least some of block-Based Delta Pulse Code Modulation (BDPCM) availability information intra _ BDPCM _ chroma _ flag and BDPCM prediction direction information intra _ BDPCM _ chroma _ dir _ flag.
intra _ bdpcmp _ chroma _ flag may be information indicating whether BDPCM is applied to the current chroma coding block. The current chroma coding block may represent the chroma component of the target block.
A value of intra _ bdpcmm _ chroma _ flag equal to "1" may specify that BDPCM is applied to the current chroma coding block.
A value of intra _ bdpcmm _ chroma _ flag equal to "1" may specify that BDPCM is not applied to the current chroma coding block.
The intra _ bdpcmp _ chroma _ dir _ flag may indicate a prediction direction of the BDPCM.
A value of intra _ bdpcmm _ chroma _ dir _ flag equal to "0" may specify that the prediction direction of the BDPCM is the horizontal direction.
A value of intra _ bdpcmm _ chroma _ dir _ flag equal to "1" may specify that the prediction direction of the BDPCM is the vertical direction.
The second information may include at least some of cross-component linear model (CCLM) mode information CCLM _ mode _ flag, CCLM mode index information CCLM _ mode _ idx, and intra chroma prediction mode information intra _ chroma _ pred _ mode.
CCLM _ mode _ flag may be information indicating whether intra prediction across component linear models (CCLMs) is used for the target block.
The CCLM _ mode _ flag may be information indicating whether one of the INTRA _ LT _ CCLM chroma INTRA prediction mode, the INTRA _ L _ CCLM chroma INTRA prediction mode, and the INTRA _ T _ CCLM chroma INTRA prediction mode is applied.
A value of CCLM _ mode _ flag equal to "1" may specify that one of the INTRA _ LT _ CCLM chroma INTRA prediction mode, the INTRA _ L _ CCLM chroma INTRA prediction mode, and the INTRA _ T _ CCLM chroma INTRA prediction mode is applied to the target block.
A value of CCLM _ mode _ flag equal to "0" may specify that none of the INTRA _ LT _ CCLM chroma INTRA prediction mode, the INTRA _ L _ CCLM chroma INTRA prediction mode, and the INTRA _ T _ CCLM chroma INTRA prediction mode is applied to the target block.
The CCLM _ mode _ idx may be information indicating which one of the INTRA _ LT _ CCLM chroma INTRA prediction mode, the INTRA _ L _ CCLM chroma INTRA prediction mode, and the INTRA _ T _ CCLM chroma INTRA prediction mode is applied to the target block.
The value of CCLM _ mode _ idx may be one of INTRA _ LT _ CCLM, INTRA _ L _ CCLM, and INTRA _ T _ CCLM.
intra _ chroma _ pred _ mode may be information indicating an intra prediction mode for a chroma component (or chroma sampling point) of a target block.
The information to be written to the bitstream may be selected based on the following equation 15.
[ formula 15]
cbWidth/SubWidthC<=MaxTsSize&&cbHeight/SubHeightC<=MaxTsSize&&sps_bdpcm_enabled_flag
Maxttssize may be the maximum block size for transform skipping.
The sps _ bdplcm _ enabled _ flag may be information indicating whether intra BDPCM luminance information intra _ bdplcm _ luma _ flag and intra BDPCM chrominance information intra _ bdpcmp _ chroma _ flag, which will be described later, may exist in a coding unit syntax for an intra coding unit.
A value of sps _ bdpcmp _ enabled _ flag equal to "1" may specify that intra-BDPCM luminance information intra _ bdpcmp _ luma _ flag and intra-BDPCM chrominance information intra _ bdpcmp _ chroma _ flag, which will be described later, may exist in a coding unit syntax for an intra-coded unit.
A value of sps _ bdpcmm _ enabled _ flag equal to "0" may specify that there are no intra _ bdpcmm _ luma _ flag and intra _ bdpcmm _ chroma _ flag in the coding unit syntax for the intra coding unit.
When the value of equation 15 is "1," the first information may be selected to be written to the bitstream, and steps 1940 and 1945 may be performed.
When the value of equation 15 is "0", the second information may be selected to be written into the bitstream, and steps 1960 and 19070 may be performed through step 1955.
In step 1940, intra _ bdplcm chroma flag may be written to the bitstream.
The first information may include intra _ bdplcm _ chroma _ flag.
At step 1945, it may be determined whether the intra _ bdplcm _ chroma _ dir _ flag is to be written to the bitstream.
If it is determined that intra _ bdpcmm _ chroma _ dir _ flag is to be written to the bitstream, step 1950 may be performed.
If it is determined that intra _ bdpcmm _ chroma _ dir _ flag is not to be written to the bitstream, the process may be terminated, or alternatively, steps not shown may be performed.
Whether the intra _ bdpcm _ chroma _ dir _ flag is to be written to the bitstream may be determined based on the following equation 16.
[ formula 16]
intra_bdpcm_chroma_flag
When the value of equation 16 is "0", it may be determined that intra _ bdplcm _ chroma _ dir _ flag is not to be written to the bitstream and the process may be terminated, or alternatively, a step not shown may be performed.
When the value of equation 16 is "1", it may be determined that intra _ bdpcmm _ chroma _ dir _ flag is to be written to the bitstream, and step 1950 may be performed.
At step 1950, intra _ bdpcmm _ chroma _ dir _ flag may be written to the bitstream.
The first information may further include intra _ bdpcmm _ chroma _ dir _ flag.
When step 1950 is performed, the process may be terminated, or alternatively, other steps not shown may be performed.
At step 1960, it may be determined whether cclm _ mode _ flag is to be written to the bitstream.
If it is determined that cclm _ mode _ flag is to be written to the bitstream, step 1965 may be performed.
If it is determined that cclm _ mode _ flag is not to be written to the bitstream, step 1970 may be performed.
Whether cclm _ mode _ flag is to be written to the bitstream may be determined based on the following equation 17.
[ formula 17]
CclmEnabled
CclmEnabled may be a variable that indicates whether CCLM is enabled.
When the value of equation 17 is "0", it may be determined that cclm _ mode _ flag will not be written to the bitstream, and step 1970 may be performed.
When the value of equation 17 is "1", it may be determined that cclm _ mode _ flag is to be written to the bitstream, and step 1965 may be performed.
At step 1965, cclm _ mode _ flag may be written to the bitstream.
The second information may include cclm _ mode _ flag.
At step 1970, it may be determined whether cclm _ mode _ idx is to be written to the bitstream.
If it is determined that cclm _ mode _ idx is to be written to the bitstream, step 1975 may be performed.
If it is determined that cclm _ mode _ idx will not be written to the bitstream, step 1985 may be performed by step 1980.
Whether cclm _ mode _ idx is to be written to the bitstream may be determined based on equation 18 below.
[ formula 18]
cclm_mode_flag
When the value of equation 18 is "0", it may be determined that cclm _ mode _ idx will not be written to the bitstream, and step 1985 may be performed through step 1980.
When the value of equation 18 is "1," it may be determined that cclm _ mode _ idx is to be written to the bitstream, and step 1975 may be performed.
At step 1975, cclm _ mode _ idx may be written to the bitstream.
The second information may also include cclm _ mode _ idx.
When step 1975 is performed, the process may be terminated, or alternatively, other steps not shown may be performed.
In step 1985, intra _ chroma _ pred _ mode may be written to the bitstream.
The second information may further include intra _ chroma _ pred _ mode.
When step 1985 is performed, the process may be terminated, or alternatively, other steps not shown may be performed.
Image coding using modified syntax for target blocks
Fig. 20 shows a modified second syntax for a target block, in accordance with an embodiment.
One line of the code shown in fig. 20 may correspond to one step of the encoding method and one step of the decoding method.
The second syntax described above with reference to fig. 19 may be replaced with a modified second syntax to be described with reference to fig. 20. For example, steps 1910, 1915, 1920, 1925, 1930, 1935, 1940, 1945, 1950, 1955, 1960, 1965, 1970, 1975, 1980, and 1985 of fig. 19 may be replaced with steps 2010, 2015, 2020, 2025, 2030, 2035, 2040, 2045, 2050, 2055, 2060, 2065, 2070, 2075, 2080, and 2085 of fig. 20, respectively.
For example, step 1910 of FIG. 19 may be replaced with step 1920 of FIG. 20.
For example, the description of steps 1910, 1915, 1920, 1925, 1930, 1935, 1940, 1945, 1950, 1955, 1960, 1965, 1970, 1975, 1980, and 1985 of fig. 19 may be applied to steps 2010, 2015, 2020, 2025, 2030, 2035, 2040, 2045, 2050, 2055, 2060, 2065, 2070, 2075, 2080, and 2085 of fig. 20, respectively. Duplicate description will be omitted here.
At step 2015, it may be determined whether palette coding is to be performed on the target block.
Palette coding may be an operation of performing coding on a target block using a palette mode.
If it is determined that palette coding is to be performed on the target block, step 2020 may be performed.
Step 2025 may be performed if it is determined that palette coding will not be performed on the target block.
Whether palette coding is to be performed on the target block may be determined based on equation 13 above.
When the value of equation 13 is "0", it may be determined that palette coding is to be performed on the target block, and step 2020 may be performed.
When the value of equation 13 is "1", palette coding may not be performed on the target block, and step 2025 may be performed.
At step 2025, it may be determined whether information about the prediction mode is to be written to the bitstream.
The information on the prediction mode may be information on an additional prediction mode other than the palette prediction mode.
The information on the prediction mode may be the first information or the second information.
If it is determined that information regarding the prediction mode is to be written to the bitstream, step 2030 may be executed.
If it is determined that information on the prediction mode will not be written to the bitstream, the process may be terminated, or alternatively, a step not shown may be performed.
Whether information on a prediction mode is to be written into a bitstream may be determined based on the following equation 19.
[ formula 19]
!pred_mode_plt_flag
When the value of equation 19 is "0", it may be determined that information on the prediction mode will not be written to the bitstream, and the process may be terminated, or alternatively, a step not shown may be performed.
When the value of equation 19 is "1", it may be determined that information on the prediction mode is to be written in the bitstream, and step 2030 may be performed.
Fig. 21 is a flowchart of an image encoding method according to an embodiment.
At step 2110, a prediction mode for the chroma components of the target block may be determined.
The prediction mode for the chroma component of the target block may specify the method to be used for prediction for the chroma component of the target block.
For example, the prediction mode may be one of an intra mode, an inter mode, an Intra Block Copy (IBC) mode, and a palette mode.
For example, the prediction modes may include at least some of an intra sub-partition (ISP) mode, a block-Based Delta Pulse Code Modulation (BDPCM) mode, and a CCLM mode.
For example, the prediction mode may include the prediction method described in conjunction with the previous embodiments.
The mode may indicate whether each of an intra mode, an inter mode, an IBC mode, and a palette mode is used.
For example, the prediction mode may indicate whether a palette mode is used for the chroma components of the target block.
For example, the prediction mode may specify whether at least some of the modes indicated by the pieces of written information described above with reference to fig. 19 to 21 are used.
For example, the modes indicated by the pieces of written information may include an Intra Sub Partition (ISP) mode, a BDPCM mode, and a CCLM mode.
For example, the prediction mode may specify whether at least some of the prediction methods described in connection with the above embodiments are used.
The determination of the prediction mode may include determination of the values of the elements described above with reference to fig. 18 to 30. The element may be plural.
Each element may specify information to be written to the bitstream.
Each element may be a syntax element.
Each element may specify the variables used in the above formula.
Further, each element may be information described in the above-described embodiments.
When the value of each element is determined, prediction corresponding to the method specified by the value may be performed.
In step 2120, encoding of the chroma components of the target block may be performed according to the prediction mode.
Encoding the chrominance components of the target block may include writing the corresponding elements described above with reference to fig. 18 to 30 into a bitstream. In other words, encoding the chroma components of the target block may include generation of a bitstream.
Encoding the chroma component of the target block may include generation of encoding information regarding the chroma component of the target block. For example, the coding information may include quantized transform coefficient levels.
As described above with respect to steps 1830, 1835, 1840, 1910, 1915, 1920, 2010, 2015 and 2020, when the treeType of the target block is SINGLE _ TREE (i.e., when the partition structure of the luma component and the partition structure of the chroma component of the target block are identical to each other) and the value of pred _ mode _ plt _ flag of the target block is "1" (i.e., when the prediction mode of the target block is the palette mode), prediction using the palette mode may be performed with respect to both the luma component and the chroma component of the target block. Pixels reconstructed by prediction using this palette mode may be generated. Further, since the palette mode is used for prediction, information about the palette mode may be written in the bitstream.
In an embodiment, step 2015 of FIG. 20 may be performed when treeType is SINGLE _ TREE or when treeType is DUAL _ TREE _ CHROMA, as described in step 2010 of FIG. 20.
Next, when the value of pred _ mode _ plt _ flag is "1" and treeType is DUAL _ TREE _ CHROMA, step 2020 of fig. 20 may be performed. Thus, step 2020 may be an operation performed on the chroma components of the target block.
At step 2020, palette coding may be performed on the chroma components of the target block.
As described with respect to step 2010, step 2015 may be performed when the partition structure of the luminance component of the target block and the partition structure of the chrominance component of the target block are identical to each other (i.e., when treeType ═ SINGLE _ TREE).
Step 2020 can be performed when the TREE _ type of the target block is DUAL _ TREE _ CHROMA, as described above with respect to step 2015. Therefore, when the TREE _ type of the target block is SINGLE _ TREE, step 2020 may not be performed, and step 2025 may be performed. In other words, when the prediction mode of the chrominance component of the target block is determined, if the partition structure of the luminance component of the target block and the partition structure of the chrominance component of the target block are identical to each other (i.e., when TREE _ type ═ SINGLE _ TREE), step 2025 is performed after step 2015.
Next, in step 2025, based on whether the palette mode is used for the chroma component of the target block (i.e., the value of | pred _ mode _ plt _ flag), it may be determined whether information regarding the prediction mode of the chroma component of the target block is encoded.
The information on the prediction mode may include one or more of the above-described intra _ bdpcmm _ chroma _ flag, intra _ bdpcmm _ chroma _ dif _ flag, cclm _ mode _ idx, and intra _ chroma _ pred _ mode.
The information on the prediction mode may be information on an additional prediction mode other than the palette mode. In other words, the information about the prediction mode may be information to be signaled when the palette mode (for the chroma components of the target block) is not used.
The encoding of the information regarding the prediction mode of the chroma component of the target block may specify operations corresponding to steps 2030, 2035, 2040, 2045, 2050, 2055, 2060, 2065, 2070, 2075, 2080 and 2085 described with reference to fig. 20.
As described above with reference to step 2025 of fig. 20, when the value of equation 11 is "1", the value of equation 12 is "0", and the value of equation 13 is "1", steps 2030, 2035, 2040, 2045, 2050, 2055, 2060, 2065, 2070, 2075, 2080 and 2085 may be performed. Accordingly, encoding of information regarding a prediction mode of a chrominance component of a target block may be performed only when the following conditions 1, 2, and 3 are satisfied.
Condition 1) the TREE _ type of the target block is SINGLE _ TREE or DUAL _ TREE _ CHROMA
Condition 2) the value of pred _ mode _ plt _ flag is "0", or treeType is not DUAL _ TREE _ CHROMA
Condition 3) the value of pred _ mode _ plt _ flag is "0"
In other words, in the case where the chroma component of the target block is encoded, when the palette mode is not used for the chroma component of the target block (i.e., when the value of pred _ mode _ plt _ flag is "0"), the encoding of information on the prediction mode of the chroma component of the target block may be performed (i.e., steps 2030, 2035, 2040, 2045, 2050, 2055, 2060, 2065, 2070, 2075, 2080, and 2085). Further, in the case where the chroma component of the target block is encoded, when the palette mode is used for the chroma component of the target block (i.e., when the value of pred _ mode _ plt _ flag is "1"), the encoding of information on the prediction mode of the chroma component of the target block may be skipped (i.e., steps 2030, 2035, 2040, 2045, 2050, 2055, 2060, 2065, 2070, 2075, 2080, and 2085).
For example, when the partition structure of the luminance component of the target block and the partition structure of the chrominance component of the target block are identical to each other, information on the prediction mode may be encoded when the palette mode is not used for the chrominance component of the target block.
Here, the encoding of the information on the prediction mode may be performed at steps 2030, 2035, 2040, 2045, 2050, 2055, 2060, 2065, 2070, 2075, 2080, and 2085.
Accordingly, when information on a prediction mode for a chroma component of a target block is encoded, first, at step 1820, pred _ mode _ plt _ flag indicating whether palette prediction is used for the target block may be encoded. Next, when the partition structure of the luminance component of the palette prediction target block and the partition structure of the chrominance component of the target block are identical to each other (i.e., when the TREE _ type is SINGLE _ TREE) for the target block and the palette prediction for the target block is not used, information for an additional prediction mode other than the palette mode may be neither written nor transmitted.
Image decoding using syntax for target block
The code shown in fig. 18 may represent image decoding in addition to image encoding.
Reference is made back to fig. 18. One line of the code shown in fig. 18 may correspond to one step of an encoding method and one step of a decoding method.
Hereinafter, the code lines of fig. 18 will be described as individual steps from the viewpoint of decoding.
Steps 1810, 1815, 1820, 1825, 1830, 1835, 1840, 1845, 1850, 1855, 1860, and 1865 of fig. 18 may be performed by the processing unit 1710 of the decoding apparatus 1700.
The target block may be a CU.
At step 1810, decoding of the target block may be initiated.
x0 may be the x coordinate of the upper left portion of the target block.
y0 may be the y coordinate of the upper left portion of the target block.
cbWidth may be the width of the target block.
cbheight may be the height of the target block.
The cqtDepth may be the depth of the currently encoded quadtree.
the treeType may be information indicating whether a Coding Tree Unit (CTU) is a single tree or a double tree used to partition the CTU.
Further, when a dual tree is used for the CTU, treeType may be information indicating whether a luminance component or a chrominance component is currently being processed.
the value of treeType may be one of SINGLE _ TREE, DUAL _ TREE _ LUMA, and DUAL _ TREE _ CHROMA.
SINGLE TREE may indicate that a SINGLE TREE is used for the CTU. SINGLE _ TREE may represent that the partition structure of the luminance component of the CTU (i.e., the partition structure of the luminance component of the target block) is the same as the partition structure of the chrominance component of the target block.
DUAL TREE LUMA may indicate that a DUAL TREE is used for the CTU and that the LUMA component of the CTU is currently being processed.
DUAL _ TREE _ CHROMA may indicate that the CHROMA component of the CTU is currently being processed.
The modeType may be information indicating the following for a CU in the CTU: 1) whether all of the intra prediction mode, the IBC mode, the palette mode, and the inter prediction mode can be used, 2) whether only the intra prediction mode, the IBC mode, and the palette mode can be used, and 3) whether only the inter mode can be used.
Hereinafter, the terms "CTU" and "coding tree node" may be used to have the same meaning and may be used interchangeably with each other.
The value of modeType may be one of MODE _ TYPE _ ALL, MODE _ TYPE _ INTRA, and MODE _ TYPE _ INTER.
MODE _ TYPE _ ALL may indicate for a CU in the CTU that an intra prediction MODE, an IBC MODE, a palette MODE, and an inter prediction MODE may ALL be used.
MODE _ TYPE _ INTRA may indicate for a CU in the CTU that an INTRA prediction MODE, an IBC MODE, and a palette MODE may be used.
MODE _ TYPE _ INTER may indicate for a CU in the CTU that only INTER MODE may be used.
In step 1815, it may be determined whether pred _ mode _ plt _ flag is to be read from the palette information bitstream.
pred _ mode _ plt _ flag may be information indicating whether the target block is a palette block. The palette block may be a block that is decoded using the palette mode.
If it is determined that pred _ mode _ plt _ flag is to be read from the bitstream, step 1820 may be performed.
If it is determined that pred _ mode _ plt _ flag will not be read from the bitstream, step 1820 may be performed.
Whether pred _ mode _ plt _ flag is to be read from the bitstream can be determined based on the following equation 20.
[ formula 20]
CuPredMode[chType][x0][y0]==MODE_INTRA&&sps_palette_enabled_flag&&cbWidth<=64&&cbHeight<=64&&cu_skip_flag[x0][y0]==0&&modeType!=MODE_TYPE_INTER
The chType may be information indicating whether a value of treeType, which will be described later, is DUAL _ TREE _ CHROMA.
CuPredMode [ chType ] [ x0] [ y0] may be information indicating a prediction mode of a target block.
The value of CuPredMode [ chType ] [ x0] [ y0] may be one of MODE _ INTRA, MODE _ INTER, MODE _ PLT and MODE _ IBC.
MODE _ INTRA may indicate that the prediction MODE for the target block is an INTRA MODE.
MODE INTER may indicate that the prediction MODE for the target block is INTER MODE.
The MODE _ PLT may indicate that the prediction MODE for the target block is a palette MODE.
MODE _ IBC may indicate that the prediction MODE for the target block is IBC MODE.
The SPS _ palette _ enabled _ flag may be information indicating whether or not the palette mode is enabled for SPS.
cu _ skip _ flag may be information indicating whether a skip mode is applied to the target block.
When the value of equation 20 is "0", it may be determined that pred _ mode _ plt _ flag will not be read from the bitstream, and step 1825 may be performed. In an embodiment, the result or information value of "0" of equation (la) may be replaced with the first value or with "false". Hereinafter, a repetitive explanation will be omitted.
When the value of equation 20 is "1", it may be determined that pred _ mode _ plt _ flag is to be read from the bitstream, and step 1820 may be performed. In an embodiment, the result or information value of "1" of equation (la) may be replaced with a second value or with "true". Hereinafter, a repetitive explanation will be omitted.
Equation 20 above may be replaced with equation 21 below.
[ formula 21]
CuPredMode[chType][x0][y0]==MODE_INTRA&&sps_palette_enabled_flag&&cbWidth<=64&&cbHeight<=64&&cu_skip_flag[x0][y0]==0&&modeType!=MODE_TYPE_INTER&&((cbWidth*cbHeight)>(treeType!=DUAL_TREE_CHROMA16:16*SubWidthC*SubHeightC))&&(modeType!=MODE_TYPE_INTRA||treeType!=DUAL_TREE_CHROMA))
The SubWidthC may be information indicating the number of samples to be sampled in the width direction of the chrominance component. For example, the width of the chrominance component of the target block may be calculated using equation 22 below.
[ formula 22]
cbWidth/SubWidthC
The subheight c may be information indicating the number of samples to be sampled in the height direction of the chrominance component. For example, the height of the chrominance component of the target block may be calculated using equation 23 below.
[ formula 23]
cbHeight/SubHeightC
In step 1820, pred _ mode _ plt _ flag may be read from the bitstream.
When step 1820 is performed, step 1825 may be performed subsequently.
At step 1825, it may be determined whether step 1830 is to be performed.
If it is determined that step 1830 is to be performed, step 1830 may be performed.
If it is determined that step 1830 is not to be performed, step 1910 may be performed.
Whether step 1830 is to be performed may be determined based on equation 24 below.
[ formula 24]
CuPredMode[chType][x0][y0]==MODE_INTRA||CuPredMode[chType][x0][y0]==MODE_PLT
When the value of equation 24 is "0," step 1830 may not be performed and step 1910 may be performed.
When the value of equation 24 is "1," step 1830 may be performed.
At step 1830, it may be determined whether step 1835 is to be performed.
If it is determined that step 1835 is to be performed, step 1835 may be performed.
If it is determined that step 1835 is not to be performed, step 1910 may be performed.
Whether step 1835 is to be performed may be determined based on equation 25 below.
[ formula 25]
treeType==SINGLE_TREEtreeType==DUAL_TREE_LUMA
When the value of equation 25 is "0", step 1835 may not be performed and step 1910, which will be described later, may be performed.
When the value of equation 25 is "1," step 1835 may be performed.
At step 1835, it may be determined whether palette coding is to be performed on the target block.
Palette coding may be an operation of performing decoding on a target block using a palette mode.
If it is determined that palette coding is to be performed on the target block, step 1840 may be performed.
If it is determined that palette coding will not be performed on the target block, steps 1850 and 1860 may be performed by step 1845.
Whether palette coding is to be performed on the target block may be determined based on equation 26 below.
[ formula 26]
pred_mode_plt_flag
When the value of equation 26 is "0," it may be determined that palette coding is to be performed on the target block, and step 1840 may be performed.
When the value of equation 26 is "1", palette coding may not be performed on the target block, and steps 1850 and 1860 may be performed through step 1845.
At step 1840, palette coding may be performed on the target block.
palette _ coding may specify a palette coding process for a target block.
In step 1850, it may be determined whether intra _ sub partitions _ mode _ flag is to be read from the bitstream.
The intra _ sub _ modes _ flag may be information indicating an intra sub partition mode for the target block.
A value of intra _ subportions _ mode _ flag equal to "1" may specify that the current intra-coded unit is partitioned into numintrasubportions x0 y0 sub-partitions of transform blocks.
A value of intra _ sub partitions _ mode _ flag equal to "0" may specify that the current intra coding unit is not partitioned into transform block sub-partitions.
If it is determined that intra _ sub _ modes _ flag is to be read from the bitstream, step 1855 may be performed.
If it is determined that intra _ sub _ partitions _ mode _ flag is not to be read from the bitstream, step 1860 may be performed.
Whether intra _ subbands _ mode _ flag is to be read from the bitstream may be determined based on the following equation 27.
[ formula 27]
sps_isp_enabled_flag&&intra_luma_ref_idx==0&&(cbWidth<=MaxTbSizeY&&cbHeight<=MaxTbSizeY)&&(cbWidth*cbHeight>MinTbSizeY*MinTbSizeY)&&!cu_act_enabled_flag
The sps _ isp _ enabled _ flag may be information indicating whether intra prediction with sub-partitions is enabled for a Coded Layer Video Sequence (CLVS).
A value of sps _ isp _ enabled _ flag equal to "1" may specify that intra prediction with sub-partitions is enabled for CLVS.
A value of sps _ isp _ enabled _ flag equal to "0" may specify that intra prediction with sub-partitions is disabled for CLV.
intra _ luma _ ref _ idx may be index information that specifies a reference sample line to be used for intra prediction.
MaxTbSizeY may be the maximum transform size (maximum size of transform unit) in the luminance component.
MinTbSizeY may be the minimum transform size (minimum size of transform unit) in the luminance component.
cu _ act _ enabled _ flag may be information indicating whether color space conversion is used for a decoded residual of a target block. This may be information indicating whether color space conversion is enabled.
A value of cu act enabled flag equal to "1" may specify that the decoding residual of the target block is applied in case of using color space conversion.
A value of cu act enabled flag equal to "0" may specify that the decoding residual of the target block is applied without using color space conversion.
When the value of equation 27 is "0", it may be determined that intra _ sub _ modes _ flag will not be read from the bitstream, and step 1860 may be performed.
When the value of equation 27 is "1", it may be determined that intra _ sub _ modes _ flag is to be read from the bitstream, and step 1855 may be performed.
In step 1855, intra _ sub _ modes _ flag may be read from the bitstream.
When step 1855 is performed, step 1860 may then be performed.
In step 1860, it may be determined whether intra _ sub partitions _ split _ flag is read from a bitstream.
The intra _ sub _ partitions _ split _ flag may be information specifying an intra sub-partition division type for the target block.
When the value of intra _ subpartitions _ mode _ flag is "0", the value of intrasubpartitionsplittype may be 0.
When the value of intra _ subpartitions _ mode _ flag is "1", the value of intrasubpartitionspatitype can be determined to be represented by the following equation 28.
[ formula 28]
IntraSubPartitionsSplitType=intra_subpartitions_mode_flag+intra_subpartitions_split_flag
A value of intrasubpartitionsplit type equal to "0" may specify that the target block is not partitioned.
A value of intrasubpartitionsplit type equal to "1" may specify that the target block is divided in the horizontal direction.
A value of intrasubpartitionsplit type equal to "2" may specify that the target block is divided in the vertical direction.
If it is determined that intra _ sub _ partitions _ split _ flag is to be read from the bitstream, step 1865 may be performed.
If it is determined that the intra _ sub _ partitions _ split _ flag is not to be read from the bitstream, step 1910 may be performed.
Whether intra _ sub _ modes _ flag is to be read from the bitstream may be determined based on the following equation 29.
[ formula 29]
intra_subpartitions_mode_flag==1
When the value of equation 29 is "0", it may be determined that intra _ sub _ partitions _ split _ flag will not be read from the bitstream, and step 1910 may be performed.
When the value of equation 29 is "1", it may be determined that intra _ sub _ partitions _ split _ flag is to be read from the bitstream, and step 1865 may be performed.
In step 1865, intra _ sub _ splits _ flag may be read from the bitstream.
When step 1865 is performed, step 1910 may then be performed.
The code shown in fig. 19 may represent image decoding in addition to image encoding
Reference is made back to fig. 19. One line of the code shown in fig. 19 may correspond to one step of the encoding method and one step of the decoding method.
Hereinafter, the code lines of fig. 19 will be described as respective steps from the viewpoint of decoding.
Steps 1910, 1915, 1920, 1925, 1930, 1935, 1940, 1945, 1950, 1955, 1960, 1965, 1970, 1975, 1980, and 1985 of fig. 19 may be performed by the processing unit 1710 of the decoding apparatus 1700.
The target block may be a CU.
The second syntax, which will be described with reference to fig. 19, may be performed after the first syntax described above with reference to fig. 18.
At step 1910, it may be determined whether step 1915 is to be performed.
If it is determined that step 1915 is to be performed, step 1915 may be performed.
If it is determined that step 1915 is not to be performed, the process may be terminated, or alternatively, other steps not shown may be performed.
Based on equation 30 below, it may be determined whether step 1915 is to be performed.
[ formula 30]
(treeType==SINGLE_TREE||treeType==DUAL_TREE_CHROMA)&&ChromaArrayType!=0
The ChromaArrayType may be information that specifies the chroma format of the target block.
When the value of equation 30 is "0", step 1815 may not be performed and the process may be terminated, or alternatively, other steps not shown may be performed.
When the value of equation 30 is "1," step 1915 may be performed.
Formula 30 above may be replaced with formula 31 below.
[ formula 31]
(treeType==SINGLE_TREE||treeType==DUAL_TREE_CHROMA)&&sps_chroma_format_idc!=0
The sps _ chroma _ format _ idc may be information indicating a chroma format of the target block, a value of subwidtc, and a value of subwight c.
When the value of sps _ chroma _ format _ idc is "0", the chroma format may be monochrome, the value of SubWidthC may be "1", and the value of subwight c may be "1".
When the value of sps _ chroma _ format _ idc is "1", the chroma format may be "4: 2: 0", the value of SubWidthC may be "2", and the value of subwight c may be "2".
When the value of sps _ chroma _ format _ idc is "2", the chroma format may be "4: 2: 2", the value of SubWidthC may be "2", and the value of subwight c may be "1".
When the value of sps _ chroma _ format _ idc is "3", the chroma format may be "4: 4: 4", the value of SubWidthC may be "1", and the value of subwight c may be "1".
When the value of equation 31 is "0", step 1815 may not be performed and the process may be terminated, or alternatively, other steps not shown may be performed.
When the value of equation 31 is "1," step 1915 may be performed.
At step 1915, it may be determined whether palette coding is to be performed on the target block.
Palette coding may be an operation of performing decoding on a target block using a palette mode.
If it is determined that palette coding is to be performed on the target block, step 1920 may be performed.
If it is determined that palette coding is not to be performed on the target block, step 1930 may be performed by step 1925.
Whether palette coding is to be performed on the target block may be determined based on equation 32 below.
[ formula 32]
pred_mode_plt_flag&&treeType==DUAL_TREE_CHROMA
When the value of equation 32 is "0," it may be determined that palette coding is to be performed on the target block, and step 1920 may be performed.
When the value of equation 32 is "1", palette coding may not be performed on the target block, and step 1930 may be performed through step 1925.
At step 1920, palette coding may be performed on the target block.
palette _ coding may specify a palette coding process for a target block.
When step 1920 is performed, the process may be terminated, or alternatively, other steps not shown may be performed.
At step 1930, it may be determined whether step 1935 is to be performed.
If it is determined that step 1935 is to be performed, step 1935 may be performed.
If it is determined that step 1935 is not to be performed, the process can be terminated, or alternatively, other steps not shown can be performed.
Whether step 1935 is to be performed may be determined based on equation 33 below.
[ formula 33]
!cu_act_enabled_flag
When the value of equation 33 is "0," step 1935 may not be performed and the process may be terminated, or alternatively, other steps not shown may be performed.
When the value of equation 33 is "1," step 1935 may be performed.
In step 1935, information to be read from the bitstream can be selected from the first information and the second information.
Steps 1940 and 1945 may be performed when the first information is selected as information to be read from the bitstream.
When the second information is selected as the information to be read from the bitstream, steps 1960 and 19070 may be performed by 1955.
The first information may include at least some of block-Based Delta Pulse Code Modulation (BDPCM) availability information intra _ bdpcmc _ chroma _ flag and BDPCM prediction direction information intra _ bdpcmc _ chroma _ dir _ flag.
intra _ bdpcmp _ chroma _ flag may be information indicating whether BDPCM is applied to the current chroma coding block. The current chroma coding block may represent the chroma component of the target block.
A value of intra _ bdpcmm _ chroma _ flag equal to "1" may specify that BDPCM is applied to the current chroma coding block.
A value of intra _ bdpcmm _ chroma _ flag equal to "1" may specify that BDPCM is not applied to the current chroma coding block.
The intra _ bdpcmp _ chroma _ dir _ flag may specify a prediction direction of the BDPCM.
A value of intra _ bdpcmm _ chroma _ dir _ flag equal to "0" may specify that the prediction direction of the BDPCM is the horizontal direction.
A value of intra _ bdpcmm _ chroma _ dir _ flag equal to "1" may specify that the prediction direction of the BDPCM is the vertical direction.
The second information may include at least some of cross-component linear model (CCLM) mode information CCLM _ mode _ flag, CCLM mode index information CCLM _ mode _ idx, and intra chroma prediction mode information intra _ chroma _ pred _ mode.
CCLM _ mode _ flag may be information indicating whether intra prediction of a cross-component linear model (CCLM) is used for a target block.
The CCLM _ mode _ flag may be information indicating whether one of the INTRA _ LT _ CCLM chroma INTRA prediction mode, the INTRA _ L _ CCLM chroma INTRA prediction mode, and the INTRA _ T _ CCLM chroma INTRA prediction mode is applied.
A value of CCLM _ mode _ flag equal to "1" may specify that one of the INTRA _ LT _ CCLM chroma INTRA prediction mode, the INTRA _ L _ CCLM chroma INTRA prediction mode, and the INTRA _ T _ CCLM chroma INTRA prediction mode is applied to the target block.
A value of CCLM _ mode _ flag equal to "0" may specify that the INTRA _ LT _ CCLM chroma INTRA prediction mode, the INTRA _ L _ CCLM chroma INTRA prediction mode, and the INTRA _ T _ CCLM chroma INTRA prediction mode are not all applied to the target block.
The CCLM _ mode _ idx may be information indicating which one of the INTRA _ LT _ CCLM chroma INTRA prediction mode, the INTRA _ L _ CCLM chroma INTRA prediction mode, and the INTRA _ T _ CCLM chroma INTRA prediction mode is applied to the target block.
The value of CCLM _ mode _ idx may be one of INTRA _ LT _ CCLM, INTRA _ L _ CCLM, and INTRA _ T _ CCLM.
intra _ chroma _ pred _ mode may be information indicating an intra prediction mode for a chroma component (or chroma sampling point) of a target block.
The information to be read from the bitstream may be selected based on the following equation 34.
[ formula 34]
cbWidth/SubWidthC<=MaxTsSize&&cbHeight/SubHeightC<=MaxTsSize&&sps_bdpcm_enabled_flag
Maxttssize may be the maximum block size for transform skipping.
The sps _ bdplcm _ enabled _ flag may be information indicating whether intra BDPCM luminance information intra _ bdplcm _ luma _ flag and intra BDPCM chrominance information intra _ bdpcmp _ chroma _ flag, which will be described later, may exist in a coding unit syntax for an intra coding unit.
A value of sps _ bdpcmp _ enabled _ flag equal to "1" may specify that intra-BDPCM luminance information intra _ bdpcmp _ luma _ flag and intra-BDPCM chrominance information intra _ bdpcmp _ chroma _ flag, which will be described later, may exist in a coding unit syntax for an intra-coded unit.
A value of sps _ bdpcmm _ enabled _ flag equal to "0" may specify that there are no intra _ bdpcmm _ luma _ flag and intra _ bdpcmm _ chroma _ flag in the coding unit syntax for the intra coding unit.
When the value of equation 34 is "1," it may be selected to read the first information from the bitstream, and steps 1940 and 1945 may be performed.
When the value of equation 34 is "0", the second information may be selected to be read from the bitstream, and steps 1960 and 19070 may be performed through step 1955.
In step 1940, intra _ bdpcmm _ chroma _ flag may be read from the bitstream.
The first information may include intra _ bdplcm _ chroma _ flag.
In step 1945, it may be determined whether intra _ bdpcmm _ chroma _ dir _ flag is to be read from the bitstream.
If it is determined that intra _ bdpcm _ chroma _ dir _ flag is to be read from the bitstream, step 1950 may be performed.
If it is determined that intra _ bdpcm _ chroma _ dir _ flag is not to be read from the bitstream, the process may be terminated, or alternatively, a step not shown may be performed.
Whether intra _ bdcpcm _ chroma _ dir _ flag is to be read from the bitstream may be determined based on the following equation 35.
[ formula 35]
intra_bdpcm_chroma_flag
When the value of equation 35 is "0", it may be determined that intra _ bdpcmm _ chroma _ dir _ flag will not be read from the bitstream and the process may be terminated, or alternatively, a step not shown may be performed.
When the value of equation 35 is "1", it may be determined that intra _ bdpcmm _ chroma _ dir _ flag is to be read from the bitstream, and step 1950 may be performed.
At step 1950, intra _ bdpcmm _ chroma _ dir _ flag may be read from the bitstream.
The first information may further include intra _ bdpcmm _ chroma _ dir _ flag.
When step 1950 is performed, the process may terminate, or alternatively, other steps not shown may be performed.
At step 1960, it may be determined whether cclm _ mode _ flag is to be read from the bitstream.
If it is determined that cclm _ mode _ flag is to be read from the bitstream, step 1965 may be performed.
Step 1970 may be performed if it is determined that cclm _ mode _ flag will not be read from the bitstream.
Whether cclm _ mode _ flag is to be read from the bitstream may be determined based on the following equation 36.
[ formula 36]
CclmEnabled
CclmEnabled may be a variable that indicates whether CCLM is enabled.
When the value of equation 36 is "0," it may be determined that cclm _ mode _ flag will not be read from the bitstream, and step 1970 may be performed.
When the value of equation 36 is "1", it may be determined that cclm _ mode _ flag is to be read from the bitstream, and step 1965 may be performed.
At step 1965, cclm _ mode _ flag may be read from the bitstream.
The second information may include cclm _ mode _ flag.
At step 1970, it may be determined whether cclm _ mode _ idx is to be read from the bitstream.
If it is determined that cclm _ mode _ idx is to be read from the bitstream, step 1975 may be performed.
If it is determined that cclm _ mode _ idx will not be read from the bitstream, step 1985 may be performed by step 1980.
Whether cclm _ mode _ idx is to be read from the bitstream may be determined based on equation 37 below.
[ formula 37]
cclm_mode_flag
When the value of equation 37 is "0", it may be determined that cclm _ mode _ idx will not be read from the bitstream, and step 1985 may be performed through step 1980.
When the value of equation 37 is "1," it may be determined that cclm _ mode _ idx is to be read from the bitstream, and step 1975 may be performed.
At step 1975, cclm _ mode _ idx may be read from the bitstream.
The second information may also include cclm _ mode _ idx.
When step 1975 is performed, the process may be terminated, or other steps not shown may alternatively be performed.
In step 1985, intra _ chroma _ pred _ mode may be read from the bitstream.
The second information may further include intra _ chroma _ pred _ mode.
When step 1985 is performed, the process may be terminated, or alternatively, other steps not shown may be performed.
Image decoding using modified syntax for target blocks
The code shown in fig. 20 may represent image decoding in addition to image encoding.
Reference is made back to fig. 20. One line of the code shown in fig. 20 may correspond to one step of the encoding method and one step of the decoding method.
The second syntax described above with reference to fig. 19 may be replaced with a modified second syntax to be described with reference to fig. 20. For example, steps 1910, 1915, 1920, 1925, 1930, 1935, 1940, 1945, 1950, 1955, 1960, 1965, 1970, 1975, 1980, and 1985 of fig. 19 may be replaced with steps 2010, 2015, 2020, 2025, 2030, 2035, 2040, 2045, 2050, 2055, 2060, 2065, 2070, 2075, 2080, and 2085 of fig. 20, respectively.
For example, step 1910 of FIG. 19 may be replaced with step 1920 of FIG. 20.
For example, the description of steps 1910, 1915, 1920, 1925, 1930, 1935, 1940, 1945, 1950, 1955, 1960, 1965, 1970, 1975, 1980, and 1985 of fig. 19 may be applied to steps 2010, 2015, 2020, 2025, 2030, 2035, 2040, 2045, 2050, 2055, 2060, 2065, 2070, 2075, 2080, and 2085 of fig. 20, respectively. Duplicate description will be omitted here.
At step 2015, it may be determined whether palette coding is to be performed on the target block.
Palette coding may be an operation of performing decoding on a target block using a palette mode.
If it is determined that palette coding is to be performed on the target block, step 2020 may be performed.
Step 2025 may be performed if it is determined that palette coding will not be performed on the target block.
Whether palette coding is to be performed on the target block may be determined based on equation 32 above.
When the value of equation 32 is "0", it may be determined that palette coding is to be performed on the target block, and step 2020 may be performed.
When the value of equation 32 is "1," palette coding may not be performed on the target block, and step 2025 may be performed.
At step 2025, it may be determined whether information about the prediction mode is to be read from the bitstream.
The information on the prediction mode may be information on an additional prediction mode other than the palette prediction mode.
The information on the prediction mode may be the first information or the second information.
If it is determined that information regarding the prediction mode is to be read from the bitstream, step 2030 may be executed.
If it is determined that information about the prediction mode will not be read from the bitstream, the process may be terminated, or alternatively, steps not shown may be performed.
Whether or not information on the prediction mode will be read from the bitstream may be determined based on the following equation 38.
[ formula 38]
!pred_mode_plt_flag
When the value of equation 38 is "0", it may be determined that information on the prediction mode will not be read from the bitstream and the process may be terminated, or alternatively, a step not shown may be performed.
When the value of equation 38 is "1", it may be determined that information on the prediction mode will be read from the bitstream, and step 2030 may be performed.
Fig. 22 is a flowchart of an image decoding method according to an embodiment.
At step 2210, a prediction mode for the chroma component of the target block may be determined.
The prediction mode for the chroma component of the target block may specify the method to be used for prediction for the chroma component of the target block.
For example, the prediction mode may be one of an intra mode, an inter mode, an Intra Block Copy (IBC) mode, and a palette mode.
For example, the prediction modes may include at least some of an intra sub-partition (ISP) mode, a block-Based Delta Pulse Code Modulation (BDPCM) mode, and a CCLM mode.
For example, the prediction mode may include the prediction method described in conjunction with the previous embodiments.
The mode may indicate whether each of an intra mode, an inter mode, an IBC mode, and a palette mode is used.
For example, the prediction mode may indicate whether a palette mode is used for the chroma components of the target block.
For example, the prediction mode may specify whether at least some of the modes indicated by the pieces of read information described above with reference to fig. 19 to 21 are used.
For example, the modes indicated by the pieces of read information may include an Intra Sub Partition (ISP) mode, a BDPCM mode, and a CCLM mode.
For example, the prediction mode may specify whether at least some of the prediction methods described in connection with the above embodiments are used.
The determination of the prediction mode may include determination of values of elements described above with reference to fig. 18 to 30. The element may be plural.
Each element may specify information to be read from the bitstream.
Each element may be a syntax element.
Each element may specify the variables used in the above formula.
Further, each element may be information described in the above-described embodiments.
When the value of each element is determined, prediction corresponding to the method specified by the value may be performed.
In step 2220, decoding of the chroma components of the target block may be performed according to the prediction mode.
The determination of the prediction mode for the chrominance components of the target block may include reading the elements described above with reference to fig. 18 to 30 from the bitstream. In other words, the determination of the prediction mode may include an operation of reading a specified element from the bitstream based on the above syntax. Decoding the chroma component of the target block may include reconstruction of the chroma component for the target block.
Decoding the chroma component of the target block may include generation of decoding information regarding the chroma component of the target block. For example, the decoding of the chrominance component of the target block may include an operation of performing decoding on encoding information regarding the chrominance component of the target block in the bitstream. The encoding information about the target block may include quantized transform coefficient levels.
As described above with respect to steps 1830, 1835, 1840, 1910, 1915, 1920, 2010, 2015, and 2020, when the treettype of the target block is SINGLE _ TREE (i.e., when the partition structure of the luma component of the target block and the partition structure of the chroma component of the target block are identical to each other) and the value of pred _ mode _ plt _ flag of the target block is "1" (i.e., when the prediction mode of the target block is the palette mode), prediction using the palette mode may be performed for both the luma component and the chroma component of the target block. Pixels reconstructed by prediction using this palette mode may be generated. Further, since the palette mode is used for prediction, information on the palette mode may be read from the bitstream.
In an embodiment, step 2015 of FIG. 20 may be performed when treeType is SINGLE _ TREE or when treeType is DUAL _ TREE _ CHROMA, as described in step 2010 of FIG. 20.
Next, when the value of pred _ mode _ plt _ flag is "1" and treeType is DUAL _ TREE _ CHROMA, step 2020 of FIG. 20 may be performed. Thus, step 2020 may be an operation performed on the chroma components of the target block.
At step 2020, palette coding may be performed on the chroma components of the target block.
As described with respect to step 2010, step 2015 may be performed when the partition structure of the luminance component and the partition structure of the chrominance component of the target block are identical to each other (i.e., when treeType ═ SINGLE _ TREE).
As described above with respect to step 2015, when the TREE _ type of the target block is DUAL _ TREE _ CHROMA, step 2020 may be performed. Therefore, when the TREE _ type of the target block is SINGLE _ TREE, step 2020 may not be performed, and step 2025 may be performed. In other words, when the prediction mode of the chrominance component of the target block is determined, if the partition structure of the luminance component of the target block and the partition structure of the chrominance component of the target block are identical to each other (i.e., when TREE _ type ═ SINGLE _ TREE), step 2025 is performed after step 2015.
Next, in step 2025, based on whether the palette mode is used for the chroma component of the target block (i.e., the value of | pred _ mode _ plt _ flag), it may be determined whether information on the prediction mode of the chroma component of the target block is decoded.
The information on the prediction mode may include one or more of the above-described intra _ bdpcmm _ chroma _ flag, intra _ bdpcmm _ chroma _ dif _ flag, cclm _ mode _ idx, and intra _ chroma _ pred _ mode.
The information on the prediction mode may be information on an additional prediction mode other than the palette mode. In other words, the information on the prediction mode may be information to be signaled when the palette mode (for the chroma components of the target block) is not used.
The decoding of the information regarding the prediction mode of the chroma component of the target block may specify operations corresponding to steps 2030, 2035, 2040, 2045, 2050, 2055, 2060, 2065, 2070, 2075, 2080, and 2085 described with reference to fig. 20.
As described above with reference to step 2025 of fig. 20, when the value of equation 30 is "1", the value of equation 31 is "0", and the value of equation 32 is "1", steps 2030, 2035, 2040, 2045, 2050, 2055, 2060, 2065, 2070, 2075, 2080, and 2085 may be performed. Accordingly, decoding of information regarding a prediction mode of a chroma component of a target block may be performed only when the following conditions 1, 2, and 3 are satisfied.
Condition 1) the TREE _ type of the target block is SINGLE _ TREE or DUAL _ TREE _ CHROMA
Condition 2) the value of pred _ mode _ plt _ flag is "0", or treeType is not DUAL _ TREE _ CHROMA
Condition 3) value of pred _ mode _ plt _ flag is "0"
In other words, in the case where the chroma component of the target block is decoded, when the palette mode is not used for the chroma component of the target block (i.e., when the value of pred _ mode _ plt _ flag is "0"), decoding of information related to the prediction mode of the chroma component of the target block may be performed (i.e., steps 2030, 2035, 2040, 2045, 2050, 2055, 2060, 2065, 2070, 2075, 2080, and 2085). Further, in the case where the chroma component of the target block is decoded, when the palette mode is used for the chroma component of the target block (i.e., when the value of pred _ mode _ plt _ flag is "1"), decoding of information related to the prediction mode of the chroma component of the target block may be skipped (i.e., steps 2030, 2035, 2040, 2045, 2050, 2055, 2060, 2065, 2070, 2075, 2080, and 2085).
For example, when the partition structure of the luminance component of the target block and the partition structure of the chrominance component of the target block are identical to each other, information on the prediction mode may be decoded when the palette mode is not used for the chrominance component of the target block.
Here, the decoding of the information on the prediction mode may be performed at steps 2030, 2035, 2040, 2045, 2050, 2055, 2060, 2065, 2070, 2075, 2080, and 2085.
Accordingly, when information on a prediction mode for a chroma component of a target block is decoded, first, at step 1820, pred _ mode _ plt _ flag indicating whether palette prediction is used for the target block may be decoded. Next, when the partition structure of the luminance component of the palette prediction target block and the partition structure of the chrominance component of the target block are identical to each other (i.e., when the TREE _ type is SINGLE _ TREE) for the target block and the palette prediction for the target block is not used, information for an additional prediction mode other than the palette mode may be neither received nor read.
The computer readable storage medium may store the above-described bitstream. The bitstream may include encoding information regarding a chrominance component of the target block.
The processing unit 1710 may perform decoding on the chrominance component of the target block using encoding information regarding the chrominance component of the target block.
In the above-described embodiments, the expression may include a plurality of elements, and the value of the expression may be calculated based on a logical expression using the plurality of elements. One or more of the plurality of elements of formula (la) may be designated as having respective specific values. For example, the specific value may be "0" or "1".
In the case where a specific element is assumed to have a specific value, the value of the corresponding expression in the above-described embodiment may be interpreted as being determined by the remaining elements other than the specific element. In other words, some particular elements of the plurality of elements in the formula may be considered constant values, such as "0" or "1".
For example, assuming that B is "1" in the sentence "a & & B" in the formula, the sentence may be considered to have a value of "a". Alternatively, assuming B is 0, the statement may be considered to have a value of "0".
For example, assuming that B is "0" in the sentence "a | | B" in the formula, the sentence may be considered to have the value "a". Alternatively, assuming B is 1, the statement may be considered to have a value of "1".
Alternatively, in the formula, a specific sentence may be regarded as having a value of "0" or "1".
The above embodiments may be performed by the encoding apparatus 1600 and the decoding apparatus 1700 using methods identical to and/or corresponding to each other. Further, for encoding and/or decoding of images, combinations of one or more of the above embodiments may be used.
In the encoding apparatus 1600 and the decoding apparatus 1700, the application orders of the embodiments may be different from each other. Alternatively, in the encoding apparatus 1600 and the decoding apparatus 1700, the application orders of the embodiments may be (at least partially) identical to each other.
The application orders of the embodiments may be different from each other in the encoding apparatus 1600 and the decoding apparatus 1700, or the application orders of the embodiments may be the same as each other in the encoding apparatus 1600 and the decoding apparatus 1700.
The embodiment may be performed on each of the luminance signal and the chrominance signal. The embodiment may be equally performed on the luminance signal and the chrominance signal.
The form of the block to which embodiments of the present disclosure are applied may have a square or non-square shape.
The embodiments of the present disclosure may be applied according to the size of at least one of a target block, an encoding block, a prediction block, a transform block, a current block, an encoding unit, a prediction unit, a transform unit, a unit, and a current unit. Here, the size may be defined as a minimum size and/or a maximum size such that the embodiment is applied, and may be defined as a fixed size to which the embodiment is applied. Further, in the embodiments, the first embodiment may be applied to the first size, and the second embodiment may be applied to the second size. That is, the embodiments can be compositely applied according to the size. Further, the embodiments of the present disclosure may be applied only to the case where the size is equal to or greater than the minimum size and less than or equal to the maximum size. That is, embodiments may only be applied to cases where block sizes fall within a particular range.
Further, the embodiments of the present disclosure may be applied only to the case where the condition that the size is equal to or greater than the minimum size and the condition that the size is less than or equal to the maximum size are satisfied, where each of the minimum size and the maximum size may be the size of one of the blocks described above in the embodiments and the units described above in the embodiments. That is, a block that is a target of the minimum size may be different from a block that is a target of the maximum size. For example, embodiments of the present disclosure may be applied only to the case where the size of the target block is equal to or greater than the minimum size of the block and less than or equal to the maximum size of the block.
For example, the embodiment can be applied only to the case where the size of the target block is equal to or larger than 8 × 8. For example, the embodiment can be applied only to the case where the size of the target block is equal to or larger than 16 × 16. For example, the embodiment can be applied only to the case where the size of the target block is equal to or larger than 32 × 32. For example, the embodiment can be applied only to the case where the size of the target block is equal to or larger than 64 × 64. For example, the embodiment can be applied only to the case where the size of the target block is equal to or larger than 128 × 128. For example, the embodiment can be applied only to the case where the size of the target block is 4 × 4. For example, the embodiments may be applied only to the case where the size of the target block is less than or equal to 8 × 8. For example, the embodiments may be applied only to the case where the size of the target block is less than or equal to 16 × 16. For example, the embodiments can be applied only to the case where the size of the target block is equal to or larger than 8 × 8 and smaller than or equal to 16 × 16. For example, the embodiments can be applied only to the case where the size of the target block is equal to or larger than 16 × 16 and smaller than or equal to 64 × 64.
Embodiments of the present disclosure may be applied according to temporal layers. To identify the temporal layer to which an embodiment is applicable, a separate identifier may be signaled, and an embodiment may be applied to a temporal layer specified by the corresponding identifier. Here, the identifier may be defined as the lowest (bottom) layer and/or the highest (top) layer to which the embodiment is applicable, and may be defined to indicate a specific layer to which the embodiment is applied. In addition, fixed time layers for application embodiments may also be defined.
For example, the embodiment may be applied only to a case where the temporal layer of the target image is the lowermost layer. For example, the embodiments may be applied only to the case where the temporal layer identifier of the target image is equal to or greater than 1. For example, the embodiment may be applied only to a case where the temporal layer of the target image is the highest layer.
A stripe type or a parallel block group type of an embodiment of the present invention to which the embodiment is applied may be defined, and the embodiment of the present invention may be applied according to the corresponding stripe type or parallel block group type.
In the above-described embodiments, it can be explained that, during application of a specific process to a specific target, assuming that a specific condition may be required and the specific process is performed under a specific determination, when it has been described that whether the specific condition is satisfied is determined based on a specific encoding parameter or the specific determination is made based on the specific encoding parameter, the specific encoding parameter may be replaced with an additional encoding parameter. In other words, encoding parameters that affect a particular condition or a particular determination may be considered merely exemplary, and it is understood that a combination of one or more additional encoding parameters, in addition to a particular encoding parameter, is used as a particular encoding parameter.
In the above-described embodiments, although the method has been described based on the flowchart as a series of steps or units, the present disclosure is not limited to the order of the steps, and some steps may be performed in an order different from that of the described steps or simultaneously with other steps. Furthermore, those skilled in the art will understand that: the steps shown in the flowcharts are not exclusive and may also include other steps, or one or more steps in the flowcharts may be deleted without departing from the scope of the present disclosure.
The above-described embodiments include examples of various aspects. Although not all possible combinations for indicating the various aspects may be described, a person skilled in the art will appreciate that other combinations are possible than those explicitly described. Accordingly, it is to be understood that the present disclosure includes other substitutions, alterations, and modifications as fall within the scope of the appended claims.
The above-described embodiments according to the present disclosure may be implemented as a program that can be executed by various computer apparatuses, and may be recorded on a computer-readable storage medium. The computer readable storage medium may include program instructions, data files, and data structures, either alone or in combination. The program instructions recorded on the storage medium may be specially designed and configured for the present disclosure, or may be known or available to those having ordinary skill in the computer software art.
Computer-readable storage media may include information used in embodiments of the present disclosure. For example, a computer-readable storage medium may include a bitstream, and the bitstream may include information described above in embodiments of the present disclosure.
The computer-readable storage medium may include a non-transitory computer-readable medium.
Examples of the computer-readable storage medium may include all types of hardware devices specifically configured to record and execute program instructions, such as magnetic media (such as hard disks, floppy disks, and magnetic tapes), optical media (such as Compact Disk (CD) -ROMs and Digital Versatile Disks (DVDs)), magneto-optical media (such as floppy disks, ROMs, RAMs, and flash memories). Examples of program instructions include both machine code, such as created by a compiler, and high-level language code that may be executed by the computer using an interpreter. The hardware devices may be configured to operate as one or more software modules to perform the operations of the present disclosure, and vice versa.
As described above, although the present disclosure has been described based on specific details (such as detailed components and a limited number of embodiments and drawings), which are provided only for easy understanding of the entire disclosure, the present disclosure is not limited to these embodiments, and those skilled in the art will practice various changes and modifications according to the above description.
Therefore, it is to be understood that the spirit of the present embodiments is not limited to the above-described embodiments, and that the appended claims and their equivalents and modifications fall within the scope of the present disclosure.

Claims (20)

1. An image encoding method comprising:
determining a prediction mode for a chroma component of a target block; and is provided with
Encoding chroma components of the target block according to the prediction mode.
2. The image encoding method of claim 1, wherein the performing encoding is configured to: determining whether information regarding the prediction mode is to be encoded based on whether a palette mode is used for a chroma component of the target block when a partition structure of a luma component of the target block is the same as a partition structure of a chroma component of the target block.
3. The image encoding method of claim 2, wherein the information on the prediction mode comprises information indicating whether block-Based Delta Pulse Code Modulation (BDPCM) is applied to the chrominance component of the target block.
4. The image encoding method of claim 2, wherein the information on the prediction mode comprises information indicating whether cross-component linear model (CCLM) intra prediction is used for a chroma component of the target block.
5. The image encoding method according to claim 2, wherein the information on the prediction mode includes information indicating an intra prediction mode for chroma sampling of the target block.
6. The image encoding method of claim 1, wherein the performing encoding is configured to: encoding information on the prediction mode if a palette mode is not used for a chroma component of the target block when a partition structure of a luma component of the target block is the same as a partition structure of a chroma component of the target block.
7. The image encoding method of claim 1, wherein the performing encoding is configured to: when the partition structure of the luminance component of the target block is the same as the partition structure of the chrominance component of the target block, if a palette mode is used for the chrominance component of the target block, information on the prediction mode is not encoded.
8. A storage medium storing a bitstream generated by the image encoding method of claim 1.
9. An image decoding method, comprising:
determining a prediction mode for a chroma component of a target block; and is
Decoding chroma components of the target block according to the prediction mode.
10. The image decoding method of claim 9, wherein the step of determining the prediction mode is configured to: determining whether information regarding the prediction mode is to be decoded based on whether a palette mode is used for a chroma component of the target block when a partition structure of a luma component of the target block is the same as a partition structure of a chroma component of the target block.
11. The image decoding method of claim 10, wherein the information on the prediction mode comprises information indicating whether block-Based Delta Pulse Code Modulation (BDPCM) is applied to the chrominance components of the target block.
12. The image decoding method of claim 10, wherein the information on the prediction mode comprises information indicating whether cross-component linear model (CCLM) intra prediction is used for a chroma component of the target block.
13. The image decoding method according to claim 10, wherein the information on the prediction mode includes information indicating an intra prediction mode for chroma sampling of the target block.
14. The image decoding method of claim 9, wherein the step of determining the prediction mode is configured to: decoding information on the prediction mode if a palette mode is not used for a chroma component of the target block when a partition structure of a luma component of the target block is the same as a partition structure of a chroma component of the target block.
15. The image decoding method of claim 9, wherein the step of determining the prediction mode is configured to: when a partition structure of a luminance component of the target block is the same as a partition structure of a chrominance component of the target block, if a palette mode is used for the chrominance component of the target block, information on the prediction mode is not decoded.
16. A computer-readable storage medium storing a bitstream, wherein:
the bitstream includes encoding information regarding chrominance components of the target block,
decoding is performed on the chroma components of the target block using the coding information,
a prediction mode for a chroma component of the target block is determined, and
encoding is performed on chroma components of the target block according to the prediction mode.
17. The computer-readable storage medium of claim 16, wherein the determination of the prediction mode is configured to: determining whether information regarding the prediction mode is to be encoded based on whether a palette mode is used for a chroma component of the target block when a partition structure of a luma component of the target block is the same as a partition structure of a chroma component of the target block.
18. The computer-readable storage medium of claim 17, wherein the information on the prediction mode comprises information indicating an intra prediction mode for chroma sampling of the target block.
19. The computer-readable storage medium of claim 16, wherein the determination of the prediction mode is configured to: encoding information on the prediction mode if a palette mode is not used for a chroma component of the target block when a partition structure of a luma component of the target block is the same as a partition structure of a chroma component of the target block.
20. The computer-readable storage medium of claim 16, wherein the determination of the prediction mode is configured to: when the partition structure of the luminance component of the target block is the same as the partition structure of the chrominance component of the target block, if a palette mode is used for the chrominance component of the target block, information on the prediction mode is not encoded.
CN202080095770.2A 2019-12-05 2020-12-07 Method and apparatus for encoding/decoding image by using palette mode, and recording medium Pending CN115066895A (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
KR20190160513 2019-12-05
KR10-2019-0160513 2019-12-05
PCT/KR2020/017794 WO2021112651A1 (en) 2019-12-05 2020-12-07 Method and device for encoding/decoding image by using palette mode, and recording medium

Publications (1)

Publication Number Publication Date
CN115066895A true CN115066895A (en) 2022-09-16

Family

ID=76412154

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202080095770.2A Pending CN115066895A (en) 2019-12-05 2020-12-07 Method and apparatus for encoding/decoding image by using palette mode, and recording medium

Country Status (2)

Country Link
KR (1) KR102505301B1 (en)
CN (1) CN115066895A (en)

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10097839B2 (en) * 2014-12-19 2018-10-09 Qualcomm Incorporated Palette mode for subsampling format

Also Published As

Publication number Publication date
KR102505301B1 (en) 2023-03-06
KR20210070941A (en) 2021-06-15

Similar Documents

Publication Publication Date Title
CN110463201B (en) Prediction method and apparatus using reference block
CN112740685B (en) Image encoding/decoding method and apparatus, and recording medium storing bit stream
CN111567045A (en) Method and apparatus for using inter prediction information
CN111699682A (en) Method and apparatus for encoding and decoding using selective information sharing between channels
KR20190114853A (en) Method and apparatus for encoding/decoding image, recording medium for stroing bitstream
US11985325B2 (en) Method, apparatus, and recording medium for encoding/decoding image by using geometric partitioning
CN112740697A (en) Image encoding/decoding method and apparatus, and recording medium storing bit stream
US20230013063A1 (en) Method and device for encoding/decoding image by using palette mode, and recording medium
US20220312009A1 (en) Method and apparatus for image encoding and image decoding using area segmentation
CN112740694A (en) Method and apparatus for encoding/decoding image and recording medium for storing bitstream
CN111684801A (en) Bidirectional intra prediction method and apparatus
CN114450946A (en) Method, apparatus and recording medium for encoding/decoding image by using geometric partition
CN116325730A (en) Method, apparatus and recording medium for encoding/decoding image by using geometric partition
US20220201295A1 (en) Method, apparatus and storage medium for image encoding/decoding using prediction
CN113228681A (en) Image encoding/decoding method and apparatus, and recording medium storing bit stream
CN111919448A (en) Method and apparatus for image encoding and image decoding using temporal motion information
US20220272321A1 (en) Method, device, and recording medium for encoding/decoding image using reference picture
CN114270865A (en) Method, apparatus and recording medium for encoding/decoding image
CN113841404A (en) Video encoding/decoding method and apparatus, and recording medium storing bitstream
CN113545052A (en) Image encoding/decoding method and apparatus, and recording medium storing bit stream
US20220295059A1 (en) Method, apparatus, and recording medium for encoding/decoding image by using partitioning
CN114270828A (en) Method and apparatus for image encoding and image decoding using block type-based prediction
KR20210063276A (en) Method, apparatus and recoding medium for video processing using motion prediction model
KR20210070942A (en) Method, apparatus and recording medium for region differential image encoding/decoding
CN113924773A (en) Image encoding/decoding method and apparatus, and recording medium for storing bit stream

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination