KR20140129414A - Method for encoding and decoding image using transform, and apparatus thereof - Google Patents

Method for encoding and decoding image using transform, and apparatus thereof Download PDF

Info

Publication number
KR20140129414A
KR20140129414A KR1020130047116A KR20130047116A KR20140129414A KR 20140129414 A KR20140129414 A KR 20140129414A KR 1020130047116 A KR1020130047116 A KR 1020130047116A KR 20130047116 A KR20130047116 A KR 20130047116A KR 20140129414 A KR20140129414 A KR 20140129414A
Authority
KR
South Korea
Prior art keywords
transform
scanning
unit
residual signal
walsh
Prior art date
Application number
KR1020130047116A
Other languages
Korean (ko)
Inventor
이영렬
유영환
홍성욱
고재현
곽재희
권령희
Original Assignee
인텔렉추얼디스커버리 주식회사
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 인텔렉추얼디스커버리 주식회사 filed Critical 인텔렉추얼디스커버리 주식회사
Priority to KR1020130047116A priority Critical patent/KR20140129414A/en
Publication of KR20140129414A publication Critical patent/KR20140129414A/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/129Scanning of coding units, e.g. zig-zag scan of transform coefficients or flexible macroblock ordering [FMO]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/18Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/91Entropy coding, e.g. variable length coding [VLC] or arithmetic coding

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

A method and an apparatus for encoding / decoding an image according to the present invention are characterized in that at least one of a Hadamard transform and a Walsh transform is applied to a residual signal to convert it into a transform coefficient ; Quantizing the transform coefficient; Scanning the quantized transform coefficients; Performing entropy coding on the scanned transform coefficients; And signaling information to the decryption device about the transform applied in the Hadamard transform and the Walsh transform.

Description

TECHNICAL FIELD The present invention relates to a method and apparatus for encoding / decoding images using a transform,

The present invention relates to an image encoding / decoding method and apparatus, and more particularly, to a method of performing a transform using a transform having +1 and -1 during encoding and decoding.

Generally, in video coding, intra prediction and inter prediction are used to generate a residual signal. The reason why the residual signal is obtained is that when the data is coded with the residual signal, the amount of data is small, so that the data compression rate is high, and the better the prediction, the smaller the value of the residual signal is.

The intraprediction method predicts the data of the current block by using the pixels around the current block. The difference between the actual value and the predicted value is called a residual signal block. In the case of HEVC, the intra prediction method is increased to 35 prediction modes as shown in FIG. 1 in nine prediction modes used in the existing H.264 / AVC, and is further segmented and predicted (the planar prediction mode and the DC prediction mode 1).

In the case of the inter prediction method, the current block is compared with the blocks in the neighboring pictures to find the closest block. At this time, the position information (Vx, Vy) of the found block is referred to as a motion vector. The difference between the intra-block pixel values of the current block and the prediction block predicted by the motion vector is called a residual-signal block (motion-compensated residual block).

In this way, intra prediction and inter prediction are further subdivided so that the amount of data of the residual signal is reduced, and a video coding and decoding method with a small amount of computation is required without degrading the codec performance using an efficient transform.

An embodiment of the present invention provides a video encoding and decoding method having good performance with a small amount of computation in a transcoding process of a video codec and an apparatus therefor.

It is to be understood, however, that the technical scope of the present invention is not limited to the above-described technical problems, and other technical problems may be present.

According to an aspect of the present invention, there is provided a method for encoding an image, the method comprising: applying at least one of a Hadamard transform and a Walsh transform to a residual signal, Conversion into coefficients; Quantizing the transform coefficients; Scanning the quantized transform coefficients; Performing entropy coding on the scanned transform coefficients; And signaling to the decoding device information about transforms applied in the transforming step of the Hadamard transform and the Walsh transform.

The image encoding apparatus according to an embodiment of the present invention includes a transform unit for transforming a residual signal by applying at least one of a Hadamard transform and a Walsh transform to a transform coefficient; A quantization unit for quantizing the transform coefficient; A scanning unit scanning the quantized transform coefficients; An entropy coding unit for performing entropy coding on the scanned transform coefficients; And a signaling unit for signaling information on transforms applied in the conversion step among the Hadamard transform and the Walsh transform, to the decoding apparatus.

According to another aspect of the present invention, there is provided a method of decoding an image, the method comprising: constructing a quantized residual signal block by performing inverse scanning on an image signal to be decoded; Performing inverse quantization on the quantized residual signal block; And performing inverse transform using at least one of a Hadamard transform and a Walsh transform for the inverse quantized residual signal block according to transform information transmitted from an encoding device.

According to another aspect of the present invention, there is provided an apparatus for decoding an image, comprising: an inverse scanning unit configured to inverse-scan an image signal to be decoded to form a quantized residual signal block; A dequantizer for dequantizing the quantized residual signal block; And an inverse transform unit performing inverse transform using at least one of a Hadamard transform and a Walsh transform for the inverse quantized residual signal block according to transform information transmitted from the encoding apparatus.

Meanwhile, the image encoding and decoding method may be embodied as a computer-readable recording medium on which a program to be executed by a computer is recorded.

According to the present invention, it is possible to provide a video encoding and decoding method having a good performance with a small amount of computation in the process of transform coding a video codec, and an apparatus therefor.

In addition, the Hadamard transform and the Walsh transform are adaptively used according to the characteristics of the image, thereby further improving the video coding efficiency.

1 is a diagram showing examples of intra prediction modes.
2 is a block diagram showing a configuration of an encoding apparatus according to an embodiment of the present invention.
3 is a view for explaining a first embodiment of a scanning method suitable for a transform according to the present invention.
4 is a view for explaining a second embodiment of a scanning method suitable for a transform according to the present invention.
5 is a view for explaining a third embodiment of a scanning method suitable for a transform according to the present invention.
6 is a view for explaining a fourth embodiment of a scanning method suitable for a transform according to the present invention.
FIG. 7 is a view for explaining an embodiment of a quantization method suitable for a transform according to the present invention.

Hereinafter, embodiments of the present invention will be described in detail with reference to the accompanying drawings so that those skilled in the art can easily carry out the present invention. It should be understood, however, that the present invention may be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. In the drawings, the same reference numbers are used throughout the specification to refer to the same or like parts.

Throughout this specification, when a part is referred to as being "connected" to another part, it is not limited to a case where it is "directly connected" but also includes the case where it is "electrically connected" do.

Throughout this specification, when a member is " on " another member, it includes not only when the member is in contact with the other member, but also when there is another member between the two members.

Throughout this specification, when an element is referred to as "including " an element, it is understood that the element may include other elements as well, without departing from the other elements unless specifically stated otherwise. The terms "about "," substantially ", etc. used to the extent that they are used throughout the specification are intended to be taken to mean the approximation of the manufacturing and material tolerances inherent in the stated sense, Accurate or absolute numbers are used to help prevent unauthorized exploitation by unauthorized intruders of the referenced disclosure. The word " step (or step) "or" step "used to the extent that it is used throughout the specification does not mean" step for.

Throughout this specification, the term " combination thereof " included in the expression of the machine form means one or more combinations or combinations selected from the group consisting of the constituents described in the expression of the machine form, And the like.

As an example of a method of encoding an actual image and its depth information map, the Moving Picture Experts Group (MPEG) and the Video Coding Experts Group (VCEG) having the highest coding efficiency among the video coding standards developed so far jointly standardize Encoding can be performed using HEVC (High Efficiency Video Coding).

The image encoding and decoding method according to the embodiment of the present invention can apply a transform having a kernel of +1 and -1 to the residual block.

More specifically, the embodiment of the present invention uses a Hadamard Transform having a kernel of +1 and -1 and a Walsh Transform, or a video encoding and decoding method using only one of them and N level quantization thereof And an apparatus therefor.

In HEVC, interpolation is performed using a DCT-based interpolation filter (DCT-IF) when interpolating sub-sample values for color difference signals. The reason for this is that in order to further reduce the residual signal, the motion prediction using an 8-point FIR filter in units of sub-integer pixels, which is finer than the motion prediction compensation of an integer pixel, (For chroma blocks, motion prediction compensation is performed using a 4-point FIR filter of 1/8 pixel unit in YUV 4: 2: 0 color format).

Also, 1/8 or 1/16 interpolation can be used for Luma for more sophisticated motion prediction and compensation.

Currently, intraprediction and inter prediction are further subdivided to reduce the amount of residual signal data, and more sophisticated prediction methods are being tried. Therefore, the video encoding and decoding method according to an embodiment of the present invention uses a transform having kernels of + 1 / -1 rather than using an existing integer transform using DCT or DST . There is no difference in the performance of the codec, and the use of kernels of + 1 / -1 reduces the amount of computation compared to other kernels.

An example of a transform of k + 1 / -1 kernels is the Hadamard Transform. Hadamard transform is a matrix of +1 and -1. The Hadamard transform form is constructed as shown in Equation 1 below.

Figure pat00001

Equation (1) means a one-dimensional forward Hadam transform and an inverse Hadamard transform. Because Hadamard transforms are separable transforms, 2D transforms can be easily implemented by applying one-dimensional transforms horizontally and then vertically (or vertically applying them horizontally).

The following equation (2) represents a one-dimensional 4-point, 8-point Hadamard transform (N = 4,8).

Figure pat00002

Equation (2) represents an orthogonal vector in which each row (base vector) is denoted by an inner product. That is, A-1 = AT. Also, the kernels are made up of +1 and -1 except for constants.

In the following Equation (3), the order of the rows may be changed to 0-4-6-2-3-7-5-1 (arrangement order of basis vectors), and used as it is.

In the equation (2), the value of N can be scaled only after (N = 4, 8 in equation (2)). The above example is an example of transform application for 4x4 residual block and 8x8 residual block, but the proposed transform can be applied to 16x16, 32x32 residual block as above (N = 2n, n is a positive integer) .

As another example of a transform, it is possible to use a Walsh Transform as a transform similar to a Hadamard transform. The Wilcian transform can be used in the equation (1) by changing k (i) to k (n-1-i). The one-dimensional Wilcian transform is constructed as shown in Equation 3 below.

Figure pat00003

Equation 3 implies a one-dimensional forward Walsh transform and an inverse Walsh transform. Because Wilcy transformer is a separable transform, it is easy to implement a two-dimensional transform by applying one-dimensional transforms horizontally and vertically (or horizontally after vertical application).

The following equation (4) represents a one-dimensional 4-point, 8-point Walsh transform kernel (N = 4,8).

Figure pat00004

In Equation 4, the Walsh transform has the same structure as the Hadamard transform and the basis vector, but the order of the basis vectors is different (orthogonal transform, i.e., A-1 = AT).

The value of N in Equation (4) (N = 4, 8 in Equation (4)) needs only to be scaled afterwards. The above example is an example of transform application for 4x4 residual block and 8x8 residual block, but the proposed transform can be applied to 16x16, 32x32 residual block as above (N = 2n, n is a positive integer) .

According to an embodiment of the present invention, when a transform having kernels of + 1 / -1 is used in a transcoding process in a video codec, a good performance can be achieved with a small amount of calculation.

FIG. 2 is a block diagram illustrating a configuration of an image encoding apparatus according to an embodiment of the present invention. Referring to FIG.

Generally, the encoding apparatus includes an encoding process and a decoding process, and the decoding apparatus has a decoding process. The decoding process of the decoding apparatus is the same as the decoding process of the encoding apparatus. Therefore, the encoding apparatus will be mainly described below.

2, an image encoding apparatus according to an exemplary embodiment of the present invention includes an encoding unit and structure, Inter prediction, Intra prediction, Interpolation, Filtering, Transform ) Method and so on.

2, the image encoding apparatus includes an encoding mode determination unit 110, an intra prediction unit 120, a motion compensation unit 130, a motion estimation unit 131, a transcoding / quantization unit 140, A dequantization / conversion decoding unit 160, a deblocking filtering unit 170, a picture storage unit 180, a subtracting unit 190, and an adding unit 200. The dequantization /

The encoding mode determination unit 110 analyzes an input video signal to divide a picture into a predetermined size of an encoding block, and determines a coding mode for the divided predetermined size of the encoding block. The encoding mode includes intraprediction encoding and inter prediction encoding.

The picture is composed of a plurality of slices, and the slice is composed of a plurality of maximum coding units (LCU). The LCU can be divided into a plurality of coding units (CUs), and the encoder can add information indicating whether or not to be divided to a bit stream. The decoder can recognize the position of the LCU by using the address (LcuAddr). The coding unit CU in the case where division is not allowed is regarded as a prediction unit (PU), and the decoder can recognize the position of the PU using the PU index.

The prediction unit PU may be divided into a plurality of partitions. Also, the prediction unit PU may be composed of a plurality of conversion units (TUs).

The encoding mode determination unit 110 sends the image data to the subtraction unit 190 in units of blocks of a predetermined size (for example, in units of PU or TU) according to the determined encoding mode.

The transform coding / quantizing unit 140 transforms the residual block calculated by the subtracting unit 190 from the spatial domain to the frequency domain. For example, two-dimensional discrete cosine transform (DCT) or discrete cosine transform (DST) -based transform is performed on the residual block.

In addition, the transcoding / quantization unit 140 determines a quantization step size for quantizing the transform coefficient, and quantizes the transform coefficient using the determined quantization step size. The quantization matrix can be determined according to the determined quantization step size and encoding mode.

The quantized two-dimensional transform coefficients are transformed into one-dimensional quantized transform coefficients by one of the predetermined scanning methods. The transformed one-dimensional sequence of quantization transform coefficients is supplied to the entropy encoding unit 150.

The inverse quantization / conversion decoding unit 160 dequantizes the quantization coefficients quantized by the transcoding / quantization unit 140. Further, the inverse quantization coefficient obtained by inverse quantization is inversely transformed. Accordingly, the residual block transformed into the frequency domain can be restored into the residual block in the spatial domain.

The deblocking filtering unit 170 receives the inverse quantized and inverse transformed image data from the inverse quantization / inverse transform coding unit 160 and performs filtering to remove a blocking effect.

The picture storage unit 180 receives the filtered image data from the deblocking filtering unit 170 and restores and restores the image in picture units. The picture may be a frame-based image or a field-based image. The picture storage unit 180 has a buffer (not shown) capable of storing a plurality of pictures. A plurality of pictures stored in the buffer are provided for intra prediction and motion estimation.

The pictures provided for intra prediction or motion estimation are referred to as reference pictures.

The motion estimation unit 131 receives the at least one reference picture stored in the picture storage unit 180 and performs motion estimation to output motion data including an index indicating a motion vector and a reference picture and a block mode do.

In order to optimize the prediction precision, a motion vector is determined with a fractional pixel precision, for example, 1/2 or 1/4 pixel accuracy. Since the motion vector can have a fractional pixel precision, the motion compensation unit 130 applies the interpolation filter for calculating the pixel value of the fractional pixel position to the reference picture so that the pixel value of the fractional pixel position .

The motion compensation unit 130 is configured to perform motion compensation on a block to be coded from a reference picture used for motion estimation among a plurality of reference pictures stored in the picture storage unit 180 according to the motion data input from the motion estimation unit 131 And outputs the extracted prediction block.

The motion compensation unit 130 determines a filter characteristic of the adaptive interpolation filter necessary for motion compensation with a decimal precision. The filter characteristic is, for example, information indicating the filter type of the adaptive interpolation filter and information indicating the size of the adaptive interpolation filter.

The size of the filter is, for example, the number of taps, which is the number of filter coefficients of the adaptive interpolation filter.

Specifically, the motion compensation unit 130 may determine either a separate type or a non-separable type adaptive filter as an adaptive interpolation filter. Then, the number of taps of the determined adaptive interpolation filter and the value of each filter coefficient are determined. The value of the filter coefficient can be determined differently for each position of the fractional pixel relative to the integer pixel. Also, the motion compensation unit 130 may use a plurality of non-adaptive interpolation filters with fixed filter coefficients.

The motion compensation unit 130 can set the characteristics of the interpolation filter in a predetermined processing unit. For example, it can be set in a fractional pixel unit, a coding basic unit (encoding unit), a slice unit, a picture unit, or a sequence unit. In addition, one characteristic may be set for one video data.

Therefore, since the same filter characteristic is used in a predetermined processing unit, the motion compensation unit 130 has a memory that temporarily holds the filter characteristic. This memory maintains filter characteristics, filter coefficients, and the like as needed. For example, the motion compensation unit 130 can determine the filter characteristic for each I picture and determine the filter coefficient for each slice.

The motion compensation unit 130 receives a reference picture from the picture storage unit 180 and applies a filter process using the determined adaptive interpolation filter to generate a prediction reference picture of a decimal precision.

Then, based on the generated reference picture and the motion vector determined by the motion estimation unit 131, motion compensation is performed with a small number of pixels to generate a prediction block.

The subtractor 190 receives the block in the reference picture corresponding to the input block from the motion compensator 130 and performs a difference operation with the input macroblock in the case of performing inter picture prediction coding on the input block to be coded, and outputs a residue signal.

The intraprediction unit 120 performs intraprediction encoding using the reconstructed pixel values in a picture to be predicted. The intra prediction unit receives the current block to be predictively encoded and performs intra prediction by selecting one of a plurality of intra prediction modes preset according to the size of the current block. The intra predictor 120 determines the intra prediction mode of the current block using the previously coded pixels adjacent to the current block, and generates a prediction block corresponding to the determined mode.

The previously encoded region of the current picture is decoded again for use by the intra prediction unit 120 and stored in the picture storage unit 180. [ The intra prediction unit 120 generates a prediction block of a current block using pixels neighboring the current block or non-adjacent but applicable pixels in the previously coded area of the current picture stored in the picture storage unit 180. [

The intra prediction unit 120 may adaptively filter adjacent pixels to predict an intra block. For the same operation in the decoder, it is possible to transmit information indicating whether or not filtering is performed in the encoder. Or the intra-prediction mode of the current block and the size information of the current block.

The prediction type used by the image coding apparatus depends on whether the input block is coded in the intra mode or the inter mode by the coding mode determination unit.

The switching between the intra mode and the inter mode is controlled by the intra / inter selector switch.

The entropy encoding unit 150 entropy-codes the quantization coefficients quantized by the transcoding / quantization unit 140 and the motion information generated by the motion estimation unit 131. [ Also, an intra prediction mode, control data (e.g., quantization step size, etc.), and the like can be coded. Also, the filter coefficient determined by the motion compensation unit 130 is encoded and output as a bit stream.

Referring to FIG. 2, the encoder according to an exemplary embodiment of the present invention may perform a transform of an 8 × 8 DCT, a 16 × 16 DCT, a 32 × 32 DCT, or a 4 × 4 DST, which is used in a transform process of a conventional HEVC encoder, It is configured using the proposed transforms with configured 4x4, 8x8, 16x16, 32x32 kernels.

Also, when the proposed transform is applied to the encoder, the decoder applies the inverse transform (A-1 = AT) with the proposed kernels of + 1 / -1 in the process of inversion of the transform.

Here, when the transformer proposed in the encoding apparatus according to FIG. 2 is used, the inverse transform proposed in the decoding process is applied.

As another example, when using a transform proposed by any video encoder, the proposed inverse transform is applied to the decoding process.

As described above, the configuration of the image decoding apparatus according to the embodiment of the present invention can be derived from the configuration of the image encoding apparatus shown in FIG. 2, and for example, the inverse of the encoding process as described with reference to FIG. 2 So that the image can be decoded.

An embodiment of the present invention relates to a scanning method applied when a Walsh transform is used in a transform (Transform) method applied to a moving picture encoding / decoding.

In detail, the scanning order and method are applied differently according to the characteristics of the Walsh transform, and it is predicted by using a decoding / decoding method and neighboring mode information, It includes all methods that can be decoded without signaling.

The coefficient using the Hadamard transform or the Walsh transform is quantized, and a bitstream is generated through entropy coding according to a scanning method. Here, a method of performing scanning through a Hadamard or Walsh transform can be performed as shown in the following example.

FIG. 3 is a view for explaining a first embodiment of a scanning method suitable for a transform according to the present invention, and shows an example of a scanning method for an 8x8 block. This can be applied to various block sizes.

As shown in FIG. 3, the method includes a scanning method from the upper left to the lower right and a scanning method from the lower right to the upper left, and scanning can be applied in various methods according to the mode information.

The above four scanning methods are simple examples, and the scanning methods in various orders can be equally applied to coding and decoding.

Hadamard or Walsh transforms would have less transformed coefficients in quantization than in integer and DCT types, so the Hadamard or Walsh transforms would have 1, -1 The amount of computation can be reduced by using the kernel.

That is, the Hadamard or Walsh transforms can have similar characteristics, such as integer DCT and DST, and thus have DC and AC split characteristics, but the reason for using the Hadamard or Walsh transform is that the predicted value The assumption is that there is very little to eliminate multiply operations and to use the improved transforms.

When the residual signal is small, the transformed coefficients may be distributed in a very small number. In this case, it is necessary to select an optimal scanning method by varying the scanning methods applied in entropy coding through various scanning methods.

For example, in the intra block, the best scanning method is determined in terms of vertical scanning in the horizontal prediction mode, horizontal scanning in the vertical prediction mode, and rate-distortion in the vertical prediction mode, and the information about the scanning method can be signaled to the decoder have.

4 shows an example of a method of dividing the inside of the 8x8 block into four scanning methods.

This method also includes a method of transmitting a form by signaling with a decryptor / decoder or a method of scanning in a promised form using peripheral information.

FIG. 5 is a diagram for explaining an embodiment of a method of dividing the inside of a block of 8x8 into four sub-blocks and scanning. In the case where a coefficient exists at a position as shown in FIG. 5, An example of dividing a block into 4 blocks and performing 4x4 scanning is shown below.

The distribution point in FIG. 5 indicates the positions where the transformed and quantized coefficients are present. When the coefficients are distributed from the upper left corner to the coefficients from the upper left corner, the scanning is performed by the dividing method, The performance can be improved.

In this case, the entropy coding plays a very important role in the performance.

FIG. 6 shows a case where all the coefficients are distributed in the upper left corner. In this case, when the entire block size is scanned to 8x8 without dividing the block, the performance is improved or the performance is better than the dividing method Includes all of the options you can use selectively.

That is, the present invention refers to a method of improving performance through a scanning method that does not divide or divide according to the distribution of coefficients.

The present invention relates to a scanning method applicable when a Hadamard or Walsh transform is used. As described above, all the scanning methods include a method of transmitting information to a decoder / decoder through signaling, (Mode information, MV information, etc.).

After the Transform, the encoding proceeds in the order of Quantization, Scanning, and Entropy Encoding. Transform coefficients generally tend to have large values at the upper left and lower values for the remaining coefficients.

Scanning includes Diagonal Scanning, Horizontal Scanning, and Vertical Scanning. In all three methods, the upper left corner with large transform coefficients is scanned first.

Therefore, if the quantization is divided into three sections in the order of scanning as described below, the encoding compression ratio may be increased.

FIG. 7 illustrates an N-level quantization method according to an embodiment of the present invention, and may be divided into N (= 2, 3, ...) intervals according to a scanning order.

After Hadamard or Wahlsh Transform, N-level quantization and N-value indicating the quantization of several sections are signaled to the decoder.

Referring to FIG. 7, the conversion coefficient denoted by white is the weakest part of the quantization. The gray-scale conversion factor is a part of the quantization that is stronger than the white conversion factor. The conversion factor with dark gray is the strongest part of the quantization. Quantization is performed weakly in a portion with a low scanning order and strongly quantized in a portion with a high scanning order.

In general, scanning first scans the portion with the largest quantized transform coefficients.

In the case of the quantization as shown in the left side of FIG. 7, it is used when a scanning method is performed by dividing 8x8 blocks into 4 blocks. For example, the vicinity of the white conversion coefficient is first scanned, the vicinity of the gray conversion coefficient is scanned, and the portion of the dark gray conversion coefficient is scanned.

For example, as shown in FIG. 4, in the case of dividing the inside of the 8x8 block into four, a quantization method as shown in the left side of FIG. 7 may be used.

With the same principle, the quantization method as shown in the center of FIG. 7 can be applied to the case of using the horizontal scanning method as shown in the bottom right of FIG.

In addition, the quantization method as shown in the right side of FIG. 7 can be applied to the case of using the vertical scanning method as shown in the lower left of FIG.

In the case of N-level quantization as described above, the following three cases may occur.

(1) When one flag is set to 0, N level quantization is not used. That is, all transform coefficients in the block are quantized to the same size. When the flag is 1, N level quantization is performed.

(2) Transmit the number N of intervals divided on the basis of quantization to the decoder. On the encoder side, the number N of quantization intervals is determined on the basis of the rate-distortion value, and on the decoder side, N is received to perform N-level quantization.

(3) Always perform predetermined N-level quantization.

On the other hand, the quantization method differs according to the scanning order, and can be applied in the same manner in the case of Hadamard transform or Walsh transform.

According to one embodiment of the present invention, either Hadamard Transform and Walsh Transform can be used together, or only one of them can be used.

For example, using Hadamard Transform and Walsh Transform might be:

(1) When using only Hadamard transform foam

(2) When using only Walsh transforms

(3) If you want to signal whether you want to use a Walsh transform or leave a flag.

(4) Hadamard transforms, Walsh transforms, and other transforms (eg, DCT, DST), but signaling what to write

In addition, when at least one of the Hadamard transform and the Wolsh transform is selected and used adaptively as described above, the binaural quantization described with reference to FIG. 7 is adaptively changed according to the transform type, It may not be.

The method according to the present invention may be implemented as a program for execution on a computer and stored in a computer-readable recording medium. Examples of the computer-readable recording medium include a ROM, a RAM, a CD- , A floppy disk, an optical data storage device, and the like, and may also be implemented in the form of a carrier wave (for example, transmission over the Internet).

The computer readable recording medium may be distributed over a networked computer system so that computer readable code can be stored and executed in a distributed manner. And, functional programs, codes and code segments for implementing the above method can be easily inferred by programmers of the technical field to which the present invention belongs.

While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it is to be understood that the invention is not limited to the disclosed exemplary embodiments, but, on the contrary, It should be understood that various modifications may be made by those skilled in the art without departing from the spirit and scope of the present invention.

Claims (8)

A method of encoding an image,
Converting at least one of a Hadamard transform and a Walsh transform to a transform coefficient for the residual signal;
Quantizing the transform coefficients;
Scanning the quantized transform coefficients;
Performing entropy coding on the scanned transform coefficients; And
And signaling to the decoding device information on transforms applied in the transforming step of the Hadamard transform and the Walsh transform.
2. The method of claim 1, wherein the scanning step
And the quantized transform coefficients are scanned based on positions where the quantized transform coefficients exist.
2. The method of claim 1, wherein the quantizing step comprises:
Wherein the residual signal block is divided into a plurality of portions according to a scanning order, and a different quantization level is applied.
The method of claim 1, wherein the transform information
And a flag indicating a transform applied in the transforming step.
5. The method of claim 4, wherein the transform information
Wherein the Hadamard Transform and the Walsh Transform are sequentially applied.
An apparatus for encoding an image, the apparatus comprising:
A transform unit for transforming the residual signal by applying at least one of a Hadamard transform and a Walsh transform to transform coefficients;
A quantization unit for quantizing the transform coefficient;
A scanning unit scanning the quantized transform coefficients;
An entropy coding unit for performing entropy coding on the scanned transform coefficients; And
And a signaling unit for signaling, from the Hadamard transform and the Walsh transform, information on a transform applied in the transforming step to a decoding apparatus.
A method for decoding an image,
Performing inverse scanning on a video signal to be decoded to form a quantized residual signal block;
Performing inverse quantization on the quantized residual signal block; And
And performing inverse transform using at least one of a Hadamard transform and a Walsh transform for the inverse quantized residual signal block according to transform information transmitted from an encoding device.
An apparatus for decoding an image, the apparatus comprising:
An inverse scanning unit for inversely scanning a video signal to be decoded to construct a quantized residual signal block;
A dequantizer for dequantizing the quantized residual signal block; And
And an inverse transform unit for performing an inverse transform using at least one of a Hadamard transform and a Walsh transform for the inverse quantized residual signal block according to transform information transmitted from an encoder.
KR1020130047116A 2013-04-29 2013-04-29 Method for encoding and decoding image using transform, and apparatus thereof KR20140129414A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
KR1020130047116A KR20140129414A (en) 2013-04-29 2013-04-29 Method for encoding and decoding image using transform, and apparatus thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
KR1020130047116A KR20140129414A (en) 2013-04-29 2013-04-29 Method for encoding and decoding image using transform, and apparatus thereof

Publications (1)

Publication Number Publication Date
KR20140129414A true KR20140129414A (en) 2014-11-07

Family

ID=52454687

Family Applications (1)

Application Number Title Priority Date Filing Date
KR1020130047116A KR20140129414A (en) 2013-04-29 2013-04-29 Method for encoding and decoding image using transform, and apparatus thereof

Country Status (1)

Country Link
KR (1) KR20140129414A (en)

Similar Documents

Publication Publication Date Title
US9451279B2 (en) Method for decoding a moving picture
JP5982612B2 (en) Intra prediction mode decoding method
KR20190016983A (en) Method and apparatus for encoding intra prediction information
JP2024020503A (en) Method for coding or decoding video signal and apparatus used for the same
KR20120012383A (en) Prediction block generating apparatus
KR20120012401A (en) Intra prediction decoding apparatus
KR20140124443A (en) Method for encoding and decoding video using intra prediction, and apparatus thereof
KR20180041578A (en) Image encoding method/apparatus, image decoding method/apparatus and and recording medium for storing bitstream
KR20130045151A (en) Method and apparatus for generating reconstruction block
KR20140129423A (en) Method for intra-prediction, and apparatus thereof
KR20130105114A (en) Method of decoding moving pictures in intra prediction
KR20140124448A (en) Method for encoding and decoding video using intra prediction, and apparatus thereof
KR20140129417A (en) Method for encoding and decoding image using a plurality of transforms, and apparatus thereof
KR20140124442A (en) Method for encoding and decoding video using intra prediction, and apparatus thereof
KR20150105348A (en) Method and apparatus for encoding/decoding images using transform
WO2014084674A2 (en) Intra prediction method and intra prediction apparatus using residual transform
KR101204385B1 (en) Method of decoding moving pictures in intra prediction
KR20210156256A (en) Method for intra-prediction, and apparatus thereof
KR102217225B1 (en) Method for intra-prediction, and apparatus thereof
KR20140129421A (en) Method for intra-prediction, and apparatus thereof
KR20140129414A (en) Method for encoding and decoding image using transform, and apparatus thereof
KR102189259B1 (en) Method for intra-prediction, and apparatus thereof
KR20140129411A (en) Method for encoding and decoding image using transform, and apparatus thereof
KR20140129410A (en) Method for encoding and decoding image using transform, and apparatus thereof
KR20140129409A (en) Method for encoding and decoding image using transform, and apparatus thereof

Legal Events

Date Code Title Description
N231 Notification of change of applicant
WITN Withdrawal due to no request for examination