[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

KR20130023444A - Apparatus and method for video encoding/decoding using multi-step inter prediction - Google Patents

Apparatus and method for video encoding/decoding using multi-step inter prediction Download PDF

Info

Publication number
KR20130023444A
KR20130023444A KR1020110086201A KR20110086201A KR20130023444A KR 20130023444 A KR20130023444 A KR 20130023444A KR 1020110086201 A KR1020110086201 A KR 1020110086201A KR 20110086201 A KR20110086201 A KR 20110086201A KR 20130023444 A KR20130023444 A KR 20130023444A
Authority
KR
South Korea
Prior art keywords
block
reference block
inter prediction
prediction
inter
Prior art date
Application number
KR1020110086201A
Other languages
Korean (ko)
Inventor
정제창
김기백
정동진
유호선
김재훈
Original Assignee
한양대학교 산학협력단
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 한양대학교 산학협력단 filed Critical 한양대학교 산학협력단
Priority to KR1020110086201A priority Critical patent/KR20130023444A/en
Publication of KR20130023444A publication Critical patent/KR20130023444A/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/533Motion estimation using multistep search, e.g. 2D-log search or one-at-a-time search [OTS]

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Disclosed are an apparatus and method for encoding / decoding an image using multi-step inter prediction. A first inter prediction step of finding a first reference block on a first reference picture from the encoding target block on the current picture, and a second inter prediction step and a first inter prediction finding a second reference block on the second reference picture from the first reference block. The present invention provides a method for encoding an image using multi-step inter prediction, including a prediction block generation step of generating a prediction block from a value obtained by adding weights to each of the reference blocks found by the step and the second inter prediction step. Further, in the image decoding method, a decoding step of decoding a coded bitstream to calculate a residual value and inter-screen prediction information, and a first reference picture on the first reference picture from the encoding target block on the current picture based on the inter-screen prediction information. An inter prediction step for finding a reference block, finding a second reference block on a second reference picture from the first reference block, and generating a predictive block, and reconstructing a reconstructed block by adding the reconstructed residual values and the predicted block. An image decoding method using multi-step inter prediction includes an image generation step.

Description

Apparatus and Method for Video Encoding / Decoding using Multi-Step Inter Prediction}

The present invention relates to encoding and decoding of an image, and more particularly, to an apparatus and method for encoding / decoding an image using multi-step inter prediction in inter prediction.

Recently, the emergence of smartphones and smart TVs has led to explosive use of video dating via wired and wireless communication networks. Video data has a higher information transfer capability than plain text data, but has a large capacity, making it difficult to transmit, play, and store data in a network channel having limited bandwidth. In addition, since a large amount of moving picture information must be appropriately processed according to an application request, a system for processing moving pictures also requires high specifications.

In order to solve these problems, video encoding algorithms, which are techniques for compressing video data into small information, have been actively studied. Representative international standards related to video compression studied so far include ISO / IEC MPEG (MPEG) series and ITU-T H.26x series.

The video data is characterized by having temporal, spatial, and statistical redundancy. Temporal redundancy means redundancy between successive frames, and pixel values of successive frames have a very high correlation. Spatial redundancy means redundancy existing in a frame, and the brightness value of one pixel has a high correlation with the brightness value of neighboring pixels. Finally, statistical redundancy means redundancy between coded data, which means redundancy by probability distribution of brightness values of pixels.

For video encoding, by removing the three redundancies it is possible to compress a large amount of video data into a smaller amount of data. For example, motion estimation and compensation techniques remove temporal redundancy, and transform coding and quantization techniques remove spatial redundancy. Entropy coding techniques, such as variable length coding, are adopted and used in international standards for video codecs for the purpose of removing statistical redundancy.

Among the international standards organizations, the MPEG standard has three types of prediction: intra coded frame (I frame), predictive coded frame (P frame), and bi-directionally-predicitve coded frame (B frame). A coding scheme is used, and P frames and B frames achieve high compression efficiency by using motion estimation and compensation techniques.

An I frame is encoded using neighboring pixel values within the screen without reference to other frames. P frames are encoded with reference to past I frames or P frames. The motion information is estimated using the predicted motion vector and the predictive block by estimating the motion information in the unit of blocks in the past I frame or the P frame. Reduction and high compression ratio can be achieved.

The B frame performs prediction by referring not only to past I frames and P frames but also to future I frames and P frames. The B frame uses the motion estimation and compensation technique like the P frame. However, the compression rate is the highest because two reference frames are used and a better prediction performance is selected among them. However, the B frame does not become a reference frame for other frames.

In inter-picture prediction, since H.264 / AVC can also predict from two frames in the past, and can predict from two frames in the future, this is called bi-predictive prediction-picture. Call. In order to perform such pair prediction, since two pieces of motion information are required, the amount of information to be sent to the decoding apparatus increases.

An object of the present invention for solving the above problems is to provide an image encoding / decoding apparatus that can generate a prediction block more accurately by allowing the reference block to be predicted again from the reference block of the encoding target block.

Another object of the present invention for solving the above problems is to provide an image encoding / decoding method that can generate a prediction block more accurately by allowing a reference block to be predicted again from a reference block of an encoding target block.

Another object of the present invention for solving the above problems is to use template matching in predicting a reference block from a block to be encoded, and to reduce the information on the motion vector by predicting the reference block from the reference block again. An image encoding / decoding apparatus and method for improving encoding efficiency are provided.

According to an aspect of the present invention, there is provided a video encoding method comprising: a first inter prediction step of finding a first reference block on a first reference picture from a block to be encoded on a current picture and a second reference picture on the second reference picture; An image using multi-step inter prediction, including a prediction block generation step of generating a prediction block based on a second inter prediction step of finding a second reference block, and a reference block found by the first inter prediction step and the second inter prediction step. Provides an encoding method.

Here, the predictive block generation step may generate a predictive block by weighting and adding each of the first reference block and the second reference block.

Here, the first inter prediction step may find the first reference block using motion estimation or template matching.

Here, the second inter prediction step may be performed by dividing the first reference block into at least two sub blocks each having a predetermined size, and then searching for a second reference block on the second reference picture for each of the divided sub blocks. can do.

According to another aspect of the present invention for achieving the above object, a decoding step of decoding a coded bitstream to calculate a reconstructed residual value and inter-screen prediction information and on the current picture based on the inter-screen prediction information. An inter prediction step and a reconstructed residual value and prediction for finding a first reference block on a first reference picture from an encoding target block and generating a prediction block by finding a second reference block on a second reference picture from the first reference block. An image decoding method using multi-step inter prediction includes a step of generating a reconstructed image by summing blocks.

Here, the inter prediction information may be characterized by including the partition information of the block, the number of predictions, and the weight.

According to an aspect of the present invention for achieving the above object, a first inter prediction unit for searching for a first reference block on a first reference picture from a block to be encoded on a current picture and a second reference from the first reference block A video encoding apparatus using multi-step inter prediction includes a second inter prediction unit that finds a second reference block on a picture.

According to another aspect of the present invention for achieving the above object, a first inter prediction unit for searching for a first reference block on a first reference picture from a block to be encoded on a current picture and a second from the first reference block The reconstructed image is generated by adding the reconstructed residual value and the prediction block generated based on the second inter prediction unit for finding the second reference block on the reference picture and the reference blocks found by the first inter prediction unit and the second inter prediction unit. Provided is an image decoding apparatus using multi-step inter prediction, which includes an adder.

When using the apparatus and method for image encoding / decoding using multi-step inter prediction according to the present invention as described above, a prediction block can be generated more accurately.

In addition, the apparatus and method for encoding / decoding an image using multi-step inter prediction according to the present invention has an advantage of improving coding efficiency by omitting information on a motion vector.

The effects of the present invention are not limited to the effects mentioned above, and other effects not mentioned can be clearly understood by those skilled in the art from the description of the claims.

1 is a conceptual diagram illustrating multi-step inter prediction using template matching according to the present invention.
2 is a conceptual diagram illustrating multi-step inter prediction according to another embodiment of the present invention.
3 is a conceptual diagram illustrating multi-step inter prediction using motion estimation according to the present invention.
4 is a conceptual diagram illustrating multi-step inter prediction according to another embodiment of the present invention.
5 is a conceptual view illustrating multi-step inter prediction according to another embodiment of the present invention.
6 is a block diagram of a video encoding apparatus using multi-step inter prediction according to the present invention.
7 is a block diagram of an image decoding apparatus using multi-step inter prediction according to the present invention.

While the invention is susceptible to various modifications and alternative forms, specific embodiments thereof are shown by way of example in the drawings and will herein be described in detail. It should be understood, however, that the invention is not intended to be limited to the particular embodiments, but includes all modifications, equivalents, and alternatives falling within the spirit and scope of the invention. Like reference numerals are used for like elements in describing each drawing.

The terms first, second, A, B, etc. may be used to describe various elements, but the elements should not be limited by the terms. The terms are used only for the purpose of distinguishing one component from another. For example, without departing from the scope of the present invention, the first component may be referred to as a second component, and similarly, the second component may also be referred to as a first component. And / or < / RTI > includes any combination of a plurality of related listed items or any of a plurality of related listed items.

When a component is referred to as being "connected" or "connected" to another component, it may be directly connected to or connected to that other component, but it may be understood that other components may be present in between. Should be. On the other hand, when an element is referred to as being "directly connected" or "directly connected" to another element, it should be understood that there are no other elements in between.

The terminology used herein is for the purpose of describing particular example embodiments only and is not intended to be limiting of the present invention. Singular expressions include plural expressions unless the context clearly indicates otherwise. In this application, the terms "comprise" or "have" are intended to indicate that there is a feature, number, step, operation, component, part, or combination thereof described in the specification, and one or more other features. It is to be understood that the present invention does not exclude the possibility of the presence or the addition of numbers, steps, operations, components, components, or a combination thereof.

Unless defined otherwise, all terms used herein, including technical or scientific terms, have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. Terms such as those defined in commonly used dictionaries are to be interpreted as having a meaning consistent with the contextual meaning of the related art and are to be interpreted as either ideal or overly formal in the sense of the present application Do not.

A video encoding apparatus (Video Encoding Apparatus), a video decoding apparatus (Video Decoding Apparatus) to be described below is a personal computer (PC), notebook computer, personal digital assistant (PDA), portable multimedia player (PMP) : Portable Multimedia Player (PSP), PlayStation Portable (PSP: PlayStation Portable), Wireless Communication Terminal (Wireless Communication Terminal), Smart Phone, Smart Phone, TV Application Server, Service Server, etc. A user terminal or a communication device such as a communication modem for communicating with a wired / wireless communication network, a memory for storing various programs and data for encoding or decoding an image or predicting between screens or in-screen for encoding or decoding. It is equipped with a microprocessor for execution and operation and control. It can mean a variety of devices.

In addition, the image encoded in the bitstream by the video encoding apparatus is real-time or non-real-time through the wired or wireless communication network, such as the Internet, local area wireless communication network, wireless LAN network, WiBro network, mobile communication network, or the like, or a cable, universal serial bus (USB: Universal) It may be transmitted to an image decoding apparatus through various communication interfaces such as a serial bus, and may be decoded by the image decoding apparatus to restore and reproduce the image.

The moving picture may be generally composed of a series of pictures, and each picture may be divided into a predetermined area such as a frame or a block. In the case where an image area is divided into blocks, the divided blocks can be classified into an intra block and an inter block according to a coding method. The in-picture block refers to a block that is coded using an Intra Prediction Coding scheme. The intra-picture prediction coding refers to a method of coding a block of pixels of previously decoded and decoded blocks in a current picture, A prediction block is generated by predicting the pixels of the current block and a difference value between the current block and the pixel of the current block is encoded. Inter-block refers to a block that is coded using Inter Prediction Coding. Inter-prediction coding refers to one or more past pictures or a future picture to generate a prediction block by predicting a current block in the current picture, And the difference value is encoded. Here, a frame to be referred to in encoding or decoding a current picture is referred to as a reference frame. It is also to be understood that the term "picture" described below may be used in place of other terms having equivalent meanings such as image, frame, etc., If you are a child, you can understand.

Hereinafter, preferred embodiments according to the present invention will be described in detail with reference to the accompanying drawings.

template Matching  Multilevel used Between screens  Forecast Example #One

1 is a conceptual diagram illustrating multi-step inter prediction using template matching according to the present invention. Referring to FIG. 1, a first reference block on a first reference picture is found from a block to be encoded on a current picture, and a second reference block on a second reference picture is again found from a first reference block. The weighted weights of the first reference block and the second reference block may be added to each other to generate (calculate) a prediction block. Here, motion estimation or template matching may be used to find the first reference block, and FIG. 1 illustrates a method of using template matching. That is, a template most similar to this is found in the forward or reverse picture (picture) using a template around the encoding target block of the current picture. The second reference block finds the block most similar to the first reference block found using the template using motion estimation. In finding the first reference block from the encoding target block, since the prediction is performed using a template around the encoding target block, the prediction may be performed without information of a motion vector. In addition, even in finding the second reference block from the first reference block, since the information about the prediction target block (the first reference block) is included, no additional motion vector information is required. That is, it may not need information about V t (motion information of the template) and V m (motion information of the first reference block).

Therefore, in performing the multi-step inter prediction using template matching, information about a motion vector can be omitted, thereby improving coding efficiency. In addition, information previously promised for selecting a reference picture may be used. For example, the first reference picture is the picture immediately before the current picture, and the second reference picture is the picture immediately before the first reference picture. In this case, information about the reference picture may be omitted.

In addition, each weight applied to the predicted first reference block and the second reference block may apply different weights for each pixel unit, and may be preferably determined according to the distance between the reference picture and the current picture.

2 is a conceptual diagram illustrating multi-step inter prediction according to another embodiment of the present invention. Referring to FIG. 2, when a block similar to the first reference block found using template matching is found by using motion estimation in a second reference picture, which is another reference picture, the first reference block is found in at least two sub-blocks having a preset size. After splitting each other, inter prediction may be performed through motion estimation on each of the divided subblocks. In this case, the method of dividing into sub-blocks, the number of sub-blocks, information on weights, etc. may be transmitted to the decoding apparatus in units of a sequence, a picture, a slice, a block, and the like.

In the inter prediction, the prediction block is divided into smaller sub-blocks to perform prediction to increase the accuracy of motion estimation and to improve coding efficiency.

Multistep with motion estimation Between screens  Forecast Example #2

3 is a conceptual diagram illustrating multi-step inter prediction using motion estimation according to the present invention. Referring to FIG. 3, a motion estimation is performed from an encoding target block on a current picture to find a first reference block on a first reference picture. After dividing the first reference block into at least two subblocks having a predetermined size, at least two second reference blocks on the second reference picture may be found through motion estimation for each of the divided subblocks. In this case, in order to perform prediction on the first reference block, information about a motion vector should be encoded and transmitted to the decoding apparatus. However, in finding the second reference block on the second reference picture, since the information about the prediction target block (the first reference block) is included, no additional motion vector information is required. That is, reference blocks on two reference pictures may be found using only information about one motion vector.

4 is a conceptual diagram illustrating multi-step inter prediction according to another embodiment of the present invention. Referring to FIG. 4, in performing multi-step inter prediction, the second reference block on the second reference picture is further divided into at least two sub blocks, and the inter prediction is performed on the divided sub blocks of the second reference block. You can do more. That is, the present invention can perform at least two inter predictions, and by applying adaptive weights to the predicted reference blocks through performing the inter predictions, it is possible to calculate more accurate prediction blocks and improve coding efficiency. have.

Pair prediction  Multilevel utilized Between screens  Forecast Example # 3

5 is a conceptual view illustrating multi-step inter prediction according to another embodiment of the present invention. Referring to Fig. 5, a first reference block on a first reference picture is found using a template around the encoding target block on the current picture. After splitting the first reference block into at least two sub-blocks having a predetermined size, inter-screen prediction may be performed on each of the divided sub-blocks. Here, in performing inter prediction on at least two divided sub-blocks, prediction may be performed in the reverse direction. According to the present invention, prediction in the forward or reverse direction may be performed in the direction of inter-screen prediction. In addition, as in Embodiment # 2, motion estimation may be used to find the first reference block.

Accordingly, the present invention can be characterized by performing inter prediction on the second reference block from the first reference block in the forward or reverse direction.

The present invention provides a video encoding method comprising: a first inter prediction step of finding a first reference block on a first reference picture from a block to be encoded on a current picture and a second finding of a second reference block on a second reference picture from a first reference block An image using multi-step inter prediction, including a prediction block generation step of generating a prediction block from a weighted value of each of the reference blocks found by the two-inter prediction step, the first inter prediction step, and the second inter prediction step. Provides an encoding method.

Particularly, in the present invention, the second inter prediction step may find a second reference block on the second reference picture from the first reference block in order to perform at least two inter predictions to produce a more accurate prediction block. You can do The second inter prediction step may further find a third reference block on the third reference picture from the second reference block.

In addition, the first inter prediction step may find the first reference block using motion estimation or template matching. That is, when template matching is used in the first inter prediction step, information on a motion vector may be omitted.

In addition, in order to improve the accuracy of inter prediction, the prediction target block may be divided into sub blocks, and then inter prediction may be performed on each of the divided sub blocks. That is, in the second inter prediction step, the present invention divides the first reference block into at least two sub-blocks each having a predetermined size, and then divides the second reference block on the second reference picture for each of the divided sub-blocks. Provided are an image encoding method using multi-step inter prediction, which may be found. In addition, inter prediction may be performed from the first reference block to the second reference block in a forward or reverse direction.

The present invention may be characterized by performing multi-step inter prediction.

The encoded bitstream is transmitted from the encoding device to the decoding device. The decoding apparatus decodes the encoded bitstream received from the encoding apparatus through the entropy decoding unit 210, the inverse quantization unit 220, and the inverse transform unit 230 to calculate a reconstructed residual value and inter-screen prediction information. The inter prediction information is information including a prediction number, split information of sub blocks, weights, and the number of sub blocks.

Accordingly, in the image decoding method, a decoding step of decoding a coded bitstream and calculating a residual value and inter-screen prediction information, and a first reference picture on the first reference picture from the encoding target block on the current picture based on the inter-screen prediction information. An inter prediction step for finding a reference block, finding a second reference block on a second reference picture from the first reference block, and generating a predictive block, and reconstructing a reconstructed block by adding the reconstructed residual values and the predicted block. An image decoding method using multi-step inter prediction includes an image generation step.

6 is a block diagram of a video encoding apparatus using multi-step inter prediction according to the present invention. Referring to FIG. 6, the image encoding apparatus using the multi-step inter prediction according to the present invention includes a subtractor 110, a transform unit 120, a quantizer 130, an entropy coder 140, and an inverse quantizer 131. , An inverse transform unit 121, an adder 150, and a memory 160, which will be briefly described as a general component of an image encoding apparatus.

The subtractor 110 generates a residual image between the current image and the predicted image by subtracting the predicted image generated by the predictor from the target image to be encoded (the current image), which is the provided input image.

The converter 120 converts the residual image generated by the subtractor 110 from the spatial domain to the frequency domain. Here, the transform unit 120 may convert the residual image into the frequency domain by using a technique of transforming an image signal of a spatial axis into a frequency axis, such as a Hadamard transform, a discrete cosine transform, or the like.

The quantization unit 130 performs quantization on the transformed data (frequency coefficients) provided from the transformation unit 120. That is, the quantization unit 130 calculates a quantization result by approximating the frequency coefficients, which are data converted by the transformation unit 120, by quantization step size.

The entropy coder 140 generates a bitstream by entropy encoding a quantization result value calculated by the quantizer 130. In addition, the entropy coder 140 may entropy-encode the quantization result calculated by the quantization unit 130 using a context-adaptive variable length coding (CAVLC) or context-adaptive binary arithmetic coding (CABAC) technique. In addition to the quantization result, entropy encoding may be performed on index information of a reference frame, information of a motion vector, inter prediction information, and the like, which are information (decoding information) necessary for decoding an image.

The inverse quantization unit 131 dequantizes the quantization result value calculated by the quantization unit 130. [ That is, the inverse quantization unit 131 restores the value (frequency coefficient) of the frequency domain from the quantization result value.

The inverse transformer 121 restores the residual image by converting a value (frequency coefficient) of the frequency domain provided from the inverse quantizer 131 from the frequency domain to the spatial domain, and the adder 150 predicts the prediction generated by the predictor. The reconstructed image of the input image is generated by adding the residual image reconstructed by the inverse change unit 121 to the image, and provides the reconstructed image to the memory 160. The prediction unit may perform inter prediction, and may include a motion estimation unit and a motion compensation unit.

In particular, the predictor for performing multi-step inter prediction according to the present invention is characterized by including a first inter predictor 171, a block divider 172, and a second inter predictor 173.

The first inter prediction unit 171 finds the first reference block on the first reference picture from the encoding target block on the current picture, and the second inter prediction unit 173 determines the second reference on the second reference picture from the first reference block. Find the block. In addition, the second inter prediction unit 173 may apply a weight to the first reference block found by the first inter prediction unit 171 and the second reference block found by the second inter prediction unit 173, and add a weight to the prediction block. Can be formed. That is, the first inter predictor 171 or the second inter predictor 173 may include a motion estimator and a motion compensator to perform inter prediction.

The prediction block generated by the second inter prediction unit 173 is sent to the subtractor 110 to calculate a residual image (residual value), and the prediction block sent to the adder 150 is a reconstructed residual image (residual value). ) And the restored image may be stored in the memory 160.

In addition, the encoding apparatus using the multi-step inter prediction according to the present invention may perform at least two inter predictions, and the second inter prediction unit 173 may perform a second reference on the second reference picture from the first reference block. The block can be predicted (finded).

That is, in the video encoding apparatus, the first inter prediction unit 171 which finds the first reference block on the first reference picture from the encoding target block on the current picture, and the second on the second reference picture from the first reference block A video encoding apparatus using multi-step inter prediction includes a second inter prediction unit 173 for finding a reference block.

In addition, the first inter prediction unit 171 may find the first reference block by using motion estimation or template matching, and divides the first reference block into at least two sub-blocks each having a preset size. 172 may be further included, and the motion estimation may be performed from the divided subblocks.

7 is a block diagram of an image decoding apparatus using multi-step inter prediction according to the present invention. Referring to FIG. 7, an apparatus for decoding an image using multi-step prediction according to the present invention includes an entropy decoder 210, an inverse quantizer 220, an inverse transformer 230, an adder 240, and a memory 250. It may be configured to include, which will be briefly described as a general component of the image decoding apparatus.

The entropy decoder 210 entropy decodes the bitstream generated from the image encoding apparatus 100 shown in FIG. 6, and the inverse quantizer 220 inversely quantizes the data received from the entropy decoder 210. Restore the value (frequency coefficient) of the domain. In addition, the inverse transformer 230 inversely converts a value (frequency coefficient) in the frequency domain from the frequency domain to the spatial domain. The adder 240 adds the residual image (residual value) reconstructed by the inverse transformer 230 to the prediction image generated by the prediction unit including the first inter prediction unit 261 and the second inter prediction unit 263. In addition, a reconstruction image of the current image is generated.

In particular, the decoding apparatus using the multi-step inter prediction according to the present invention is characterized by including a first inter prediction unit 261, a block splitter 262, and a second inter prediction unit 263.

Here, the first inter prediction unit 171 or the second inter prediction unit 173 receives the index information of the reference frame, the information of the motion vector, and the inter prediction information, which is information (decoding information) necessary for decoding, and is inter-prediction. Can be performed. Here, the inter prediction information may include partition information of a block, a prediction number, a weight, and the like.

That is, in the video decoding apparatus, the first inter prediction unit 261 looking for the first reference block on the first reference picture from the encoding target block on the current picture, and the second on the second reference picture from the first reference block. The prediction block generated based on the reference block found by the second inter prediction unit 263, the first inter prediction unit 261, and the second inter prediction unit 263, which finds the reference block, is summed. An image decoding apparatus using multi-step inter prediction includes an adder 240 for generating a reconstructed image.

According to the above-described present invention, a reference block is again found from a reference block for an encoding target block, and an apparatus and method for generating a prediction block by weighted sum of a plurality of reference blocks are provided to obtain a more accurate prediction block. In addition, the present invention provides an apparatus and method for encoding / decoding an image having improved coding efficiency.

It will be apparent to those skilled in the art that various modifications and variations can be made in the present invention without departing from the spirit or scope of the present invention as defined by the following claims It can be understood that

100: encoding apparatus 110: subtractor
120: transform unit 130: quantization unit
140: entropy code part 121, 230: inverse transform part
131 and 220: inverse quantization unit 150 and 240: adder
160 and 250: memory 171 and 261: first inter prediction unit
172, 262: block division unit 173, 263: second inter prediction unit
200: decoding apparatus 210: entropy decoding unit

Claims (18)

In the video encoding method,
A first inter prediction step of finding a first reference block on a first reference picture from a block to be encoded on the current picture;
A second inter prediction step of finding a second reference block on a second reference picture from the first reference block; And
And a prediction block generation step of generating a prediction block based on the reference blocks found by the first inter prediction step and the second inter prediction step.
The method of claim 1, wherein the prediction block generating step
And weighting each of the first reference block and the second reference block to add the prediction block to generate the prediction block.
The method of claim 1, wherein the first inter prediction step
The method of claim 1, wherein the first reference block is found using motion estimation or template matching.
The method of claim 1, wherein the second inter prediction step
And dividing the first reference block into at least two subblocks having a predetermined size, and then finding the second reference block on the second reference picture for each of the divided subblocks. Image coding method used.
The method of claim 2, wherein the second inter prediction step
And the second reference block is found in the forward or reverse direction from the first reference block.
In the image decoding method,
A decoding step of decoding the encoded bitstream to calculate a reconstructed residual value and inter-screen prediction information;
Based on the inter prediction information, a first reference block on a first reference picture may be found from a block to be encoded on a current picture, and a prediction block may be generated by finding a second reference block on a second reference picture from the first reference block. Inter-screen prediction step; And
And a reconstructed image generating step of generating a reconstructed block by summing the reconstructed residual value and the predictive block.
The method of claim 6, wherein the inter prediction step
And decoding the first reference block using motion estimation or template matching.
The method of claim 6, wherein the inter prediction step
And dividing the first reference block into at least two subblocks having a predetermined size, and then finding the second reference block on the second reference picture for each of the divided subblocks. Image decoding method using.
9. The method of claim 8,
The inter-prediction information includes image segmentation information, prediction counts, and weights of blocks.
A video encoding apparatus comprising:
A first inter predictor for finding a first reference block on a first reference picture from a block to be encoded on the current picture; And
And a second inter prediction unit for finding a second reference block on a second reference picture from the first reference block.
The method of claim 10, wherein the second inter prediction unit
And encoding and adding weights to each of the first reference block and the second reference block to generate a prediction block.
The method of claim 10, wherein the first inter prediction unit
And a first reference block using motion estimation or template matching to find the first reference block.
The apparatus of claim 10, wherein the video encoding apparatus
And a block divider for dividing the first reference block into at least two sub-blocks each having a predetermined size.
The method of claim 13, wherein the second inter prediction unit
And the second reference block on the second reference picture is found for each of the at least two subblocks.
In the image decoding apparatus,
A first inter predictor for finding a first reference block on a first reference picture from a block to be encoded on the current picture;
A second inter predictor for finding a second reference block on the second reference picture from the first reference block; And
Image decoding using multi-step inter prediction, including an adder for generating a reconstructed image by summing a predicted block generated based on a reference block found by the first inter predictor and the second inter predictor, and a reconstructed residual value Device.
The method of claim 15, wherein the first inter prediction unit
And a first reference block using motion estimation or template matching to find the first reference block.
The image decoding apparatus of claim 15,
And a block divider for dividing the first reference block into at least two sub-blocks each having a predetermined size.
The method of claim 17, wherein the second inter prediction unit
And the second reference block on the second reference picture is found for each of the at least two sub-blocks.
KR1020110086201A 2011-08-29 2011-08-29 Apparatus and method for video encoding/decoding using multi-step inter prediction KR20130023444A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
KR1020110086201A KR20130023444A (en) 2011-08-29 2011-08-29 Apparatus and method for video encoding/decoding using multi-step inter prediction

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
KR1020110086201A KR20130023444A (en) 2011-08-29 2011-08-29 Apparatus and method for video encoding/decoding using multi-step inter prediction

Publications (1)

Publication Number Publication Date
KR20130023444A true KR20130023444A (en) 2013-03-08

Family

ID=48175670

Family Applications (1)

Application Number Title Priority Date Filing Date
KR1020110086201A KR20130023444A (en) 2011-08-29 2011-08-29 Apparatus and method for video encoding/decoding using multi-step inter prediction

Country Status (1)

Country Link
KR (1) KR20130023444A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018056763A1 (en) * 2016-09-23 2018-03-29 엘지전자(주) Method and apparatus for performing prediction using template-based weight
CN111050177A (en) * 2018-10-15 2020-04-21 腾讯科技(深圳)有限公司 Video encoding method, video decoding method, video encoding apparatus, video decoding apparatus, computer device, and storage medium
US10972751B2 (en) 2013-10-17 2021-04-06 Nippon Telegraph And Telephone Corporation Video encoding apparatus and method, and video decoding apparatus and method
CN112997487A (en) * 2018-11-15 2021-06-18 北京字节跳动网络技术有限公司 Coordination between affine mode and other inter-frame coding tools

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10972751B2 (en) 2013-10-17 2021-04-06 Nippon Telegraph And Telephone Corporation Video encoding apparatus and method, and video decoding apparatus and method
WO2018056763A1 (en) * 2016-09-23 2018-03-29 엘지전자(주) Method and apparatus for performing prediction using template-based weight
US10805631B2 (en) 2016-09-23 2020-10-13 Lg Electronics Inc. Method and apparatus for performing prediction using template-based weight
CN111050177A (en) * 2018-10-15 2020-04-21 腾讯科技(深圳)有限公司 Video encoding method, video decoding method, video encoding apparatus, video decoding apparatus, computer device, and storage medium
WO2020078211A1 (en) * 2018-10-15 2020-04-23 腾讯科技(深圳)有限公司 Video encoding and video decoding methods, apparatus, computer device, and storage medium
CN111050177B (en) * 2018-10-15 2021-07-20 腾讯科技(深圳)有限公司 Video encoding method, video decoding method, video encoding apparatus, video decoding apparatus, computer device, and storage medium
US11445205B2 (en) 2018-10-15 2022-09-13 Tencent Technology (Shenzhen) Company Limited Video encoding method and apparatus, video decoding method and apparatus, computer device, and storage medium
CN112997487A (en) * 2018-11-15 2021-06-18 北京字节跳动网络技术有限公司 Coordination between affine mode and other inter-frame coding tools
CN112997487B (en) * 2018-11-15 2024-07-09 北京字节跳动网络技术有限公司 Coordination between affine patterns and other inter-frame codec tools

Similar Documents

Publication Publication Date Title
US20210160508A1 (en) Method and apparatus for encoding and decoding video using skip mode
KR101997604B1 (en) Method and Apparatus for Video Encoding/Decoding
KR101943049B1 (en) Method and Apparatus for Image Encoding/Decoding
KR20200026863A (en) Method for intra prediction and apparatus thereof
US9264731B2 (en) Method and an apparatus for decoding a video
KR101681303B1 (en) Method and Apparatus for Encoding/Decoding of Video Data Using Partitioned-Block Prediction
KR101955374B1 (en) Method and Apparatus for Image Encoding/Decoding By Fast Coding Unit Mode Decision
KR20130085838A (en) Video coding method and apparatus using weighted prediction
KR20090095012A (en) Method and apparatus for encoding and decoding image using consecutive motion estimation
KR20120025174A (en) Method and apparatus for encoding/decoding of video data using efficient selection of intra prediction mode set
KR101427229B1 (en) Apparatus and Method for Video Encoding/Decoding using Adaptive Coding Order
US20130070846A1 (en) Motion vector encoding/decoding method and apparatus using a motion vector resolution combination, and image encoding/decoding method and apparatus using same
KR20140007097A (en) Method and apparatus for intra prediction using adaptive filtering
KR101596085B1 (en) Apparatus and method for video coding/decoding using adaptive intra prediction
KR101427516B1 (en) Method of prediction based on distribution of transform coefficient and apparatus thereof
KR101943058B1 (en) Method and Apparatus for Video Encoding/Decoding
KR20130023444A (en) Apparatus and method for video encoding/decoding using multi-step inter prediction
KR101911587B1 (en) Apparatus and method for video coding/decoding using adaptive intra prediction
KR20110090841A (en) Apparatus and method for encoding/decoding of video using weighted prediction
WO2012081895A1 (en) Intra prediction method and apparatus
KR101307406B1 (en) Encoding/decoding apparatus with reference frame compression
KR101401857B1 (en) Encoder/decoder for performing intra-prediction and methods for performing the same
KR101997655B1 (en) Method and Apparatus for Image Encoding/Decoding
KR101766600B1 (en) Encoder/decoder for performing intra-prediction and methods for performing the same
KR101672085B1 (en) Encoder/decoder for performing intra-prediction and methods for performing the same

Legal Events

Date Code Title Description
A201 Request for examination
E902 Notification of reason for refusal
E601 Decision to refuse application