throbber
US007903735B2
`
`(12) United States Patent
`Cha et al.
`
`(10) Patent No.:
`(45) Date of Patent:
`
`US 7,903,735 B2
`Mar. 8, 2011
`
`(54) METHOD OF EFFECTIVELY PREDICTING
`MULTI-LAYER BASED VIDEO FRAME, AND
`VIDEO CODING METHOD AND APPARATUS
`USING THE SAME
`
`(75) Inventors: Sang-chang Cha, Hwaseong-si (KR);
`Woo-jin Han, Suwon-si (KR)
`
`(73) Assignee: Samsung Electronics Co., Ltd.,
`Suwon-si (KR)
`
`(*) Notice:
`
`Subject to any disclaimer, the term of this
`patent is extended or adjusted under 35
`U.S.C. 154(b) by 1477 days.
`
`(21) Appl. No.: 11/311,384
`
`(22) Filed:
`
`Dec. 20, 2005
`
`(65)
`
`Prior Publication Data
`US 2006/O1651.71 A1
`Jul. 27, 2006
`
`Related U.S. Application Data
`(60) Provisional application No. 60/646,534, filed on Jan.
`25, 2005.
`
`(30)
`
`Foreign Application Priority Data
`
`Feb. 26, 2005 (KR). 10-2005-OO16270
`
`(51) Int. Cl.
`(2006.01)
`H04N 7/2
`(52) U.S. Cl. .......... 375/240.12; 375/240.15: 375/240.16;
`375/240.24; 375/240.08
`(58) Field of Classification Search ........................ None
`See application file for complete search history.
`
`(56)
`
`References Cited
`
`U.S. PATENT DOCUMENTS
`2003, OO12279 A1*
`1/2003 Chaddha .................. 375,240.12
`11/2003 Zhou
`2003/0206594 A1
`
`
`
`2004/0042549 A1* 3/2004 Huang et al. ............. 375,240.11
`2005. O157784 A1* 7, 2005 Tanizawa et al. ........ 375,240.03
`2005, 0163211 A1* 7/2005 Shanableh ................. 375,240.1
`2005/01693.74 A1* 8/2005 Marpe et al. ...
`375,240.16
`2005/0195896 A1* 9/2005 Huang et al. ............. 375,240.03
`FOREIGN PATENT DOCUMENTS
`O7-222 165
`8, 1995
`08-237.599
`9, 1996
`11-275585
`10, 1999
`1 429 564 A1
`6, 2004
`2001-0080644 A
`8, 2001
`10-2004-0054747. A
`6, 2004
`
`JP
`JP
`JP
`JP
`KR
`KR
`
`OTHER PUBLICATIONS
`Ascenso, Joao; Pereira, Fernando; Drift Reduction for a H.264/AVC
`Fine Grain Scalability with Motion Compensation Architecture;
`2004; International Conference on Image Processing: pp. 2259
`2262.* Yuwen He, Xuejun Zhao, Yuzhuo Zhong, Shiqiang Yang;
`Improved Fine Granular Scalable Coding with Inter-Layer Predic
`tion; 2002; Data Compression Conference; pp. 1-10.*
`* cited by examiner
`Primary Examiner — Mark KZimmerman
`Assistant Examiner — Miya J Cato
`(74) Attorney, Agent, or Firm — Sughrue Mion, PLLC
`(57)
`ABSTRACT
`A method and apparatus are provided for effectively predict
`ing a video frame that use all of the advantages of an intra
`mode and an intra BL mode in multi-layer structure based
`Video coding. The method includes reconstructing the intra
`block of a lower layer using the previously reconstructed first
`neighboring blocks of the intra block; subtracting the first
`neighboring blocks from previously stored the second neigh
`boring blocks of an upper layer corresponding to the first
`neighboring blocks; creating a differential predicted block
`based on a predetermined intra prediction mode by perform
`ing intra prediction using virtual differential neighboring
`blocks that are created as a result of the Subtraction; adding
`the differential predicted block and the reconstructed intra
`block; and Subtracting a predicted block, which is created as
`a result of the addition, from a block of the upper layer
`corresponding to the intra block.
`
`25 Claims, 11 Drawing Sheets
`
`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 1
`
`

`

`U.S. Patent
`
`Mar. 8, 2011
`
`Sheet 1 of 11
`
`US 7,903,735 B2
`
`F.G. 1
`
`8 directions
`
`(8)
`
`(1)
`
`(6)
`
`+ DC (2)
`
`(3)
`
`(7)
`
`(O)
`
`(5)
`
`(4)
`
`
`
`FIG 2
`
`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 2
`
`

`

`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 3
`
`

`

`U.S. Patent
`
`US 7,903,735 B2
`
`
`
`
`
`CN
`
`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 4
`
`

`

`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 5
`
`

`

`U.S. Patent
`
`Mar. 8, 2011
`
`Sheet 5 of 11
`
`US 7,903,735 B2
`
`9
`
`* ?I),
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 6
`
`

`

`U.S. Patent
`
`Mar. 8, 2011
`
`Sheet 6 of 11
`
`US 7,903,735 B2
`
`FG 7
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`inter
`prediction
`unit 62
`
`intra
`prediction
`unit (63)
`
`intra B
`prediction
`unit (64)
`
`
`
`
`
`
`
`
`
`
`
`WeSe
`quantization
`Unit (71)
`
`VerSe
`transform
`unit (72)
`
`
`
`
`
`
`
`
`
`
`
`VerSe inter
`prediction
`unit (74)
`
`in VerSe
`prediction
`Unit (75)
`
`
`
`Verse intra
`B prediction
`unit 76)
`
`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 7
`
`

`

`U.S. Patent
`
`Mar. 8, 2011
`
`Sheet 7 of 11
`
`US 7,903,735 B2
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`(£11
`
`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 8
`
`

`

`U.S. Patent
`
`Mar. 8, 2011
`
`Sheet 8 of 11
`
`US 7,903,735 B2
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`0T º ?I ?I
`
`(DGI)
`
`Fl-ºf
`
`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 9
`
`

`

`U.S. Patent
`
`Mar. 8, 2011
`
`Sheet 9 of 11
`
`US 7,903,735 B2
`
`FG 11
`
`y
`f
`- ->,
`l
`(5)
`
`(O)
`
`y
`(7)
`
`Upper
`layer
`
`OWer
`layer
`
`O)
`
`F.G. 12
`
`
`
`(8)
`
`(1)
`
`(6)
`
`3)
`
`(7)
`
`(O)
`
`(5)
`
`(4)
`
`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 10
`
`

`

`U.S. Patent
`
`Mar. 8, 2011
`
`Sheet 10 of 11
`
`US 7,903,735 B2
`
`F.G. 13
`82
`
`81
`
`Upper
`layer
`
`83
`
`
`
`OWer
`layer
`
`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 11
`
`

`

`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 12
`
`

`

`US 7,903,735 B2
`
`1.
`METHOD OF EFFECTIVELY PREDICTING
`MULTI-LAYER BASED VIDEO FRAME, AND
`VIDEO CODNG METHOD AND APPARATUS
`USING THE SAME
`
`CROSS-REFERENCE TO RELATED
`APPLICATIONS
`
`This application claims priority from Korean Patent Appli
`cation No. 10-2005-0016270 filed on Feb. 26, 2005 in the
`Korean Intellectual Property Office, and U.S. Provisional
`Patent Application No. 60/646,534 filed on Jan. 25, 2005 in
`the United States Patent and Trademark Office, the disclo
`sures of which are incorporated herein by reference in their
`entirety.
`
`10
`
`15
`
`BACKGROUND OF THE INVENTION
`
`1. Field of the Invention
`Apparatuses and methods consistent with the present
`invention relate generally to video coding, and more particu
`larly, to effectively predicting a video frame that use all of the
`advantages of an intra mode and an intra BL mode in multi
`layer structure based-video coding.
`2. Description of the Related Art
`As information and communication technology, including
`the Internet, develops, image-based communication as well
`as text-based communication and Voice-based communica
`tion is increasing. The existing text-based communication is
`insufficient to satisfy various consumers demands. Therefore,
`the provision of multimedia service capable of accommodat
`ing various types of information, such as text, images and
`music, is increasing. Since the amount of multimedia data is
`large, multimedia data require high-capacity storage media
`and require broad bandwidth at the time of transmission.
`Therefore, to transmit multimedia data, including text,
`images and audio, it is essential to use a compression coding
`technique.
`The fundamental principle of data compression is to elimi
`nate redundancy in data. Data can be compressed by elimi
`nating spatial redundancy Such as the case where an identical
`color or object is repeated in an image, temporal redundancy
`Such as the case where there is little change between neigh
`boring frames or identical audio Sound is repeated, or psy
`chovisual redundancy in which the fact that humans visual
`and perceptual abilities are insensitive to high frequencies is
`taken into account.
`For such a moving image compression method, H.264/
`Advanced Video Coding (AVC), which has higher compres
`sion efficiency than Moving Picture Experts Group (MPEG)-
`4, has attracted attention recently. H.264 uses directional
`intra-prediction, which eliminates spatial similarity in each
`frame, as one of the schemes for improving compression
`efficiency.
`Directional intra-prediction is a method of predicting the
`values of a current Sub-block and encoding only a difference
`in Such a way as to perform copying in predetermined direc
`tions using neighboring pixels on the upper and left sides of a
`sub-pixel with respect to the sub-block.
`In H.264, a predicted block with respect to a current block
`is generated based on other blocks having preceding sequen
`tial positions. The difference between the current block and
`the predicted block is encoded. For a luminance component,
`each predicted block is generated on a 4x4 block or 6x16
`macroblock basis. There area total of nine optional prediction
`modes for each 4x4 block, whereas there are a total of four
`optional prediction modes for each 16x16 block. An H.264
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`2
`based video encoder selects the one prediction mode that
`minimizes the difference between the current block and the
`predicted block, from among the prediction modes, for each
`block.
`For prediction modes for the 4x4 block, H.264, as shown in
`FIG. 1, employs a total of nine prediction modes, including a
`total of nine directional modes (modes 0, 1, 3 to 8), and a DC
`mode (mode 2) that uses an average of the values of nine
`neighboring pixels.
`FIG. 2 illustrates an example of labeling to illustrate the
`nine prediction modes. In this example, a predicted block
`(including regions “a” to “p') with respect to a current block
`is generated using previously decoded samples A to M. If
`regions E, F, G and H cannot be previously decoded, regions
`E, F, G and H can be virtually created by copying region D to
`the locations of the regions E, F, G and H.
`With reference to FIG. 3, the nine prediction modes are
`respectively described in detail below. In the case of mode 0.
`the pixels of a predicted block are extrapolated using upper
`samples A, B, C and D in a vertical direction, and in the case
`of mode 1, the pixels are extrapolated using left samples I, J.
`Kand L in a horizontal direction. Furthermore, in the case of
`mode 2, the pixels of the predicted block are uniformly
`replaced by the averages of upper samples A, B, C and Dand
`left samples I, J, K and L.
`In the case of mode 3, the pixels of the predicted block are
`interpolated between a lower-left position and an upper-right
`position in a direction that is inclined at an angle of 45°, and
`in the case of mode 4, the pixels are extrapolated in a direction
`that is inclined toward an upper-left position at an angle of
`45°. Furthermore, in the case of mode 5, the pixels of the
`predicted block are extrapolated in a direction that is inclined
`rightward from a vertical direction at an angle of about 26.6°
`(width/height=/2).
`In the case of mode 6, the pixels of the predicted block are
`extrapolated in a direction that is inclined downward from a
`horizontal direction at an angle of about 26.6°, and in the case
`of mode 7, the pixels are extrapolated in a direction that is
`inclined leftward from a vertical directionatan angle of about
`26.6°. Finally, in the case of mode 8, the pixels of the pre
`dicted block are interpolated in a direction that is inclined
`upward from a horizontal direction at an angle of about 26.6°.
`The arrows of FIG. 3 indicate prediction directions in
`respective modes. In modes 3 to 8, the samples of the pre
`dicted block can be generated from the weighted averages of
`previously decoded reference samples A to M. For example,
`in the case of mode 4, Sampled, which is located in the upper
`left, can be predicted as expressed by the following Equation
`1. In this Equation, the round.(...) function is a function that
`rounds off an input value to an integer position.
`
`Meanwhile, a 16x16 prediction model for luminance com
`ponents includes four modes, that is, mode 0, mode 1, mode
`2 and mode 3. In the case of mode 0, the pixels of a predicted
`block are extrapolated from upper samples H, and in the case
`of mode 1, the pixels of a predicted block are extrapolated
`from left samples V. Furthermore, in the case of mode 2, the
`pixels of a predicted block are calculated using the averages
`ofupper samples H and left samples V. Finally, in the case of
`mode 3, a “plane' function suitable for upper samples H and
`left samples V is used. This mode is more suitable for a region
`in which luminance Smoothly changes.
`Meanwhile, in addition to efforts to improve the efficiency
`of video coding, research into video coding that allows the
`resolution of transmitted video data, a frame rate, and a Sig
`
`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 13
`
`

`

`3
`nal-to-Noise Ratio (SNR) to be adjusted, that is, that supports
`Scalability, is actively being carried out.
`With regard to this scalable video coding technique, stan
`dardization work is in progress in the Moving Picture Experts
`Group (MPEG)-21 PART-13. Of these methods for support
`ing scalability, a multi-layered video coding method is con
`sidered a prominent method. For example, multiple layers,
`including a base layer, a first enhanced layer and a second
`enhanced layer 2, are provided and respective layers have
`different resolutions QCIF, CIF and 2CIF or different frame
`rates.
`In the scalable video coding standard currently in progress,
`besides interprediction and directional intra prediction (here
`inafter simply referred to as intra prediction) used in existing
`H.264 to predict a current block or macroblock, a method of
`predicting a layer to which a current block belongs using the
`correlation between the current block and a corresponding
`lower layer block is additionally introduced. This prediction
`method is referred to as “intra BL (intra BL) prediction' in
`the standard, and the case of performing encoding using Such
`prediction is referred to as “intra BL mode.”
`FIG. 4 is a schematic diagram showing the three prediction
`methods, which illustrates the case of performing intra pre
`diction on a macroblock of a current frame 1 (CD), the case of
`performing interprediction using a frame 2 placed at a tem
`25
`poral location different from that of the current frame 1 (2)),
`and the case of performing intra BL prediction using texture
`data about the region 6 of the frame of a base layer corre
`sponding to a macroblock ((3)).
`As described above, in the scalable video coding standard,
`the advantageous one of the three prediction methods is
`selected for each macroblock, and a corresponding macrob
`lock is encoded using the selected method. That is, for one
`macroblock, inter prediction, intra prediction and intra BL
`prediction are selectively used. However, a differential block
`35
`created using intra BL prediction still has considerable cor
`relation with neighboring differences. Accordingly, it is nec
`essary to develop a prediction technique that takes the advan
`tages of both intra BL prediction and intra prediction into
`account. Although a prediction method that takes the advan
`tages of intra prediction, intra BL prediction and interpredic
`tion into account may be considered, the characteristics of
`intra BL prediction and intra prediction are considerably dif
`ferent from those of inter prediction, so this method is not
`desirable.
`
`10
`
`15
`
`30
`
`40
`
`45
`
`US 7,903,735 B2
`
`4
`According to an aspect of the present invention, there is
`provided a method of efficiently predicting a multi-layer
`based video frame, including reconstructing the intra block of
`a lower layer using previously reconstructed the first neigh
`boring blocks of the intra block; subtracting the first neigh
`boring blocks from the previously stored second neighboring
`blocks of an upper layer corresponding to the first neighbor
`ing blocks; creating a differential predicted block based on a
`predetermined intra prediction mode by performing intra pre
`diction using virtual differential neighboring blocks that are
`created as a result of the subtraction; adding the differential
`predicted block and the reconstructed intra block; subtracting
`a predicted block, which is created as a result of the addition,
`from a block of the upper layer corresponding to the intra
`block; and encoding a residual block that is created as a result
`of the subtraction.
`According to an aspect of the present invention, there is
`provided a method of efficiently predicting a multi-layer
`based video frame, including reconstructing the intra block of
`a lower layer using the previously reconstructed first neigh
`boring blocks of the intra block; subtracting the first neigh
`boring blocks from the previously stored second neighboring
`blocks of an upper layer corresponding to the first neighbor
`ing blocks; creating a differential predicted block based on a
`predetermined intra prediction mode by performing intra pre
`diction using virtual differential neighboring blocks that are
`created as a result of the subtraction; adding the differential
`predicted block and the reconstructed intra block; recon
`structing a block of the upper layer corresponding to the intra
`block; and adding the reconstructed block of the upper layer
`and a predicted block that is acquired as a result of the addi
`tion.
`According to an aspect of the present invention, there is
`provided a multi-layer based video encoder, including a
`means for reconstructing the intra block of a lower layer using
`the previously reconstructed first neighboring blocks of the
`intra block; a means for Subtracting the first neighboring
`blocks from the previously stored second neighboring blocks
`of an upper layer corresponding to the first neighboring
`blocks; a means for creating a differential predicted block
`based on a predetermined intra prediction mode by perform
`ing intra prediction using virtual differential neighboring
`blocks that are created as a result of the Subtraction; a means
`for adding the differential predicted block and the recon
`structed intra block; a means for Subtracting a predicted
`block, which is created as a result of the addition, from a block
`of the upper layer corresponding to the intra block; and a
`means for encoding a residual block that is created as a result
`of the subtraction.
`According to an aspect of the present invention, there is
`provided a multi-layer based video encoder, including a
`means for reconstructing the intra block of a lower layer using
`the previously reconstructed first neighboring blocks of the
`intra block; a means for Subtracting the first neighboring
`blocks from the previously stored second neighboring blocks
`of an upper layer corresponding to the first neighboring
`blocks; a means for creating a differential predicted block
`based on a predetermined intra prediction mode by perform
`ing intra prediction using virtual differential neighboring
`blocks that are created as a result of the Subtraction; a means
`for adding the differential predicted block and the recon
`structed intra block; a means for reconstructing a block of the
`upper layer corresponding to the intra block; and a means for
`adding the reconstructed block of the upper layer and a pre
`dicted block that is acquired as a result of the addition.
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`
`The above and other aspects of the present invention will be
`more clearly understood from the following detailed descrip
`
`SUMMARY OF THE INVENTION
`
`The present invention provides a prediction method that
`uses both intra BL prediction and intra prediction.
`Furthermore, the present invention may improve video
`coding efficiency using the prediction method.
`According to an aspect of the present invention, there is
`provided a method of efficiently predicting a multi-layer
`based video frame, including reconstructing the intra block of
`a lower layer using the previously reconstructed first neigh
`boring blocks of the intra block; subtracting the first neigh
`boring blocks from previously stored the second neighboring
`blocks of an upper layer corresponding to the first neighbor
`ing blocks; creating a differential predicted block based on a
`predetermined intra prediction mode by performing intra pre
`diction using virtual differential neighboring blocks that are
`created as a result of the subtraction; adding the differential
`predicted block and the reconstructed intra block; and sub
`tracting a predicted block, which is created as a result of the
`addition, from a block of the upper layer corresponding to the
`intra block.
`
`50
`
`55
`
`60
`
`65
`
`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 14
`
`

`

`US 7,903,735 B2
`
`10
`
`15
`
`5
`tion of the exemplary embodiments taken in conjunction with
`the accompanying drawings, in which:
`FIG. 1 is a diagram illustrating prediction directions for
`respective modes defined by H.264;
`FIG. 2 is a diagram illustrating an example of labeling that
`is used to illustrate the intra prediction modes of FIG. 1;
`FIG. 3 is a diagram illustrating the intra prediction modes
`of FIG. 1 in detail;
`FIG. 4 is a schematic diagram illustrating conventional
`three prediction methods;
`FIG. 5 is a diagram illustrating the concept of a prediction
`method according to an exemplary embodiment of the present
`invention;
`FIG. 6 is a block diagram illustrating the construction of a
`Video encoder according to an exemplary embodiment of the
`present invention;
`FIG. 7 is a block diagram illustrating the schematic con
`struction of an encoding unit that is included in the video
`encoder of FIG. 6;
`FIG. 8 is a block diagram illustrating the schematic con
`struction of a decoding unit that is included in the video
`encoder of FIG. 6;
`FIG. 9 is a block diagram illustrating the construction of a
`Video encoder according to another exemplary embodiment
`of the present invention;
`25
`FIG. 10 is a block diagram illustrating the construction of
`a video encoder according to still another exemplary embodi
`ment of the present invention;
`FIG. 11 is a view illustrating neighboring direction with
`respect to a vertical mode;
`FIG. 12 is a view illustrating neighboring directions with
`respect to 8 intra prediction modes having directionality:
`FIG. 13 is a diagram illustrating the corresponding regions
`between layers when the resolutions of the layers do not
`coincide with each other; and
`35
`FIG. 14 is a block diagram illustrating the construction of
`a video decoder according to an exemplary embodiment of
`the present invention.
`
`6
`which is produced between a video encoder and a video
`decoder, using such closed loop encoding, so that recon
`structed images, which are decoded after being encoded, are
`used as the images of existing blocks that are used to predict
`a current block in the present invention. However, it is appar
`ent to those skilled in the art that the present invention can also
`be applied to an open loop method using the original images
`of existing blocks.
`In a lower layer, a block corresponding to the current block
`10 (hereinafter referred to as a “corresponding block”) refers
`to an intra block 20 that has been encoded and decoded using
`intra prediction as shown in FIG. 3. Furthermore, the neigh
`boring blocks 25 of the intra block 20 can be encoded/de
`coded using any prediction method in the same manner as the
`neighboring blocks 15 of the upper layer. The neighboring
`blocks 25 of the lower layer exist at locations corresponding
`to those of the neighboring blocks 15.
`Meanwhile, the resolution of the upper layer may be iden
`tical to or different from the resolution of the lower layer. If
`the resolutions of both layers are identical to each other, the
`size of the blocks 10 and 15 of the upper layer is identical to
`the size of the blocks 20 and 25 of the lower layer. Otherwise,
`the blocks 10 and 15 of the upper layer may be larger than
`those of the lower layer. For example, if the resolution of the
`upper layer is two times the resolution of the lower layer and
`the blocks 20 and 25 of the lower layer have a 4x4 size, the
`blocks of the upper layer may have an 8x8 size.
`Furthermore, although the present invention is described
`with the term “neighboring block” referring to four blocks,
`that is, left, upper left, upper and upper right blocks, which are
`referred to for intra prediction, the number and locations of
`neighboring blocks may vary if the type of blocks, which are
`referred to for intra prediction, varies.
`As described above, when the neighboring blocks 15 of the
`upper layer and the intra block 20 and neighboring blocks 25
`of the lower layer have been all reconstructed, the predicted
`block 40 of the current block 10 can be created using them. A
`process of creating the predicted block 40 is described in
`detail below.
`First, virtual differential neighboring blocks 35 are created
`by subtracting the reconstructed neighboring blocks 25 of the
`lower layer from the neighboring blocks 15 of the upper layer.
`The subtraction is performed between the corresponding
`blocks of both layers. When directional intra prediction illus
`trated in FIG. 3 is applied to the neighboring blocks 35 cre
`ated as described above, differential predicted blocks 30 can
`be created for nine modes.
`Thereafter, each of the nine created differential predicted
`blocks 30 is added to the intra block 20, and then an optimal
`candidate is selected from nine candidates obtained for the
`predicted block 40. For example, the selection may be per
`formed in Such a way as to select one that minimizes the
`image different between each of the nine candidates and the
`current block 10. Alternatively, for another example, a
`method of predicting the current block 10 using the nine
`candidates for the predicted block 40, encoding prediction
`results and selecting an optimal candidate using the Rate
`Distortion (R-D) cost function may be used. The former
`method has the advantages of a small amount of calculation
`and simple performance, while the latter method has the
`advantage of more accurate selection between candidates.
`When the predicted block 40 is selected from the candi
`dates for the predicted block 40, the current block 10 is
`predicted using the predicted block 40. That is, a residual
`block 50 is created by subtracting the predicted block 40 from
`
`30
`
`DESCRIPTION OF THE EXEMPLARY
`EMBODIMENTS
`
`Reference now should be made to the drawings, in which
`the same reference numerals are used throughout the different
`drawings to designate the same or similar components.
`FIG. 5 is a diagram illustrating the concept of a prediction
`method according to an exemplary embodiment of the present
`invention. In the method, how to create a predicted block 40
`for the original image of a current block 10 (hereinafter sim
`ply referred to as a current block) that belongs to an upper
`layer is an essential point. In the present invention, the term
`“block” may be defined as a block having a size identical to
`that of an intra block of the H.264 standard. Meanwhile, for a
`conventional intra block, 4x4 mode and 16x16 mode exist for
`luminance components and 8x8 mode exists for chrominance
`components and, accordingly, the “block” of the present
`invention may have one of the various sizes.
`In an upper layer, before the prediction of a current block
`10, the neighboring blocks of the current block 10 have
`already been encoded/decoded using a certain prediction
`method. The neighboring blocks 15 may be encoded/decoded
`using any prediction method, Such as inter-prediction, intra
`prediction, or intra BL prediction.
`In order to predict the current block 10 as described above,
`images passed through encoding and decoding (closed loop),
`rather than the original images of the neighboring blocks 15.
`are used. Currently, most codecs eliminate drifting error,
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`Unified Patents, LLC v. Elects. & Telecomm. Res. Inst., et al.
`
`Ex. 1026, p. 15
`
`

`

`US 7,903,735 B2
`
`7
`the created current block 10. Thereafter, the fundamental
`operation of the present invention is completed by encoding
`the residual block 50.
`The schematic operation of the present invention is
`described above. The conventional method using an intra BL
`mode is schematically compared to the method of the present
`invention from a conceptual point of view. When the image of
`the current block10 is C1 and the reconstructed intra block 20
`of the lower layer is CO, CO is used as the predicted block of
`C1 in the conventional intra BL mode on the assumption that
`CO is similar to C1, as shown in Equation 2.
`(2)
`C1-COsO
`In contrast, the present invention does not use C0 as the
`predicted block, but instead uses the correlation of the fol
`lowing Equation 3. In Equation 3, INT(...) refers to a function
`that performs directional intra prediction, N1 refers to the
`neighboring blocks of the upper layer and No refers to the
`neighboring blocks of the lower layer.
`
`10
`
`15
`
`8
`forming the inverse of the process of the encoding unit 105,
`that is, inverse prediction, on inversely transformed signals.
`The reconstructed neighboring blocks are input to a Subtrac
`tor 135.
`An intra encoding unit 115 encodes the corresponding
`block image, and the intra decoding unit 120 decodes
`encoded results. It should be noted that in the present inven
`tion, the corresponding block must be encoded using intra
`prediction. However, the neighboring blocks of the current
`block or the neighboring blocks of the corresponding block
`may be encoded using any prediction method.
`Accordingly, the intra encoding unit 115 may beformed of
`the intra prediction unit 63, the transform unit 65 and the
`quantization unit 66, as shown in the block diagram of FIG. 7,
`and the intra decoding unit 120 may be formed of the inverse
`quantization unit 71, the inverse transform unit 72 and the
`inverse intra prediction unit 75, as shown in the block diagram
`of FIG. 8.
`A signal output from the intra decoding unit 120, that is, a
`reconstructed corresponding block (intra block), is input to an
`adder 160 selectively through an up-sampler 150. The up
`sampler 150 is not used when the resolutions of the upper
`layer and the lower layer coincide with each other, and per
`forms up-sampling so that the resolution of the reconstructed
`intra block coincides with that of the upper layer when the
`resolutions do not coincide with each other.
`Meanwhile, the encoding unit 125 encodes the images of
`the neighboring blocks of the corresponding block, and the
`decoding unit 130 decodes the encoded results. The encoding
`unit 125 may have a construction identical to that of FIG. 7,
`and the decoding unit 130 may have a construction identical
`to that of FIG. 8.
`A signal output from the decoding unit 130, that is, recon
`structed neighboring blocks, is input to the subtractor 160
`selectively through the up-sampler 140. The up-sampler 140
`is not used when the resolutions of the upper layer and the
`lower layer coincide with each other, and performs up-Sam
`pling so that the resolution of the reconstructed neighboring
`blocks coincides with that of the upper layer when the reso
`lutions do not coincide with each other.
`The subtractor 135 acquires differential neighboring
`blocks by Subtracting a signal, which is input from the decod
`ing unit 130 or up-sampler 140, from a signal, which is input
`from the decoding unit 110.
`A predicted block creation unit 145 creates a predeter
`mined number of differential predicted blocks for respective
`modes by performing directional intra prediction using the
`differential neighboring blocks. For example, as shown in
`FIG.3, in H.264, a total of nine modes, including eight modes
`having directionality and one DC mode, can be used. Nine
`differential predicted blocks are created for the respective
`modes, and the created differential predicted blocks are pro
`vided to the adder 160.
`The adder 160 adds the differential predicted blocks for the
`respective modes to the signal input from the intra decoding
`unit 120 or from the up-sampler 150. As a result, a number of
`the predicted block candidates equal to the number of modes
`are created.
`The mode selection unit 190 selects an optimal mode (intra
`prediction mode) from the modes that the predicted block
`candidates have, and selects an optimal predicted block from
`the predicted block candidates based on the selected mode.
`The mode selection is performed in Such away as to select the
`mode that minimizes the difference between the predicted
`block candidates and the current block, as described above. In
`
`25
`
`30
`
`35
`
`40
`
`As shown in Equation3, the difference between the current
`block Co and the corresponding lower laye

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket