throbber
United States Patent [191
`Tahara
`
`HllllIlllliilllllllllll|l|llllllllllllllllllll||||l|l|||illlllllllllllllll
`US005473380A
`[11] Patent Number:
`5,473,380
`[45] Date of Patent:
`Dec. 5, 1995
`
`[54] PICTURE SIGNAL TRANSMITTING
`METHOD AND APPARATUS
`
`5,260,783 11/1993 Dixit ..................................... .. 348/401
`FOREIGN PATENT DOCUMENTS
`
`[75] Inventor: Katsumi Tahara, Kanagawa, Japan
`
`0349347 1/1990 European Pat. 01f. .
`
`[73] Assignee: Sony Corporation, Tokyo, Japan
`
`[21] Appl. No.: 219,472
`[22] Filed:
`Mar. 29, 1994
`_
`_
`‘V
`_
`_
`Forelgn Apphcatlon Pnonty Data
`
`[30]
`
`OTHER PUBLICATIONS
`AT&T Technical Journal, vol. 72, No. 1, Feb. 1993, New
`York, pp. 67-89, Aravind et al., “Image and Video Coding
`St (1 d ”.
`an at S
`Primary Examiner-Howard W. Britton
`Attorney, Agent, or Firm-—Wi1lia.m S. Frommer; Alvin
`
`Mar. 29, 1993
`
`[JP]
`
`Japan
`
`. .. .
`
`...... .. 5-069829
`
`Smderbrand
`
`'
`
`[51] Int. Cl.6 ..................................................... .. H04N 7/50
`[52] US. Cl. ......................... .. 348/423; 358/335; 358/339
`[58] Field of Search ................................... .. 348/423, 402,
`348/416; 370/1101; 358/335, 339; H04N 7/133,
`7/137
`
`[56]
`
`_
`References Clted
`Us PATENT DOCUMENTS
`
`ABSTRACT
`[571
`A picture type identi?er, indicating one of intra-picture
`coding (an I-picture), forward or backward predictive cod
`ing (a P-picture) and bi-directionally predictive coding (a
`B-picture), is included with a picture signal when the signal
`is encoded and when the signal is decoded. Each of initial
`and subsequent encoding and decoding is a function of the
`included picture type.
`
`5,148,272
`
`9/1992 Acampora .......................... .. 370/1101
`
`20 Claims, 18 Drawing Sheets
`
`INPUT
`PICTURE
`SIGNAL
`
`PFIE-
`PROCESSING
`CIRCUIT
`
`LUMINANCE
`SIGNAL
`COLQR
`
`,2
`3
`ND
`
`[O
`
`SEPARATION —_>I_T%_I1
`(
`ll
`,3
`
`38
`r/
`PRE-
`PROCESSING
`
`Y/C
`SYNTHESIS
`
`OUTPUT
`2:21.112
`
`3
`O
`
`“£13191? E
`$2505
`IF
`D S'GNA'"
`
`36
`S
`
`D/A
`
`D/A
`
`S
`37
`
`FRAME MEMORY
`I5
`
`I4
`'/
`
`—-'
`
`I6
`,
`COLOR DIFFERENCE
`FRAME MEMORY
`
`3
`
`[7 .
`~
`co?l'é'g?m
`
`,
`
`FRAME
`l
`
`B"
`STREAM
`
`I
`
`RECORDING
`MEDIUM
`BIT
`
`E
`
`FRAM
`'
`
`34
`J
`1_-
`
`_
`
`DECODER
`
`3/ 2
`
`FORMAT
`609.1911“
`BLOCK
`
`‘A \35 32 \_
`coLoR DIFFERENCE 33
`FRAME MEMORY N/
`
`FRAME
`
`Apple Exhibit 1012
`Page 1 of 29
`
`

`
`US. Patent
`
`Dec. 5, 1995
`
`Sheet 1 of 18
`
`5,473,380
`
`maul 1k¥
`
`P01)
`
`Apple Exhibit 1012
`Page 2 of 29
`
`

`
`US. Patent
`
`Dec. 5, 1995
`
`5,473,380
`
`$
`
`,
`
`if
`
`T
`
`m
`E
`5“ I
`
`'1'
`
`5E
`
`*-
`
`E
`
`1*
`
`at
`
`“E
`
`Apple Exhibit 1012
`Page 3 of 29
`
`

`
`U.S. Patent
`
`Dec. 5, 1995
`
`Sheet 3 of 18
`
`5,473,380
`
`@X
`
`OX
`
`<m.0_..._
`
`Apple Exhibit 1012
`Page 4 of 29
`
`Apple Exhibit 1012
`Page 4 of 29
`
`

`
`U.S. Patent
`
`Dec. 5, 1995
`
`Sheet 4 of 18
`
`5,473,380
`
`10E
`
`Apple Exhibit 1012
`Page 5 of 29
`
`

`
`US. Patent
`
`Dec. s, 1995
`
`Sheet 5 of 18
`
`5,473,380
`
`H DOTS
`
`SLICE 1
`
`SLICE 2
`
`SLICE 3
`
`SLICE N
`
`(A) FRAME
`
`8 DOTS 8 DOTS
`
`8 DOTS
`
`Y [ 1 ]
`
`Y [ 2 ]
`
`“DOTS
`
`Y[31
`
`Y[4l
`
`VLINES
`
`SDQTS Cb[5]
`
`CI'[6]
`
`xN
`
`XM
`
`(C) MACRQ BLOCK
`
`16
`
`1612345678
`
`(B) SLICE
`
`Apple Exhibit 1012
`Page 6 of 29
`
`

`
`U.S. Patent
`
`Dec. 5, 1995
`
`Sheet 6 of 18
`
`5,473,380
`
`._.Dm._.DO
`
`mm_rs_<mm»m
`
` _I _s_<mEm:m4_Iv_o<mammu_
`
`zo_mm__2
`
`mmumam
`
`_____
`
`
`
`:=om_ozo_m_omo
`
`
`
`mmtamVmzoFo_nmm._wm:oI_mm_.._1._<zoFomm_n_m\momm=._<>_mmnm<zE9=<Ez_
`mS._omm<_nosawus.__
`
`____
`
`mm:5_a
`
`
`
`_IImo_>mo5%.IIIIIIL.IIIIIIIIIIII_wn_>._.
`
`Ben
`
`>mos_m_2m_2<E
`
`em
`
`2m
`
`._<2_O_mO
`
`mmD._.O_..._
`
`._<2_0EO
`
`mmaoa
`
`III;w<.EEn__o._m:m_:<E_
`
`nE<_sv_o<m
`
`zo_._.o_nmmu
`
`
`
`nmm'lmE:.o_._zo=<m-
`
`
`
`P8II..,n_.,__$_,m._.,.__._vmmmmm>mw_wm<_._o
`>mo_zms_ms_<mu_muImmapofi-
`
`
`
`zmn_s_oo
`zofiozh
`
`G_.:._
`
`zo:.os_
`
`_o._m_u_
`
`
`
`
`
`n._mEms_<E
`
`zo:o_omm._
`
`IIIL
`
`_.0<._u_zo_»o_omEa
`
`_n._u_n=m_2<Efl_
`
`NDOE
`
`:_._om_o
`
`IE».wM___ommn___w¢._<&%a
`.
`mm>o-
`mm>omoz<:oND
`
`zofiowpmn
`
`._.Dn_z_
`
`$.25...
`
`_______
`
`._._=O_..=O
`
`on
`
`mv
`
`Apple Exhibit 1012
`Page 7 of 29
`
`Apple Exhibit 1012
`Page 7 of 29
`
`
`
`
`
`
`
`
`
`
`
`

`
`US. Patent
`
`Dec. 5, 1995
`
`Sheet 7 of 18
`
`5,473,380
`
`
`
`2:6 I! llll |- 111:!!! $30
`
`
`
`woos wZBEmE 3mm
`
`
`
`
`
`21> HHHHHHHM. |||||||l.1||||||l. 72> 21> ||||||||||||||| I Z:
`
`
`
`
`
`
`
`Z: I: Z; |||||||||||||||| 1:;
`
`
`
`HHMHHHH, Hnnnnmn 2:5 ||||||||||||||| E718
`
`
`
`
`
`
`
`Emu-m _
`
`Apple Exhibit 1012
`Page 8 of 29
`
`

`
`US. Patent
`
`Dec. 5, 1995
`
`Sheet 8 of 18
`
`5,473,380
`
`21>
`
`Z;
`Q:
`
`.llllnllllllllll
`
`[l‘lllll'll
`
`.Illllllll
`
`5;
`
`3:
`3:
`
`I:
`
`2:5
`
`$30
`H36
`
`$30
`
`
`
`MODE PUO DAME
`
`
`
`MODE FOO N572“?
`
`QUE
`
`Apple Exhibit 1012
`Page 9 of 29
`
`

`
`U.S. Patent
`
`5,473,380
`
`.::om_oNezaoomaom
`
`
`
`HrIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIH
`
`_
`
`
`s_<mEm_.H,HH._...__%H3c_-mm.._..5me.:z<:oczaoomo.Dmm=.5EHmmmm>z_3>Hzo_Emomm
`
`nmo:ooEmmHH:¢._mz_
`.5._SoHvm_HmmLHrIIIIInNHmHm._<ommHHzo_»<~Fz<:oH
`
`
`
`0<:_u_zoFo_nmE1HHHH.mHm_..=:o_._$.25...HHHo._m:ms_<EHH.HwHzo:.oauE H_HH.HH_HHI..u<._u_Boo._u_...m_2<E4:HH_mczHHlllllllllllllInHHBHxuqmxmHHH.
`
`
`
`
`
` zo_5amEHHH.1mw¢o%.H_,HHHwHn:<2Eon_om<_sxo<mH_HHnnnnnnnnnnHHH_H%HH_HHmoo:J_HHHHHHHHHzo=o_omEHH_HHH_HHH_H__H_H:.._oEoHHHHHH5Hzo:<mzm._s_ooHHHHHmHzo:.os_HHHHHH
`
`
`Apple Exhibit 1012
`Page 10 of 29
`
`Apple Exhibit 1012
`Page 10 of 29
`
`
`

`
`US. Patent
`
`Dec. s, 1995
`
`Sheet 10 of 18
`
`5,473,380
`
`SNR +
`
`PPPPPP PPPPPP
`
`TIME
`
`Apple Exhibit 1012
`Page 11 of 29
`
`

`
`U.S. Patent
`
`Dec. 5, 1995
`
`Sheet 11 of 18
`
`5,473,380
`
`ON
`
`pamhao
`
`0._<zo.m
`
`_._...0_n_
`
`MONEE-
`
`N\N
`
`._.Dn_._.DO
`
`2._<zo_m
`
`EN
`
`N\N
`
`IN
`
`.0O._<Z<
`
`omn_>
`
`._<zo_m
`
`NONm.4N
`
`._<:a_o
`
`omn_>
`
`._<zc_w
`
`HI.
`
`‘ONIN
`
`5&2.
`
`m._<zw_m
`
`._<z.o_m
`
`omo_>
`
`co._<z<
`OON
`
`Apple Exhibit 1012
`Page 12 of 29
`
`Apple Exhibit 1012
`Page 12 of 29
`
`
`
`
`
`

`
`US. Patent
`
`Dec. 5, 1995
`
`Sheet 12 of 18
`
`5,473,380
`
`sun‘
`
`PPPPPP PPPPPP
`
`OUTPUT
`SIGNAL b
`
`B
`IBBBBBB| BBBBBBB
`
`PPPPPP PPPPPP
`
`OUTPUT
`SIGNAL 6
`
`BBBBBBBBBBBBBB
`
`TIME
`
`FIG.12
`
`Apple Exhibit 1012
`Page 13 of 29
`
`

`
`U
`
`a
`
`0
`
`5,473,380
`
`.owns3wo._<z<
`
`P5mH._<z.o_m.
`
`t§_>....um.._<.:c_a"uwmum__mcom
`
`_
`
`mo<uEm»z_3an:-.5
`§_<mEmtn.
`
`NB
`
`
`
`
`mo<uEm»z_—mo<uEm:z_M.M.2<.mw__,_»%_m_m.._<.:_o_n.=_E.o_a
`
`%ommmmmasooD1Now,
`5,an32m._<:.c_n,D:m
`
`m%fi_%__»_z_am
`
`B1W...%smammanm.mom1
`
`._.Dn_._.DO
`
`o._<z.o_m
`
`26.".
`
`Apple Exhibit 1012
`Page 14 of 29
`
`Apple Exhibit 1012
`Page 14 of 29
`
`
`
`

`
`SU
`
`
`\O\.oo._<z<P:momo_>
`
`unim._<z.u_mm§u__>mMn9M9_2<mEm
`.._<.:.o_oOS09
`
`TE
`
`,.\.czaoomnczaoo.
`
`fifimmm3.§._%\\3.Momn.__»=_mw_..m<z<HN28mnmxm._._=.S_2%\:o:2555N9
`mm..__.ww_n_
`
`
`
`
`
`
`&m.__.,_._.?
`
`81
`
`5,473,380
`
`
`
`m‘._oo._<z<T
`
`mO\
`
`mo‘
`
`E
`
`v_..G_u_
`
`In.1:
`
`mE.Fo_._Smat fo._<zo_mo._.:n_...:OM.._<z_u_maomo_>a3<z<....Izo:<mEmwmomxm._._:..5s_
`
`
`
`
`
`
`NEDFOE:_<Z.0_w
`
`Apple Exhibit 1012
`Page 15 of 29
`
`Apple Exhibit 1012
`Page 15 of 29
`
`
`

`
`US. Patent
`
`Dec. 5, 1995
`
`Sheet 15 of 18
`
`5,473,380
`
`SNR
`
`PPPPPP PPPPPPOUTPUT
`I
`SlGNALb
`
`I
`
`PPPPPP PPPPPP
`
`BBBBBBB BBBBBIBOUTPUT
`B
`SIGNAL c
`
`TIME
`
`FIG.15
`
`Apple Exhibit 1012
`Page 16 of 29
`
`

`
`U.S. Patent
`
`Dec. 5, 1995
`
`Sheet 16 of 18
`
`5,473,380
`
`\®‘
`
`D3
`
`mE:.o_._
`
`mat.
`
`._<z.o_m
`
`§<mE.m:3«.9
`
`._.:n_._.3O
`
`..._<zo_m
`
`._<.._.o_o
`
`mo<nEmpz_
`
`._<.—.__U_D
`
`omo_>
`
`._<z.o_m
`
`mED._.O_n_
`
`NV‘
`
`0O._<2<
`
`omo_>
`
`._<z.o_m
`
`._.3nE.3O
`
`o._<zo.m
`
`m;
`
`._<._._G_D
`
`omo_>
`
`._<zc_w
`
`._<.—._0_D
`
`mo<"EE.z_
`
`2.0.".
`
`m9
`
`zo_.::.._<._mm
`
`._<za_m
`
`
`
`ommmmmmsoomm_«um._om_mz_
`
`
`
`omn_>..<.:o_o
`
`wzaoomomi%
`
`omxm._.__H5=
`
`ma>»
`
`
`
`$0..mm:._.o_..
`
`mm.
`
`._<:.o_a
`
`omo_>
`
`._<z.o_m
`
`‘V9
`
`._<:¢_o
`
`ND.-62.000Z1mo<uEm»z_
`mmaozo
`
`mn_>._.
`
`.9
`
`a._<zo_m
`
`Ow;
`
`mi
`
`Apple Exhibit 1012
`Page 17 of 29
`
`.oo._<z<
`
`omo_>
`
`ommmmmmzoo
`
`._<z.o_w
`
`:1
`
`
`§_<mm.5to.N3
`
`xtlé._<z0_m.
`...:._z_
`
`Apple Exhibit 1012
`Page 17 of 29
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`

`
`US. Patent
`
`Dec. 5, 1995
`
`Sheet 17 of 18
`
`5,473,380
`
`
`
`Eosms msEE N. P. w
`
`
`
`WEE. mmEoE éfOk
`
`Apple Exhibit 1012
`Page 18 of 29
`
`

`
`U.S. Patent
`
`Dec. 5, 1995
`
`Sheet 18 of 18
`
`5,473,380
`
`uzaoumaom
`:=om_o\
`
`
`
`IIIIIIIIIII|'\4l.|]n'|I-I||l'|l||||'lI|lI|.’||l||IIlI'll|I
`
`._.Dn:.:O
`
`amoanomnmm
`
`mm=..._.U_n_
`
`mmvmrIIIIIII
`
`-:z<:c
`
`zo:<~
`
`m._<om
`
`m._m<_m<>
`
`IhO2m4
`
`wzaoomo
`
`\
`
`m
`
`
`
`oz_>_momm_2<mm_._.m
`
`
`
`mmuuam..
`
`._.Dn_2_
`
`.:m
`
`Dm<>>mOu_
`
`m>:o_nmm..
`
`mmD._.0_n_
`
`om<2o_o<m
`
`m>:o_nmE
`
`
`
`mE:.o_..__
`
`v_z<m
`
`.uz__._oEsm
`
`ms_<E
`
`>mos_m_2
`
`m>_5_omEn_
`
`mo.5m>
`
`o._m_n_ms_<E
`
`
`
`G<._..._u>_»o_omE
`
`o._m_n=m_2<E
`
`0<;_n_poo
`
`m_..5»o_n_
`
`Apple Exhibit 1012
`Page 19 of 29
`
`Apple Exhibit 1012
`Page 19 of 29
`
`
`
`
`
`
`
`
`
`
`
`

`
`5,473,380
`
`1
`PICTURE SIGNAL TRANSMITTING
`METHOD AND APPARATUS
`
`BACKGROUND OF THE INVENTION
`
`2
`ing P-pictures, while FIG. 2B shows the reference pictures
`used for encoding B-pictures.
`As shown in FIGS. 3A and 3B, there are four methods for
`encoding the macro-blocks (discussed below) of a picture.
`When multiple methods are suitable, the method which will
`give the smallest amount of encoded data is employed on a
`macro-block by macro-block basis. Blocks F1 to F5 in FIG.
`3A represent data for frames of moving picture signals,
`whereas blocks FlX to FSX in FIG. 38 represent data for
`encoded frames. The solid line arrows in FIG. 3A show the
`frames to which motion vectors x1 .
`.
`. x6 relate.
`The ?rst method, shown as SP1, is to not use predictive
`encoding, that is, to use only intra-frarne correlation. This is
`suitable for any macro-blocks of an I-picture a P-picture and
`a B-picture. In other words, if less encoded data is produced
`without predictive encoding, then this method is selected.
`The second method, shown as SP2, is to predictively
`encode relative to a picture which temporally succeeds the
`current picture, referred to as backward predictive encoding.
`The third method, shown as SP3, is to predictively encode
`relative to a picture which temporally precedes the current
`picture, referred to as forward predictive encoding. The
`second method is suitable for macro-blocks of only B-pic
`tures. The third method is suitable for macro-blocks of
`P-pictures and B-pictures.
`The fourth method, shown as SP4, is to predictively
`encode relative to the mean value of two pictures, one
`temporally preceding and one temporally succeeding the
`current picture. This method is suitable for macro-blocks of
`only B-pictures.
`The encoding sequence will now be described.
`The ?rst frame F1 is encoded as an I-picture using the ?rst
`method SP1 so that it is directly transmitted over a trans
`mission channel as encoded data FlX.
`The third frame F3 is encoded as a P-picture. When the
`third method SP3, forward predictive coding, is used for a
`macro-block, difference signals from the temporally preced
`ing frame F1 used as the reference picture, as indicated by
`a broken-line arrow SP3, and a motion vector x3 between
`the reference picture F1 and the current picture F3, are
`calculated and encoded as data F3X for that macro-block.
`Alternatively, in this or another macro~block of the P picture,
`if a smaller amount of encoded data is produced for a
`macro-block of the P picture being encoded, the ?rst method
`SP1 can be used wherein the data of the original frame F3
`are directly utilized as the transmission data F3X for that
`macro-block.
`The second frame F2 is encoded as a B-picture.
`When the fourth method SP4 is used to encode a macro
`block of the frame F2, a difference between the mean value
`of the temporally preceding frame F1 and the temporally
`succeeding frame F3 is calculated, on a pixel by pixel basis.
`The difference data and the motion vectors x1 and x2 are
`encoded as data F2X.
`When the ?rst processing method SP1 is used to encode
`a macro-block of the frame F2, the data of the original frame
`F2 forms the encoded data F2X.
`When one of the second or third methods SP2, SP3 is used
`to encode a macro-block of the frame F2, one of the
`difference between the temporally succeeding frame F3 and
`the current frame F2, and the difference between the tem
`porally preceding frame F1 and the current frame F2 is
`calculated. The difference data and one of the motion vectors
`x1, x2 are encoded as the data F2X.
`The frame F4 for the B-picture and the frame F5 for the
`
`The present invention relates to coding and decoding of a
`picture signal for transmission, and, more particularly, is
`directed to matching the type of predictive coding applied to
`pictures of the picture signal.
`In, for example, a teleconferencing system or a video
`telephone system, moving picture signals are compressed
`and encoded by taking advantage of intra-frame and inter
`frame correlation so that they can be more e?iciently trans
`mitted over a communication channel to a remote location.
`Intra-frame correlation can be utilized by an orthogonal
`transformation, such as a discrete cosine transformation
`(DCT).
`Inter-frame correlation can be utilized by predictive
`encoding between successive pictures. As used herein, a
`picture generally refers to an image represented by a frame.
`When the ?elds of a frame are coded in a non~interlaced
`manner, that is, separately, each ?eld may be referred'to as
`a picture.
`As shown in FIG. 1A, for example, frame pictures PCI,
`PC2 and PC3 are generated at time points t1, t2 and t3. As
`shown by shading in FIG. 1B, the difference between the
`frame pictures PCl and PC2 is obtained as diiTerence picture
`data PC12, and the difference between the frame pictures
`PC2 and PC3 is obtained as diiference picture data PC23.
`Since there is a fairly small change between signals of
`temporally neighboring frames, transmission of only the
`difference picture data utilizes the transmission channel
`more e?iciently than transmission of the original pictures.
`That is, using the difference picture data as encoded picture
`signals reduces the amount of data to be transmitted.
`However, if only the difference signals are transmitted,
`the original picture cannot be restored. It is advantageous to
`occasionally transmit a picture which is not predictively
`encoded as a reference for difference picture data, and
`because it is sometimes more e?icient than transmitting the
`picture as a predictively encoded picture.
`Pictures which are encoded utilizing only intra-frame
`correlation and not inter-frame correlation, are referred to
`herein as intra-pictures or l-pictures.
`Pictures which are encoded with predictive encoding
`relative to one previously encoded picture are referred to
`herein as predictive pictures or P-pictures. The previously
`encoded picture may be an I-picture or a P-picture, and
`temporally succeeds the P-picture.
`Pictures which are encoded with predictive encoding
`relative to at most two pictures, a temporally preceding and
`a temporally succeeding picture, are referred to herein as
`bi-directionally predictive coded pictures or B-pictures. The
`two pictures may each be an I-picture or a P-picture. When
`both are used, the mean value of the two pictures is obtained
`and used as a reference picture for the picture to be encoded.
`A series of pictures may be considered as groups of
`pictures having a predetermined number of frames such as
`F1 .
`.
`. F17. The luminance and chrominance picture signals
`of the leading frame F1 are encoded as an I-picture, the
`picture signals of the second frame F2 are encoded as a
`B-picture, and the picture signals of the third frame F3 are
`encoded as a P-picture. The fourth and the following frames
`F4 to F17 are encoded alternately as B-pictures and P-pic
`tures. FIG. 2A shows the reference pictures used for encod
`
`l0
`
`15
`
`25
`
`35
`
`40
`
`45
`
`50
`
`55
`
`65
`
`Apple Exhibit 1012
`Page 20 of 29
`
`

`
`3
`
`4
`
`5,473,380
`
`P-picture are processed in a similar manner as described
`above to generate transmitted data F4X and F5X.
`FIG. 4 illustrates an arrangement for encoding and decod-
`ing moving picture signals in accordance with the above-
`described predictive encoding scheme. As shown in FIG. 4,
`an encoding device 1 encodes input picture signals and
`transmits the encoded signals to a recording medium 3 as a
`transmission channel for recording. A decoding device 2
`reproduces the signals recorded on the recording medium 3
`and decodes these as output signals.
`The encoding device 1 includes an input terminal 10, a
`pre-processing circuit 11, A/D converters 12 and 13, a frame
`memory 14 including a lurrrinance signal frame memory 15
`and a color difference signal frame memory 16, a format
`converting circuit 17 and an encoder 18.
`Input terminal 10 is adapted to receive a video signal VD
`and to supply the signal VD to pre-processing circuit 11
`which functions to separate the video signal VD into lurni—
`nance signals and color signals, herein chrominance or color
`difference signals, that are applied to analog-to-digital (A/D)
`converters 12 and 13, respectively. The video signals, digi-
`tized by analog-to-digital conversion by the A/D converters
`12 and 13, are supplied to frame memory 14 having memo-
`ries 15, 16 which function to store the luminance signals and
`the color difference signals, respectively, and to read out the
`signals stored therein to format converting circuit 17.
`The converter 17 is operative to convert frame format
`signals stored in the frame memory section 14 into block
`format signals. As shown in FIG. SA, pictures are stored in
`the frame memory section 14 as frarne-forrnat data having V
`lines each consisting of H dots. The converting circuit 17
`divides each frame into N slices, each slice comprising a
`multiple of 16 lines. As shown, in FIG. 5B, the converter 17
`divides each slice into M macro-blocks. As shown in FIG.
`5C, each macro-block represents luminance signals Y cor-
`responding to l6><16 pixels or dots, and associated chrorni-
`nance Cr, Cb signals. These luminance signals are sub-
`divided into blocks Y1 to Y4, each consisting of 8x8 dots.
`The 16x16 dot luminance signals are associated with 8x8
`dot Cb signals and 8x8 dot Cr signals. The converter 17 is
`also operative to supply the block format signals to the
`encoder 18, which is described in detail below with refer-
`ence to FIG. 6.
`
`The encoder 18 operates to encode the block format
`signals and to supply the encoded signals as a bitstrearn over
`a transmission channel for recording on the recording
`medium 3.
`
`The decoding device 2 includes a decoder 31, a format
`converting circuit 32, a frame memory section 33 including
`a luminance signal frame memory 34 and a color difference
`signal frame memory 35, digital-to-analog converters 36 and
`37, a post-processing circuit 38 and an output terminal 30.
`The decoder 31 is operative to reproduce encoded data
`from the recording medium 3 and to decode the encoded
`data, as described in detail below with reference to FIG. 9,
`and to supply decoded data signals to format converting
`circuit 32 which is operative to convert the decoded data
`signals into frame format data signals and to supply the
`frame format data signals as luminance signals and color
`difference signals to the memory 33. The memories 34, 35
`of the memory 33 function to store the luminance and
`chrominance signals, respectively, and to apply these signals
`to D/A converters 36 and 37, respectively. The analog
`signals from converters 36, 37 are synthesized by a post-
`processing circuit 38 which functions to form output picture
`signals and to output them to output terminal 30, and thence
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`to a display unit, such as a CRT, not shown, for display.
`FIG. 6 illustrates the encoder 18 shown in FIG. 4.
`
`Generally, the encoder 18 stores three pictures, the current
`picture and the pictures temporally preceding and succeed-
`ing the current picture. Based on the sequential position of
`the current picture in the group of pictures, the picture
`coding type (I, P or B) is selected for each picture. The
`picture type sequence is determined by a user using picture
`type input device 65, independent of the pictures applied to
`an input terminal 49.
`The encoder 18 also chooses one of frarne-based and
`field-based predictive encoding as will be explained with
`reference to FIG. 7, and further chooses one of frarne-based
`and field-based DCT encoding as will be explained with
`reference to FIG. 8. For each picture, appropriate motion
`vectors are obtained and the picture is predictively encoded
`relative to zero, one or two previously encoded pictures
`which have been locally decoded and which are referred to
`as reference pictures to form a difference data signal. The
`dilference data signal
`is orthogonally transformed into
`blocks of coefficient data which are quantized, variable
`length encoded and transmitted as encoded data.
`At the encoder 18, the quantized data are dequantized,
`inverse orthogonally transformed, and stored as the refer-
`ence pictures. The predictive encoding applies the motion
`vector(s) obtained for the current picture to the reference
`picture(s) to produce a prediction picture which is subtracted
`from the current picture to yield the difference data.
`The elements of the encoder 18 will now be explained in
`detail.
`
`Picture data for encoding is supplied macro-block by
`macro-block to the input terminal 49 and thence to a motion
`vector detection circuit 50 which is operative to process the
`picture data of respective frames as I-pictures, P-pictures or
`as B-pictures, in accordance with a predetermined sequence
`for each group of pictures, as shown for example, in FIGS.
`2A, 2B. The circuit 50 applies the picture data of the current
`frame to a frame memory 51 having frame memories 51a,
`51b, 51c used for storing a temporally preceding picture, the
`current picture and a temporally succeeding picture, respec-
`tively.
`More specifically, the frames F1, F2, F3 are stored in the
`memories 51a, 51b, 51c, respectively. Then the picture
`stored in memory 51c is transferred to memory 51a. The
`frames F4, F5 are stored in the memories 51b, 51c, respec-
`tively. The operations of transferring the picture in memory
`51c to memory 51a and storing the next two pictures in
`memories 51b, 51c are repeated for the remaining pictures
`in the group of pictures.
`After the pictures are read into the memory and tempo-
`rarily stored, they are read out and supplied to a prediction
`mode changeover circuit 52 which is adapted to process the
`current picture for one of frame based and field based
`predictive encoding. After processing the first frame picture
`data in a group of pictures as an I-picture and before
`processing the second frame picture as a B-picture, the
`motion vector detection circuit 50 processes the third frame
`P-picture. The processing sequence is different from the
`sequence in which the pictures are supplied because the
`B-picture may involve backward prediction, so subsequent
`decoding may require that the P-picture temporally succeed-
`ing the B-picture have been previously decoded.
`The motion vector detection circuit 50 calculates as an
`estimated value for intra-coding for each macro-block, the
`sum of absolute values of prediction errors for the frame
`prediction mode for each macro-block and the sum of
`
`Apple Exhibit 1012
`Page 21 of 29
`
`Apple Exhibit 1012
`Page 21 of 29
`
`

`
`5,473,380
`
`5
`
`absolute values of prediction errors for the field prediction
`mode for each macro-block and supplies these sums to the
`prediction decision circuit 54 which compares these sums
`and selects frame prediction mode or field prediction mode
`in accordance with the smallest of these values and provides
`the selected mode to the prediction mode changeover circuit
`52.
`
`If the frame prediction mode is selected, the prediction
`mode changeover circuit 52 outputs the four luminance
`blocks Y1 to Y4 and the two chrominance or color difference
`blocks Cb, Cr of each macro-block received from the motion
`vector detection circuit 50 without processing. As shown in
`FIG. 7A, odd or first field line data, indicated by solid lines,
`and even or second field line data, indicated by dashed lines,
`alternate in each luminance and color difference block as
`received from the motion vector detection circuit 50. In FIG.
`
`7A, a indicates units for motion compensation. In the frame
`prediction mode, motion compensation is performed with
`four luminance blocks (macro-blocks) as a unit and a single
`motion vector is associated with the four luminance blocks
`Y1 to Y4.
`
`If the field prediction mode is selected, the prediction
`mode changeover circuit 52 processes the signals received
`from the motion vector detection circuit 50 so that each of
`
`the four luminance blocks comprises data from a single field
`and the two color difference blocks have non-interlaced odd
`and even field data. Specifically, as shown in FIG. 7B, the
`luminance blocks Y1 and Y2 have odd-field data and the
`luminance blocks Y3 and Y4 have even-field data, while the
`upper halves of the color difference blocks Cb, Cr represent
`odd field color diiference data for the luminance blocks Y1
`and Y2 and the lower halves of the color difference blocks
`
`Cb, Cr represent even field color difference data for the
`luminance blocks Y3 and Y4. In FIG. 7B, b indicates units
`for motion compensation. In the field prediction mode,
`motion compensation is performed separately for the odd-
`field blocks and even-field blocks so that one motion vector
`is associated with the two luminance blocks Y1 and Y2 and
`another motion vector is associated with the two luminance
`blocks Y3 and Y4.
`
`The prediction mode changeover circuit 52 supplies the
`current picture, as processed for frame based or field based
`predictive encoding, to arithmetic unit 53 of FIG. 6. The
`arithmetic unit 53‘ functions to perform one of intra-picture
`prediction, forward prediction, backward prediction or bi-
`directional prediction. A prediction decision circuit 54 is
`adapted to select the best type of prediction in dependence
`upon the prediction error signals associated with the current
`picture signals.
`The motion vector detection circuit 50 calculates, for the
`current picture, the sum of absolute values of the diiferences
`between each Aij and the average value of the Aij in each
`macro-block 2lAij—(average of Aij)! and supplies the sum as
`an estimated value for intra-coding to the prediction decision
`circuit 54.
`The motion vector detection circuit 50 calculates the sum
`of absolute values (or sum of squares) of the dilference (Aij
`—Bij) between signals Aij of the macro-blocks of the current
`picture, and signals Bij of the macro-blocks of the prediction
`picture )3|Aij—Bij l in each of frame prediction mode and field
`prediction mode. As explained above, the motion vector(s)
`for the current picture are applied to the reference picture(s)
`to generate the prediction picture. When the reference pic-
`ture temporally precedes the current picture, the quantity
`2|Aij— Bij! is referred to as a forward prediction error signal,
`and when the reference picture temporally succeeds the
`
`6
`current picture, the quantity 2‘.IAij—Bij| is referred to as a
`backward prediction error signal. When the prediction pic-
`ture is the mean of a temporally preceding and a temporally
`succeeding reference picture, as motion-compensated, the
`quantity 2lAij—Bij| is referred to as a bi-directional predic-
`tion error signal.
`The circuit 50 supplies the forward frame prediction, the
`forward field prediction, the backward frame prediction, the
`backward field prediction, the bi-directional frame predic-
`tion and the bi-directional field prediction error signals to the
`prediction decision circuit 54.
`The prediction decision circuit 54 selects one of intra-
`coding, forward inter-picture prediction, backward inter-
`picture prediction or bi-directional inter-picture prediction
`and one of frame and field prediction mode in accordance
`with the smallest of the estimated value for intra-coding and
`the forward frame, the forward field, the backward frame,
`the backward field,
`the bi-directional frame and the bi-
`directional field prediction error signals. The arithmetic unit
`53 predictively encodes the current picture, as processed by
`the frame or field changeover circuit 52, in accordance with
`the prediction mode selected by the prediction decision
`circuit 54.
`The motion vector detection circuit 50 serves to calculate
`
`and supply the motion vector(s) associated with the selected
`prediction mode to a variable length encoding circuit 58 and
`a motion compensation circuit 64, explained later.
`The sums of the absolute values of the inter-frarne dif-
`
`ferences (prediction errors) on the macro-block basis are
`supplied from the motion vector detection circuit 50 to the
`prediction mode changeover circuit 52 and to the prediction
`decision circuit 54, in the manner as described above.
`The arithmetic unit 53 supplies predictively encoded data,
`also referred to as difference data, for the current picture to
`a DCT mode changeover circuit 55 which is adapted to
`process the current picture for one of frame based and field
`based orthogonal transformation.
`The DCT changeover circuit 55 functions to compare the
`encoding efliciency when the DCT operations for the macro-
`blocks in a picture are performed with the odd field data
`alternating with the even field data, that is, for frame based
`orthogonal transformation, as shown in FIG. 8A, with the
`encoding efliciency when the DCT operations for the macro-
`blocks in a picture are performed with the odd field data
`separated from the even field data, that is, for field based
`orthogonal transformation, as shown in FIG. 8B. The circuit
`55 functions to select the mode with the higher encoding
`efliciency.
`To evaluate the encoding efiiciency for frame based
`orthogonal transfonnation, the DCT mode changeover cir-
`cuit 55 places the luminance macro-block data into inter-
`laced form, as shown in FIG. 8A, and calculates the differ-
`ences between the odd field line signals and even field line
`signals vertically adjacent to each other, and finds the sum
`of absolute values of the diiferences EFM, or the sum of
`squared values of the differences.
`
`10
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`16 15
`EFM= _2‘.
`_}L‘
`1:1 1:1
`
`|o(i,j)— e(i,j)l +
`
`Eq. 1
`
`l6 l5
`_2 _2 le(i,j)—o(i+ l,j)l
`]=1t=l
`
`65
`
`To evaluate the encoding efficiency for field based
`orthogonal transformation, the DCT mode changeover cir-
`
`Apple Exhibit 1012
`Page 22 of 29
`
`Apple Exhibit 1012
`Page 22 of 29
`
`

`
`5,473,380
`
`7
`cuit 55 places the luminance macro-block data into non-
`interlaced form, as shown in FIG. 8B, and calculates the
`differences between vertically adjacent odd field line signals
`and the differences between vertically adjacent even field
`line signals, and finds the sum of absolute values of the
`differences EFD, or the sum of squared values of the
`differences.
`
`16 15
`EFD= .2 _Z (|o(i,]) — o(i+ l,j)l + |e(i,j) — e(i+ l,})l)
`1:1 t=1
`
`Eq.2
`
`l0
`
`The DCT changeover circuit 55 compares the difference
`between the frame based and field based sums of the
`absolute values with a predetermined threshold and selects
`frame based DCT transformation if the difference EFM —
`EFD is less than the predetermined threshold.
`If the frame prediction mode is selected in the prediction
`mode changeover circuit 52, the probability is high that the
`frame DCT mode will be selected in the DCT mode
`
`changeover circuit 55. If the field prediction mode is
`selected in the prediction mode changeover circuit 52, the
`probability is high that the field DCT mode will be selected
`in the DCT mode changeover circuit 55. However, since this
`is not necessarily the case, the prediction mode changeover
`circuit 52 sets the mode which will give the least value of the
`sum of the absolute values of prediction errors, while the
`DCT mode changeover circuit 55 sets the mode which will
`give the optimum orthogonal transformation encoding efli-
`ciency.
`transformation mode, also
`If frame based orthogonal
`referred to as frame DCT mode, is selected, the DCT mode
`changeover circuit 55 functions to ensure that the four
`luminance blocks Y1 to Y4 and two color difference blocks
`Cb, Cr represent alternating or interlaced odd and even field
`lines, as shown in FIG. 8A.
`
`transformation mode, also
`If field based orthogonal
`referred to as field DCT mode, is selected, the DCT mode
`changeover circuit 55 functions to ensure that each of the
`luminance blocks represents only one field, and that each of
`the color difference blocks has segregated or non-interlaced
`odd and even field lines, as shown in FIG. 8B.
`
`The DCT mode changeover circuit 55 functions to output
`the data having the configuration associated with the
`selected DCT mode, and to output a DCT flag indicating the
`selected DCT mode to the variable length encoding circuit
`58 and the motion compensation circuit 64.
`The DCT mode changeover circuit 55 supplies appropri-
`ately configured diiference picture data to a DCT circuit 56
`shown in FIG. 6 which is operative to orthogonally trans-
`form it using a discrete cosine transformation into DCT
`coeflicients, and to supply the DCT coeflicient data to a
`quantization circuit 57 that functions to quantize the

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket