`
`SAMSUNG EXHIBIT 1007
`Samsung v. Image Processing Techs.
`
`
`
`FOR THE PURPOSES OF INFORMATION ONLY
`
`Codes used to identify States party to the PCT on the front pages of pamphlets publishing international applications under the PCT.
`
`Zimbabwe
`
`Albania
`Armenia
`Austria
`Australia
`Azerbaijan
`Bosnia and Herzegovina
`Barbados
`Belgium
`Burkina Faso
`Bulgaria
`Benin
`Brazil
`Belarus
`Canada
`Central African Republic
`Congo
`Switzerland
`Céte d’Ivoire
`Cameroon
`China
`Cuba
`Czech Republic
`Germany
`Denmark
`Estonia
`
`SI
`SK
`SN
`SZ
`TD
`TG
`TJ
`
`Slovenia
`Slovakia
`Senegal
`Swaziland
`Chad
`Togo
`Tajikistan
`Turkmenistan
`Turkey
`Trinidad and Tobago
`Ukraine
`Uganda
`United States of America
`Uzbekistan
`Viet Nam
`Yugoslavia
`
`™T
`
`R
`TT
`UA
`UG
`US
`UZ
`VN
`YU
`ZW
`
`ES
`FI
`FR
`GA
`GB
`GE
`GH
`GN
`GR
`HU
`TE
`iL
`Is
`It
`JP
`KE
`KG
`KP
`
`KR
`KZ
`Lc
`LI
`LK
`LR
`
`Spain
`Finland
`France
`Gabon
`United Kingdom
`Georgia
`Ghana
`Guinea
`Greece
`Hungary
`Treland
`Tsrael
`Iceland
`Italy
`Japan
`Kenya
`Kyrgyzstan
`Democratic People’s
`Republic of Korea
`Republic of Korea
`Kazakstan
`Saint Lucia
`Liechtenstein
`Sri Lanka
`Liberia
`
`LS
`LT
`LU
`LV
`MC
`MD
`MG
`MK
`
`ML
`MN
`MR
`MW
`Mx
`NE
`NL
`NO
`NZ
`PL
`PT
`RO
`RU
`SD
`SE
`SG
`
`Lesotho
`Lithuania
`Luxembourg
`Latvia
`Monaco
`Republic of Moldova
`Madagascar
`The former Yugoslav
`Republic of Macedonia
`Mali
`Mongolia
`Mauritania
`Malawi
`Mexico
`Niger
`Netherlands
`Norway
`New Zealand
`Poland
`Portugal
`Romania
`Russian Federation
`Sudan
`Sweden
`Singapore
`
`SAMSUNG EXHIBIT 1007
`Page 2 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 2 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`DESCRIPTION
`
`SYSTEM FOR HUMAN FACE TRACKING
`
`BACKGROUND OF THE INVENTION
`
`The present invention relates to a system for
`locating a human face within an image, and more
`particularly to a system suitable for real-time tracking
`of a human face in video sequences.
`
`Numerous systems have been developed for the
`detection of a target with an input image.
`In
`particular, human face detection within an image is of
`considerable importance. Numerous devices benefit from
`automatic determination of whether an image (or video
`frame) contains a human face, and if so where the human
`face is in the image.
`Such devices may be, for example,
`a video phone or a human computer interface.
`A human
`computer interface identifies the location of a face, if
`any,
`identifies the particular face, and understands
`facial expressions and gestures.
`
`Traditionally,
`face detection has been
`performed using correlation template based techniques
`which compute similarity measurements between a fixed
`target pattern and multiple candidate image locations.
`If any of the similarity measurements exceed a threshold
`value then a "match" is declared indicating that a face
`has been detected and its location thereof. Multiple
`correlation templates may be employed to detect major
`facial sub-features.
`A related technique is known as
`"view-based eigen-spaces," and defines a distance metric
`based on a parameterizable sub-space of the original
`image vector space.
`If the distance metric is below a
`
`threshold value then the system indicates that a face has
`
`been detected.
`
`An alternative face detection technique
`involves using spatial image invariants which rely on
`compiling a set of image invariants particular to facial
`images.
`The input image is then scanned for positive
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`SAMSUNG EXHIBIT 1007
`Page 3 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 3 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`occurrences of these invariants at all possible locations
`to identify human faces.
`
`in a paper entitled A Real-Time
`Yang et al.
`Face Tracker discloses a real-time face tracking system.
`The system acquires a red-green-blue (RGB)
`image and
`filters it to obtain chromatic colors (r and g) known as
`"pure" colors,
`in the absence of brightness.
`The
`transformation of red-green-blue to chromatic colors is a
`transformation from a three dimensional space (RGB)
`to a
`two dimensional space (rg).
`The distribution of facial
`colors within the chromatic color space is primarily
`clustered in a small region. Yang et al. determined
`after a detailed analysis of skin-color distributions
`
`that the skin color of different people under different
`lighting conditions in the chromatic color space have
`similar Guassian distributions.
`To determine whether a
`particular red-green-blue pixel maps onto the region of
`the chromatic color space indicative of a facial color,
`Yang et al.
`teaches the use of a two-dimensional Guassian
`
`10
`
`15
`
`20
`
`model. Based on the results of the two-dimensional
`
`the
`Guassian model for each pixel within the RGB image,
`facial region of the image is determined. Unfortunately,
`the two-dimensional Guassian model is computationally
`intensive and thus unsuitable for inexpensive real-time
`
`the system taught by Yang et al. uses
`systems. Moreover,
`a simple tracking mechanism which results in the position
`of the tracked face being susceptible to jittering.
`Eleftheriadis et al.,
`in a paper entitled
`"Automatic Face Location Detection and Tracking for
`Model-Assisted Coding of Video Teleconferencing Sequences
`at Low Bit-Rate," teaches a system for face location
`detection and tracking.
`The system is particularly
`designed for video data that includes head-and-shoulder
`sequences of people which are modeled as elliptical
`regions of interest.
`The system presumes that the
`outline of people’s heads are generally elliptical and
`have high temporal correlation from frame to frame.
`
`2
`
`25
`
`' 30
`
`35
`
`SAMSUNG EXHIBIT1007
`Page 4 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 4 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`the system calculates the
`Based on this premise,
`difference between consecutive frames and thresholds the
`
`result to identify regions of significant movement, which
`are indicated as non-zero. Elliptical non-zero regions
`are located and identified as facial regions.
`Unfortunately,
`the system taught by Eleftheriadis et al.
`is computationally intensive and is not suitable for
`real-time applications. Moreover, shadows or partial
`occlusions of the person’s face results in non-zero
`regions that are not elliptical and therefore the system
`may fail to identify such regions as a face.
`In
`addition, if the orientation of the person’s face is away
`from the camera then the resulting outline of the
`
`person’s head will not be elliptical and therefore the
`system may fail to identify the person’s head. Also, if
`there is substantial movement within the background of
`the image the facial region may be obscured.
`Hager et al.
`in a paper entitled, Real-Time
`Tracking of Image Regions with Changes in Geometry and
`Illumination, discloses a face tracking system that
`analyzes the brightness of an image within a window.
`pattern of the brightness within the window is used to
`track the face between frames. The system taught by
`Hager et al.
`is sensitive to face orientation changes and
`partial occlusions and shadows which obscure the pattern
`of the image.
`The system is incapable of initially
`determining the position of the face(s).
`What is desired,
`therefore, is a face tracking
`system that is insensitive to partial occlusions and
`shadows,
`insensitive to face orientation and/or scale
`changes,
`insensitive to changes in lighting conditions,
`easy to calibrate, and can determine the initial position
`of the facé(s).
`In addition,
`the system should be
`computationally simple so that it is suitable for
`real-time applications.
`
`The
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`SAMSUNG EXHIBIT1007
`Page 5 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 5 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`SUMMARY OF THE INVENTION
`
`The present invention overcomes the
`aforementioned drawbacks of the prior art by providing a
`system for detecting a face within an image that receives
`the image which includes a plurality of pixels, where a
`plurality of the pixels of the image is represented by
`respective groups of at least three values.
`The image is
`filtered by transforming a plurality of the respective
`groups of the at least three values to respective groups
`of less than three values, where the respective groups of
`the less than three values has less dependency on
`brightness than the respective groups of the at least
`
`three values. Regions of the image representative of
`skin-tones are determined based on the filtering.
`A
`first distribution of the regions of the image
`representative of the skin-tones in a first direction is
`calculated.
`A second distribution of the regions of the
`image representative of the skin-tones in a second
`
`direction is calculated, where the first direction and
`
`10
`
`15
`
`20
`
`the second direction are different.
`
`The face within the
`
`image is located based on the first distribution and the:
`
`second distribution.
`
`25
`
`30
`
`35
`
`Using a system that determines skin-tone
`
`regions based on a color representation with reduced
`brightness dependency together with first and second
`distributions permits the face tracking system to be
`insensitive to partial occlusions and shadows,
`insensitive to face orientation and/or scale changes,
`insensitive to changes in lighting conditions, and can
`determine the initial position of the face(s).
`In
`addition,
`the decomposition of the image using first and
`‘second distributions allows the system to be
`computationally simple so that it is suitable for real-
`
`time applications.
`
`In the preferred embodiment the estimated face
`
`location may also be used for tracking the face between
`frames of a video.
`For simplicity the face motion may be
`4
`
`SAMSUNG EXHIBIT 1007
`Page 6 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 6 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`modeled as a piece-wise constant two-dimensional
`translation within the image plane.
`A linear Kalman
`filter may be used to predict and correct the estimation
`of the two-dimensional translation velocity vector. The
`estimated (filtered) velocity may then also be used to
`determine the tracked positions of faces.
`
`The foregoing and other objectives, features,
`and advantages of the invention will be more readily
`understood upon consideration of the following detailed
`description of the invention,
`taken in conjunction with
`the accompanying drawings.
`
`BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS
`
`FIG. 1 is a block diagram of an exemplary
`embodiment of a face detection and tracking system of the
`present invention.
`
`FIG. 2 is a graph of the distributions of the
`
`skin-colors of different people in chromatic color space
`with the grey-scale reflecting the magnitude of the color
`
`10
`
`15
`
`20
`
`concentration.
`
`3 is a circle centered generally within
`FIG.
`the center of the distribution shown in FIG. 2.
`
`FIG. 4 is an image with a face.
`
`FIG. 5 is a binary image of the face of FIG. 4.
`FIG. 6 is a pair of histograms of the binary
`image of FIG. 5 together with medians and variances for
`each histogran.
`
`DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
`
`Referring to FIG. 1, a face detection and
`tracking system 6 includes an image acquisition device 8,
`such as a still camera or a video camera.
`A frame
`
`grabber 9 captures individual frames from the acquisition
`device 8 for face detection and tracking. An image
`processor 11 receives an image 10 from the frame grabber
`9 with each pixel represented by a red value, a green
`value, and a blue value, generally referred to as an RGB
`
`5
`
`25
`
`30
`
`35
`
`SAMSUNG EXHIBIT1007
`Page 7 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 7 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`image.
`
`The image 10 may alternatively be represented by
`
`other color formats, such as for example; cyan, magenta,
`and yellow;
`luminance,
`intensity, and chromaticity
`generally referred to as the YIQ color model; hue,
`saturation,
`intensity; hue,
`lightness, saturation; and
`
`hue, value, chroma. However,
`
`the RGB format is not
`
`necessarily the preferred color representation for
`characterizing skin-color.
`In the RGB color space the
`
`10
`
`three colors [R, G, B] represent not only the color but
`also its brightness.
`For example, if the corresponding
`elements of two pixels,
`[R1l, Gl, Bl] and [R2, G2, B2],
`are proportional (i.e., R1/R2=G1/G2=B1/B2)
`then they
`
`characterize the same color albeit at different
`
`The human visual system adapts to
`brightnesses.
`different brightness and various illumination sources
`
`15
`
`such that a perception of color constancy is maintained
`within a wide range of: environmental lighting conditions.
`Therefore it is desirable to reduce the brightness
`information from the color representation, while
`preserving accurate low dimensional color information.
`Since brightness is not important for characterizing skin
`colors under the normal lighting conditions,
`the image 10
`is transformed by a transformation 12 (filter) to the
`chromatic color space. Chromatic colors (r, g), known as
`"pure" colors in the absence of brightness, are generally
`defined by a normalization process:
`
`r=R/ (R+G+B)
`
`g=G/ (R+G+B)
`
`The effect of the transformation 12 is to map the three
`dimensional RGB image 10 to a two dimensional rg
`chromatic color space representation.
`The color blue is
`redundant after the normalization process because
`
`r+gt+b=1. Any suitable transformation 12 may be used
`which results in a color space where the dependence on
`brightness is reduced, especially in relation to the RGB
`
`color space.
`
`20
`
`25
`
`30
`
`35
`
`SAMSUNG EXHIBIT 1007
`Page 8 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 8 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`It has also been found that the distributions
`
`of the skin-colors of different people are clustered in
`chromatic color space, as shown in FIG. 2.
`The grey-
`
`scale in FIG.
`2 reflects the magnitude of the color
`concentration. Although skin colors of different people
`
`they differ much less
`appear to vary over a wide range,
`in color than in brightness.
`In other words,
`the skin-
`
`colors of different people are actually quite similar,
`
`while mainly differing in intensities.
`
`The two primary purposes of the transformation
`12 are to (1) facilitate distinguishing skin from other
`objects of an image, and (2)
`to detect skin tones
`irrespective of the particular color of the person’s skin
`which differs from person to person and differs for the
`same person under different lighting conditions.
`
`Accordingly, a suitable transformation 12 facilitates the
`
`ability to track the face(s) of an image equally well
`under different lightning conditions even for people with
`
`different ethnic backgrounds.
`the present inventor
`Referring to FIG. 3,
`determined that a straightforward characterization of the
`chromaticity distribution of the skin tones may be a
`circle 20 centered generally within the center of the
`
`distribution shown in FIG. 2. Alternatively, any
`suitable regular or irregular polygonal shape (including
`a circle) may be used, such as a square, a pentagon, a
`hexagon, etc.
`The use of a polygonal shape permits
`simple calibration of the system by adjusting the radius
`of the polygonal shape.
`The region encompassed by the
`polygonal shape therefore defines whether or not a
`particular pixel is a skin tone.
`In addition, it is
`computationally simple to determine whether or not a
`particular set of rg values is within the region defined
`by the polygonal shape.
`If the rg values are within the
`polygonal shape, otherwise referred to as the skin-tone
`region,
`then the corresponding pixel of the image 10 is
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`SAMSUNG EXHIBIT 1007
`Page 9 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 9 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`considered to be a facial feature, or otherwise having a
`skin tone.
`
`Based on whether each pixel of the image 10 is
`within the skin tone region the system generates a binary
`image 14 corresponding to the image 10.
`The binary image
`14 has a value of 1 for each pixel of the image 10 that
`is identified as a skin tone.
`In contrast,
`the binary
`image 14 has a value of 0 for each pixel of the image
`that is not identified as a skin tone.
`It is to be
`
`understood that groups of pixels may likewise be compared
`on a group by group basis,
`instead of a pixel by pixel
`basis, if desired.
`The result is a binary image 14 that
`contains primarily 1’s in those portions of the image 10
`that contain skin tones, such as the face, and primary
`0’s in the remaining portions of the image.
`It is noted
`that some portions of non-facial regions will have skin
`tone colors and therefore the binary image 14 will
`include a few 1’s at non-face locations.
`The opposite is
`also true, facial regions may include pixels that are
`indicative of non~skin tones and will therefore be
`
`Such regions may include beards,
`indicated by 0’s.
`moustaches, and hair.
`For example,
`the image 10 as shown
`in FIG.
`4 may be mapped to the binary image 14 as shown
`in FIG. 5.
`
`the representation of the 0’s
`Alternatively,
`and 1’s may be reversed, if desired. Moreover, any other
`suitable representation may be used to distinguish those
`portions that define skin-tones from those portions that
`do not define skin tones. Likewise,
`the results of the
`transformation 12 may result in weighted values that are
`indicative of the likelihood that a pixel (or region of
`pixels) are indicative of skin tones.
`As shown in FIG. 5,
`the facial region of the
`image is generally indicated by the primary grouping of
`1’s.
`The additional 1’s scattered throughout the binary
`image 14 do not indicate a facial feature, and are
`generally referred to as noise.
`In addition,
`the facial
`
`8
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`SAMSUNG EXHIBIT1007
`Page 10 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 10 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`region also includes some 0’s, generally referred to as
`noise.
`
`The present inventor came to the realization
`that the two dimensional binary image 14 of skin tones
`may further be decomposed into a pair of one dimensional
`models using a face locator 16. The reduction of the two
`dimensional representation to a pair of one dimensional
`representations reduces the computational requirements
`necessary to calculate the location of the face.
`Referring to FIG. 6,
`the mean of the distribution of the
`1’s (skin-tones) is calculated in both the x and y
`directions.
`The distribution is a histogram of the
`number of 1’s in each direction.
`The mean may be
`calculated by p=(1/N)Ux,.
`The approximate central
`location 38 of the face is determined by projecting the
`x-mean 30 and the y-mean 32 onto the binary image 14.
`The variance of the distribution in each of the x and y
`directions is also calculated. The variance may be
`calculated by o*=(1/N)2(x,-y)?. The variances 34a-34d
`indicate the width of the facial feature in its
`|
`
`respective directions. Projecting the variances 34a-34d
`onto the binary image 14 defines a rectangle around the
`facial region.
`The mean and variance are generally
`insensitive to variations for random distributions of
`
`10
`
`15
`
`20
`
`25
`
`noise.
`
`In other words,
`
`the mean and variance are robust
`
`for which such additional 1’s and 0’s are not
`
`statistically important. Under different lighting
`conditions for the same person and for different persons,
`the mean and variance technique defines the facial
`region. Moreover,
`the mean and variance are techniques
`merely requiring the summation of values which is
`computationally efficient.
`
`The system may alternatively use other suitable
`statistical techniques on the binary image 14 in the x
`and y direction to determine a location indicative of the
`central portion of the facial feature and/or its size, if
`desired. Also, a more complex calculation may be
`
`30
`
`35
`
`9
`
`SAMSUNG EXHIBIT 1007
`Page 11 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 11 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`The system may
`employed if the data has weighted values.
`also decompose the two-dimensional binary image into
`directions other than x and y.
`The face locator and tracker 16 provides the
`general location of the center of the face and its size.
`
`The output of image processor 11 provides data to a
`communication module 40 which may transmit or display the
`image in any suitable format.
`The face tracking system 6
`may enhance the bit rate for the portion of the image
`containing the face, as suggested by Eleftheriadis.
`The estimated face location may also be used
`for tracking the face between frames of a video.
`For
`simplicity the face motion may be modeled as a piece-wise
`constant two-dimensional translation within the image
`plane.
`A linear Kalman filter may be used to predict and
`correct the estimation of the two-dimensional translation
`velocity vector.
`The estimated (filtered) velocity may
`then also be used to determine the tracked positions of
`faces.
`
`The preferred system model for tracking the
`
`motion is:
`
`x (K+1) =F (kK) x(k) +w(k)
`
`2 (K+1) =H (k+1) x (k+1)+v(k+1)
`where x(k)
`is the true velocity vector to be estimated,
`z(k)
`is the observed instantaneous velocity vector, w(k),
`v(k) are white noise, and F(k)=I, H(k)=I for piece-wise
`constant motion.
`The Kalman predictor is:
`R(k+1|k)=F(k)2(k|k), %(0]0)=0
`@(k+1|k)=H(k+1)x(K+1]k)
`The Kalman corrector is:
`S(k+1]k+1) =(k+1]k) +K(k+1) Az(K+1| k)
`Az(k+1|k) =z (k+1) -2 (k+1]k)
`where K(k+1)
`is the Kalman gain.
`computed as:
`K(k+1)=P(k+1|k)H"(k+1) [H(k+1) P(K+1[k) H"(k+1)+R(k+1) ]7!
`
`The Kalman gain is
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`10
`
`SAMSUNG EXHIBIT 1007
`Page 12 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 12 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`A
`
`The covariances are computed as:
`P(k+1[k)=F(k) P(k|k) F"(k)+Q(k), P(0|0)=P,
`P(k+1[k+1)=[I-K(k+1)H(k+1) ]P(kK+1|k)
`where Q(k)=E({w(k)w'(k)], R(k)=E(v(k)v'(k)] and
`P,=E[x(0)x"(0)].
`|
`In the presence of lighting fluctuation and
`image noise,
`the tracked face image may be jittering.
`nonlinear filtering module therefore may be included in
`the tracking system to remove the undesirable jittering.
`A simple implementation of the nonlinear filtering module
`is to cancel any movement of the tracked face which is
`smaller in magnitude than a prescribed threshold and
`shorter in duration than another prescribed threshold.
`A particular application suitable for the face
`detection and tracking system described herein involves a
`video phone. Other suitable device may likewise be used.
`An image of the background without a person present is
`obtained by the system. Thereafter images are obtained
`in the presence of the person. Each image obtained is
`compared against the background image to distinguish the
`foreground portion of the image from the background image
`previously obtained.
`The recipient’s video phone has a
`nice background image displayed thereon.
`The foreground,
`which is presumably the person,
`is transmitted to and
`overlayed on the nice background image of the recipient’s
`video phone on a frame-by-frame manner.
`The location of
`the face is determined by the face tracking system to
`smooth out the movement of the person and remove jitter.
`Alternatively,
`the nice background image may be
`transmitted to the recipient’s video phone, and is
`preferably transmitted only once per session. This
`provides the benefit of disguising the actual background
`environment and potentially reducing the bandwidth
`requirements.
`
`The system may be expanded using the same
`teachings to locate and track multiple faces within an
`image.
`
`11
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`SAMSUNG EXHIBIT 1007
`Page 13 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 13 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`The terms and expressions which have been
`employed in the foregoing specification are used therein
`as terms of description and not of limitation, and there
`is no intention,
`in the use of such terms and
`expressions, of excluding equivalents of the features
`shown and described or portions thereof, it being
`recognized that the scope of the invention is defined and
`limited only by the claims which follow.
`
`5
`
`12
`
`SAMSUNG EXHIBIT 1007
`Page 14 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 14 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`CLAIMS
`
`1.
`
`A method of detecting a face within an
`image comprising the steps of:
`receiving said image including a plurality
`of pixels, where a plurality of said
`pixels of said image is represented by
`respective groups of at least three
`
`(a)
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`values;
`
`(b)
`
`filtering said image by transforming a
`plurality of said respective groups of
`said at least three values to respective
`groups of less than three values, where
`said respective groups of said less than
`
`(c)
`
`(4)
`
`(e)
`
`three values has less dependency on
`brightness than said respective groups of
`said at least three values;
`determining regions of said image
`representative of skin-tones based on said
`filtering of step (b);
`calculating a first distribution of said
`regions of said image representative of
`said skin-tones in a first direction;
`
`calculating a second distribution of said
`regions of said image representative of
`said skin-tones in a second direction,
`where said first direction and said second
`
`direction are different; and
`
`(f£)
`
`locating said face within said image based
`on said first distribution and said second
`
`distribution.
`
`The method of claim 1 where said image
`includes from a video containing multiple images.
`
`2.
`
`The method of claim 1 where said image
`includes a human face.
`
`3.
`
`13
`
`SAMSUNG EXHIBIT 1007
`Page 15 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 15 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`4.
`
`The method of claim 1 where said at least
`
`three values includes a red value, a green value, and a
`blue value.
`
`5.
`
`The method of claim 4 where said
`
`respective groups of less than three values includes, ar
`value defined by said red value divided by the summation
`of said red value, said green value, and said blue value,
`and a g value defined by said green value divided by the
`summation of said red value, said green value, and said
`blue value.
`
`6.
`
`The method of claim 1 wherein at least one
`
`of said regions is an individual pixel of said image.
`
`7.
`
`‘The method of claim 1 wherein said
`
`determining of step (c)
`
`is based on a polygonal shape.
`
`10
`
`15
`
`20
`
`determining of step (c)
`
`is based on a circle.
`
`8.
`
`The method of claim 1 wherein said
`
`9.
`
`The method of claim 1 wherein at least one
`
`of said first distribution and said second distribution
`
`is a histogram.
`
`25
`
`10.
`
`The method of claim 1 wherein said first
`
`distribution is in a x-direction.
`
`11.
`
`The method of claim 10 wherein said second
`
`30
`
`distribution is in a y-direction.
`
`12.
`
`The method of claim 11 wherein said first
`
`distribution and said second distribution are in
`orthogonal directions.
`
`14
`
`SAMSUNG EXHIBIT 1007
`Page 16 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 16 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`13.
`The method of claim 1 wherein said first
`distribution and said second distribution are independent
`of each other.
`
`14.
`the steps of:
`(a)
`
`(b)
`
`The method of claim 1 further comprising
`
`calculating a first generally central
`location of said first distribution;
`calculating a first generally central
`location of said second distribution; and
`
`(c)
`
`locating said face based on said first
`
`generally central location of said first
`distribution and said first generally
`central location of said second
`
`distribution.
`
`15.
`
`The method of claim 14 wherein at least
`
`one of said first generally central location of said
`first distribution and said first generally central
`location of said second distribution is a mean.
`
`10
`
`15
`
`20
`
`16.
`
`The method of claim 14 wherein the size of
`
`said face is based on the variance of said first
`
`distribution and the variance of said second
`
`25
`
`distribution.
`
`17.
`
`The method of claim 1 wherein said face is
`
`tracked between subsequent frames.
`
`30
`
`The method of claim 17 wherein jitter
`18.
`movement of said face is reduced between said subsequent
`frames.
`
`15
`
`SAMSUNG EXHIBIT 1007
`Page 17 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 17 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`FIG.
`
`1
`
`AND TRACKER
`
`40
`
`COMMUNICATION
`MODULE
`
`TRANSFORMATION
`
`BINARY
`IMAGE
`
`FACE
`LOCATOR
`
`175
`
`SAMSUNG EXHIBIT 1007
`Page 18 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 18 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`°
`eee
`
`°
`tevee 0
`
`:yesspiessie
`«GIES
`L, Si in
`:%aes see" Lawl, .
`f0*
`-
`mn
`
`meatnte.e, ? , 1
`
`FIG. 3
`
`2/5
`
`SAMSUNG EXHIBIT 1007
`Page 19 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 19 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`$2526
`
`BRS
`SeSSHSCS
`
`Soseuego
`
`SSF
`oSS664
`
`ehatatetachststctete
`
`
`5552505050525$5252529
`Ox$882
`SSSSs
`
`S255
`oS52505
`52525
`
`SSS5555522
`SOSoSP5558
`52525250525
`
`xOqss
`
`
`25250535555550505
`550505050805
`erataratacatecerere,
`SSSS505
`rereratataracatates
`S2526255
`650505
`S250QOQREOOOY
`SSL
`SORRY
`59
`SOY
`eso
`
`
`
`VAAAAAA/ZWAAAAAAAAAHE25250505$2525.5525252552525
`
`S$S282SKS552528282528252
`
`Sex$62552605250505
`
`erereteterereteteren
`5555555054
`555255050505655555085
`erererers:
`2525055
`S550
`
`westanstatetgtetseaecececs
`Veretetatetateretetetete:82S
`
`D82525232829292
`BSS505252
`55.5.2525.505.05
`5060005
`355505050505
`525505250505
`
`OSS555.582
`55505250505
`95250550505
`305055050505
`
`3/5
`
`SAMSUNG EXHIBIT 1007
`Page 20 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 20 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`
`00000000000001000001000000010Lo0000000000000
`00001010100010000000000000000001001001010L00
`
`bhLELEL
`
`00
`
`G‘SIs
`
`LLOl
`Olt
`LOLELL
`
`bEELL
`
`LL
`LLL
`LLELELE
`
`bt
`bh
`BLEEEL
`
`LL
`
`LL
`
`LL
`
`bt
`
`coo0000c00NNONONONNDO}o0000000000000
`
`0000000LO0000000000LoL000000000001000
`
`
`
`
`00000000100000000000001000000000000000000100
`
`
`
`
`00000lt00000000
`
`Oo0000LOoCooOL
`
`00LOOO000L00
`
`OLLEELLEE
`
`OLLLBEE
`OLLbh0
`
`bEL0000000100000
`000000010000
`00LO000001L0
`
`Lf
`ae)
`bhEELYL
`
`oO}
`
`OODOODOOOCOOOOLLILLbLL-ELOOOOOONONN0N
`
`00
`O-Lt-
`LELLO
`
`00
`
`
`bLELLLOOOOODOOLOOOOOOODDOLOOOLLILILOOOOCODOODOOOOLELILLLbLLELLLoOOOOOOODND
`
`
`
`
`
`OL
`bb
`LOLS
`
`LLOt
`
`bh
`LOL
`bLEELL
`
`bh
`LE
`bLELSL
`
`bt
`
`bf
`bb
`OLLO}L
`
`bb
`
`Lb
`bE}
`LLOLL
`
`Lt
`
`bLELLLOOOOOOOLObLLELELLOLOOOOOODOb}LOLLLOOOOOOOOLOOODDOOODOOOLLOLLLEOGLOOOLOOOLLELELELELLhLLELELLOOOLOOOOOOOLOOOOOOLLLOLLLCOOOOODOOOLOLLELELE
`
`
`
`
`
`
`
`
`
`
`
`
`
`OOCOOOCOOOOOLILELLELEbLELELELOOOOOOOOL
`
`
`
`
`LOL-L-LLOOOOOOONOOONO‘btELEOLLOOOOOONONOODCOODOOOCOOOOLLELEf}LELtLLOOOOOOONNDOOOCOOOOOOOLLLLILECOOOODODOOCOOLLILOLOL
`
`
`
`
`
`
`
`
`
`
`00L!Oo0000000000001LO0000LOOO0ON0O0OL
`
`
`
`LL
`LEE
`LLELLO
`
`LL
`
`
`
`oo00000000000000LtOoo00000000000
`
`0001000010000}
`bELLL
`LOOOLOOOOLOOA
`
`LE
`LEE
`LOLLE
`
`LL
`
`Lb
`LEI
`bhLELEE
`
`LL
`
`000001000000001
`
`bbL|bt
`
`LOO0000LOOOOL
`
`Lb
`LEE
`bhLELEEL
`
`LL
`
`
`
`oo000000000NNN0NNNNNLoo00000000000000
`
`bE
`LOLLE
`
`
`
`000000000000000000000Lo000000000000000000000
`
`
`
`
`
`
`
`00000000010000000000010000000000Lo0000000010
`
`
`
`
`
`00000000000000000000010000000000Lo0000000000
`
`LI
`LEE
`OLOLL
`
`0000000001001
`bbbBLES
`LOOO0O0000010
`
`bELt
`
`LOLtL|LL
`000
`000
`O-LL
`bE
`00000
`00000
`LLEELE
`LOLS
`
`Lt
`
`Lt
`
`bt
`
`4/5
`
`SAMSUNG EXHIBIT 1007
`Page 21 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 21 of 25
`
`
`
`WO 99/35606
`
`PCT/JP99/00010
`
`FIG. 6
`
`38
`
`
`0010101007001 00000000000009Q60001 000101010000
`000000000900001 01 00090001 900001 go00000000000
`00100000090000000000071 007000000G0001 co000000
`
`An
`5
`0-5-6-4
`00010000090000060111741191 000000900001 0000000
` 00000100000001111111
`
`010000001901111111171
`TUVVTTTI11¢
`
`
`
`
`00000000001111110111
`
`00100000090111110001141110000111
`
`
`
`
`00000000091111110111901111101111
`
`OLO0CCOOOOTIWAAVITAT TTT TOmtt1titt
`
`
`
`
`1900000000
` TTOTTTIVVOTTOUTITATIT11¢
`
`
`
`000010001
`DUVTTAT TV OFTTtttttitt
`
`100000000
`TUVTTTTT ATTA Ttttatitt
`
`
`
`600000000411710111111
`
`00000000041011111000
`00000000061111111000
`OL COCOOCOGTIIITTTTTOPOWTTVAT ttt
`
`11111100¢000000100000
`000001000000
`001000000100
`ooco00c00000
`000100000000
`1000C00000000
`101000000000
`
`110000000000
`110001000100
`1T1eqoca0ccg0N
`1190000000000
`t1raodco000g000na
`ooo0o0000c000
`ootac0cdqcaco0o0a
`
`
`10000100090011111111 111171111 aa000007 00000 \
`
`000000000G000111111141111111100e@0000Cc0000000
`100000001 00000111011111111110000000000000100
`00000000090000001111901111000000G000000000000
`oodcdcndcccOGgooonooNoNNgO! Conoco NooNaGgoOGOOCN0N0N0NN
`010000000G01 OG000000001 000000009001 coco00000
`000000000G01 0O000000Gd1 cdoddCadGgooON0CON0N00000
`
`34A
`
`32
`
`34B
`
`34D
`
`36
`
`34C
`
`5/5
`
`SAMSUNG EXHIBIT 1007
`Page 22 of 25
`
`SAMSUNG EXHIBIT 1007
`Page 22 of 25
`
`
`
`
`
`B. FIELDS SEARCHED
`
`Minimum documentation searched (classification system followed by classification symbols)
`IPC 6
`GOQ6K
`
`INTERNATIONAL SEARCH REPORT
` Interna:
`| Application No
`
`
`PCT/JP 99/00010
`
`
`A. CLASSIFICATION OF SUBJECT MATTER
`
`
`IPC 6
`G06K9/00
`
`
` According to international Patent Classification (IPC) or to both nationalclassification and IPC
`
`
`
`
`
`
`Documentation searchedother than minimum documentation to the extent that such documentsare included in the fields searched
`
`
`Electronic data base consulted during the international search (name of data base and, where practical, search terms used)
`
`
`
`Category °|Citation of document, with indication, where appropriate, of the relevant passages Relevant to claim No.
`
`line 1
`line 9 - column 9,
`line 15 - line 23; figures
`
`
`
`
`
` C. DOCUMENTS CONSIDERED TO BE RELEVANT
`a
`
`
`
`
`
` WONG C ET AL:
`
` >
`
`
`
`EP 0 654 749 A (HITACHI EUROP LTD)
`24 May 1995
`see column 8,
`see column 10,
`10-20
`
`"A MOBILE ROBOT THAT
`RECOGNIZES PEOPLE"
`PROCEEDINGS OF THE 7TH INTERNATIONAL
`CONFERENCE ON TOOLS WITH ARTIFICIAL
`INTELLIGENCE, HERNDON, VA., NOV.
`5 - 8
`1995,5 November 1995, pages 346-353,
`XP000598377
`INSTITUTE OF ELECTRICAL AND ELECTRONICS
`ENGIN