throbber

`
`
`UNITED STATES PATENT AND TRADEMARK OFFICE
`____________
`
`BEFORE THE PATENT TRIAL AND APPEAL BOARD
`____________
`
`
`GOOGLE LLC,
`Petitioner
`
`v.
`
`GESTURE TECHNOLOGY PARTNERS LLC,
`Patent Owner
`____________
`
`
`Case No. IPR2022-00360
`U.S. Patent No. 8,553,079
`____________
`
`
`
`
`
`
`
`
`PETITION FOR INTER PARTES REVIEW
`OF U.S. PATENT NO. 8,553,079
`
`

`

`
`
`TABLE OF CONTENTS
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`INTRODUCTION ......................................................................................... 1
`I.
`SUMMARY OF THE ’079 PATENT .......................................................... 1
`II.
`The ’079 Patent’s Alleged Invention ............................................................... 1
`A.
`The ’079 Patent’s Prosecution ......................................................................... 3
`B.
`A Person Having Ordinary Skill in the Art ..................................................... 4
`C.
`III. REQUIREMENTS FOR IPR UNDER 37 C.F.R. § 42.104 ....................... 4
`A.
`Standing Under 37 C.F.R. § 42.104(A) ........................................................... 4
`B.
`Challenge Under 37 C.F.R. § 42.104(B) and Relief Requested ...................... 4
`C.
`Claim Construction Under 37 C.F.R. § 42.104(B)(3) ..................................... 5
`IV. THE CHALLENGED CLAIMS ARE UNPATENTABLE ....................... 6
`A.
`Ground 1: Claims 1, 2, 4-14, 17, 19, 21-22, 24-28, and 30 are obvious under
`pre-AIA 35 U.S.C. § 103 over Numazaki in view of the knowledge of a
`PHOSITA ........................................................................................................ 6
`Ground 2: Claims 3, 15, and 23 are obvious under pre-AIA 35 U.S.C. § 103
`over Numazaki in view of Numazaki ’863 ....................................................34
`Ground 3: Claims 16 and 29 are obvious under pre-AIA 35 U.S.C. § 103
`over Numazaki in view of DeLuca ................................................................41
`Ground 4: Claim 18 is obvious under pre-AIA 35 U.S.C. § 103 over
`Numazaki in view of DeLeeuw ......................................................................48
`Ground 5: Claim 20 is obvious under pre-AIA 34 U.S.C. § 103 over
`Numazaki in view of Maruno ........................................................................55
`DISCRETIONARY CONSIDERATIONS ................................................68
`V.
`The Fintiv Factors Favor Institution ..............................................................68
`A.
`1. Whether the court granted a stay or evidence exists that one may be granted
`if a proceeding is instituted. .........................................................................69
`Proximity of the court’s trial date to the Board’s projected statutory
`deadline for a final written decision. ............................................................70
`Investment in the parallel proceeding by the court and the parties. ............74
`Overlap between issues raised in the petition and in the parallel proceeding.
` ......................................................................................................................75
`5. Whether the petitioner and the defendant in the parallel proceeding are the
`same party. ...................................................................................................76
`Other circumstances that impact the Board’s exercise of discretion,
`including the merits. .....................................................................................76
`
`B.
`
`C.
`
`D.
`
`E.
`
`2.
`
`3.
`4.
`
`6.
`
`
`
`i
`
`

`

`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`VI. CONCLUSION ............................................................................................77
`VII. MANDATORY NOTICES UNDER 37 C.F.R. § 42.8(A)(1) ...................77
`A.
`Real Party-In-Interest.....................................................................................77
`B.
`Related Matters ..............................................................................................77
`C.
`Lead and Back-Up Counsel ...........................................................................78
`APPENDIX OF EXHIBITS ..................................................................................80
`
`
`
`
`
`ii
`
`

`

`INTRODUCTION
`Petitioner Google LLC (“Petitioner”) requests an Inter Partes Review
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`I.
`
`(“IPR”) of claims 1–30 (the “Challenged Claims”) of U.S. Patent No. 8,553,079
`
`(“the ’079 Patent”). This petition is substantively the same as IPR2021-00922
`
`(which is instituted), and is being filed concurrently with a motion for joinder
`
`with respect to that proceeding.
`
`II.
`
`SUMMARY OF THE ’079 PATENT
`The ’079 Patent’s Alleged Invention
`A.
`The ’079 Patent generally describes computer input devices employing
`
`cameras and lights to observe points on the human body and optically sense human
`
`positions and/or orientations.’079 Patent (Ex. 1001), 1:54-2:6. Examples of input
`
`devices contemplated by the patent include a computer keyboard, puzzle toy, and
`
`handheld computer. Id. at 2:15-31. Fig. 2 below illustrates one exemplary
`
`embodiment implemented in a laptop computer:
`
`
`
`1
`
`

`

`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`
`Id. at Fig. 2. As illustrated, a laptop 138 may include camera locations 100, 101, 105,
`
`106, 108, 109; keyboard surface 102; screen housing 107; light 122; light emitting
`
`diodes (LEDs) 210 and 211, and work volume area 170 within which a user’s
`
`movements are detected. Id. at 2:39-53. The system can detect a user’s finger alone
`
`or the user may employ external objects such as ring 208 to help detect and recognize
`
`gestures performed in work volume area 170. Id. at 2:54-3:8. The ’079 Patent
`
`
`
`2
`
`

`

`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`describes detecting point, pinch, and grip gestures using this configuration. Id. at
`
`3:48-51.
`
`The ’079 Patent’s Prosecution
`B.
`The Application that resulted in the ’079 Patent was filed on December 14,
`
`2012. The Application claims priority to provisional patent application No.
`
`60/107,652, filed November 9, 1998, through a succession of continuation
`
`applications 09/433,297; 10/866,191; and 12/700,055. Id. For purposes of this
`
`petition and without waiving its right to challenge priority in this or any other
`
`proceeding, Petitioner adopts November 9, 1998 as the invention date for the
`
`Challenged Claims.
`
`Applicant canceled twenty originally filed claims (1–20) by preliminary
`
`amendment and replaced them with thirty new claims (21–50). ’079 File History
`
`(Ex. 1002), 134-138. A Notice of Allowance issued on July 24, 2013, in which
`
`the Examiner allowed all thirty claims (renumbered 1–30) in a first office action.
`
`Id. at 150, 175. Examiner amended the abstract and reasoned that none of Naoi et
`
`al. (US 5459793), Platzker et al. (US 5528263), Sellers (US 5864334 A), nor
`
`Fukushima et al. (US 6346929 B1) taught or suggested the independently claimed
`
`elements: (1) providing a camera oriented to observe a gesture performed in the
`
`work volume; (2) the camera being fixed relative to the light source; and (3)
`
`
`
`3
`
`

`

`determining, using the camera, the gesture performed in the work volume and
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`illuminated by the light source. Id. at 153.
`
`C. A Person Having Ordinary Skill in the Art
`A person having ordinary skill in the art (“PHOSITA”) at the time of the
`
`’079 Patent would have had at least a bachelor’s degree in electrical engineering
`
`or equivalent with at least one year of experience in the field of human computer
`
`interaction. Additional education or experience might substitute for the above
`
`requirements. Bederson Dec. (Ex. 1010), ¶¶ 29-31.
`
`III. REQUIREMENTS FOR IPR UNDER 37 C.F.R. § 42.104
`Standing Under 37 C.F.R. § 42.104(A)
`A.
`Petitioner certifies that the ’079 Patent is available for IPR and that
`
`Petitioner is not barred or estopped from requesting an IPR challenging the claims
`
`of the ’079 Patent. Specifically, (1) Petitioner is not the owner of the ’079 Patent,
`
`(2) Petitioner has not filed a civil action challenging the validity of any claim of
`
`the ’079 Patent, and (3) Petitioner has not been served with a complaint more than
`
`one year ago alleging infringement of the ’079 patent.
`
`B. Challenge Under 37 C.F.R. § 42.104(B) and Relief Requested
`In view of the prior art and evidence presented, claims 1-30 of the ’079
`
`Patent are unpatentable and should be cancelled. 37 C.F.R. § 42.104(b)(1).
`
`
`
`4
`
`

`

`Further, based on the prior art references identified below, IPR of the Challenged
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`Claims should be granted. 37 C.F.R. § 42.104(b)(2).
`
`Proposed Ground of Unpatentability
`Ground 1: Claims 1, 2, 4-14, 17, 19, 21-22, 24-28, and 30 are
`obvious under pre-AIA 35 U.S.C. § 103 over U.S. Patent No.
`6,144,366 (“Numazaki”) in view of the knowledge of a PHOSITA
`Ground 2: Claims 3, 15, and 23 are obvious under pre-AIA 35 USC
`§ 103 over Numazaki in view of U.S. Patent No. 5,900,863
`(“Numazaki ’863”)
`Ground 3: Claims 16 and 29 are obvious under pre-AIA 35 U.S.C §
`103 over Numazaki in view of U.S. Patent No. 6,064,354 (“DeLuca”)
`Ground 4: Claim 18 is obvious under pre-AIA 35 U.S.C § 103 over
`Numazaki in view of U.S. Patent No. 6,088,018 (“DeLeeuw”)
`Ground 5: Claim 20 is obvious under pre-AIA 35 USC § 103 over
`Numazaki in view of U.S. Patent No. 6,191,773 (“Maruno”)
`
`Exhibits
`Ex. 1004
`
`Ex. 1004,
`Ex. 1005
`
`Ex. 1004,
`Ex. 1006
`Ex. 1004,
`Ex. 1007
`Ex. 1004,
`Ex. 1008
`
`Section IV identifies where each element of the Challenged Claims is
`
`found in the prior art. 37 C.F.R. § 42.104(b)(4). The exhibit numbers of the
`
`evidence relied upon to support the challenges are provided above and the
`
`relevance of the evidence to the challenges raised is provided in Section IV. 37
`
`C.F.R. § 42.104(b)(5). Exhibits 1001-1017 are also attached.
`
`C. Claim Construction Under 37 C.F.R. § 42.104(B)(3)
`In this proceeding, claims are interpreted under the same standard applied
`
`by Article III courts (i.e., the Phillips standard). See 37 C.F.R § 42.100(b); see
`
`also 83 Fed. Reg. 197 (Oct. 11, 2018); Phillips v. AWH Corp., 415 F.3d 1303,
`
`1312 (Fed. Cir. 2005) (en banc). Under this standard, words in a claim are given
`
`their plain meaning which is the meaning understood by a person of ordinary skill
`
`
`
`5
`
`

`

`in the art in view of the patent and file history. Phillips, 415 F.3d 1303, 1212-13.
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`For purposes of the proposed grounds below, Petitioner proposes no terms require
`
`express construction.
`
`IV. THE CHALLENGED CLAIMS ARE UNPATENTABLE
`A. Ground 1: Claims 1, 2, 4-14, 17, 19, 21-22, 24-28, and 30 are
`obvious under pre-AIA 35 U.S.C. § 103 over Numazaki in view of
`the knowledge of a PHOSITA
`Overview of Numazaki
`
`U.S. Patent No. 6,144,366 to Numazaki et al. (“Numazaki”) (Ex. 1004) was
`
`filed on October 17, 1997 and is prior art to the ’079 Patent under at least 35 U.S.C.
`
`§ 102(e) (pre-AIA). Numazaki was not cited or considered during prosecution of
`
`the ’079 Patent or its parent, U.S. Patent No. 6,750,848. ’079 Patent (Ex. 1001);
`
`’848 Patent (Ex. 1003).
`
`Numazaki is generally directed to a method for detecting a gesture or the
`
`movement of a user’s hand. (“Numazaki”) (Ex. 1004), Abstract, 4:9-40.
`
`Numazaki purports to have improved upon prior methods by using a controlled
`
`light source to illuminate the target object (e.g., the user’s hand), a first camera
`
`
`
`6
`
`

`

`unit (referred to by Numazaki as a “photo-detection unit”),1 and a second camera
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`unit. Id. at 11:9-23. This arrangement is illustrated in Fig. 2 below:
`
`
`
`Id. at Fig. 2. A timing control unit is used to turn lighting unit 101 on (i.e.,
`
`illuminating the target object) when the first camera unit is active and off when the
`
`second camera unit is active. Id. at 11:20-32. The result of this light control is the
`
`first camera unit captures an image of the target object illuminated by both natural
`
`light and the lighting unit 101 and the second camera unit captures an image of the
`
`
`1 A PHOSITA would have considered Numazaki’s photo-detection units to be
`
`camera units. Bederson Dec. (Ex. 1010), ¶ 36 (explaining that Numazaki describes
`
`using CMOS or CCD sensor units, which were two of the more common optical
`
`sensors used in camera units at the time).
`
`
`
`7
`
`

`

`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`target object illuminated by only natural light. Id. at 11:33-39. The difference
`
`between the two images—obtained by difference calculation unit 111—represents
`
`the “reflected light from the object resulting from the light emitted by the lighting
`
`unit 101.” Id. at 11:43-51. This information is then used by feature data generation
`
`unit 103 to determine gestures, pointing, etc. of the target object that may be
`
`converted into commands executed by a computer. Id. at 10:57-66.
`
`In its eighth embodiment, Numazaki describes implementing this structure
`
`in a computer such that a user can point or gesture with an index finger while
`
`typing on the keyboard “with[] hardly any shift of the hand position.” Id. at 50:25-
`
`43. This arrangement is illustrated in Fig. 74 below:
`
`
`
`8
`
`
`
`

`

`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`Id. at Figure 74 (annotated to indicate light source 701 and photo-detection sensor
`
`unit 702). Numazaki teaches the entirety of the operator’s hand is illuminated within
`
`the range depicted by the dashed circle such that the user’s pointing and gestures can
`
`be captured and converted to commands. Id. at 50:38-48.
`
`Numazaki
`
`teaches
`
`that
`
`its eighth embodiment
`
`incorporates “the
`
`information input generation apparatus of the present invention as described in
`
`the above embodiments.” Id. at 50:21-24. A PHOSITA would have understood
`
`that the referenced information input generation apparatus is that illustrated in
`
`Fig. 2 and described in the corresponding disclosure. Bederson Dec. (Ex. 1010),
`
`¶¶ 42-43 (explaining that Numazaki describes its controlled light and two-camera
`
`configuration as key to its invention and noting that Numazaki at 53:22-36 teaches
`
`that the eighth embodiment uses the precise image difference calculation taught
`
`by Fig. 2 and its corresponding disclosure).
`
`Because Numazaki, like the ’079 Patent, discloses a method and apparatus
`
`for generating computer input information by capturing hand gestures, Numazaki
`
`is in the same field of endeavor as the ’079 Patent. Compare Numazaki (Ex.
`
`1004), 50:29-37 (“In this computer of FIG. 74, a lighting unit 701 and a photo-
`
`detection sensor unit 702 of the information input generation apparatus are
`
`provided at positions beyond the keyboard” such that the “entire hand of the
`
`operator is illuminated . . . [in] a range of illumination”) with ’079 Patent (Ex
`
`
`
`9
`
`

`

`1001), Abstract (describing “[a] method for determining a gesture illuminated by
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`a light source” that “utilizes the light source to provide illumination through a
`
`work volume above the light source” where a “camera is positioned to observe
`
`and determine the gesture performed in the work volume”). Numazaki is therefore
`
`analogous art to the ’079 Patent. Bederson Dec. (Ex. 1010), ¶ 41.
`
`i.
`
`Claim 1.
`
`1[P] A computer implemented method comprising:
`
`To the extent the preamble is limiting, Numazaki teaches a computer
`
`implemented method. In particular, Numazaki teaches a “method . . . for
`
`generating information input . . . capable of realizing a direct command type
`
`information input scheme by which the gesture or the motion can be inputted
`
`easily.” Numazaki (Ex. 1004), 4:9-13.
`
`Numazaki’s eighth embodiment teaches a computer implemented method
`
`for controlling functions on a laptop device through gestures or pointing:
`
`In this configuration, the operator operating the keyboard can make the
`
`pointing or gesture input by slightly raising and moving the index
`
`finger. The user’s convenience is remarkably improved here because
`
`the keyboard input and the pointing or gesture input can be made
`
`without hardly any shift of the hand position.
`
`Id. at 50:38-43. This arrangement is illustrated in Fig. 74 below:
`
`
`
`10
`
`

`

`
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`
`
`Id. at Fig. 74.
`
`[1(a)] providing a light source adapted to direct illumination through a work
`volume above the light source;
`Numazaki teaches a lighting unit 701 (i.e., light source) that is adapted to
`
`illuminate a user’s hand (i.e., human body part) within a work volume generally
`
`above the light source. For example, Numazaki teaches “the entire hand of the
`
`operator is illuminated, as can be seen from a dashed line circle indicating a
`
`range of illumination.” Id. at 50:35-37 (emphasis added). As depicted in Fig. 74
`
`below, the lighting unit 701 is adapted to illuminate the user’s hand by positioning
`
`
`
`11
`
`

`

`it “beyond the keyboard” to illuminate a work volume indicated by the dashed
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`line circle above it:
`
`
`Id. at Fig. 74 (annotated to illustrate light source and range of hand illumination).
`
`[1(b)] providing a camera oriented to observe a gesture performed in the work
`volume, the camera being fixed relative to the light source; and determining,
`using the camera, the gesture performed in the work volume and illuminated by
`the light source.
`Numazaki discloses a photo-detection sensor unit 702 (i.e., camera2) that
`
`is positioned next to lighting unit 701 (i.e., in fixed location relative to the light
`
`
`2 As described in the overview section above, a PHOSITA would have considered
`
`Numazaki’s photo-detection units to be camera units. Bederson Dec. (Ex. 1010), ¶
`
`
`
`12
`
`

`

`source) and arranged to have the optical axis of the photo-detection sections
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`pointing obliquely upward towards the operator side to observe a gesture
`
`performed by the user (i.e., observe a gesture) within a “dashed line circle” (i.e.,
`
`work volume) depicted below. Id. at 50:30-43. This arrangement is illustrated in
`
`Fig. 74 below:
`
`
`Id. at Fig. 74 (annotated to show camera).
`
`As described with reference to Fig. 2, which is incorporated into the eighth
`
`embodiment as noted above, Numazaki uses this light and camera arrangement to
`
`
`36 (explaining that Numazaki describes using CMOS or CCD sensor units, which
`
`were two of the more common optical sensors used in camera units at the time).
`
`
`
`13
`
`

`

`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`illuminate the target object (e.g., the user’s hand) in a controlled manner such that
`
`a precise image of the user’s hand and hand movement can be ascertained. Id. at
`
`11:9-23. Specifically, a timing control unit is used to turn lighting unit 101 on (i.e.,
`
`illuminating the target object) when the first camera unit is active and off when the
`
`second camera unit is active. Id. at 11:20-32. The result of this light control is the
`
`first camera unit captures an image of the target object illuminated by both natural
`
`light and the lighting unit 101 and the second camera unit captures an image of the
`
`target object illuminated by only natural light. Id. at 11:33-39. The difference
`
`between the two images – obtained by difference calculation unit 111 – represents
`
`the “reflected light from the object resulting from the light emitted by the lighting
`
`unit 101.” Id. at 11:43-51. This information is then used by feature data generation
`
`unit 103 to determine gestures, pointing, etc. of the target object that may be
`
`converted into commands executed by a computer. Id. at 10:57-66. Through this
`
`arrangement, a PHOSITA would have understood that component 702 illustrated
`
`in Fig. 2 is “oriented to observe a gesture performed in the work volume”
`
`illuminated by lighting unit 701. Bederson Dec. (Ex. 1010), ¶¶ 42-43.
`
`
`
`14
`
`

`

`ii.
`
`Claim 2
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`2. The method according to claim 1 wherein the light source includes a light
`emitting diode.
`
`As discussed above with reference to limitation 1[a], Numazaki teaches light
`
`source 701 in the embodiment depicted in Fig. 74 below:
`
`
`
`Numazaki (Ex. 1004), Fig. 74; see also limitation 1[a], supra. Although Numazaki
`
`does not discuss the specific lighting technology contemplated for light source 701,
`
`with reference to the first embodiments, Numazaki teaches that an “LED can be used
`
`as the light source” since “the LED has a property that it can emit more intense light
`
`instantaneously” while “reduc[ing] the power required for the light emission.” Id. at
`
`14:49-56; see also id. at 68:13-20. Based on Numazaki’s teachings, a PHOSITA
`
`
`
`15
`
`

`

`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`would have been motivated to implement light source 701 in Fig. 74 using LED
`
`technology for a number of reasons. Bederson Dec. (Ex. 1010), ¶ 44. First, the
`
`benefit of emitting more intense light instantaneously described with reference to
`
`the first embodiment would have improved Fig. 74’s apparatus such that it could
`
`quickly and accurately detect pointing and gestures. Id; see also KSR Int’l Co. v.
`
`Teleflex Inc., 550 U.S. 398, 417 (2007) (obvious to use known techniques to improve
`
`similar devices in the same way). Second, a PHOSITA would have understood that
`
`the “note PC” depicted in Fig. 74 is a self-contained and portable unit that would
`
`have benefitted from the power reduction of an LED light source discussed with
`
`reference to the first embodiment. Bederson Dec. (Ex. 1010), ¶ 44 (noting that
`
`Numazaki’s eighth embodiment contemplates other portable implementations and,
`
`at col. 52, ln. 33 – col. 53, ln. 7, expressly discusses the desire to conserve power
`
`and techniques for doing so by controlling the lighting unit).
`
`iv.
`
`Claim 4
`
`
`4. The method according to claim 1 wherein detecting a gesture includes
`analyzing sequential images of the camera.
`Numazaki’s eighth embodiment analyzes the sequential movement of
`
`“click and drag” operations by detecting a combination of “pointing or gesture
`
`input” and “keyboard input” via a “button for use in conjunction with the pointing
`
`
`
`16
`
`

`

`or gesture input.” Numazaki (Ex. 1004), 50:38-47. This permits “selecting and
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`moving icons on the screen.” Id. at 50:45-47.
`
`A PHOSITA would have understood that using gesture recognition to
`
`implement a “click and drag” operation would have been implemented by
`
`capturing a sequence of images to determine the user’s hand movement, which
`
`dictates the “drag” operation to be performed. Bederson Dec. (Ex. 1010), ¶¶ 45-
`
`46. Indeed, in its second embodiment, Numazaki expressly describes a process
`
`through which the system tracks lateral finger movements by detecting the center
`
`of gravity of a finger, where “finger tip movement and the center of gravity
`
`movement can be smoothly correlated” using pixel values. Numazaki (Ex. 1004),
`
`19:43-20:25. Fig. 10 below illustrates the process by which the reflected light
`
`image of a hand and finger are mapped to a pixelated target space, and Fig. 11
`
`illustrates how coordinate-based finger movement can be tracked on the basis of
`
`pixel value changes:
`
`
`
`17
`
`

`

`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`Id. at Fig. 10, 19:13-42 (describing the same).
`
`
`
`
`
`Id. at Fig. 11A, 19:43-20:25 (describing a center of gravity tracking that enables a
`
`fingertip to be tracked). Using this technique, Numazaki teaches “the cursor on [a]
`
`screen can be controlled” so that “when the finger is moved, the cursor is also
`
`moved” and a cursor position depends on a fingertip position. Id. at 26:8-14, 26:23-
`
`
`
`18
`
`

`

`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`25. A PHOSITA would have understood that processing such cursor control is
`
`similar to processing the “click and drag” functionality described with reference to
`
`the eighth embodiment and would have been motivated to implement the eighth
`
`embodiment to use the same processing functionality described in the second
`
`embodiment. Bederson Dec. (Ex. 1010), ¶¶ 45-46.
`
`v.
`
`Claim 5
`
`
`5. The method according to claim 1 wherein the detected3 gesture includes at
`least one of a pinch gesture, a pointing gesture, and a grip gesture.
`In the eighth embodiment, Numazaki teaches “the operator operating the
`
`keyboard can make [a] pointing or gesture input by slightly raising and moving
`
`the index finger” and consequently conduct a “click and drag” operation for
`
`“selecting and moving icons on the screen.” Numazaki (Ex. 1004), 50:38-48.
`
`Figure 74 illustrates such an operation:
`
`
`3 Claim 5 recites “the detected gesture” of claim 1, but the term “detected” lacks
`
`antecedent basis. For purposes of this petition, Petitioner assumes “the detected
`
`gesture” of claim 5 is a reference to Claim 1’s “determining, using the camera, the
`
`gesture performed in the work volume and illuminated by the light source.”
`
`
`
`19
`
`

`

`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`
`
`Id. at Fig. 74 (annotated to show pointing gesture).
`
`vi. Claim 6
`
`
`6. The method according to claim 1 further including determining the pointing
`direction of a finger in the work volume.
`In the eighth embodiment, Numazaki teaches “the operator operating the
`
`keyboard can make [a] pointing or gesture input by slightly raising and moving
`
`the index finger” and consequently conduct a “click and drag” operation for
`
`“selecting and moving icons on the screen.” Id. at 50:38-48. Any operator hand
`
`movements used as computer inputs are detectable within the volume of space
`
`
`
`20
`
`

`

`Numazaki refers to as a “range of illumination.” Figure 74 illustrates this
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`operation:
`
`
`
`
`Id. at Fig. 74 (annotated to show pointing gesture, icon selection, and range of
`
`illumination). A PHOSITA would have understood that determining the pointing
`
`direction of a finger in a work volume is necessary to implement the described “click
`
`and drag” feature. Bederson Dec. (Ex. 1010), ¶ 47 (explaining that the pointing
`
`direction of a finger is necessary to determine which icon a user intends to select and
`
`how far and in which direction the icon is to be moved).
`
`
`
`21
`
`

`

`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`vii. Claim 7
`
`
`7. The method according to claim 1 further including providing a target
`positioned on a user that is viewable in the work volume.
`As depicted in Figures 74 and 77, Numazaki’s eighth embodiment also
`
`teaches “the entire hand of [an] operator is illuminated, as can be seen from a
`
`dashed line circle indicating a range of illumination” Numazaki (Ex. 1004),
`
`50:35-37 (emphasis added).
`
`
`
`
`
`Id. at Figs. 74, 77 (annotated to illustrate the expanding range of illumination for
`
`capturing hand gestures). In this configuration, the user’s hand operates within, and
`
`is detectable within, an illuminated work volume. Numazaki discloses the user’s
`
`hand itself as a target object in the range of illumination:
`
`When the target object is a hand, it becomes possible to obtain the
`information regarding a gesture or a pointing according to the feature
`data extracted from the reflected light image of the hand . . . and it
`
`
`
`22
`
`

`

`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`becomes possible to operate a computer by using this obtained
`information.
`Id. at 10:57-66 (emphasis added). Numazaki discloses additional hand attributes as
`
`subsidiary targets, including color, material, and shape information. Id. at 16:45-61.
`
`The orientation of the hand also creates different targets with different distances
`
`extracted from the hand as a target object to help extract its overall “3D shape.” Id.
`
`at 12:27-45.
`
`Although the above disclosures in Numazaki utilize natural characteristics
`
`of a user’s hand to improve target detection, Numazaki also notes that it was
`
`known in the prior art to position a target on a user (i.e., something that is added
`
`to a user’s person) in order to improve target detection. For example, Numazaki
`
`notes that it was known to paint a fingertip or to wear a ring in a particular color
`
`to improve detection. Id. at 3:4-11. Numazaki, however, cautions that requiring
`
`users to wear or mount some external component may negatively impact the
`
`user’s convenience and may bring with it durability issues. Id. at 3:32-38. A
`
`PHOSITA would have understood, however, that the Fig. 74 arrangement
`
`described in the eighth embodiment is particularly well suited to a ring or other
`
`small target mounted on a user’s finger. Bederson Dec. (Ex. 1010), ¶¶ 48-49.
`
`Given the option of improved accuracy in exchange for the minor inconvenience
`
`of wearing a small ring or other hand-based target when using gesture recognition
`
`
`
`23
`
`

`

`while typing, a PHOSITA would have understood that many users would accept
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`this tradeoff. Id. Indeed, the durability concerns are implicated by a ring target,
`
`and many adults wear rings routinely while typing with no ill effect, which
`
`suggests that such a tradeoff would be acceptable to many users. Id.
`
`viii. Claim 8
`
`
`8. The method according to claim 1 further including determining the three-
`dimensional position of a point on a user.
`As noted above, Numazaki’s eighth embodiment uses the “information
`
`input generation apparatus” of Fig. 2 to detect and process a user’s gestures.
`
`Within this Fig. 2 apparatus, feature data generation unit 103 determines three-
`
`dimensional information representing the user’s hand:
`
`The feature data generation unit 103 extracts various feature data from
`the reflected light image. Many different types of the feature data and
`their extraction methods can be used in the present invention, as will be
`described in detail in the later embodiments. When the target object
`is a hand, it becomes possible to obtain the information regarding a
`gesture or a pointing according to the feature data extracted from the
`reflected light image of the hand, for example, and it becomes possible
`to operate a computer by using this obtained information. It is also
`possible to extract the 3D information on the target object for
`further utilization.
`Numazaki (Ex. 1004), 10:57-67 (emphasis added). A PHOSITA would have
`
`understood that detecting a three-dimensional representation of a user’s hand
`
`involves determining the three-dimensional position of at least one “point on a user”
`
`as claimed. Bederson Dec. (Ex. 1010), ¶ 50 (explaining that representing a hand in
`
`
`
`24
`
`

`

`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`three-dimensions necessarily involves ascertaining the three-dimensional position of
`
`a plurality of points on the user’s hand).
`
`Further, Numazaki uses distance- and region-based extraction methods to
`
`track the uniformity or homogeneity of a surface reflecting light to extract a 3D
`
`shape of a user’s hand (e.g., detect the uneven palm of a hand). Numazaki (Ex.
`
`1004), 16:34-44.
`
`ix. Claim 9
`
`
`9. The method according to claim 1 wherein the camera and the light source are
`positioned in fixed relation relative to a keypad.
`Numazaki’s eighth embodiment places the light source and camera above
`
`a keyboard as depicted in Figure 74 below:
`
`
`
`
`
`25
`
`

`

`Id. at Fig. 74 (annotated to show keyboard in relation to light source and camera).
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`Numazaki further teaches that the lighting and camera units are mounted
`
`in a fixed location just above the keyboard:
`
`This computer of FIG. 74 is a portable computer generally called note
`PC in which a keyboard and a display are integrally provided with the
`computer body. In this computer of FIG. 74, a lighting unit 701 and a
`photo-detection sensor unit 702 of the information input generation
`apparatus are provided at positions beyond the keyboard when
`viewed from an operator side, and arranged to have the optical axis of
`the photo-detection sections pointing obliquely upward towards the
`operator side.
`Id. at 50:25-35. This positioning ensures “the entire hand of the operator is
`
`illuminated” within the “dashed line circle . . . range of illumination.” Id. at 50:35-
`
`37.
`
`Even when using a keyboard separate from the display as illustrated in Fig.
`
`75, Numazaki teaches keeping a lighting unit 703 and photo-detection sensor unit
`
`704 at “positions in such a positional relationship with the keyboard that the
`
`light is irradiated onto the hand when t

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket