`
`
`UNITED STATES PATENT AND TRADEMARK OFFICE
`____________
`
`BEFORE THE PATENT TRIAL AND APPEAL BOARD
`____________
`
`
`GOOGLE LLC,
`Petitioner
`
`v.
`
`GESTURE TECHNOLOGY PARTNERS LLC,
`Patent Owner
`____________
`
`
`Case No. IPR2022-00360
`U.S. Patent No. 8,553,079
`____________
`
`
`
`
`
`
`
`
`PETITION FOR INTER PARTES REVIEW
`OF U.S. PATENT NO. 8,553,079
`
`
`
`
`
`TABLE OF CONTENTS
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`INTRODUCTION ......................................................................................... 1
`I.
`SUMMARY OF THE ’079 PATENT .......................................................... 1
`II.
`The ’079 Patent’s Alleged Invention ............................................................... 1
`A.
`The ’079 Patent’s Prosecution ......................................................................... 3
`B.
`A Person Having Ordinary Skill in the Art ..................................................... 4
`C.
`III. REQUIREMENTS FOR IPR UNDER 37 C.F.R. § 42.104 ....................... 4
`A.
`Standing Under 37 C.F.R. § 42.104(A) ........................................................... 4
`B.
`Challenge Under 37 C.F.R. § 42.104(B) and Relief Requested ...................... 4
`C.
`Claim Construction Under 37 C.F.R. § 42.104(B)(3) ..................................... 5
`IV. THE CHALLENGED CLAIMS ARE UNPATENTABLE ....................... 6
`A.
`Ground 1: Claims 1, 2, 4-14, 17, 19, 21-22, 24-28, and 30 are obvious under
`pre-AIA 35 U.S.C. § 103 over Numazaki in view of the knowledge of a
`PHOSITA ........................................................................................................ 6
`Ground 2: Claims 3, 15, and 23 are obvious under pre-AIA 35 U.S.C. § 103
`over Numazaki in view of Numazaki ’863 ....................................................34
`Ground 3: Claims 16 and 29 are obvious under pre-AIA 35 U.S.C. § 103
`over Numazaki in view of DeLuca ................................................................41
`Ground 4: Claim 18 is obvious under pre-AIA 35 U.S.C. § 103 over
`Numazaki in view of DeLeeuw ......................................................................48
`Ground 5: Claim 20 is obvious under pre-AIA 34 U.S.C. § 103 over
`Numazaki in view of Maruno ........................................................................55
`DISCRETIONARY CONSIDERATIONS ................................................68
`V.
`The Fintiv Factors Favor Institution ..............................................................68
`A.
`1. Whether the court granted a stay or evidence exists that one may be granted
`if a proceeding is instituted. .........................................................................69
`Proximity of the court’s trial date to the Board’s projected statutory
`deadline for a final written decision. ............................................................70
`Investment in the parallel proceeding by the court and the parties. ............74
`Overlap between issues raised in the petition and in the parallel proceeding.
` ......................................................................................................................75
`5. Whether the petitioner and the defendant in the parallel proceeding are the
`same party. ...................................................................................................76
`Other circumstances that impact the Board’s exercise of discretion,
`including the merits. .....................................................................................76
`
`B.
`
`C.
`
`D.
`
`E.
`
`2.
`
`3.
`4.
`
`6.
`
`
`
`i
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`VI. CONCLUSION ............................................................................................77
`VII. MANDATORY NOTICES UNDER 37 C.F.R. § 42.8(A)(1) ...................77
`A.
`Real Party-In-Interest.....................................................................................77
`B.
`Related Matters ..............................................................................................77
`C.
`Lead and Back-Up Counsel ...........................................................................78
`APPENDIX OF EXHIBITS ..................................................................................80
`
`
`
`
`
`ii
`
`
`
`INTRODUCTION
`Petitioner Google LLC (“Petitioner”) requests an Inter Partes Review
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`I.
`
`(“IPR”) of claims 1–30 (the “Challenged Claims”) of U.S. Patent No. 8,553,079
`
`(“the ’079 Patent”). This petition is substantively the same as IPR2021-00922
`
`(which is instituted), and is being filed concurrently with a motion for joinder
`
`with respect to that proceeding.
`
`II.
`
`SUMMARY OF THE ’079 PATENT
`The ’079 Patent’s Alleged Invention
`A.
`The ’079 Patent generally describes computer input devices employing
`
`cameras and lights to observe points on the human body and optically sense human
`
`positions and/or orientations.’079 Patent (Ex. 1001), 1:54-2:6. Examples of input
`
`devices contemplated by the patent include a computer keyboard, puzzle toy, and
`
`handheld computer. Id. at 2:15-31. Fig. 2 below illustrates one exemplary
`
`embodiment implemented in a laptop computer:
`
`
`
`1
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`
`Id. at Fig. 2. As illustrated, a laptop 138 may include camera locations 100, 101, 105,
`
`106, 108, 109; keyboard surface 102; screen housing 107; light 122; light emitting
`
`diodes (LEDs) 210 and 211, and work volume area 170 within which a user’s
`
`movements are detected. Id. at 2:39-53. The system can detect a user’s finger alone
`
`or the user may employ external objects such as ring 208 to help detect and recognize
`
`gestures performed in work volume area 170. Id. at 2:54-3:8. The ’079 Patent
`
`
`
`2
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`describes detecting point, pinch, and grip gestures using this configuration. Id. at
`
`3:48-51.
`
`The ’079 Patent’s Prosecution
`B.
`The Application that resulted in the ’079 Patent was filed on December 14,
`
`2012. The Application claims priority to provisional patent application No.
`
`60/107,652, filed November 9, 1998, through a succession of continuation
`
`applications 09/433,297; 10/866,191; and 12/700,055. Id. For purposes of this
`
`petition and without waiving its right to challenge priority in this or any other
`
`proceeding, Petitioner adopts November 9, 1998 as the invention date for the
`
`Challenged Claims.
`
`Applicant canceled twenty originally filed claims (1–20) by preliminary
`
`amendment and replaced them with thirty new claims (21–50). ’079 File History
`
`(Ex. 1002), 134-138. A Notice of Allowance issued on July 24, 2013, in which
`
`the Examiner allowed all thirty claims (renumbered 1–30) in a first office action.
`
`Id. at 150, 175. Examiner amended the abstract and reasoned that none of Naoi et
`
`al. (US 5459793), Platzker et al. (US 5528263), Sellers (US 5864334 A), nor
`
`Fukushima et al. (US 6346929 B1) taught or suggested the independently claimed
`
`elements: (1) providing a camera oriented to observe a gesture performed in the
`
`work volume; (2) the camera being fixed relative to the light source; and (3)
`
`
`
`3
`
`
`
`determining, using the camera, the gesture performed in the work volume and
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`illuminated by the light source. Id. at 153.
`
`C. A Person Having Ordinary Skill in the Art
`A person having ordinary skill in the art (“PHOSITA”) at the time of the
`
`’079 Patent would have had at least a bachelor’s degree in electrical engineering
`
`or equivalent with at least one year of experience in the field of human computer
`
`interaction. Additional education or experience might substitute for the above
`
`requirements. Bederson Dec. (Ex. 1010), ¶¶ 29-31.
`
`III. REQUIREMENTS FOR IPR UNDER 37 C.F.R. § 42.104
`Standing Under 37 C.F.R. § 42.104(A)
`A.
`Petitioner certifies that the ’079 Patent is available for IPR and that
`
`Petitioner is not barred or estopped from requesting an IPR challenging the claims
`
`of the ’079 Patent. Specifically, (1) Petitioner is not the owner of the ’079 Patent,
`
`(2) Petitioner has not filed a civil action challenging the validity of any claim of
`
`the ’079 Patent, and (3) Petitioner has not been served with a complaint more than
`
`one year ago alleging infringement of the ’079 patent.
`
`B. Challenge Under 37 C.F.R. § 42.104(B) and Relief Requested
`In view of the prior art and evidence presented, claims 1-30 of the ’079
`
`Patent are unpatentable and should be cancelled. 37 C.F.R. § 42.104(b)(1).
`
`
`
`4
`
`
`
`Further, based on the prior art references identified below, IPR of the Challenged
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`Claims should be granted. 37 C.F.R. § 42.104(b)(2).
`
`Proposed Ground of Unpatentability
`Ground 1: Claims 1, 2, 4-14, 17, 19, 21-22, 24-28, and 30 are
`obvious under pre-AIA 35 U.S.C. § 103 over U.S. Patent No.
`6,144,366 (“Numazaki”) in view of the knowledge of a PHOSITA
`Ground 2: Claims 3, 15, and 23 are obvious under pre-AIA 35 USC
`§ 103 over Numazaki in view of U.S. Patent No. 5,900,863
`(“Numazaki ’863”)
`Ground 3: Claims 16 and 29 are obvious under pre-AIA 35 U.S.C §
`103 over Numazaki in view of U.S. Patent No. 6,064,354 (“DeLuca”)
`Ground 4: Claim 18 is obvious under pre-AIA 35 U.S.C § 103 over
`Numazaki in view of U.S. Patent No. 6,088,018 (“DeLeeuw”)
`Ground 5: Claim 20 is obvious under pre-AIA 35 USC § 103 over
`Numazaki in view of U.S. Patent No. 6,191,773 (“Maruno”)
`
`Exhibits
`Ex. 1004
`
`Ex. 1004,
`Ex. 1005
`
`Ex. 1004,
`Ex. 1006
`Ex. 1004,
`Ex. 1007
`Ex. 1004,
`Ex. 1008
`
`Section IV identifies where each element of the Challenged Claims is
`
`found in the prior art. 37 C.F.R. § 42.104(b)(4). The exhibit numbers of the
`
`evidence relied upon to support the challenges are provided above and the
`
`relevance of the evidence to the challenges raised is provided in Section IV. 37
`
`C.F.R. § 42.104(b)(5). Exhibits 1001-1017 are also attached.
`
`C. Claim Construction Under 37 C.F.R. § 42.104(B)(3)
`In this proceeding, claims are interpreted under the same standard applied
`
`by Article III courts (i.e., the Phillips standard). See 37 C.F.R § 42.100(b); see
`
`also 83 Fed. Reg. 197 (Oct. 11, 2018); Phillips v. AWH Corp., 415 F.3d 1303,
`
`1312 (Fed. Cir. 2005) (en banc). Under this standard, words in a claim are given
`
`their plain meaning which is the meaning understood by a person of ordinary skill
`
`
`
`5
`
`
`
`in the art in view of the patent and file history. Phillips, 415 F.3d 1303, 1212-13.
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`For purposes of the proposed grounds below, Petitioner proposes no terms require
`
`express construction.
`
`IV. THE CHALLENGED CLAIMS ARE UNPATENTABLE
`A. Ground 1: Claims 1, 2, 4-14, 17, 19, 21-22, 24-28, and 30 are
`obvious under pre-AIA 35 U.S.C. § 103 over Numazaki in view of
`the knowledge of a PHOSITA
`Overview of Numazaki
`
`U.S. Patent No. 6,144,366 to Numazaki et al. (“Numazaki”) (Ex. 1004) was
`
`filed on October 17, 1997 and is prior art to the ’079 Patent under at least 35 U.S.C.
`
`§ 102(e) (pre-AIA). Numazaki was not cited or considered during prosecution of
`
`the ’079 Patent or its parent, U.S. Patent No. 6,750,848. ’079 Patent (Ex. 1001);
`
`’848 Patent (Ex. 1003).
`
`Numazaki is generally directed to a method for detecting a gesture or the
`
`movement of a user’s hand. (“Numazaki”) (Ex. 1004), Abstract, 4:9-40.
`
`Numazaki purports to have improved upon prior methods by using a controlled
`
`light source to illuminate the target object (e.g., the user’s hand), a first camera
`
`
`
`6
`
`
`
`unit (referred to by Numazaki as a “photo-detection unit”),1 and a second camera
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`unit. Id. at 11:9-23. This arrangement is illustrated in Fig. 2 below:
`
`
`
`Id. at Fig. 2. A timing control unit is used to turn lighting unit 101 on (i.e.,
`
`illuminating the target object) when the first camera unit is active and off when the
`
`second camera unit is active. Id. at 11:20-32. The result of this light control is the
`
`first camera unit captures an image of the target object illuminated by both natural
`
`light and the lighting unit 101 and the second camera unit captures an image of the
`
`
`1 A PHOSITA would have considered Numazaki’s photo-detection units to be
`
`camera units. Bederson Dec. (Ex. 1010), ¶ 36 (explaining that Numazaki describes
`
`using CMOS or CCD sensor units, which were two of the more common optical
`
`sensors used in camera units at the time).
`
`
`
`7
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`target object illuminated by only natural light. Id. at 11:33-39. The difference
`
`between the two images—obtained by difference calculation unit 111—represents
`
`the “reflected light from the object resulting from the light emitted by the lighting
`
`unit 101.” Id. at 11:43-51. This information is then used by feature data generation
`
`unit 103 to determine gestures, pointing, etc. of the target object that may be
`
`converted into commands executed by a computer. Id. at 10:57-66.
`
`In its eighth embodiment, Numazaki describes implementing this structure
`
`in a computer such that a user can point or gesture with an index finger while
`
`typing on the keyboard “with[] hardly any shift of the hand position.” Id. at 50:25-
`
`43. This arrangement is illustrated in Fig. 74 below:
`
`
`
`8
`
`
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`Id. at Figure 74 (annotated to indicate light source 701 and photo-detection sensor
`
`unit 702). Numazaki teaches the entirety of the operator’s hand is illuminated within
`
`the range depicted by the dashed circle such that the user’s pointing and gestures can
`
`be captured and converted to commands. Id. at 50:38-48.
`
`Numazaki
`
`teaches
`
`that
`
`its eighth embodiment
`
`incorporates “the
`
`information input generation apparatus of the present invention as described in
`
`the above embodiments.” Id. at 50:21-24. A PHOSITA would have understood
`
`that the referenced information input generation apparatus is that illustrated in
`
`Fig. 2 and described in the corresponding disclosure. Bederson Dec. (Ex. 1010),
`
`¶¶ 42-43 (explaining that Numazaki describes its controlled light and two-camera
`
`configuration as key to its invention and noting that Numazaki at 53:22-36 teaches
`
`that the eighth embodiment uses the precise image difference calculation taught
`
`by Fig. 2 and its corresponding disclosure).
`
`Because Numazaki, like the ’079 Patent, discloses a method and apparatus
`
`for generating computer input information by capturing hand gestures, Numazaki
`
`is in the same field of endeavor as the ’079 Patent. Compare Numazaki (Ex.
`
`1004), 50:29-37 (“In this computer of FIG. 74, a lighting unit 701 and a photo-
`
`detection sensor unit 702 of the information input generation apparatus are
`
`provided at positions beyond the keyboard” such that the “entire hand of the
`
`operator is illuminated . . . [in] a range of illumination”) with ’079 Patent (Ex
`
`
`
`9
`
`
`
`1001), Abstract (describing “[a] method for determining a gesture illuminated by
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`a light source” that “utilizes the light source to provide illumination through a
`
`work volume above the light source” where a “camera is positioned to observe
`
`and determine the gesture performed in the work volume”). Numazaki is therefore
`
`analogous art to the ’079 Patent. Bederson Dec. (Ex. 1010), ¶ 41.
`
`i.
`
`Claim 1.
`
`1[P] A computer implemented method comprising:
`
`To the extent the preamble is limiting, Numazaki teaches a computer
`
`implemented method. In particular, Numazaki teaches a “method . . . for
`
`generating information input . . . capable of realizing a direct command type
`
`information input scheme by which the gesture or the motion can be inputted
`
`easily.” Numazaki (Ex. 1004), 4:9-13.
`
`Numazaki’s eighth embodiment teaches a computer implemented method
`
`for controlling functions on a laptop device through gestures or pointing:
`
`In this configuration, the operator operating the keyboard can make the
`
`pointing or gesture input by slightly raising and moving the index
`
`finger. The user’s convenience is remarkably improved here because
`
`the keyboard input and the pointing or gesture input can be made
`
`without hardly any shift of the hand position.
`
`Id. at 50:38-43. This arrangement is illustrated in Fig. 74 below:
`
`
`
`10
`
`
`
`
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`
`
`Id. at Fig. 74.
`
`[1(a)] providing a light source adapted to direct illumination through a work
`volume above the light source;
`Numazaki teaches a lighting unit 701 (i.e., light source) that is adapted to
`
`illuminate a user’s hand (i.e., human body part) within a work volume generally
`
`above the light source. For example, Numazaki teaches “the entire hand of the
`
`operator is illuminated, as can be seen from a dashed line circle indicating a
`
`range of illumination.” Id. at 50:35-37 (emphasis added). As depicted in Fig. 74
`
`below, the lighting unit 701 is adapted to illuminate the user’s hand by positioning
`
`
`
`11
`
`
`
`it “beyond the keyboard” to illuminate a work volume indicated by the dashed
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`line circle above it:
`
`
`Id. at Fig. 74 (annotated to illustrate light source and range of hand illumination).
`
`[1(b)] providing a camera oriented to observe a gesture performed in the work
`volume, the camera being fixed relative to the light source; and determining,
`using the camera, the gesture performed in the work volume and illuminated by
`the light source.
`Numazaki discloses a photo-detection sensor unit 702 (i.e., camera2) that
`
`is positioned next to lighting unit 701 (i.e., in fixed location relative to the light
`
`
`2 As described in the overview section above, a PHOSITA would have considered
`
`Numazaki’s photo-detection units to be camera units. Bederson Dec. (Ex. 1010), ¶
`
`
`
`12
`
`
`
`source) and arranged to have the optical axis of the photo-detection sections
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`pointing obliquely upward towards the operator side to observe a gesture
`
`performed by the user (i.e., observe a gesture) within a “dashed line circle” (i.e.,
`
`work volume) depicted below. Id. at 50:30-43. This arrangement is illustrated in
`
`Fig. 74 below:
`
`
`Id. at Fig. 74 (annotated to show camera).
`
`As described with reference to Fig. 2, which is incorporated into the eighth
`
`embodiment as noted above, Numazaki uses this light and camera arrangement to
`
`
`36 (explaining that Numazaki describes using CMOS or CCD sensor units, which
`
`were two of the more common optical sensors used in camera units at the time).
`
`
`
`13
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`illuminate the target object (e.g., the user’s hand) in a controlled manner such that
`
`a precise image of the user’s hand and hand movement can be ascertained. Id. at
`
`11:9-23. Specifically, a timing control unit is used to turn lighting unit 101 on (i.e.,
`
`illuminating the target object) when the first camera unit is active and off when the
`
`second camera unit is active. Id. at 11:20-32. The result of this light control is the
`
`first camera unit captures an image of the target object illuminated by both natural
`
`light and the lighting unit 101 and the second camera unit captures an image of the
`
`target object illuminated by only natural light. Id. at 11:33-39. The difference
`
`between the two images – obtained by difference calculation unit 111 – represents
`
`the “reflected light from the object resulting from the light emitted by the lighting
`
`unit 101.” Id. at 11:43-51. This information is then used by feature data generation
`
`unit 103 to determine gestures, pointing, etc. of the target object that may be
`
`converted into commands executed by a computer. Id. at 10:57-66. Through this
`
`arrangement, a PHOSITA would have understood that component 702 illustrated
`
`in Fig. 2 is “oriented to observe a gesture performed in the work volume”
`
`illuminated by lighting unit 701. Bederson Dec. (Ex. 1010), ¶¶ 42-43.
`
`
`
`14
`
`
`
`ii.
`
`Claim 2
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`2. The method according to claim 1 wherein the light source includes a light
`emitting diode.
`
`As discussed above with reference to limitation 1[a], Numazaki teaches light
`
`source 701 in the embodiment depicted in Fig. 74 below:
`
`
`
`Numazaki (Ex. 1004), Fig. 74; see also limitation 1[a], supra. Although Numazaki
`
`does not discuss the specific lighting technology contemplated for light source 701,
`
`with reference to the first embodiments, Numazaki teaches that an “LED can be used
`
`as the light source” since “the LED has a property that it can emit more intense light
`
`instantaneously” while “reduc[ing] the power required for the light emission.” Id. at
`
`14:49-56; see also id. at 68:13-20. Based on Numazaki’s teachings, a PHOSITA
`
`
`
`15
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`would have been motivated to implement light source 701 in Fig. 74 using LED
`
`technology for a number of reasons. Bederson Dec. (Ex. 1010), ¶ 44. First, the
`
`benefit of emitting more intense light instantaneously described with reference to
`
`the first embodiment would have improved Fig. 74’s apparatus such that it could
`
`quickly and accurately detect pointing and gestures. Id; see also KSR Int’l Co. v.
`
`Teleflex Inc., 550 U.S. 398, 417 (2007) (obvious to use known techniques to improve
`
`similar devices in the same way). Second, a PHOSITA would have understood that
`
`the “note PC” depicted in Fig. 74 is a self-contained and portable unit that would
`
`have benefitted from the power reduction of an LED light source discussed with
`
`reference to the first embodiment. Bederson Dec. (Ex. 1010), ¶ 44 (noting that
`
`Numazaki’s eighth embodiment contemplates other portable implementations and,
`
`at col. 52, ln. 33 – col. 53, ln. 7, expressly discusses the desire to conserve power
`
`and techniques for doing so by controlling the lighting unit).
`
`iv.
`
`Claim 4
`
`
`4. The method according to claim 1 wherein detecting a gesture includes
`analyzing sequential images of the camera.
`Numazaki’s eighth embodiment analyzes the sequential movement of
`
`“click and drag” operations by detecting a combination of “pointing or gesture
`
`input” and “keyboard input” via a “button for use in conjunction with the pointing
`
`
`
`16
`
`
`
`or gesture input.” Numazaki (Ex. 1004), 50:38-47. This permits “selecting and
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`moving icons on the screen.” Id. at 50:45-47.
`
`A PHOSITA would have understood that using gesture recognition to
`
`implement a “click and drag” operation would have been implemented by
`
`capturing a sequence of images to determine the user’s hand movement, which
`
`dictates the “drag” operation to be performed. Bederson Dec. (Ex. 1010), ¶¶ 45-
`
`46. Indeed, in its second embodiment, Numazaki expressly describes a process
`
`through which the system tracks lateral finger movements by detecting the center
`
`of gravity of a finger, where “finger tip movement and the center of gravity
`
`movement can be smoothly correlated” using pixel values. Numazaki (Ex. 1004),
`
`19:43-20:25. Fig. 10 below illustrates the process by which the reflected light
`
`image of a hand and finger are mapped to a pixelated target space, and Fig. 11
`
`illustrates how coordinate-based finger movement can be tracked on the basis of
`
`pixel value changes:
`
`
`
`17
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`Id. at Fig. 10, 19:13-42 (describing the same).
`
`
`
`
`
`Id. at Fig. 11A, 19:43-20:25 (describing a center of gravity tracking that enables a
`
`fingertip to be tracked). Using this technique, Numazaki teaches “the cursor on [a]
`
`screen can be controlled” so that “when the finger is moved, the cursor is also
`
`moved” and a cursor position depends on a fingertip position. Id. at 26:8-14, 26:23-
`
`
`
`18
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`25. A PHOSITA would have understood that processing such cursor control is
`
`similar to processing the “click and drag” functionality described with reference to
`
`the eighth embodiment and would have been motivated to implement the eighth
`
`embodiment to use the same processing functionality described in the second
`
`embodiment. Bederson Dec. (Ex. 1010), ¶¶ 45-46.
`
`v.
`
`Claim 5
`
`
`5. The method according to claim 1 wherein the detected3 gesture includes at
`least one of a pinch gesture, a pointing gesture, and a grip gesture.
`In the eighth embodiment, Numazaki teaches “the operator operating the
`
`keyboard can make [a] pointing or gesture input by slightly raising and moving
`
`the index finger” and consequently conduct a “click and drag” operation for
`
`“selecting and moving icons on the screen.” Numazaki (Ex. 1004), 50:38-48.
`
`Figure 74 illustrates such an operation:
`
`
`3 Claim 5 recites “the detected gesture” of claim 1, but the term “detected” lacks
`
`antecedent basis. For purposes of this petition, Petitioner assumes “the detected
`
`gesture” of claim 5 is a reference to Claim 1’s “determining, using the camera, the
`
`gesture performed in the work volume and illuminated by the light source.”
`
`
`
`19
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`
`
`Id. at Fig. 74 (annotated to show pointing gesture).
`
`vi. Claim 6
`
`
`6. The method according to claim 1 further including determining the pointing
`direction of a finger in the work volume.
`In the eighth embodiment, Numazaki teaches “the operator operating the
`
`keyboard can make [a] pointing or gesture input by slightly raising and moving
`
`the index finger” and consequently conduct a “click and drag” operation for
`
`“selecting and moving icons on the screen.” Id. at 50:38-48. Any operator hand
`
`movements used as computer inputs are detectable within the volume of space
`
`
`
`20
`
`
`
`Numazaki refers to as a “range of illumination.” Figure 74 illustrates this
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`operation:
`
`
`
`
`Id. at Fig. 74 (annotated to show pointing gesture, icon selection, and range of
`
`illumination). A PHOSITA would have understood that determining the pointing
`
`direction of a finger in a work volume is necessary to implement the described “click
`
`and drag” feature. Bederson Dec. (Ex. 1010), ¶ 47 (explaining that the pointing
`
`direction of a finger is necessary to determine which icon a user intends to select and
`
`how far and in which direction the icon is to be moved).
`
`
`
`21
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`vii. Claim 7
`
`
`7. The method according to claim 1 further including providing a target
`positioned on a user that is viewable in the work volume.
`As depicted in Figures 74 and 77, Numazaki’s eighth embodiment also
`
`teaches “the entire hand of [an] operator is illuminated, as can be seen from a
`
`dashed line circle indicating a range of illumination” Numazaki (Ex. 1004),
`
`50:35-37 (emphasis added).
`
`
`
`
`
`Id. at Figs. 74, 77 (annotated to illustrate the expanding range of illumination for
`
`capturing hand gestures). In this configuration, the user’s hand operates within, and
`
`is detectable within, an illuminated work volume. Numazaki discloses the user’s
`
`hand itself as a target object in the range of illumination:
`
`When the target object is a hand, it becomes possible to obtain the
`information regarding a gesture or a pointing according to the feature
`data extracted from the reflected light image of the hand . . . and it
`
`
`
`22
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`becomes possible to operate a computer by using this obtained
`information.
`Id. at 10:57-66 (emphasis added). Numazaki discloses additional hand attributes as
`
`subsidiary targets, including color, material, and shape information. Id. at 16:45-61.
`
`The orientation of the hand also creates different targets with different distances
`
`extracted from the hand as a target object to help extract its overall “3D shape.” Id.
`
`at 12:27-45.
`
`Although the above disclosures in Numazaki utilize natural characteristics
`
`of a user’s hand to improve target detection, Numazaki also notes that it was
`
`known in the prior art to position a target on a user (i.e., something that is added
`
`to a user’s person) in order to improve target detection. For example, Numazaki
`
`notes that it was known to paint a fingertip or to wear a ring in a particular color
`
`to improve detection. Id. at 3:4-11. Numazaki, however, cautions that requiring
`
`users to wear or mount some external component may negatively impact the
`
`user’s convenience and may bring with it durability issues. Id. at 3:32-38. A
`
`PHOSITA would have understood, however, that the Fig. 74 arrangement
`
`described in the eighth embodiment is particularly well suited to a ring or other
`
`small target mounted on a user’s finger. Bederson Dec. (Ex. 1010), ¶¶ 48-49.
`
`Given the option of improved accuracy in exchange for the minor inconvenience
`
`of wearing a small ring or other hand-based target when using gesture recognition
`
`
`
`23
`
`
`
`while typing, a PHOSITA would have understood that many users would accept
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`this tradeoff. Id. Indeed, the durability concerns are implicated by a ring target,
`
`and many adults wear rings routinely while typing with no ill effect, which
`
`suggests that such a tradeoff would be acceptable to many users. Id.
`
`viii. Claim 8
`
`
`8. The method according to claim 1 further including determining the three-
`dimensional position of a point on a user.
`As noted above, Numazaki’s eighth embodiment uses the “information
`
`input generation apparatus” of Fig. 2 to detect and process a user’s gestures.
`
`Within this Fig. 2 apparatus, feature data generation unit 103 determines three-
`
`dimensional information representing the user’s hand:
`
`The feature data generation unit 103 extracts various feature data from
`the reflected light image. Many different types of the feature data and
`their extraction methods can be used in the present invention, as will be
`described in detail in the later embodiments. When the target object
`is a hand, it becomes possible to obtain the information regarding a
`gesture or a pointing according to the feature data extracted from the
`reflected light image of the hand, for example, and it becomes possible
`to operate a computer by using this obtained information. It is also
`possible to extract the 3D information on the target object for
`further utilization.
`Numazaki (Ex. 1004), 10:57-67 (emphasis added). A PHOSITA would have
`
`understood that detecting a three-dimensional representation of a user’s hand
`
`involves determining the three-dimensional position of at least one “point on a user”
`
`as claimed. Bederson Dec. (Ex. 1010), ¶ 50 (explaining that representing a hand in
`
`
`
`24
`
`
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`three-dimensions necessarily involves ascertaining the three-dimensional position of
`
`a plurality of points on the user’s hand).
`
`Further, Numazaki uses distance- and region-based extraction methods to
`
`track the uniformity or homogeneity of a surface reflecting light to extract a 3D
`
`shape of a user’s hand (e.g., detect the uneven palm of a hand). Numazaki (Ex.
`
`1004), 16:34-44.
`
`ix. Claim 9
`
`
`9. The method according to claim 1 wherein the camera and the light source are
`positioned in fixed relation relative to a keypad.
`Numazaki’s eighth embodiment places the light source and camera above
`
`a keyboard as depicted in Figure 74 below:
`
`
`
`
`
`25
`
`
`
`Id. at Fig. 74 (annotated to show keyboard in relation to light source and camera).
`
`IPR2022-00360
`U.S. Patent No. 8,553,079
`
`
`Numazaki further teaches that the lighting and camera units are mounted
`
`in a fixed location just above the keyboard:
`
`This computer of FIG. 74 is a portable computer generally called note
`PC in which a keyboard and a display are integrally provided with the
`computer body. In this computer of FIG. 74, a lighting unit 701 and a
`photo-detection sensor unit 702 of the information input generation
`apparatus are provided at positions beyond the keyboard when
`viewed from an operator side, and arranged to have the optical axis of
`the photo-detection sections pointing obliquely upward towards the
`operator side.
`Id. at 50:25-35. This positioning ensures “the entire hand of the operator is
`
`illuminated” within the “dashed line circle . . . range of illumination.” Id. at 50:35-
`
`37.
`
`Even when using a keyboard separate from the display as illustrated in Fig.
`
`75, Numazaki teaches keeping a lighting unit 703 and photo-detection sensor unit
`
`704 at “positions in such a positional relationship with the keyboard that the
`
`light is irradiated onto the hand when t