throbber

`
`UNITED STATES PATENT AND TRADEMARK OFFICE
`
`BEFORE THE PATENT TRIAL AND APPEAL BOARD
`
`APPLE INC.
`GOOGLE LLC
`Petitioner
`
`v.
`
`GESTURE TECHNOLOGY PARTNERS LLC
`Patent Owner
`
`Case No. IPR2021-00921IPR2022-00362
`U.S. Patent No. 8,878,949
`
`PETITION FOR INTER PARTES REVIEW OF
`U.S. PATENT NO. 8,878,949
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`Page 1 of 77
`
`GOOGLE EXHIBIT 1018
`
`

`

`
`
`TABLE OF CONTENTS
`
`IPR2021-00921
`IPR2022-00362
`U.S. Patent No. 8,878,949
`
`I.
`II.
`
`1
`INTRODUCTION
`1
`SUMMARY OF THE ’949 PATENT
`1
`A.
`The ’949 Patent’s Alleged Invention
`2
`B.
`The ’949 Patent’s Prosecution
`4
`C. Overview of the Proposed Grounds
`5
`D. A Person Having Ordinary Skill in the Art
`6
`III. REQUIREMENTS FOR IPR UNDER 37 C.F.R. § 42.104
`6
`A.
`Standing Under 37 C.F.R. § 42.104(A)
`6
`B.
`Challenge Under 37 C.F.R. § 42.104(B) and Relief Requested
`7
`C.
`Claim Construction Under 37 C.F.R. § 42.104(B)(3)
`10
`IV. THE CHALLENGED CLAIMS ARE UNPATENTABLE
`A.
`Ground 1: Claims 1-18 are obvious under pre-AIA 35 U.S.C. § 103
`over Numazaki in view of Nonaka
`10
`B. Ground 2: Claims 6, 12, and 17 are obvious under pre-AIA 35 U.S.C.
`§ 103 over Numazaki in view of Nonaka and in further view of Aviv 50
`V. DISCRETIONARY CONSIDERATIONS
`57
`A.
`The Fintiv Factors Favor Institution
`57
`VI. CONCLUSION
`68
`VII. MANDATORY NOTICES UNDER 37 C.F.R. § 42.8(A)(1)
`69
`A.
`Real Party-In-Interest
`69
`B.
`Related Matters
`69
`C.
`Lead and Back-Up Counsel
`70
`
`
`
`
`
`
`
`
`i
`
`Page 2 of 77
`
`

`

`U.S. Patent No. 8,878,949
`I.
`INTRODUCTION
`Petitioner Apple Inc.Google LLC (“Petitioner”) requests an Inter Partes
`
`IPR2021-00921
`
`Review (“IPR”) of claims 1–18 (the “Challenged Claims”) of U.S. Patent No.
`
`8,878,949 (“the ’949 Patent”). This petition is substantively the same as IPR2021-
`
`00921 (which is instituted), and is being filed concurrently with a motion for joinder
`
`with respect to that proceeding.
`
`II.
`
`SUMMARY OF THE ’949 PATENT
`A. The ’949 Patent’s Alleged Invention
`Generally directed to digital imaging, the ’949 Patent seeks to automate the
`
`process of taking a picture by analyzing the scene and capturing an image when
`
`“certain poses of objects, sequences of poses, motions of objects, or any other states
`
`or relationships of objects are represented.” ’949 Patent (Ex. 1001), 1:50-2:8. The
`
`patent describes a number of different scenarios that, when detected, cause the
`
`camera to capture an image. Some examples include detecting (1) a “[s]ubject in a
`
`certain pose,” (2) a “[s]ubject in a sequence of poses,” (3) a “[p]ortion of [s]ubject
`
`in a sequence of poses (e.g., gestures),” (4) a “[s]ubject or portion(s) in a specific
`
`location or orientation,” (5) a “[s]ubject in position relative to another object or
`
`person” such as a “bride and groom kissing in a wedding,” and (6) “a subject
`
`undertak[ing] a particular signal comprising a position or gesture” such as “raising
`
`one’s right hand.” Id. at 5:30-49. Only gestures are claimed, however. Each of the
`
`
`
`Page 3 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`Challenged Claims requires detecting or determining a “gesture has been
`
`performed.” Id. at Independent Claims 1, 8, 13.
`
`The ’949 Patent contemplates multiple image sensors to accomplish its goal.
`
`For example, a “central camera . . . is for picture taking and has high resolution and
`
`color accuracy,” while “lower resolution” cameras “with little or no accurate color
`
`capability . . . are used to simply see object positions.” Id. at 5:1-6. Although the
`
`term is not used outside the claims, all Challenged Claims refer to the gesture-
`
`capturing sensor as an “electro-optical sensor.” Id. at Independent Claims 1, 8, 13.
`
`B.
`The ’949 Patent’s Prosecution
`The Application that resulted in the ’949 Patent was filed on August 7, 2013.
`
`The Application claims priority to provisional patent application No. 60/133,671,
`
`filed May 11, 1999. Id. at (22), (60). For purposes of this petition and without
`
`waiving its right to challenge priority in this or any other proceeding, Petitioner
`
`adopts May 11, 1999 as the invention date for the Challenged Claims.
`
`A first office action rejected all initially presented claims as anticipated or
`
`obvious over U.S. Patent No. 6,359,647 to Sengupta et al. (“Sengupta”). ’949 File
`
`History (Ex. 1002), 136-144. The examiner noted that Sengupta teaches an electro-
`
`optical sensor separate from a digital camera, which triggers an image capture when
`
`it detects movement within the sensor’s field of view. Id. at 140-141.
`
`
`
`2
`
`Page 4 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`In response, the Applicant characterized Sengupta as a system comprising
`
`multiple security cameras that transitions to an appropriate camera when an object
`
`moves from one camera’s field of view to another’s. Id. at 167-168. Focusing on
`
`structural distinctions, the Applicant argued that Sengupta did not teach “a device
`
`housing including a forward facing portion having an electro-optical sensor and a
`
`digital camera” as required by Claim 1 and its dependents. Id. at 168. The Applicant
`
`drew a functional distinction with respect to the claims that ultimately issued as
`
`independent Claims 8 and 13 (and their dependents), arguing Sengupta does not
`
`“identify a particular gesture apart from a plurality of gestures, where the particular
`
`gesture corresponds to an image capture command.” Id. at 169-170.
`
`A second office action rejected the Applicant’s alleged distinctions, finding
`
`the structural point was “not clearly defined in claim 1” and “the term ‘gesture’ []
`
`not clearly defined in the claim[s]” to support the purported distinction regarding
`
`independent Claims 8 and 13. Id. at 186. Following an examiner interview on August
`
`7, 2014 (Id. at 199), the Applicant further amended the claims to distinguish the
`
`claimed invention from Sengupta. Id. at 210-217. The Applicant noted, “[w]ith
`
`respect to [the] amended independent claims . . . , Sengupta does not disclose, teach
`
`or suggest: a) a device housing including a forward facing portion that encompasses
`
`an electro-optical sensor and a digital camera; or b) a processor to determine a
`
`
`
`3
`
`Page 5 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`gesture corresponds to an image capture command, which causes the camera to store
`
`or capture an image.” Id. at 215. Regarding the claimed “gesture,” the Applicant
`
`distinguished Sengupta’s generic movement tracking from the claimed invention,
`
`which must “identif[y] a particular gesture” that triggers the device to capture an
`
`image. Id. at 216.
`
`In response, the Examiner held the pending claims allowable, and the
`
`application was deemed allowable on September 18, 2014 after multiple double
`
`patenting issues were resolved. Id. at 254-260.
`
`C. Overview of the Proposed Grounds
`Neither the camera structure—multiple forward-facing image sensors—nor
`
`the concept of detecting a specific gesture that causes a camera to capture an image
`
`were new. As set forth in detail below, U.S. Patent 6,144,366 to Numazaki et al.
`
`(“Numazaki”) teaches portable devices equipped with multiple forward-facing
`
`image sensors, detects gestures performed by users, connects those gestures with
`
`specific commands, and captures images with a separate sensor from those used to
`
`detect gestures. Although Numazaki does not expressly contemplate detecting a
`
`gesture that causes an image capture command to be processed, this was also known
`
`in the art. Namely, JPH4-73631 to Osamu Nonaka (“Nonaka”) (a Japanese
`
`Unexamined Patent Application Publication filed by Olympus Optical Co., Ltd.)
`
`
`
`4
`
`Page 6 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`teaches a process by which “the photographer gives a release instruction” for the
`
`camera to capture an image by performing a “predetermined motion towards the
`
`camera.” Nonaka (Ex. 1005), 3:34-36. The camera “detects this motion by the
`
`subject [] and exposure is carried out.” Id. at 37-38. Nonaka contemplates multiple
`
`predetermined gestures such as holding one’s hand out toward the camera (as
`
`depicted in Fig. 3 below left) or moving one’s hand toward the camera (as depicted
`
`in Fig. 7 below right).
`
`
`
`
`Id. at Figs. 3, 7. For the reasons discussed in the proposed grounds below, a
`
`PHOSITA would have been motivated to implement such image capture gesture
`
`functionality in Numazaki’s gesture-recognizing, multi-camera devices.
`
`D. A Person Having Ordinary Skill in the Art
`A person having ordinary skill in the art (“PHOSITA”) at the time of the ’949
`
`Patent would have had at least a bachelor’s degree in electrical engineering or
`
`equivalent with at least one year of experience in the field of human computer
`
`
`
`5
`
`Page 7 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`interaction. Additional education or experience might substitute for the above
`
`requirements. Bederson Dec. (Ex. 1003), ¶¶ 29-31.
`
`III. REQUIREMENTS FOR IPR UNDER 37 C.F.R. § 42.104
`A.
`Standing Under 37 C.F.R. § 42.104(A)
`Petitioner certifies that the ’949 Patent is available for IPR and that Petitioner
`
`is not barred or estopped from requesting an IPR challenging the claims of the ’949
`
`Patent. Specifically, (1) Petitioner is not the owner of the ’949 Patent, (2) Petitioner
`
`has not filed a civil action challenging the validity of any claim of the ’949 Patent,
`
`and (3) this Petition is filed less than one year after the Petitioner washas not been
`
`served with a complaint more than one year ago alleging infringement of the ’949
`
`Patentpatent.
`
`B. Challenge Under 37 C.F.R. § 42.104(B) and Relief Requested
`In view of the prior art and evidence presented, claims 1–18 of the ’949 Patent
`
`are unpatentable and should be cancelled. 37 C.F.R. § 42.104(b)(1). Further, based
`
`on the prior art references identified below, IPR of the Challenged Claims should be
`
`granted. 37 C.F.R. § 42.104(b)(2).
`
`Proposed Ground of Unpatentability
`Ground 1: Claims 1-18 are obvious under pre-AIA 35 U.S.C. §
`103 over U.S. Patent 6,144,366 to Numazaki, et al.
`(“Numazaki”) in view of JPH4-73631 to Osamu Nonaka
`(“Nonaka”)
`
`Exhibits
`Ex. 1004,
`Ex. 1005
`
`
`
`6
`
`Page 8 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`Ground 2: Claims 6, 11, and 12 are obvious under pre-AIA 35
`U.S.C. § 103 over Numazaki in view of Nonaka and in further
`view of U.S. Patent No. 5,666,157 to David G. Aviv (“Aviv”)
`
`Ex. 1004,
`Ex. 1005,
`Ex. 1006
`
`
`
`Section IV identifies where each element of the Challenged Claims is found
`
`in the prior art. 37 C.F.R. § 42.104(b)(4). The exhibit numbers of the evidence relied
`
`upon to support the challenges are provided above and the relevance of the evidence
`
`to the challenges raised is provided in Section IV. 37 C.F.R. § 42.104(b)(5). Exhibits
`
`1001-10171018 are also attached.
`
`C. Claim Construction Under 37 C.F.R. § 42.104(B)(3)
`In this proceeding, claims are interpreted under the same standard applied by
`
`Article III courts (i.e., the Phillips standard). See 37 C.F.R § 42.100(b); see also 83
`
`Fed. Reg. 197 (Oct. 11, 2018); Phillips v. AWH Corp., 415 F.3d 1303, 1312 (Fed.
`
`Cir. 2005) (en banc). Under this standard, words in a claim are given their plain
`
`meaning which is the meaning understood by a person of ordinary skill in the art in
`
`view of the patent and file history. Phillips, 415 F.3d 1303, 1212–13. With the single
`
`exception discussed below in this section, Petitioner proposes that no terms require
`
`express construction to resolve the proposed grounds presented herein.
`
`a. Capture and store an image
`All three independent claims require capturing and/or storing “an image” in
`
`response to detecting an image capture gesture. See ’949 Patent (Ex. 1001), Claim
`
`
`
`7
`
`Page 9 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`1 (“the image capture command causes the digital camera to store an image to
`
`memory”), Claim 8 (“capturing an image to the digital camera in response to . . . the
`
`image capture command”), Claim 13 (“correlate the gesture detected . . . with an
`
`image capture function and subsequently capture an image using the digital
`
`camera”). For a number of reasons, these limitations should be construed broadly
`
`enough to encompass capturing/storing video or still images. First, the ’949 Patent
`
`repeatedly teaches that its invention applies both to still image and video capture.
`
`See, e.g., id. at 2:19-30 (describing efficiencies that can be realized in “movie
`
`making” by implementing the invention, noting it “allows more cost effective film
`
`production by giving the director the ability to expose the camera to the presence of
`
`masses of data, but only saving or taking that data which is useful), 8:10-30
`
`(describing “digital or other camera 320” that “record[s] the picture” when the
`
`subjects are posed in a predetermined manner, noting “camera 320 may be a video
`
`camera and recorder”).
`
`Second, the ’949 Patent teaches that captured video comprises a sequence of
`
`images. Id. at 8:10-30 (describing an embodiment in which “camera 320 may be a
`
`video camera and recorder which streams in hundreds or even thousands of frames
`
`of image data[,]” noting the subjects may then select individual prints from the
`
`captured image data). The Federal Circuit “has repeatedly emphasized that an
`
`
`
`8
`
`Page 10 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`indefinite article ‘a’ or ‘an’ in patent parlance carries the meaning of ‘one or more’
`
`in open-ended claims containing the transitional phrase ‘comprising.’” Baldwin
`
`Graphic Sys., Inc. v. Siebert, Inc., 512 F.3d 1338, 1342 (Fed. Cir. 2008) (quoting
`
`KCJ Corp. v. Kinetic Concepts, Inc., 223 F.3d 1351, 1356 (Fed. Cir. 2000)). Indeed,
`
`this principle “is best described as a rule, rather than merely as a presumption or
`
`even a convention” and “[t]he exceptions to this rule are extremely limited.” Id.
`
`(noting the patentee must evince a clear intent to limit “a” or “an” to “one”). Here,
`
`all
`
`independent claims are open-ended containing
`
`the
`
`transitional phrase
`
`“comprising,” and each requires capturing/storing “an image.” Applying the Federal
`
`Circuit’s well-established “rule,” the claims should be construed to capture
`
`capturing/storing one or more images to memory in response to detecting the image
`
`capture gesture. Because the ’949 Patent teaches that captured video is a sequence
`
`of still image frames, the claims should be construed to encompass capturing/storing
`
`still images or a video sequence of images.
`
`Finally, as discussed above, the prior art-based rejections during
`
`prosecution all focused on U.S. Patent No. 6,359,647 to Sengupta, et al.
`
`(“Sengupta”) (Ex. 1014). Sengupta discloses “a system for controlling multiple
`
`video cameras” that supports “an automated camera handoff for selecting and
`
`directing cameras” by “tracking a figure.” The applicant conceded that “the
`
`
`
`9
`
`Page 11 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`security system in Sengupta tracks movement of an object ‘from one camera’s
`
`field of view to another camera’s field of view,’” but disputed that Sengupta
`
`taught detecting a gesture that corresponds to an image capture command. ’949
`
`File History (Ex. 1002), 216 (arguing that Sengupta simply “’reports the location’
`
`of the moving objects”). Critically, however, the applicant did not attempt to
`
`distinguish Sengupta on the basis that it captures video, rather than still images.
`
`Accordingly, the prosecution history supports a conclusion that the applicant
`
`intended the Challenged Claims to encompass both capturing/storing still images
`
`and video.
`
`IV. THE CHALLENGED CLAIMS ARE UNPATENTABLE
`A. Ground 1: Claims 1-18 are obvious under pre-AIA 35 U.S.C. § 103
`over Numazaki in view of Nonaka
`Overview of Numazaki
`U.S. Patent No. 6,144,366 to Numazaki et al. (“Numazaki”) (Ex. 1004) was
`
`filed on October 17, 1997, issued on November 7, 2000, and is prior art to the ’949
`
`Patent under at least 35 U.S.C. § 102(e) (pre-AIA). Numazaki was not cited or
`
`considered during prosecution of the ’949 Patent. ’949 Patent (Ex. 1001).
`
`Numazaki is generally directed to a method for detecting a gesture or the
`
`movement of a user’s hand. Numazaki (Ex. 1004), Abstract, 4:9-40. Numazaki
`
`purports to have improved upon prior methods by using a controlled light source to
`
`
`
`10
`
`Page 12 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`illuminate the target object (e.g., the user’s hand), a first camera unit (referred to by
`
`Numazaki as a “photo-detection unit”),1 and a second camera unit. Id. at 11:9-23.
`
`This arrangement is illustrated in Fig. 2 below:
`
`
`
`
`
`
`Id. at Fig. 2. A timing control unit is used to turn lighting unit 101 on (i.e.,
`
`illuminating the target object) when the first camera unit is active and off when the
`
`second camera unit is active. Id. at 11:20-32. The result of this light control is the
`
`
`1 A PHOSITA would have considered Numazaki’s photo-detection units to be
`
`camera units. Bederson Dec. (Ex. 1003), ¶ 35 (explaining that Numazaki describes
`
`using CMOS or CCD sensor units at 15:24-16:19, which were two of the more
`
`common optical sensors used in camera units at the time).
`
`
`
`11
`
`Page 13 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`first camera unit captures an image of the target object illuminated by both natural
`
`light and the lighting unit 101 and the second camera unit captures an image of the
`
`target object illuminated by only natural light. Id. at 11:33-39. The difference
`
`between the two images—obtained by difference calculation unit 111—represents
`
`the “reflected light from the object resulting from the light emitted by the lighting
`
`unit 101.” Id. at 11:43-51. This information is then used by feature data generation
`
`unit 103 to determine gestures, pointing, etc. of the target object that may be
`
`converted into commands executed by a computer. Id. at 10:57-66.
`
`Throughout its lengthy disclosure, Numazaki describes various features and
`
`functionality based on this two-camera structure. Although referred to as
`
`“embodiments,” as discussed in detail below, a PHOSITA would have understood
`
`that many are complementary and would have been motivated to combine certain of
`
`these features and functionalities in a single device.
`
`Numazaki’s third embodiment is “another exemplary case of the feature data
`
`generation unit of the first embodiment, which realizes a gesture camera for
`
`recognizing the hand action easily and its application as a pointing device in the
`
`three-dimensional space.” Id. at 29:4-8. In this embodiment, data reflecting pre-
`
`registered gestures or hand positions is stored in “shape memory unit 332.” Id. at
`
`29:9-38. This stored data is compared with the output from the two-sensor “reflected
`
`
`
`12
`
`Page 14 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`light extraction unit” (i.e., component 102 from first embodiment) to identify when
`
`the user has performed a pre-registered gesture and instructs the device to implement
`
`a “command corresponding to that [gesture].” Id. at 29:22-30:5. In addition to “hand
`
`[g]esture recognition as the means for inputting the command into the three-
`
`dimensional space and the like into the computer,” the third embodiment
`
`contemplates additional gesture-based instructions such as “instructing the power
`
`ON/OFF of the TV, the lighting equipment, etc.” Id. at 31:3-10. Numazaki cautions
`
`that gestures can be erroneously detected when the user did not intend to command
`
`the system and proposes a sequence of gestures to prevent such “erroneous
`
`commanding.” Id. at 31:11-44, Fig. 29 (describing a sequence of “two fingers, fist,
`
`five fingers” that causes the system to “power on”). Such a sequence (frames a-c)
`
`and the rectangular data extracted from the images (frames d-f) are illustrated in Fig.
`
`28 below:
`
`
`
`13
`
`
`
`Page 15 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`
`
`
`Id. at Fig. 28.
`
`Numazaki’s fifth embodiment is directed to video capture and transmission
`
`for applications such as videoconferencing. Specifically, the fifth embodiment
`
`recognizes the difficulties involved with videoconference applications using a
`
`portable device, including communication costs and power consumption. Id. at 39:6-
`
`16. To combat these difficulties, Numazaki teaches a “TV telephone” that extracts
`
`and transmits only the faces of the participating users, removing extraneous
`
`background information that would otherwise consume bandwidth and battery
`
`power. Id. at 39:12-20. To accomplish this improvement, Numazaki uses the same
`
`two-camera “reflected light extraction unit 102” (i.e., the same component 102 from
`
`
`
`14
`
`Page 16 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`first embodiment) in conjunction with a separate “visible light photo-detection
`
`array 351 which is generally used as a CCD camera for taking video images.” Id.
`
`at 39:21-49. This structure is depicted in Fig. 46 below:
`
`
`
`
`
`
`Id. at Fig. 46. Using this arrangement, the fifth embodiment processes the output of
`
`two-sensor structure 102 to identify an outline of the subject and subtracts everything
`
`outside this outline from the image captured by the visible light sensor 351. Id. at
`
`39:24-60. In doing so, the fifth embodiment arrives at image information that
`
`contains only the subject without any background image information and uses much
`
`less data as a result of removing the extraneous image information. Id. at 40:32-35
`15
`
`
`
`Page 17 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`(“By recording the extracted image, it is possible to reduce the recording capacity
`
`considerably (in an order of one tenth to one hundredth)[.]”).
`
`Numazaki’s eighth embodiment describes various portable computers such as
`
`laptops and handheld devices on which the earlier-described hardware and
`
`functionalities may be implemented. Id. at 50:19-24 For example, Fig. 74 depicts a
`
`laptop computer with a lighting unit 701 and camera unit 702:
`
`
`
`
`Id. at Fig. 74, 50:25-37 (describing the same). Similar implementations in which a
`
`user can control a cursor by moving a finger are described with reference to Fig. 78
`
`(a PDA device) and Fig. 79 (a wristwatch device). Numazaki expressly teaches that
`
`these portable devices implement the information input generation apparatus
`
`described in the preceding embodiments. Id. at 50:19-24 (“This eighth embodiment
`
`
`
`16
`
`Page 18 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`is directed to a system configuration incorporating the information input generation
`
`apparatus of the present invention as described in the above embodiments.”). For
`
`example, Numazaki’s eighth embodiment devices include the controlled light and
`
`two-camera configuration described in its first embodiment. As described with
`
`respect to the first embodiment above, Numazaki’s information input generation
`
`apparatus uses controlled lighting and two camera sensors to isolate an image of an
`
`object (such as a user’s hand) by subtracting the image taken when the light is off
`
`from the image taken when the light is on, thereby obtaining an image that focuses
`
`on the illuminated object. Numazaki teaches that this precise image difference
`
`calculation is included in the eighth embodiment. Id. at 53:22-36. Accordingly, a
`
`PHOSITA would have understood that Numazaki’s eighth embodiment portable
`
`devices incorporate the controlled lighting and two-camera sensor structure
`
`described with respect to the first embodiment. Bederson Dec. (Ex. 1003), ¶ 44. As
`
`discussed below in the motivations to combine Numazaki and Nonaka section, a
`
`PHOSITA would also have been motivated to implement in the eighth embodiment
`
`laptop the third embodiment’s feature that maps specific gestures to specific
`
`commands and the fifth embodiment’s video capture feature.
`
`Because Numazaki, like the ’949 Patent, discloses a camera system that may
`
`be controlled by human gesture input, Numazaki is in the same field of endeavor as
`
`
`
`17
`
`Page 19 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`the ’949 Patent. Compare Numazaki (Ex. 1004), 31:3-10 (noting gesture recognition
`
`is used as “the means for inputting the command . . . into the computer” and noting
`
`gestures can be used to turn on/off equipment such as a TV or lighting equipment),
`
`50:25-48 (describing a portable laptop device equipped with a light source and
`
`camera such that a user can control certain functionalities using hand gestures) with
`
`’949 Patent (Ex 1001), 5:24-49 (describing the process of identifying a subject has
`
`performed a predetermined gesture and controlling the camera device to capture an
`
`image in response). Numazaki is therefore analogous art to the ’949 Patent. Bederson
`
`Dec. (Ex. 1003), ¶¶ 33-45.
`
`Overview of Nonaka
`
`Japanese Unexamined Patent Application Publication JPH4-73631 to Osamu
`
`Nonaka (“Nonaka”) (Ex. 1005) published on March 9, 1992 and is prior art to the
`
`’949 Patent under at least 35 U.S.C. § 102(b) (pre-AIA). Nonaka was not cited or
`
`considered during prosecution of the ’949 Patent. ’949 Patent (Ex. 1001).
`
`Nonaka teaches a “remote release device-equipped camera” that allows a user
`
`to signal a desire for the camera to take a picture by “mak[ing] a predetermined
`
`motion.” Nonaka
`
`(Ex. 1005), 15:11-14. Nonaka contemplates multiple
`
`predetermined gestures such as holding one’s hand out toward the camera (as
`
`
`
`18
`
`Page 20 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`depicted in Fig. 3 below left) or moving one’s hand toward the camera (as depicted
`
`in Fig. 7 below right):
`
`
`
`
`
`
`Id. at Figs. 3, 7, 3:34-4:4 (describing “the photographer giv[ing] a release instruction
`
`by means of a predetermined motion towards the camera . . . such as moving his or
`
`her hand forward”), 6:11-22 (describing the photographer “giv[ing] a release
`
`instruction . . . [by] mov[ing] his or her hand, etc. at a certain speed”).
`
`Because Nonaka, like the ’949 Patent, discloses a camera system that may be
`
`controlled by human gesture input, Nonaka is in the same field of endeavor as the
`
`’949 Patent. Compare Nonaka (Ex. 1005), 14:32-36 (“[D]uring the remote release
`
`operation of the camera according to the present exemplary embodiment, if the
`
`photographer makes a motion such as moving his or her hand . . . when he or she
`
`intends to carry out the release operation, the photographer is thus able to
`
`communicate that intent to the camera.”) with ’949 Patent (Ex 1001), 5:24-49
`
`
`
`19
`
`Page 21 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`(describing the process of identifying a subject has performed a predetermined
`
`gesture and controlling the camera device to capture an image in response). Nonaka
`
`is therefore analogous art to the ’949 Patent. Bederson Dec. (Ex. 1003), ¶¶ 46-47.
`
`Motivation to Combine Numazaki and Nonaka
`
`As discussed above, Numazaki’s first embodiment teaches a two-camera
`
`“reflected light extraction unit” through which an object such as a user’s hand can
`
`be detected. Numazaki’s third embodiment takes the output from the two-sensor
`
`“reflected light extraction unit,” compares it with stored data reflecting pre-
`
`registered gestures or hand positions in “shape memory unit 332” to detect when the
`
`user has performed a pre-registered gesture, and instructs the device to implement a
`
`“command corresponding to that [gesture].” Numazaki (Ex. 1004), 29:19-30:5. Such
`
`gesture commands, for example, can instruct the system to turn devices on or off. Id.
`
`at 31:3-10. In its fifth embodiment, Numazaki adds an additional “visible light photo-
`
`detection array 351 . . . for taking video images.” Id. at 39:21-49. Using this
`
`arrangement, the fifth embodiment processes the output of two-sensor “reflected
`
`light extraction unit” 102 to identify an outline of the subject and subtracts
`
`everything outside this outline from the image captured by the visible light sensor
`
`351, thereby decreasing the size of the ultimate video file. Id. at 39:24-60, 4:32-35.
`
`Finally, Numazaki’s eighth embodiment teaches portable devices that implement the
`
`
`
`20
`
`Page 22 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`information input generation apparatus described in the preceding embodiments. Id.
`
`at 50:19-24. Although Numazaki does not expressly describe combining all these
`
`features into a single portable device such that a user could perform a gesture
`
`command (pursuant to its third embodiment) that causes video capture to initiate
`
`(pursuant to its fifth embodiment), a PHOSITA would have been motivated to
`
`implement Numazaki’s portable device in this manner pursuant to Nonaka’s image
`
`capture command gesture teachings.
`
`First, implementing Numazaki’s features as proposed would have improved
`
`Numazaki’s portable devices in the same way Nonaka’s gesture-based image capture
`
`functionality benefits its camera device. Bederson Dec. (Ex. 1003), ¶¶ 48-49
`
`(explaining that Numazaki’s native functionality of associating hand gestures with
`
`commands would have been a natural fit as a means to initiate video capture); KSR
`
`Int’l Co. v. Teleflex Inc., 550 U.S. 398, 417 (2007) (obvious to use known techniques
`
`to improve similar devices in the same way). Nonaka explains that users desired the
`
`ability to remotely trigger image capture, but that then-existing options were limited
`
`to self-timer mechanisms and expensive wireless remote controls—both of which
`
`were undesirable. Nonaka (Ex. 1006), 2:6-25. According to Nonaka, its gesture-
`
`based image capture solution “achiev[es] a higher degree of freedom, good
`
`portability, and cost benefits.” Id. at 2:26-29. A PHOSITA would have recognized
`
`
`
`21
`
`Page 23 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`that these same benefits would be realized in Numazaki’s laptop. Namely,
`
`implementing Numazaki’s laptop as proposed would provide a greater degree of
`
`freedom than alternative means of trigger image capture such as timers or manual
`
`initiation, ensuring the user is able to get in position and prepared before the video
`
`capture begins. Bederson Dec. (Ex. 1003), ¶ 49. Using gesture-based image capture
`
`initiation also maintains Numazaki’s portability and reduces costs by employing
`
`Numazaki’s native gesture recognition functionality, rather than adding wireless or
`
`wired remote control hardware. Id.
`
`Second, a PHOSITA would have been motivated to utilize distinct sensor
`
`units to perform the gesture recognition and video capture. Specifically, the lower
`
`resolution “reflected light extraction unit” 102 is well suited to monitor the scene
`
`and detect any gestures performed by the user, while the higher resolution “visible
`
`light photo-detection array 351” is better suited for taking full quality images of the
`
`scene (when such video is requested by the user). Bederson Dec. (Ex. 1003), ¶ 50.
`
`Indeed, as Dr. Bederson explains, it was widely recognized that lower resolution
`
`monitoring enables a system to quickly respond to changes in the environment and
`
`is preferable to higher resolution monitoring:
`
`A key requirement of many image processing systems is that they
`operate quickly enough to respond to changes in the visual environment
`in a timely fashion. This is especially true for systems that need to
`
`
`
`22
`
`Page 24 of 77
`
`

`

`PR2021-00921
`PR2022-00362
`U.S. Patent No. 8,878,949
`respond to events as they happen in the physical world. Because the
`essential nature of image processing is that the pixels in the image are
`examined, the more pixels there are, the heavier the computational
`requirements are, and the slower the system can respond. A tension is
`that if an image does not hold enough detail about a scene, the image
`processing algorithm will not have enough information to correctly
`analyze that scene. Thus, it is not surprising that managing the
`resolution of images for processing is an essential part of image
`processing systems. People have for decades found ways to work with
`lower resolution images—sometimes starting with lower resolution
`images at first, and then working with higher resolution images later for
`fine-tuned processing.
`
`Id. at ¶ 50 (further discussing a 1982 textbook that describes image resolution
`
`algorithms that are used to process images from coarse-to-fine resolution as required

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket