throbber
Case IPR2016-01372
`Patent No. 8,659,571
`
`
`
`
`
`
`UNITED STATES PATENT AND TRADEMARK OFFICE
`___________________
`
`BEFORE THE PATENT TRIAL AND APPEAL BOARD
`___________________
`
`
`APPLE INC.,
`Petitioner,
`
`v.
`
`IMMERSION CORPORATION,
`Patent Owner.
`___________________
`
`Case IPR2016-01372
`Patent No. 8,659,571
`___________________
`
`
`
`
`IMMERSION CORPORATION’S
`
`PATENT OWNER PRELIMINARY RESPONSE
`
`
`
`
`
`Mail Stop “PATENT BOARD”
`Patent Trial and Appeal Board
`U.S. Patent and Trademark Office
`P.O. Box 1450
`Alexandria, VA 22313-145
`
`
`
`
`
`
`
`
`

`

`Case IPR2016-01372
`Patent No. 8,659,571
`
`
`TABLE OF CONTENTS
`
`INTRODUCTION ........................................................................................... 1 
`
`THE ’571 PATENT ......................................................................................... 2 
`
`
`I. 
`
`II. 
`
`III.  LEVEL OF ORDINARY SKILL IN THE ART ............................................. 5 
`
`IV.  CLAIM CONSTRUCTION ............................................................................ 5 
`
`A. 
`
`“gesture signal” (claims 1-7, 12-18, 23-29) .......................................... 8 
`
`1. 
`
`2. 
`
`Immersion’s Construction Tracks the Intrinsic Record .............. 8 
`
`Apple’s Construction Is Too Broad and Vitiates the
`Meaning of “Gesture” ............................................................... 12 
`
`“dynamic interaction parameter” (claims 1, 4-7, 12, 15-18,
`23, 26-29) ............................................................................................ 14 
`
`“vector signal” (claims 2, 13, 24) ........................................................ 15 
`
`“on-screen signal” (claims 3, 14, 25) .................................................. 15 
`
`“physical model” (claims 5, 16, 27) .................................................... 16 
`
`“generating a dynamic interaction parameter using . . . an
`animation” (claims 6, 17, 28) .............................................................. 16 
`
`B. 
`
`C. 
`
`D. 
`
`E. 
`
`F. 
`
`G. 
`
`“module”/ “drive module” (claims 12, 15-18) .................................... 17 
`
`V.  GROUND 1: BURROUGH DOES NOT RENDER CLAIMS 1-7,
`12-18, AND 23-29 OBVIOUS UNDER 35 U.S.C. § 103(A) ....................... 19 
`
`A. 
`
`Burrough Does Not Disclose or Render Obvious Claims 1,
`12, and 23 ............................................................................................ 21 
`
`1. 
`
`Burrough Does Not Disclose a “First Gesture Signal”
`or a “Second Gesture Signal” ................................................... 21 
`
`
`
`
`- ii -
`
`
`
`

`

`Case IPR2016-01372
`Patent No. 8,659,571
`
`
`Burrough Does Not Disclose “Generating a Dynamic
`Interaction Parameter Using the First Gesture Signal
`and the Second Gesture Signal” ................................................ 24 
`
`Burrough Does Not Disclose “Applying a Drive Signal
`to a Haptic Output Device According to the Dynamic
`Interaction Parameter” .............................................................. 27 
`
`2. 
`
`3. 
`
`B. 
`
`C. 
`
`D. 
`
`E. 
`
`F. 
`
`G. 
`
`H. 
`
`Burrough Does Not Disclose or Render Obvious Claims 2,
`13, and 24 (“the First or Second Gesture Signal Comprises a
`Vector Signal”) .................................................................................... 28 
`
`Burrough Does Not Disclose or Render Obvious Claims 3,
`14, and 25 (“the First or Second Gesture Signal Comprises
`an On-Screen Signal”) ......................................................................... 30 
`
`Burrough Does Not Disclose or Render Obvious Claims 4,
`15, and 26 (“Generating a Dynamic Interaction Parameter
`from a Difference Between the First Gesture Signal and the
`Second Gesture Signal”) ..................................................................... 31 
`
`Burrough Does Not Disclose or Render Obvious Claims 5,
`16, and 27 (“Generating a Dynamic Interaction Parameter
`Using the First Gesture Signal and the Second Gesture Signal
`and a Physical Model”) ....................................................................... 32 
`
`Burrough Does Not Disclose or Render Obvious Claims 6,
`17, and 28 (“Generating a Dynamic Interaction Parameter
`Using the First Gesture Signal and the Second Gesture Signal
`and an Animation”) ............................................................................. 35 
`
`Burrough Does Not Disclose or Render Obvious Claims 7,
`18, and 29 (Generating a Dynamic Interaction Parameter
`Based on Two Device Sensor Signals as Well as a First and
`Second Gesture Signal) ....................................................................... 37 
`
`Burrough Does Not Disclose or Render Obvious Claim 12
`(“Drive Module”) ................................................................................ 40 
`
`
`
`
`- iii -
`
`
`
`

`

`VI.  GROUND 2: ROSENBERG ’373 DOES NOT RENDER CLAIMS
`1, 2, 4-6, 12, 13, 15-18, 23, 24, AND 26-29 OBVIOUS UNDER
`PRE-AIA 35 U.S.C. § 103(A) ....................................................................... 41 
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`
`A. 
`
`B. 
`
`C. 
`
`D. 
`
`E. 
`
`F. 
`
`The Board Should Deny Institution of Grounds 2 and 3
`Because the Patent Office Considered Substantially Similar
`Material to Rosenberg ’373 During Prosecution ................................ 42 
`
`Rosenberg ’373 Does Not Disclose or Render Obvious
`Claims 1, 12, and 23 ............................................................................ 45 
`
`1. 
`
`2. 
`
`Rosenberg ’373 Does Not Disclose “Receiving a First
`Gesture Signal” or “Receiving a Second Gesture
`Signal” ....................................................................................... 45 
`
`Rosenberg ’373 Does Not Disclose “Generating a
`Dynamic Interaction Parameter Using the First Gesture
`Signal and the Second Gesture Signal” .................................... 48 
`
`Rosenberg ’373 Does Not Disclose or Render Obvious
`Claims 2, 13, and 24 (“the First or Second Gesture Signal
`Comprises a Vector Signal”) ............................................................... 50 
`
`Rosenberg ’373 Does Not Disclose or Render Obvious
`Claims 4, 15, and 26 (“Generating a Dynamic Interaction
`Parameter From a Difference Between the First Gesture
`Signal and the Second Gesture Signal”) ............................................. 52 
`
`Rosenberg ’373 Does Not Disclose or Render Obvious
`Claims 5, 16, and 27 (“Generating a Dynamic Interaction
`Parameter Using the First Gesture Signal and the Second
`Gesture Signal and a Physical Model”) .............................................. 53 
`
`Rosenberg ’373 Does Not Disclose or Render Obvious
`Claims 6, 17, and 28 (“Generating a Dynamic Interaction
`Parameter Using the First Gesture Signal and the Second
`Gesture Signal and an Animation”) .................................................... 53 
`
`G. 
`
`Rosenberg ’373 Does Not Disclose or Render Obvious
`Claims 7, 18, and 29 (Generating a Dynamic Interaction
`
`
`
`
`- iv -
`
`
`
`

`

`Parameter Based on Two Device Sensor Signals as Well as a
`First and Second Gesture Signal) ........................................................ 54 
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`
`H. 
`
`Rosenberg ’373 Does Not Disclose or Render Obvious Claim
`12 ......................................................................................................... 58 
`
`VII.  GROUND 3: THE COMBINATION OF ROSENBERG ’373 AND
`ROSENBERG ’846 DOES NOT RENDER CLAIMS 3, 14, AND
`25 OBVIOUS UNDER PRE-AIA 35 U.S.C. § 103(A) ................................ 58 
`
`VIII.  CONCLUSION .............................................................................................. 60 
`
`
`
`
`
`
`- v -
`
`
`
`

`

`TABLE OF AUTHORITIES
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`
`
`Cases 
`ActiveVideo Networks, Inc. v. Verizon Commc’ns, Inc.,
`694 F.3d 1312 (Fed. Cir. 2012) .............................................................. 35
`
`Chimie v. PPG Indus., Inc.,
`402 F.3d 1371 (Fed Cir. 2005) ............................................................... 18
`
`Cuozzo Speed Techs., LLC v. Michelle K. Lee,
`136 S. Ct. 2131 (2016) .............................................................................. 5
`
`In re Suitco Surface, Inc.,
`603 F.3d 1255 (Fed. Cir. 2010) .............................................................. 14
`
`Intelligent Bio-Sys., Inc. v. Illumina Cambridge Ltd.,
`IPR2013-00324, Paper No. 19 (P.T.A.B. Nov. 21, 2013) ...................... 42
`
`Interactive Gift Exp., Inc. v. CompuServe Inc.,
`256 F.3d 1323 (Fed. Cir. 2001) .............................................................. 18
`
`KSR Int’l Co. v. Teleflex Inc.,
`550 U.S. 398, 418 (2007) ................................................................. 37, 56
`
`Microboards Tech., LLC d/b/a Afinia v. Stratasys Inc.,
`IPR2015-00287, Paper No. 13 (P.T.A.B. May 28, 2015) ...................... 43
`
`Microsoft Corp. v. Proxyconn, Inc.,
`789 F.3d 1292 (Fed. Cir. 2015) .............................................................. 14
`
`Phillips v. AWH Corp.,
`415 F.3d 1303 (Fed. Cir. 2005) .............................................................. 18
`
`Prism Pharma Co., Ltd. v. Choongwae Pharma Corp.,
`IPR2014-00315, Paper No. 14 (P.T.A.B. July 8, 2014) ......................... 43
`
`Sanofi-Synthelabo v. Apotex, Inc.,
`550 F.3d 1075 (Fed. Cir. 2008) .......................................................... 2, 32
`
`
`
`
`- vi -
`
`
`
`

`

`
`Statutes 
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`35 U.S.C. § 103(a) ............................................................................................ 21
`
`35 U.S.C. § 314(a) ............................................................................................ 60
`
`35 U.S.C. § 325(d) ...................................................................................... 42, 45
`
`37 C.F.R. § 42.100(b ........................................................................................... 6
`
`37 C.F.R. § 42.100(b) ......................................................................................... 5
`
`
`
`
`
`
`- vii -
`
`
`
`

`

`
`
`
`Immersion
`Ex. 2001
`Immersion
`Ex. 2002
`Immersion
`Ex. 2003
`Immersion
`Ex. 2004
`Immersion
`Ex. 2005
`Immersion
`Ex. 2006
`Immersion
`Ex. 2007
`Immersion
`Ex. 2008
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`EXHIBIT LIST
`
`Declaration of Yon Visell, Ph.D. in Support of Immersion
`Corporation’s Patent Owner Preliminary Response
`Dictionary.com, http://www.dictionary.com/browse/gesture?s=t
`(last visited Oct. 11, 2016)
`August 2, 2012 Applicant Remarks in Prosecution of U.S. Patent
`No. 8,279,193
`U.S. Patent Application Publication No. US 2007/0279392
`(Rosenberg ’392)
`July 19, 2012 Non-Final Rejection in Prosecution of U.S. Patent
`No. 8,279,193
`May 16, 2012 Original Claims in Prosecution of U.S. Patent No.
`8,279,193
`Curriculum Vitae of Yon Visell, Ph.D.
`
`Oct. 11, 2016 Joint Proposed Claim Construction Chart submitted
`by Apple, Immersion, and OUII Staff in ITC Investigation Nos.
`337-TA-990 and 337-TA-1004
`
`
`
`
`
`
`
`
`- viii -
`
`
`
`

`

`
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`STATEMENT OF MATERIAL FACTS IN DISPUTE
`
`Petitioner Apple Inc. did not submit a statement of material facts in this
`
`Petition. Accordingly, no response is due pursuant to 37 C.F.R. § 42.23(a), and no
`
`facts are admitted.
`
`
`
`
`
`
`- vi -
`
`
`
`

`

`
`I.
`
`INTRODUCTION
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`Patent Owner Immersion Corporation (“Immersion”) submits that the Patent
`
`Trial and Appeal Board (“Board”) should deny the petition for inter partes review
`
`(“Petition”) submitted by Petitioner Apple Inc. (“Apple”) and not institute inter
`
`partes review of U.S. Patent No. 8,659,571 (Ex. 1001, the “’571 patent”) for at
`
`least two independent reasons.
`
`First, the Petition fails to establish that Ex. 1005, U.S. Patent Application
`
`Publication No. US 2010/0156818 (“Burrough”) or the other two prior art
`
`references submitted by Apple, Ex. 1004, U.S. Patent No. 5,734,373 (“Rosenberg
`
`’373”); and Ex. 1006, U.S. Patent No. 6,429,846 (“Rosenberg ’846”), disclose or
`
`render obvious the challenged claims of the ’571 patent. Apple particularly fails to
`
`demonstrate the obviousness of the claim limitations “receiving a first gesture
`
`signal,” “receiving a second gesture signal,” and “generating a dynamic interaction
`
`parameter using the first gesture signal and the second gesture signal” as recited in
`
`independent claims 1, 12, and 23. The Petition repeatedly attempts to read out
`
`these critical limitations from the claims by referring to snippets of raw position
`
`data at particular points in time and calling those fragments “gesture signals.” The
`
`Petition should be denied and trial should not be instituted on this basis alone.
`
`Second, although all grounds identified in the Petition are obviousness rather
`
`than anticipation, the Petition addresses obviousness in only a conclusory fashion,
`
`
`
`
`- 1 -
`
`
`
`

`

`
`without identifying any prior art teaching that would lead a person of ordinary skill
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`in the art to arrive at the inventions claimed in the ’571 patent. For example,
`
`Apple’s primary prior art reference repeatedly teaches, at best, that only a single
`
`gesture signal is used to contribute to a haptic effect. Nevertheless, the Petition
`
`argues the claims of the ’571 patent are obvious under that reference, when instead
`
`the ’571 patent claims require a different system where both a first and a second
`
`gesture signal are used to generate a dynamic interaction parameter. Apple’s
`
`obviousness analysis is also flawed because it is performed on a claim limitation
`
`by claim limitation basis, as opposed to analyzing obviousness of the claim as a
`
`whole. Sanofi-Synthelabo v. Apotex, Inc., 550 F.3d 1075, 1086 (Fed. Cir. 2008)
`
`(“The determination of obviousness is made with respect to the subject matter as a
`
`whole, not separate pieces of the claim.”). The Petition should also be denied for
`
`this reason.
`
`As explained in further detail below, and in the supporting declaration of
`
`Professor Yon Visell, Ex. 2001, a person of ordinary skill in the art would not have
`
`found the challenged claims obvious under the cited prior art. Thus, the Board
`
`should deny institution of inter partes review on all three proposed grounds.
`
`II. THE ’571 PATENT
`The ’571 patent is entitled “Interactivity Model for Shared Feedback on
`
`Mobile Devices.” The ’571 patent is directed to a novel way of producing haptic
`
`
`
`
`- 2 -
`
`
`
`

`

`
`effects in electronic devices. The fundamental insight that is described and
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`claimed in the ’571 patent is to track and analyze gesture signals to provide more
`
`compelling haptic feedback in response to a user’s input. Ex. 1001 at 1:49-62.
`
`Reflecting this focus, the claims specify that both a first and a second gesture
`
`signal (each based on a user’s gestural inputs) are used to generate something
`
`called a “dynamic interaction parameter.” Id. at claim 1 (“receiving a first gesture
`
`signal; receiving a second gesture signal; generating a dynamic interaction
`
`parameter using the first gesture signal and the second gesture signal”). The
`
`claims require more than a single gesture signal to form the dynamic interaction
`
`parameter.
`
`The dynamic interaction parameter is meant to accurately and responsively
`
`track the user’s behavior. As such, the dynamic interaction parameter changes or
`
`reacts in real time to the user’s interactions, and is used to alter the haptic effects
`
`produced by the device. This allows the device to provide responsive haptic
`
`feedback to the user. Ex. 1001 at 1:29-33 (“[V]ibrotactile haptic effects . . . may
`
`be useful in providing cues to users of electronic devices to alert the user to
`
`specific events, or provide realistic feedback to create greater sensory immersion
`
`within a simulated or virtual environment.”). The approach of the ’571 patent is an
`
`improvement over the prior art because it can improve the timing and/or nature of
`
`haptic feedback: “[B]ecause these user gestures and system animations have
`
`
`
`
`- 3 -
`
`
`
`

`

`
`variable timing, the correlation to haptic feedback [in the prior art] may be static
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`and inconsistent and therefore less compelling to the user.” Id. at 1:49-56.
`
`Other ingredients may be used in addition to a first gesture signal and a
`
`second gesture signal to generate the dynamic interaction parameter. For example,
`
`additional device sensor signals may be used. Id. at claim 7. Using these
`
`additional ingredients is another improvement over the prior art. E.g., id. at 1:56-
`
`60 (“Further, device sensor information is typically not used in combination with
`
`gestures to produce haptic feedback.”). The various ingredients may be combined
`
`and processed in several different ways to generate the dynamic interaction
`
`parameter. E.g., id. at Table 2 (listing 14 different example “methods of synthesis”
`
`that may be employed). The challenged claims require specific ingredients in
`
`specific numbers to be used to generate the dynamic interaction parameter. E.g.,
`
`id. at claim 7 (“receiving a first device sensor signal; receiving a second device
`
`sensor signal; and wherein generating a dynamic interaction parameter comprises
`
`generating a dynamic interaction parameter using the first gesture signal and the
`
`second gesture signal and the first device sensor signal and the second device
`
`sensor signal”).
`
`Once the dynamic interaction parameter has been generated using a first
`
`gesture signal, a second gesture signal, and potentially other ingredients, it is used
`
`to modify the haptic output of the system. Id. at 15:8-9 (“At 1313, a drive signal is
`
`
`
`
`- 4 -
`
`
`
`

`

`
`applied to a haptic actuator according to the interaction parameter.”); claim 1
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`(“applying a drive signal to a haptic output device according to the dynamic
`
`interaction parameter”). For example, in one embodiment, a user may scroll
`
`between different film frames on an electronic device with a touchscreen, and may
`
`receive haptic feedback for that interaction. Id. at 13:56-61.
`
`III. LEVEL OF ORDINARY SKILL IN THE ART
`A person of ordinary skill in the art for the field of the ’571 patent would
`
`have at least: (1) a Bachelor’s of Science degree in an engineering discipline such
`
`as Mechanical Engineering or Computer Science, or (2) at least two years of
`
`experience working with human machine interface systems, graphical user
`
`interfaces, haptic feedback systems, robotics, biomechanics, or mobile devices or
`
`equivalent embedded systems. Ex. 2001 at ¶ 24. The conclusions regarding claim
`
`construction and validity contained herein would be the same regardless of whether
`
`Immersion’s or Apple’s proposed level of skill is adopted by the Board. Id. at ¶
`
`25.
`
`IV. CLAIM CONSTRUCTION
`During inter partes review, claims are given their “broadest reasonable
`
`construction in light of the specification.” Cuozzo Speed Techs., LLC v. Michelle
`
`K. Lee, 136 S. Ct. 2131, 2133 (2016); 37 C.F.R. § 42.100(b). The standard for
`
`claim construction at the Patent Office is different from that used during a U.S.
`
`
`
`
`- 5 -
`
`
`
`

`

`
`District Court litigation. 37 C.F.R. § 42.100(b). Immersion expressly reserves the
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`right to argue a different claim construction in litigation for any term of the ’571
`
`patent, as appropriate in that proceeding.
`
`Apple has proposed seven different claim constructions. See Petition at 7-
`
`12. Apple’s overall approach to claim construction is seriously deficient for at
`
`least three separate reasons.
`
`First, in many situations neither Apple nor its expert alleges that the
`
`supposed “constructions” that it wants Immersion to “be held to” are the proper
`
`broadest reasonable construction that should be applied in this proceeding. See id.
`
`at 12 (refusing to state whether its proposed constructions constitute “an
`
`interpretation of the claims under the ‘broadest reasonable construction’ standard
`
`and/or under the Immersion’s [sic] infringement contentions and technical
`
`domestic industry contentions in the ITC Investigation”); id. at 9 (stating that
`
`Apple does not agree with the construction of “dynamic interaction parameter” that
`
`it asks the Board to adopt); id. at 10 (stating that “on-screen signal” and “physical
`
`model” should be construed “[b]ased on Immersion’s public contentions” but not
`
`arguing that those constructions are the broadest reasonable constructions); Ex.
`
`1002 at ¶ 46 (improperly basing proposed constructions “on the Immersion’s [sic]
`
`apparent belief about the scope of the claim terms” rather than the evidence
`
`applicable to the ’571 patent). Apple cites no authority for the proposition that a
`
`
`
`
`- 6 -
`
`
`
`

`

`
`claim construction that is not the broadest reasonable construction should be
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`applied.
`
`Second, in many situations Apple has not provided any claim construction at
`
`all, and instead has made arguments about what should be “encompassed” within a
`
`claim term. These arguments about what is “encompassed” within a term are not
`
`proper claim constructions at all. They do not provide any clarity to the Board
`
`about the actual meaning of a claim term.
`
`Third, Apple relies on Immersion’s claim charts in an ongoing ITC
`
`investigation to contend that Immersion has taken certain claim construction
`
`positions that it should be held to before the Board. This is improper. The claim
`
`charts do not purport to set forth a claim construction of any term, and no claim
`
`construction has yet been endorsed by the ITC. What Apple presents as the
`
`“constructions” in the charts are instead Apple’s speculations about what
`
`assumptions it thinks the charts must be making.
`
`Apple’s faulty claim constructions should be rejected, and Apple’s
`
`substantive arguments for obviousness should therefore be rejected as well due to
`
`the reliance on incorrect claim constructions. In addition to these global
`
`deficiencies, Immersion addresses the individual constructions below and provides
`
`its proposed claim constructions. Immersion may propose additional terms for
`
`construction in future filings.
`
`
`
`
`- 7 -
`
`
`
`

`

`
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`A.
`
`“gesture signal” (claims 1-7, 12-18, 23-29)
`1.
`The broadest reasonable construction of the term “gesture signal” in view of
`
`Immersion’s Construction Tracks the Intrinsic Record
`
`the relevant evidence is “an electronic signal, representing a recognized movement
`
`of the body that conveys meaning or user intent.” Immersion’s proposed
`
`construction is supported by every level of the claim construction hierarchy: the
`
`plain meaning, the specification, the prosecution history, and the relevant extrinsic
`
`evidence.
`
`To begin, the plain meaning of “gesture signal” suggests that it must be
`
`related to a user’s gesture. Ex. 2001 at ¶ 38.
`
`Additionally, the specification describes a “gesture” as a “movement of the
`
`body that conveys meaning or user intent.” Ex. 1001 at 3:34-35 (emphasis
`
`added); see also 3:56-59 (“A gesture can also be any form of hand movement
`
`recognized by a device having an accelerometer, gyroscope, or other motion
`
`sensor, and converted to electronic signals.”) (emphasis added). The specification
`
`thus refers to: the movement of the body; the movement conveys meaning or user
`
`intent; the device recognizes the movement; and the recognized movement is
`
`converted to electronic signals. Ex. 2001 at ¶ 39.
`
`The specification uniformly refers to full body movements—and specifically
`
`to hand movements over a period of time—as producing a gesture signal:
`
`
`
`
`- 8 -
`
`
`
`

`

`
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`It will be recognized that simple gestures may be combined to form
`more complex gestures. For example, bringing a finger into contact
`with a touch sensitive surface may be referred to as a “finger on”
`gesture, while removing a finger from a touch sensitive surface may
`be referred to as a separate “finger off” gesture. If the time between
`the “finger on” and “finger off” gestures is relatively short, the
`combined gesture may be referred to as “tapping” . . . .
`
`Id. at 3:36-52; Ex. 2001 at ¶ 40. This comports with the plain and ordinary
`
`meaning of “gesture.”
`
`Importantly, the specification discloses that there are movements of the body
`
`and/or the hand that are not full gestures and could not comprise an entire gesture
`
`signal. Id. at ¶ 41. This depends on the implementation of a particular system.
`
`For example, in one embodiment disclosed in the specification, swiping a finger
`
`across a touch screen does produce multiple position signals (i.e., “device sensor
`
`signals”) from the touchscreen hardware at different times, but those position
`
`signals are together the constituents of a single “swipe” gesture:
`
`Multiple inputs from the index finger are received from the single
`gesture. Each of the multiple inputs may occur at a different time and
`may indicate a different two dimensional position of the contact point
`of the index finger with the touch sensitive display.
`
`
`
`
`- 9 -
`
`
`
`

`

`
`Ex. 1001 at 10:36-43 (emphasis added); see also id. at 14:18-22 (a single “scrolling
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`gesture” is detected based on moving a finger across the touchscreen over time).
`
`See also Ex. 2001 at ¶ 41.
`
`The common element to these embodiments disclosed in the specification is
`
`that the gestures that are recognized by a given system and encapsulated in the
`
`claimed “gesture signals” depend on how the system is programmed and
`
`designed. Id. at ¶¶ 42-43. The specification discloses that it is possible to develop
`
`a system that recognizes separate “finger on” and “finger off” gestures and has no
`
`concept of “compound” gestures. Ex. 1001 at 3:34-62. Similarly, a different
`
`system could use the same hardware as the first, could be manipulated with the
`
`same user hand movements, and could process the same underlying sensor signals
`
`from the touchscreen, but instead may be configured to recognize only a “swiping”
`
`gesture. Id. at 3:37-52 (explaining that “finger on” and “finger off” gestures could
`
`be interpreted as a single, more complex gestures such as “tapping,” “long
`
`tapping,” “swiping,” “smearing,” “smudging,” or “flicking”); 14:18-22 (describing
`
`a single “scrolling gesture” that implicitly involves subsidiary “finger on,” “finger
`
`off,” and “swipe” actions). If the software in this different system is not equipped
`
`to track and package the underlying sensor signals into distinct “finger on” and
`
`“finger off” gestures, then the system does not detect a “gesture signal” for those
`
`gestures. Immersion’s proposed construction preserves this important aspect of the
`
`
`
`
`- 10 -
`
`
`
`

`

`
`claimed “gesture signals” by requiring the signals to represent a recognized
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`movement of the body. Ex. 2001 at ¶¶ 42-43.
`
`The prosecution history also supports Immersion’s construction. The
`
`distinction between position information and a gesture signal was made by the
`
`applicant during prosecution. See Ex. 2003 at 9 (August 2, 2012 Applicant
`
`Remarks in prosecution of U.S. Patent No. 8,279,193, a prior patent in the same
`
`family as the ’571 patent) (“[Prior art reference raised by the examiner] Marvit
`
`describes gestures in the context of motion sensor engagement for a handheld
`
`device. For example, input movement may be in the form of translation and/or
`
`gestures. Translation-based input focuses on a beginning point and endpoint of a
`
`motion and difference between such beginning points and endpoints.”) (emphasis
`
`added). A translation-based input (i.e., a change in position over time captured by
`
`sensors) is not necessarily a gesture. Ex. 2001 at ¶ 44.
`
`Immersion’s proposed construction is also consistent with the use of the
`
`term “gesture” in the art of human-computer interaction, where it represents a
`
`movement of the body that expresses a full interaction with or command to a
`
`device. See Ex. 2002 at 1 (http://www.dictionary.com/browse/gesture?s=t)
`
`(“gesture . . . 4. Digital Technology. a particular movement of the body, typically
`
`the fingers or hand, used to control or interact with a digital device (often used
`
`attributively): a gesture command; Use a two-finger pinching gesture on your
`
`
`
`
`- 11 -
`
`
`
`

`

`
`touchscreen to zoom in or out.”); Ex. 2001 at ¶ 45.
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`Professor Visell also explains that a person of ordinary skill in the art would
`
`understand that a “gesture signal” in the context of the ’571 patent would refer to
`
`“an electronic signal, representing a recognized movement of the body that
`
`conveys meaning or user intent.” Electronic devices typically include haptic
`
`feedback that is triggered once a gesture has been recognized. Id.
`
`2.
`
`Apple’s Construction Is Too Broad and Vitiates the
`Meaning of “Gesture”
`
`Apple’s construction of gesture signal (“a signal indicating user interaction
`
`with a user interface device”) is too broad. First, it is not limited to a movement of
`
`the hand, or even the body. Id. at ¶ 46. For example, a user could “interact” with a
`
`“user interface device” via electrodes attached to the scalp that measure voltage
`
`fluctuations resulting from ionic current within the neurons of the brain (i.e., using
`
`electroencephalography (EEG) to measure thoughts). Ex. 1001 at 11:13, 11:29.
`
`The ’571 patent does not contemplate referring to that as a “gesture.” Even the
`
`section of the specification that Apple hangs its proposed construction on (Ex.
`
`1001 at 3:35-36) requires a gesture to at least be a “movement of the body.”
`
`Apple’s proposed construction goes beyond that to encompass any user interaction,
`
`including interactions that do not involve hand or body movement. The claim term
`
`is “gesture signal,” not “interaction signal.”
`
`
`
`
`- 12 -
`
`
`
`

`

`Second, Apple’s proposed construction does not require that the gesture
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`
`
`signal correspond to a particular recognized gesture—i.e. to an action that
`
`demonstrates “meaning or user intent.” Ex. 1001 at 3:35-36 (“A gesture is any
`
`movement of the body that conveys meaning or user intent.”); Ex. 2001 at ¶ 47.
`
`There is no linkage in Apple’s construction to any distinct expressed user intent or
`
`to the recognition of a gesture. Compare Ex. 1001 at 3:56-59 (gesture is
`
`“recognized by a device” and “converted to electronic signals”).
`
`Third, Apple’s proposed construction conflates a “device sensor signal,”
`
`which is separately described and claimed (e.g., id. at claim 7) with a “gesture
`
`signal.” Id. at ¶ 48. There are many user interactions that could result in a “device
`
`sensor signal” that would also qualify as a “gesture signal” under Apple’s
`
`construction, but which have no direct relationship to a purposeful, gestural
`
`movement of the body. See, e.g., Ex. 1001 at 11:11-14 (device sensor signals that
`
`are based on some form of user interaction include, in addition to EEG signals,
`
`“any type of bio monitor such as skin or body temperature, blood pressure (BP),
`
`heart rate monitor (HRM), . . . or galvanic skin response (GSR)”).
`
`Fourth, Apple’s construction effectively eliminates the requirement for both
`
`a “first gesture signal” and a “second gesture signal” to be received. E.g., id. at
`
`claim 1. The sections for Burrough and Rosenberg ’373 in this preliminary
`
`response provide more details. Apple is reading the claims of the ’571 patent on
`
`
`
`
`- 13 -
`
`
`
`

`

`
`systems that detect only a single user gesture (if any gesture at all). See also Ex.
`
`Case IPR2016-01372
`Patent No. 8,659,571
`
`2001 at ¶ 49.
`
`The effect of Apple’s proposed construction, therefore, is essentially to read
`
`out “gesture signal” from the claim language and replace it with “interaction
`
`signal.” Ex. 2001 at ¶ 50. This approach directly contradicts the claims, the
`
`specification, and the prosecution history, as well as the relevant extrinsic
`
`evidence. The challenged claims use “gesture signals” that correspond to
`
`recognized movements of the body. Immersion’s proposed construction of
`
`“gesture signal” should be adopted, and Apple’s proposed construction should be
`
`rejected. Microsoft Corp. v. Proxyconn, Inc., 789 F.3d 1292, 1298 (Fed. Cir. 2015)
`
`(The Board should not “construe claims during IPR so broadly that its
`

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket