throbber
UNITED STATES PATENT AND TRADEMARK OFFICE
`
`________________
`
`BEFORE THE PATENT TRIAL AND APPEAL BOARD
`
`________________
`
`AXIS COMMUNICATIONS AB,
`CANON INC., and CANON U.S.A., INC.,
`Petitioners,
`
`v.
`
`AVIGILON FORTRESS CORPORATION,
`Patent Owner.
`____________
`
`Case No.: IPR2018-00138
`U.S. Patent No. 8,564,661
`
`Case No.: IPR2018-00140
`U.S. Patent No. 8,564,661
`
`________________
`
`DECLARATION OF DR. ALAN BOVIK, PH.D. IN SUPPORT OF PATENT
`OWNER’S RESPONSES TO INTER PARTES REVIEWS OF U.S. PATENT
`NO. 8,564,661
`
`Canon Ex. 1038 Page 1 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`
`TABLE OF CONTENTS
`
`PAGE
`INTRODUCTION...........................................................................................1
`I.
`QUALIFICATIONS AND EXPERIENCE.....................................................1
`II.
`III. MATERIALS CONSIDERED........................................................................7
`IV.
`PERSON OF ORDINARY SKILL IN THE ART........................................10
`V.
`LEGAL STANDARDS.................................................................................12
`Claim Construction .............................................................................12
`Anticipation.........................................................................................14
`Obviousness ........................................................................................15
`VI. U.S. PATENT NO. 8,564,661.......................................................................16
`Disclosure of the ’661 Patent..............................................................16
`Re-Exam of U.S. Patent No. 7,868,912 ..............................................20
`Description of Courtney............................................................21
`The ’912 Patent Overcame Courtney, Brill, and Olson............24
`The ’912 and ’661 Patents Share the Independence-Based
`Claim Elements.........................................................................27
`VII. CORRECT CLAIM CONSTRUCTION.......................................................33
`The Independence-Based Claim Elements .........................................33
`“User Rule that Defines an Event” .....................................................38
`VIII. DESCRIPTION OF ASSERTED ART.........................................................40
`Kellogg................................................................................................40
`Kellogg does not teach object or attribute detection ................41
`Kellogg does not teach the independence-based claim
`elements ....................................................................................44
`Dimitrova ............................................................................................47
`Dimitrova does not teach the independence-based claim
`elements ....................................................................................47
`Dimitrova does not teach user rules..........................................55
`Brill......................................................................................................56
`i
`
`Canon Ex. 1038 Page 2 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`I, Dr. Alan Bovik, being of legal age, hereby declare, affirm, and state the
`
`following:
`
`I.
`
`INTRODUCTION
`1.
`I have been retained as an independent expert by Patent Owner
`
`Avigilon Fortress Corporation (“Patent Owner” or “Avigilon”) for IPR2018-00138
`
`(“the 138 IPR”) and IPR2018-00140 (“the 140 IPR”), both involving U.S. Patent
`
`No. 8,564,661 (the “’661 patent”). I have personal knowledge of the facts and
`
`opinions set forth in this declaration, and, if called upon to do so, I would testify
`
`competently thereto. All of the opinions and conclusions found in this declaration
`
`are my own.
`
`2.
`
`I am being compensated at a rate of $500 per hour for my services. I
`
`am being paid regardless of the conclusions or opinions I reach. I have no personal
`
`or financial stake or interest in the outcome of the present proceedings, and my
`
`compensation is not dependent in any way upon the outcome of these proceedings.
`
`II. QUALIFICATIONS AND EXPERIENCE
`I expect to testify regarding my background, qualifications, and
`
`3.
`
`experience relevant to the issues in this litigation. I hold a Ph.D. in in Electrical and
`
`Computer Engineering from the University of Illinois, Urbana-Champaign (awarded
`
`in 1984). I also hold a Master's degree in Electrical and Computer Engineering from
`
`the University of Illinois, Urbana-Champaign (awarded in 1982).
`
`1
`
`Canon Ex. 1038 Page 3 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`I am a tenured full Professor and I hold the Cockrell Family Regents
`
`4.
`
`Endowed Chair at the University of Texas at Austin. My appointments are in the
`
`Department of Electrical and Computer Engineering, the Department of Computer
`
`Sciences, and the Department of Biomedical Engineering. I am also the Director of
`
`the Laboratory for Image and Video Engineering (“LIVE”).
`
`5.
`
`My research is in the general area of digital television, digital cameras,
`
`image and video processing, computational neuroscience, and modeling of
`
`biological visual perception. I have published over 800 technical articles in these
`
`areas and hold seven U.S. patents. I am also the author of The Handbook of Image
`
`and Video Processing, Second Edition (Elsevier Academic Press, 2005); Modern
`
`Image Quality Assessment (Morgan & Claypool, 2006); The Essential Guide to
`
`Image Processing (Elsevier Academic Press, 2009); and The Essential Guide to
`
`Video Processing (Elsevier Academic Press, 2009); and numerous other
`
`publications.
`
`6.
`
`I received the 2017 Edwin H. Land Medal from the Optical Society of
`
`America in September 2017 with citation: For substantially shaping the direction
`
`and advancement of modern perceptual picture quality computation, and for
`
`energetically engaging industry to transform his ideas into global practice.
`
`7.
`
`I received a Primetime Emmy Award for Outstanding Achievement in
`
`Engineering Development, for the Academy of Television Arts and Sciences, in
`
`2
`
`Canon Ex. 1038 Page 4 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`October 2015, for the widespread use of my video quality prediction and monitoring
`
`models and algorithms that are widely used throughout the global broadcast, cable,
`
`satellite and internet Television industries.
`
`8.
`
`Among other awards and honors, I have received the 2013 IEEE Signal
`
`Processing Society’s “Society Award,” which is the highest honor accorded by that
`
`technical society (“for fundamental contributions to digital image processing theory,
`
`technology, leadership and education”). In 2005, I received the Technical
`
`Achievement Award of the IEEE Signal Processing Society, which is the highest
`
`technical honor given by the Society, for “broad and lasting contributions to the field
`
`of digital image processing”; and in 2008 I received the Education Award of the
`
`IEEE Signal Processing Society, which is the highest education honor given by the
`
`Society, for “broad and lasting contributions to image processing, including popular
`
`and important image processing books, innovative on-line courseware, and for the
`
`creation of the leading research and educational journal and conference in the image
`
`processing field.”
`
`9.
`
`My technical articles have been widely recognized as well, including
`
`the 2009 IEEE Signal Processing Society Best Journal Paper Award for the paper
`
`“Image quality assessment: From error visibility to structural similarity,” published
`
`in IEEE Transactions on Image Processing, volume 13, number 4, April 2004; this
`
`same paper received the 2017 IEEE Signal Processing Society Sustained Impact
`
`3
`
`Canon Ex. 1038 Page 5 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`Paper Award as the most impactful paper published over a period of at least ten
`
`years; the 2013 Best Magazine Paper Award for the paper “Mean squared error:
`
`Love it or leave it?? A new look at signal fidelity measures,” published in IEEE
`
`Transactions on Image Processing, volume 26, number 1, January 2009; the IEEE
`
`Circuits and Systems Society Best Journal Paper Prize for the paper “Video quality
`
`assessment by reduced reference spatio-temporal entropic differencing,” published
`
`in the IEEE Transactions on Circuits and Systems for Video Technology, vol. 23,
`
`no. 4, pp. 684-694, April 2013; and the 2017 IEEE Signal Processing Letters Best
`
`Paper Award for the paper A. Mittal, R. Soundararajan and A.C. Bovik, “Making a
`
`‘completely blind’ image quality analyzer,” published in the IEEE Signal Processing
`
`Letters, vol. 21, no. 3, pp. 209-212, March 2013.
`
`10.
`
`I received the Google Scholar Classic Paper citation twice in 2017, for
`
`the paper “Image information and visual quality,” published in the IEEE
`
`Transactions on Image Processing, vol. 15, no. 2, pp. 430-444, February 2006 (the
`
`main algorithm developed in the paper, called the Visual Information Fidelity (VIF)
`
`Index, is a core picture quality prediction engine used to quality-assess all encodes
`
`streamed globally by Netflix), and for “An evaluation of recent full reference image
`
`quality assessment algorithms,” published in the IEEE Transactions on Image
`
`Processing, vol. 15, no. 11, pp. 3440-3451, November 2006. (the picture quality
`
`database and human study described in the paper, the LIVE Image Quality Database,
`
`4
`
`Canon Ex. 1038 Page 6 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`has been the standard development tool for picture quality research since its first
`
`introduction in 2003). Google Scholar Classic Papers are very highly-cited papers
`
`that have stood the test of time, and are among the ten most-cited articles in their
`
`area of research over the ten years since their publication.
`
`11.
`
`I have also been honored by other technical organizations, including the
`
`Society for Photo-optical and Instrumentation Engineers (SPIE), from which I
`
`received the Technology Achievement Award (2013) “For Broad and Lasting
`
`Contributions to the Field of Perception-Based Image Processing,” and the Society
`
`for Imaging Science and Technology, which accorded me Honorary Membership,
`
`which is the highest recognition by that Society given to a single individual, “for his
`
`impact in shaping the direction and advancement of the field of perceptual image
`
`processing.” I was also elected as a Fellow of the Institute of Electrical and
`
`Electronics Engineers (IEEE) “for contributions to nonlinear image processing” in
`
`1995, a Fellow of the Optical Society of America (OSA) for “fundamental research
`
`contributions to and technical leadership in digital image and video processing” in
`
`2006, and as a Fellow of SPIE for “pioneering technical, leadership, and educational
`
`contributions to the field of image processing” in 2007.
`
`12. Among other relevant research, I have worked with the National
`
`Aeronautics and Space Administration (“NASA”) to develop high compression
`
`image sequence coding and animated vision technology, on various military projects
`
`5
`
`Canon Ex. 1038 Page 7 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`for the Air Force Office of Scientific Research, Phillips Air Force Base, the Army
`
`Research Office, and the Department of Defense. These projects have focused on
`
`developing local spatio-temporal analysis in vision systems, scalable processing of
`
`multi-sensor and multi-spectral imagery, image processing and data compression
`
`tools for satellite imaging, AM-FM analysis of images and video, the scientific
`
`foundations of image representation and analysis, computer vision systems for
`
`automatic target recognition and automatic recognition of human activities, vehicle
`
`structure recovery from a moving air platform, passive optical modeling, and
`
`detection of speculated masses and architectural distortions
`
`in digitized
`
`mammograms. My research has also recently been funded by Netflix, Qualcomm,
`
`Facebook, Texas Instruments, Intel, Cisco, and the National Institute of Standards
`
`and Technology (NIST) for research on image and video quality assessment. I have
`
`also received numerous grants from the National Science Foundation for research
`
`on image and video processing and on computational vision.
`
`13.
`
`I have conducted extensive research throughout my career in the areas
`
`of image and video processing, analysis, and surveillance. For example in the time
`
`period of the ’661 patent I was funded by Freescale Semiconductor, Inc., to conduct
`
`research of video surveillance tools for industry perimeters and other outdoor areas.
`
`I was in the same time period funded by Advanced Digital Imaging Research (ADIR)
`
`to conduct research of face detection and recognition methods and systems, which
`
`6
`
`Canon Ex. 1038 Page 8 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`are common integral components of video surveillance systems. I have published
`
`extensively in these areas and authored a chapter in one of my books on image and
`
`video processing entitled “Video Surveillance.”
`
`14. Additional details about my employment history, fields of expertise,
`
`and publications are further described in my curriculum vitae, which is attached as
`
`Exhibit A to this report.
`
`III. MATERIALS CONSIDERED
`15.
`In forming the opinions set forth herein, I have considered and relied
`
`upon my education, knowledge in the relevant field, and my experience. I have also
`
`reviewed and considered the materials cited herein, including the following
`
`materials:
`
`Description
`
`Exhibit No.
`IPR2018-00138
`
`Exhibit No.
`IPR2018-00140
`
`U.S. Patent No. 8,564,661 (“the ’661
`patent”)
`Prosecution History of the ’661 patent
`Ex. 1003 “Visual Memory” by
`Christopher James Kellogg (“Kellogg”)
`“Motion Recovery for Video Content
`Classification” by N. Dimitrova et al.
`(“Dimitrova”)
`“Event Recognition and Reliability
`Improvements for the Autonomous Video
`Surveillance System” by Frank Brill et al.
`(“Brill”)
`
`Ex. 1001
`
`Ex. 1002
`
`Ex. 1003
`
`-
`
`Ex. 1001
`
`Ex. 1002
`
`-
`
`Ex. 1003
`
`Ex. 1004
`
`Ex. 1004
`
`7
`
`Canon Ex. 1038 Page 9 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`
`Description
`
`Exhibit No.
`IPR2018-00138
`
`Exhibit No.
`IPR2018-00140
`
`Ex. 1005
`
`-
`
`Ex. 1006
`
`Declaration of John R. Grindon, D.Sc.
`(138 IPR)
`Declaration of John R. Grindon, D.Sc.
`(140 IPR)
`Curriculum Vitae of John R. Grindon,
`D.Sc.
`Declaration of Emily R. Florio (138 IPR) Ex. 1007
`Declaration of Emily R. Florio (140 IPR)
`Reference Numbering for Challenged
`Claims 1-32
`’912 ex parte reexamination - Applicant
`Response of Oct. 30, 2013
`’912 inter partes reexamination -
`Applicant Response of June 11, 2012
`Declaration of Christopher James Bailey-
`Kellogg
`90/012,878 Reexamination, Final Action
`(March 27, 2014)
`U.S. Patent No. 6,628,835, to Brill
`“Moving Object Detection and Event
`Recognition Algorithms for Smart
`Cameras,” Thomas J. Olson and Frank Z.
`Brill, Proceedings of the 1997 Image
`Understanding Workshop, New Orleans,
`May 1997, pp. 159-175
`U.S. Patent No. 7,868,912
`90/012,878 Reexamination, Declaration
`of Dr. Kenneth Zeger (Oct. 30, 2013)
`U.S. Patent App. 11/828,842 (issued as
`the ’661 patent), IDS Form by Applicant
`(March 21, 2012)
`
`Ex. 1008
`
`Ex. 1009
`
`Ex. 1010
`
`Ex. 1011
`
`Ex. 2001
`
`Ex. 2002
`Ex. 2003
`
`Ex. 2004
`Ex. 2005
`
`-
`
`8
`
`-
`
`Ex. 1005
`
`Ex. 1006
`
`Ex. 1007
`
`Ex. 1008
`
`Ex. 1009
`
`Ex. 1010
`
`-
`
`Ex. 2001
`
`Ex. 2002
`Ex. 2003
`
`Ex. 2004
`Ex. 2005
`
`Ex. 2006
`
`Canon Ex. 1038 Page 10 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`
`Description
`
`Declaration of Dr. Alan Bovik in Support
`of Patent Owner’s Response
`U.S. Patent No. 5,969,755 to Courtney
`Deposition Transcript of Dr. John
`Grindon (Aug. 15, 2018)
`“Automatic video indexing via object
`motion analysis,” Pattern Recognition,
`30(4), April 1997, by J.D. Courtney
`
`Exhibit No.
`IPR2018-00138
`Ex. 2007
`
`Exhibit No.
`IPR2018-00140
`Ex. 2007
`
`Ex. 2008
`Ex. 2009
`
`Ex. 2010
`
`Ex. 2008
`Ex. 2009
`
`Ex. 2010
`
`16.
`
`I have also considered the papers filed in each of the 138 IPR and the
`
`140 IPR, including the Petition, the preliminary response, and the institution
`
`decision.
`
`17.
`
`I have also considered other background references, of which I had
`
`been previously aware, and not cited herein, that a person of ordinary skill in the art
`
`(“POSA”) at the time of the invention would have recognized as being related to the
`
`subject matter of the ’661 patent.
`
`18.
`
`I have reviewed Dr. Zeger’s declaration from the ’878 Re-Exam,
`
`submitted in these trials as Exhibit 2005. I agree with Dr. Zeger’s conclusions, and
`
`incorporate the following paragraphs from Exhibit 2005 into my declaration in their
`
`entirety: 31-35, 66-82, 88-93, 99-110, and 115.
`
`19.
`
`This declaration is made based on information currently available to
`
`me. I intend to continue my investigation and study, which may include a review of
`
`9
`
`Canon Ex. 1038 Page 11 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`documents and information that may yet be produced, as well as deposition
`
`testimony from depositions for which transcripts are not yet available or that may
`
`yet be taken in this proceeding. Therefore, I expressly reserve the right to expand or
`
`modify my opinions as my investigation and study continue, and to supplement my
`
`opinions in response to any additional information that becomes available to me, any
`
`matters raised by Patent Owner and/or other opinions provided by Patent Owner’s
`
`expert(s), or in light of any relevant orders from the Patent Trial and Appeal Board
`
`or other authoritative body.
`
`IV. PERSON OF ORDINARY SKILL IN THE ART
`20.
`I have been asked for the purposes of this declaration to opine on the
`
`knowledge and understanding of one of ordinary skill in the art as of October 24,
`
`2000, or before.
`
`21.
`
`I have been informed and understand that a person of ordinary skill in
`
`the art (“POSA”) is a hypothetical person who is presumed to have been familiar
`
`with the relevant art at the time of the invention. I have also been informed and
`
`understand that, in determining the level of skill in the art, one may consider: the
`
`type of problems encountered in the art, prior art solutions to those problems, the
`
`rapidity with which innovations are made, the sophistication of the technology, and
`
`the educational level of active workers in the field.
`
`10
`
`Canon Ex. 1038 Page 12 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`I understand that Petitioners and their retained expert Dr. Grindon have
`
`22.
`
`alleged that appropriate level of skill for a POSA in the art of the ’661 patent “would
`
`have (i) a Bachelor of Science degree in electrical engineering, computer
`
`engineering, or computer science, with approximately two years of experience or
`
`research related to video processing and/or surveillance systems or (ii) equivalent
`
`training and work experience in computer engineering and video processing and/or
`
`surveillance systems.” I understand that Petitioners and Dr. Grindon further claim
`
`that a POSA “would be knowledgeable and familiar with the video processing and
`
`information extraction concepts and techniques recited in the claims of the ’661
`
`patent, as they were well known in 2000, as demonstrated above.” Ex. 1005 (’138
`
`IPR), ¶¶98-99
`
`23.
`
`I disagree with Petitioners’ level of skill for a POSA. In my opinion, a
`
`POSA for the ’661 patent would have practical experience working with and
`
`designing video surveillance systems. Accordingly, a POSA in the art of the ’661
`
`patent would have (i) a Bachelor of Science degree in electrical engineering,
`
`computer engineering, or computer science, with approximately two years of
`
`experience or research in the field of video surveillance systems or (ii) equivalent
`
`training and work experience in the field of video surveillance systems. In addition,
`
`a POSA would be knowledgeable and familiar with video processing, including
`
`segmentation and information extraction known at the time of the ’661 patent.
`
`11
`
`Canon Ex. 1038 Page 13 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`
`V.
`
`LEGAL STANDARDS
`24.
`In preparing and expressing my opinions and considering the subject
`
`matter of the ’661 patent, I am relying on certain basic legal principles that counsel
`
`have explained to me. These principles are discussed below.
`
`Claim Construction
`I understand that the first step in determining the validity of a claim is
`
`25.
`
`for the claim to be properly construed. I have been further advised that, in an inter
`
`partes review proceeding, the claims of an unexpired patent are to be given their
`
`broadest reasonable interpretation in view of the specification. I have also been
`
`informed that claim terms are typically given their ordinary and customary meaning
`
`as would be understood by one of ordinary skill in the art in the context of the patent
`
`disclosure (specification), unless the specification or the file history of the patent
`
`provides a specific definition. I understand that a patentee can be his own
`
`lexicographer, and that any special definition for a claim term must be set forth in
`
`the specification with reasonable clarity and precision.
`
`26.
`
`I understand that claim elements may be expressed as a means for
`
`performing a recited function as set out in 35 U.S.C. § 112, ¶ 6. I understand that
`
`construing such terms is a two-step process. First the function of the means-plus-
`
`function term must be identified. Second, the corresponding structure to that
`
`function must be identified. To be corresponding structure, the specification must
`
`12
`
`Canon Ex. 1038 Page 14 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`adequately disclose this structure based on the understanding of one skilled in the
`
`art, and the disclosed structure must be clearly linked with the recited function.
`
`However it is not enough that one of skill in the art would be capable of
`
`implementing a structure but rather that one of skill in the art would understand that
`
`the specification itself discloses the structure.
`
`27.
`
`I understand that claim elements that recite the word “means” as a
`
`structural element for performing one or more recited functions are presumed to be
`
`covered by 35 U.S.C. § 112, ¶ 6. I further understand that for such a means-plus-
`
`function element, the element is to be construed to cover only the structure or
`
`structures described in the patent specification for per-forming the exact function
`
`recited by the element and structural equivalents thereof. I further understand that
`
`structures are deemed equivalent if they are insubstantially different. One way of
`
`determining whether structures are equivalent is to determine whether each performs
`
`the identical recited function in a substantially similar way to obtain a substantially
`
`similar result. I understand that a structural equivalence analysis must be supported
`
`by specific evidence, and that a conclusory statement alleging that a structure within
`
`an accused product or process is equivalent to structure disclosed in the patent for
`
`performing the identical recited function is insufficient. I further understand that a
`
`structure may be equivalent only if it was available at the time of the issuance of the
`
`claim including the means-plus-function element being considered.
`
`13
`
`Canon Ex. 1038 Page 15 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`I also understand that when construing means-plus-function limitations
`
`28.
`
`that concern a general purpose computer or a microprocessor that is specially
`
`programmed to carry out an algorithm, the corresponding structure is to be construed
`
`as the general purpose computer and the algorithm disclosed in the patent
`
`specification by which the processor performs the claimed function. I understand
`
`that in this scenario, an algorithm must be disclosed unless the claimed functions can
`
`be achieved by any general purpose computer, without special programming, which
`
`I understand is a narrow exception that only applies in rare circumstances. I
`
`understand that an algorithm must be disclosed if the claimed function requires more
`
`than merely plugging in a general purpose computer. I also understand that if the
`
`algorithm is missing, then the claim is invalid as indefinite.
`
`Anticipation
`I have been informed that a patent claim is anticipated under pre-AIA
`
`29.
`
`35 U.S.C. § 102 if each and every element of the claim, as properly construed, is
`
`found either explicitly or inherently in a single prior art reference. I also understand
`
`that an element of a patent claim is inherent in a prior art reference if the element
`
`must necessarily be present and such would be recognized by one of ordinary skill
`
`in the art. However, I understand that inherency cannot be established by mere
`
`probabilities or possibilities. I also understand that to qualify as a sufficient
`
`disclosure under 35 U.S.C. § 102, the reference must enable the claimed invention
`
`14
`
`Canon Ex. 1038 Page 16 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`such that a POSA could implement the claimed invention without undue
`
`experimentation.
`
`Obviousness
`I have been informed that a claim may be invalid as obvious if the
`
`30.
`
`subject matter described by the claim as a whole would have been obvious to a POSA
`
`at the time the claimed invention was made. I understand that the standard for
`
`obviousness in an inter partes review proceeding is by a preponderance of the
`
`evidence.
`
`31.
`
`I have also been informed that a determination of obviousness involves
`
`an analysis of the scope and content of the prior art, the similarities between the
`
`claimed invention and the prior art, and the level of ordinary skill in the art. I have
`
`been informed and understand that a prior art reference should be viewed as a whole.
`
`32.
`
`I have been informed that in considering whether an invention for a
`
`claimed combination would have been obvious, I may assess whether there are
`
`apparent reasons to combine known elements in the prior art in the manner claimed
`
`in view of interrelated teachings of multiple prior art references, the effects of
`
`demands known to the design community or present in the market place, and/or the
`
`background knowledge possessed by a POSA. I also understand that other principles
`
`may be relied on in evaluating whether a claimed invention would have been
`
`obvious.
`
`15
`
`Canon Ex. 1038 Page 17 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`I have been informed that, in making a determination as to whether or
`
`33.
`
`not the claimed invention would have been obvious to a POSA, one may consider
`
`certain objective indicators of non-obviousness if they are present, such as:
`
`commercial success of product(s) practicing the claimed invention; long-felt but
`
`unsolved need; teaching away; unexpected results; copying; and praise by others in
`
`the field. I understand that for such objective evidence to be relevant to the non-
`
`obviousness of a claim, however, there must be a causal relationship (called a
`
`“nexus”) between the claim and the evidence. I also understand that this nexus must
`
`be based on a novel element of the claim rather than something available in the prior
`
`art.
`
`34.
`
`I have also been informed and understand that when considering the
`
`obviousness of a patent claim, one should consider whether a reason or motivation
`
`existed for combining the elements of the references in the manner claimed, and that
`
`the prior art must create a reasonable expectation of success in producing the claimed
`
`subject matter.
`
`VI. U.S. PATENT NO. 8,564,661
`Disclosure of the ’661 Patent
`The ’661 patent is directed to the field of video surveillance systems.
`
`35.
`
`Video surveillance systems are used to monitor areas of interest and detect whether
`
`objects are engaged in activities, also known as events. Prior to the ’661 patent,
`
`16
`
`Canon Ex. 1038 Page 18 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`video surveillance systems used “large volumes of video imagery as the primary
`
`commodity of information exchange.” Ex. 1001, 6:19-21. These early systems
`
`required extensive human viewing, and thus these systems recorded and stored all
`
`video regardless of whether the footage contained an event so that a human could
`
`later review the video. Ultimately, these early systems required large amounts of
`
`storage and analysis time.
`
`36.
`
`Later video surveillance systems prior to the ’661 patent digitally
`
`analyzed video to determine video primitives that could be used as a short hand to
`
`describe certain aspects of what was captured on video so that event detection could
`
`become an automated process. An example of a video primitive is the color of a
`
`shirt worn by a person captured on video, where a value of “red” is referred to as an
`
`“attribute” of the shirt object. In other words, attributes are observable
`
`characteristics of an object in a video. Some examples of attributes are:
`
`size, shape, perimeter, position, trajectory, speed and direction of motion,
`motion salience and its features, color, rigidity, texture, and/or classification.
`The object descriptor may also contain some more application and type
`specific information: for humans, this may include the presence and ratio of
`skin tone, gender and race information, some human body model de-scribing
`the human shape and pose; or for vehicles, it may include type (e.g., truck,
`SUV, sedan, bike, etc.), make, model, license plate number. The object
`descriptor may al-so contain activities, including, but not limited to, carrying
`an object, running, Walking, standing up, or raising arms. Some activities,
`such as talking, fighting or colliding, may also refer to other objects. The
`object descriptor may also contain identification information, including, but
`not limited to, face or gait.
`
`17
`
`Canon Ex. 1038 Page 19 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`
`Ex. 1001, 19:10-25.
`
`37.
`
`These systems prior to the ’661 patent and its family looked specifically
`
`for primitives that were necessary to satisfy event queries regarding events that were
`
`predetermined or predefined by the system. An exemplary predefined rule is to send
`
`an alert when the event of a person wearing a red shirt enters a particular door. These
`
`systems were limited by system rules that defined events such that these systems
`
`were constrained in capability to only collect metadata (primitives) that were
`
`associated with the events that had been predefined by the system.
`
`38.
`
`The ’661 patent describes a system that detects, collects, and stores
`
`video primitives that correlate to attributes of objects in a scene. Ex. 1001, 5:51-6:2.
`
`While the ’661 patent specification discusses databases generally, the claims do not
`
`require that any attributes are stored in a database, nor do the claims attempt to claim
`
`any database specific functionality. “Attributes” are information about objects,
`
`while “events” are determined by analyzing attributes to determine what an object
`
`in a scene is doing. Id., 6:3-14. More precisely, the ’661 patent teaches that after
`
`multiple attributes have been detected and collected, they can be analyzed to
`
`determine whether or not an event has occurred, including events that were not
`
`contemplated when the attributes were detected and collected. Id., 13:51-62.
`
`39.
`
`The ’661 patent further describes the process of “tasking” the system.
`
`Specifically, tasking allows a user to create user rules, whereby the user specifies
`
`18
`
`Canon Ex. 1038 Page 20 of 161
`
`

`

`Case Nos. IPR2018-00138 and IPR2018-00140
`U.S. Patent No. 8,564,661
`one or more “event discriminators.” Ex. 1001, 16:25-30. “Event discriminators are
`
`identified with one or more objects (whose descriptions are based on video
`
`primitives), along with one or more optional spatial attributes, and/or one or more
`
`optional temporal attributes.” Id., 6:3-14. For example, a user can define an event
`
`discriminator (called a “loitering” event in this example) as a “person” object in the
`
`“automatic teller machine” space for “longer than 15 minutes” and “between 10:00
`
`p.m. and 6:00 a.m.” Id., 6:4-12; see also 18:16-19. The user can also defin

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket