`
`
`
`
`
`
`
`
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`
`
`
`
`
`UNITED STATES PATENT AND TRADEMARK OFFICE
`__________________
`
`BEFORE THE PATENT TRIAL AND APPEAL BOARD
`__________________________________________________________________
`
`SUBARU OF AMERICA, INC.; MAZDA MOTOR OF AMERICA, INC.;
`VOLVO CARS OF NORTH AMERICA, LLC; MITSUBISHI ELECTRIC
`CORPORATION; MITSUBIHSI MOTORS CORPORATION; DAIMLER AG;
`BMW OF NORTH AMERICA, LLC
`Petitioners
`v.
`
`BLITZSAFE TEXAS, LLC
`Patent Owner
`
`
`
`
`
`Patent No. 8,155,342
`Issue Date: April 10, 2012
`Title: MULTIMEDIA DEVICE INTEGRATION SYSTEM
`__________________________________________________________________
`
`DECLARATION OF DR. RICHARD STERN, PH.D.
`
`Case No. IPR2018-00090
`________________________________________________________________
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`IPR2018-00090 – Ex. 2001
`
` Blitzsafe Texas, LLC, Patent Owner
`1
`
`
`
`
`
`
`I, Richard Stern, declare as follows:
`
`
`
`
`
`
`
`
`
`
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`I.
`1.
`
`BACKGROUND AND QUALIFICATIONS
`
`A copy of my curriculum vitae that summarizes my education, work history
`
`and publications is attached to this report as Exhibit A.
`
`2.
`
`I have over 40 years of experience in the fields relevant to multimedia
`
`device integration. In that time, I have studied and researched signal
`
`processing and acoustics. Specifically, I have studied, researched, and
`
`analyzed signal processing of audio sounds, which involves the creation of a
`
`signal, the communication of data through the signal from one device to
`
`another, and the interpretation of that signal by a human listener. While my
`
`main areas of current professional activity are primarily in signal processing
`
`for robust speech recognition and auditory perception, I have additionally
`
`studied and performed research in a wide range of related fields of audition,
`
`acoustics, signal processing, and instrumentation.
`
`3.
`
`I received an S.B. in Electrical Engineering from the Massachusetts Institute
`
`of Technology (MIT) in 1970, an M.S. in Electrical Engineering and
`
`Computer Sciences from the University of California, Berkeley in 1972, and
`
`my Ph.D. in Electrical Engineering and Computer Science from the
`
`Massachusetts Institute of Technology in 1977.
`
`4. While I was a student, I worked as a Teaching and Research Assistant in the
`
`
`
`2
`
`2
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`Department of Electrical Engineering at MIT, from 1973 to 1976. My
`
`
`
`
`
`
`
`teaching experience was in the area of signal processing under the direct
`
`supervision of Professors Alan Oppenheim and Alan Willsky, and in the
`
`area of acoustics under the direct supervision of Professor Amar Bose. My
`
`research at MIT had been in the area of auditory perception.
`
`5.
`
`I am a Professor of Electrical and Computer Engineering at Carnegie Mellon
`
`University (CMU), where I have taught and carried out research since 1977.
`
`While my primary appointment is with the Department of Electrical and
`
`Computer Engineering, I am also a Professor by Courtesy in the Language
`
`Technologies Institute and Computer Science Department. I have also been
`
`a Lecturer in CMU’s School of Music since 2008, and I was a Professor in
`
`the CMU Biomedical Engineering Department from 1977 to 1995. From
`
`1995 to 2003, I was Associate Director of the CMU Information Networking
`
`Institute, where I was responsible for every aspect of its Master of Science in
`
`Information Networking, including admissions, curricular development and
`
`support, and student life. In addition to my appointments at Carnegie
`
`Mellon, I was an invited Visiting Professor at Nippon Telegraph and
`
`Telephone Laboratories in Tokyo in 1985 and at the Nara Institute of
`
`Science and Technology in Japan in 2003. I was also an Adjunct Assistant
`
`Professor of Otolaryngology at the University of Pittsburgh School of
`
`
`
`3
`
`3
`
`
`
`
`Medicine from 1979 to 1981.
`
`
`
`
`
`
`
`
`
`
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`6.
`
`As noted above, my main areas of professional activity are automatic speech
`
`recognition, auditory perception, signal processing, and acoustics. My work
`
`is well known and widely cited, as a search on the generic term “robust
`
`speech recognition” will reveal. In those fields, I have been author or co-
`
`author of more than 35 archival journal articles, 13 book chapters, and more
`
`than 30 invited conference presentations, many of which were keynote
`
`addresses at major meetings, and a large number of additional critically-
`
`reviewed conference presentations. I have been invited to present my work
`
`on speech recognition in China, Japan, Korea, the Czech Republic,
`
`Denmark, France, Germany, Italy, Portugal, Spain, the United Kingdom,
`
`Israel, India, Mexico, Argentina, Brazil, and Chile, as well as virtually every
`
`major corporate and academic center of research in speech recognition and
`
`related technologies in the United States. I have supervised 24 Ph.D. theses,
`
`and more than 25 M.S. research projects. My former doctoral and masters
`
`research students include the present or former heads of speech research at
`
`Apple, Microsoft, Google, and major Government intelligence agencies, the
`
`founders of startup companies valued at hundreds of millions of dollars, and
`
`professors in major research universities.
`
`7.
`
`I am one of the few individuals who is an elected Fellow of the Institute of
`
`
`
`4
`
`4
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`Electrical and Electronics Engineers (IEEE), the Acoustical Society of
`
`
`
`
`
`
`
`America (ASA), and the International Speech Communication Association
`
`(ISCA). I was the 2008-2009 Distinguished Lecturer of ISCA, which
`
`sponsored tours of South America and India for the purpose of disseminating
`
`my research results. I was also a co-recipient of the Allen Newell Award for
`
`Research Excellence in 1992, and I was the recipient of the CMU Electrical
`
`Engineering Department’s annual teaching award in 1979.
`
`8.
`
`I have served on numerous technical and standards committees for the IEEE,
`
`ISCA, and for the Human Language Technology Program of the U.S.
`
`Defense Department Advanced Research Projects Agency (DARPA). I
`
`served as General Chair of the 2006 Interspeech International Conference on
`
`Spoken Language Processing in Pittsburgh in 2006. The Interspeech
`
`meeting, which attracted in excess of 1100 attendees, is considered to be the
`
`major world forum (along with the IEEE International Conference on
`
`Acoustics, Speech, and Signal Processing) for the exchange and
`
`dissemination of new findings in speech processing by humans and
`
`machines. I also serve or have served as the Technical Program Committee
`
`Chair of the Meeting of the Acoustical Society of America in Pittsburgh in
`
`2002, Chair of the Nominating Committee for the IEEE James L. Flanagan
`
`Technical Field, and as a member of the ISCA International Advisory Board
`
`
`
`5
`
`5
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`and Nomination Committee for Fellows of ISCA, the Secretary for the
`
`
`
`
`
`
`
`DARPA Speech Coordinating Committee, and I have led or served as a
`
`member of other professional, academic, and governmental committees too
`
`extensive to be enumerated here.
`
`9.
`
`As a Professor, I have taught 15 courses, many of which are at the graduate
`
`level, in the general areas of signal processing, communication theory, and
`
`acoustics. The courses that I teach at CMU encompass all aspects of signal
`
`processing including adaptive filtering, array processing, short-time Fourier
`
`analysis, and multi-rate signal processing.
`
`10.
`
`I am being compensated at my customary rate of $450 per hour. My
`
`compensation is not dependent upon the substance of the opinions I offer
`
`below, the outcome of this petition or any issues involved in or related to
`
`U.S. Patent 8,155,342 (the “’342 Patent”). I have no financial interest in, or
`
`affiliation with, any of the real parties in interest or the patent owner.
`
`II. MATERIALS CONSIDERED
`11. The materials I considered include the ’342 Patent and the original
`
`prosecution history for the ’342 Patent including the applications to which
`
`the ’342 Patent claims priority. I also considered the materials that I refer to
`
`and that I cite in this declaration.
`
`12.
`
`I also considered the Petition for Inter Partes Review filed by Subaru of
`
`
`
`6
`
`6
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`America, Inc., Mazda Motor of America, Inc., Volvo Cars of North
`
`
`
`
`
`
`
`America, LLC, Mitsubishi Electric Corporation, Mitsubishi Motors
`
`Corporation, Daimler AG, and BMW of North America, LLC (IPR2018-
`
`00090) (the “Petition”) including Exhibits 1001-1022 including the
`
`declaration of Dr. Matheson.
`
`13.
`
`I also considered the full record associated with the Petition for Inter Partes
`
`Review filed by Toyota Motor Corporation (IPR2016-00418) (the “Toyota
`
`IPR”) including Exhibits 1001 – 1028, 2001- 2013, and 3001-3003.
`
`14.
`
`I have also considered the Board’s Decision on Institution of Inter Partes
`
`Review (Paper 15) in the -00418 IPR.
`
`15.
`
`I have also considered the Board’s Decisions to Deny Institution of Inter
`
`Partes Review in two other proceedings concerning the ’342 Patent:
`
`IPR2016-00419 (Paper 12) and IPR2016-00118 (Paper 11).
`
`16.
`
`I have also considered the Petitioner’s request for rehearing in the -00419
`
`proceeding (Paper 14) and the Board’s Decision denying rehearing (Paper
`
`15).
`
`17.
`
`I have also considered U.S. Patent Appl. No. 2001/0028717 to Ohmura
`
`(“Ohmura”), which was the primary reference at issue in the -00419 and -
`
`00118 denials of institution.
`
`18.
`
`In addition, I have drawn on my experience and knowledge, as discussed
`
`
`
`7
`
`7
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`above and described more fully in my CV, in the areas of audio and
`
`
`
`
`
`
`
`multimedia integration.
`
`III. LEGAL PRINCIPLES
`I understand that a patent claim is unpatentable as “obvious” if the subject
`19.
`
`matter of the claim as a whole would have been obvious to a person of
`
`ordinary skill in the art as of the time of the invention at issue.
`
`20.
`
`I understand that the use of “the person of ordinary skill” rubric is to prevent
`
`one from improperly, in the present day, using hindsight to decide whether a
`
`claim is obvious.
`
`21.
`
`I understand that the following factors must be evaluated to determine
`
`whether the claimed subject matter is obvious: (1) the scope and content of
`
`the prior art; (2) the difference or differences, if any, between the scope of
`
`the patent claim and the scope of the prior art; and (3) the level of ordinary
`
`skill in the art at the time of the invention.
`
`22.
`
`I understand that, unlike anticipation, which allows consideration of only
`
`one item of prior art, obviousness may be shown by considering more than
`
`one item of prior art. I understand that, when considering a combination of
`
`prior art references as part of an obviousness analysis, it can be important to
`
`ascertain if the references are from the same field of endeavor and also to
`
`ascertain whether there is any reason that would have prompted a person of
`
`
`
`8
`
`8
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`ordinary skill in the relevant art to combine the elements in the way the
`
`
`
`
`
`
`
`claim does. In other words, a claim generally cannot be rendered obvious by
`
`combining (i) art from across different fields, including outside the field of
`
`the claimed invention, or (ii) art that itself teaches away from combination
`
`with other art that would otherwise provide its missing limitations, or (iii) art
`
`for which there is not at least an articulable, common sense reason to bridge
`
`the gap between its disclosure and the claim at issue.
`
`23. Moreover, I understand that so-called “objective indicia of non-
`
`obviousness,” also known as “secondary considerations,” are also to be
`
`considered when assessing obviousness. I understand that this objective
`
`evidence includes at least: (1) the commercial success of the invention; (2)
`
`the long felt but unresolved need to develop the invention; and (3) any praise
`
`of the invention in the market. I also understand that evidence of objective
`
`indicia of non-obviousness must be commensurate in scope with the claimed
`
`subject matter; i.e., that there must be a nexus or connection between the
`
`criteria and the claim itself.
`
`IV. LEVEL OF ORDINARY SKILL
`I understand that I should perform my analysis from the viewpoint of a
`24.
`
`person of ordinary skill in the art. I understand that this hypothetical person
`
`of ordinary skill in the art is considered to have the normal skills of a person
`
`
`
`9
`
`9
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`in a certain technical field. I understand that factors that may be considered
`
`
`
`
`
`
`
`in determining the level of ordinary skill in the art include: (1) the education
`
`level of the inventor; (2) the types of problems encountered in the art; (3) the
`
`prior art solutions to those problems; (4) rapidity with which innovations are
`
`made; (5) the sophistication of the technology; and (6) the education level of
`
`active workers in the field.
`
`25.
`
`I do not disagree with Dr. Matheson’s proposed level of ordinary skill in the
`
`art, which he lists as a Bachelor’s degree in electrical engineering or
`
`equivalent science/engineering degree and at least two years of experience in
`
`signal processing and/or electronic system design, or at least four years of
`
`experience in signal processing and/or electronic system design.
`
`V.
`
`26.
`
` ANALYSIS
`A.
`I understand that this proceeding concerns the wireless connection of a
`
` Technical Background
`
`portable electronic device (such as a mobile phone or MP3 player) with a
`
`second device (in this case, the automobile) for the purposes of information
`
`and entertainment. I also understand that one of the issues in contention in
`
`this litigation concerns the method(s) by which audio (which is encoded or
`
`compressed into a format such as MP3) is transmitted from the portable
`
`electronic device to the automobile and converted into a format that is
`
`
`
`10
`
`10
`
`
`
`
`
`
`
`suitable for presentation to a human listener.
`
`
`
`
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`27. Audio and video coding are technologies that enable widespread access to
`
`media that we have come to enjoy in the 21st century, including video
`
`downloaded from services such as Netflix and YouTube, and audio from
`
`devices such as the iPod, iPad, iPhone, and Android phones. Typically, the
`
`motivation for video and audio coding is the reduction of file size. For
`
`example, a one-minute segment of music in stereo sampled at 44,100 Hz
`
`with 16 bits per sample (as in an audio CD) represents about 84.7 million
`
`bits of information. A representation of that segment of music using
`
`reasonably high-quality MP3 encoding with typical coding parameters
`
`would require only 11.5 million bits of storage. Clearly, this is a reduction
`
`of storage requirements by a factor of about 7.4 to 1, which concomitantly
`
`permits the storage of 7.4 times as much audio on a given device than would
`
`have been possible if the data had been stored in the original un-encoded
`
`form. (In practice the actual bit-rate reduction depends on the type of signal
`
`that is being coded as well as a number of user-selectable preferences.)
`
`28. MP3 coding is one of a class of protocols collectively referred to as
`
`perceptual audio coding, which exploits the observation that the auditory
`
`system masks certain components of sound in time and frequency. Bit-rate
`
`reduction is obtained by storing only a crude representation of those sound
`
`
`
`11
`
`11
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`components that are masked by others. Other perceptual audio coding
`
`
`
`
`
`
`
`methods include Advanced Audio Coding (AAC) and Unified Speech and
`
`Audio Coding (USAC). In this report, I use the term “MP3” to refer to any
`
`type of perceptual audio coding used by the portable electronic device,
`
`regardless of whether the actual coding method used is MP3, AAC, USAC,
`
`or some other method.
`
`29.
`
`I note that conversion from the MP3 representation on a portable electronic
`
`device, such as those implicated in this case, takes place in two separable
`
`stages: (1) decoding the MP3 parameters, which converts them into a digital
`
`representation of the original audible waveform, and (2) digital-to-analog
`
`conversion (D/A conversion), which converts the digital representation of
`
`the audible waveform into a continuous-time or “analog” representation
`
`which can be directly converted into sound by a traditional output device
`
`such as a loudspeaker or headphones.
`
`30.
`
`It is critically important to maintain a distinction between audio as
`
`represented by the audible waveform (or a digital representation of it) versus
`
`a representation of the audio by the coded MP3 parameters.
`
`31.
`
`In other words, “audio generated by the portable device” for the purpose of
`
`this proceeding specifically refers to the decoding of the MP3 on the
`
`portable device itself. Merely transmitting the MP3 file, either in whole or
`
`
`
`12
`
`12
`
`
`
`
`
`
`in part, would not satisfy that limitation.
`
`
`
`
`
`
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`32.
`
`In order to communicate with portable devices over a communications
`
`channel, it was possible to use Universal Serial Bus (USB) protocol or
`
`Bluetooth protocol. The actual communicated signal streams transmitted
`
`over these protocols, regardless of whether they represent an audible
`
`waveform or a sequence of MP3 parameters, are segmented into a set of
`
`subsequences or “packets” often referred to as a “stream.” Both the USB
`
`and Bluetooth protocols specify means by which communication is
`
`established between the devices.
`
`33.
`
`It was known that these standards permitted transfer of files, such as MP3
`
`audio files as packets of audio, referred to more commonly as a “stream” or
`
`“streaming.”
`
`34.
`
`I understand that this proceeding concerns the wireless connection of a
`
`portable electronic device (such as a mobile phone or MP3 player) with a
`
`second device (in this case, the automobile) for the purposes of information
`
`and entertainment. I also understand that one of the issues in contention in
`
`this litigation concerns the method(s) by which audio (which is encoded or
`
`compressed into a format such as MP3) is transmitted from the portable
`
`device to and received by the integration subsystem of the second device.
`
`
`
`13
`
`13
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`
`Figure 1: Comparison of Alternative Modes of Communication Between
`Portable Electronic Device and Automobile
`
`35. Figure 1 illustrates three different contrasting ways of converting the coded
`
`MP3 parameters which enable compact storage of the user’s music
`
`collection into an audible waveform that if presented to a human listener
`
`after digital-to-analog (D/A) conversion would be intelligible and
`
`informative and/or entertaining.
`
`36. The first approach, labeled Case 1, describes a process by which both the
`
`
`
`14
`
`14
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`MP3 decoding and D/A conversion takes place on the portable electronic
`
`
`
`
`
`
`
`device. In this case, the “communication” between the portable electronic
`
`device and the automobile is through an analog channel, which in fact is
`
`simply the conventional audio cable that connects the audio output of the
`
`portable electronic device to (typically) the auxiliary input of the
`
`entertainment system on the automobile.
`
`37. The second approach, labeled Case 2, describes a process by which the MP3
`
`decoding takes place on the portable electronic device and a digitally-
`
`encoded version of the audible waveform is transmitted over a digital
`
`communications channel using either wired or wireless protocols. This
`
`decoding is performed in response to a “play” command received by the
`
`portable device. The automobile receives the digitally-encoded waveform
`
`and converts it to an analog waveform suitable for direct connection to the
`
`speaker or headphones.
`
`38. The third approach, labeled Case 3, describes a process by which the coded
`
`MP3 parameters are transmitted directly over the digital communications
`
`channel using either wired or wireless protocols to the automobile. This file
`
`transfer is performed in response to a “play” command received by the
`
`portable device. In this case, both MP3 decoding and D/A conversion take
`
`place on the automobile itself.
`
`
`
`15
`
`15
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`In each of the cases, play commands may be received by the portable device.
`
`
`
`
`
`
`
`39.
`
`The play command is agnostic as to the approach for transmission and as to
`
`where the audio is generated. In Cases 2 and 3 above, there will be some
`
`encoding and decoding of the transmission for protocol purposes, e.g.,
`
`Bluetooth. The audio generated by the portable device in Case 2 may
`
`undergo additional encoding or decoding for transmission protocol purposes
`
`(e.g., Bluetooth) without affecting my analysis.
`
`40. As described in further detail below, the Clayton reference, much like the
`
`Ohmura reference, only teaches decoding of audio in the “interface”
`
`associated with the car stereo system, and not in the portable device, in
`
`response to “play” requests.
`
`The ’342 Patent
`
`B.
`41. The ’342 Patent was issued from a continuation-in-part application, claiming
`
`priority to several earlier applications including U.S. Patent Application No.
`
`11/071,667 (“the ’667 Application”). Ex. 1001. The ’342 Patent relates to a
`
`multimedia device integration system that allows a plurality of portable
`
`electronic devices to be wirelessly integrated into an existing car stereo
`
`system, via an “integration subsystem,” while allowing information to be
`
`displayed on, and control to be provided from the car stereo. Ex. 1001at
`
`2:44–54, 33:43–46; Abstract. The claims at issue in this proceeding are
`
`
`
`16
`
`16
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`directed to an integration subsystem in communication with the car
`
`
`
`
`
`
`
`audio/video system, allowing data and control signals to be exchanged
`
`between the portable device and the car audio/video system. The integration
`
`subsystem processes and formats data so that instructions and information
`
`are processed by the portable device and vice versa, and permits audio and
`
`video generated, i.e. decoded, by the portable device to be played on the car
`
`audio/video system. Id. at 33:43–35:62, Figs. 18, 19.
`
`C. Claim Construction
`I have been instructed that the Board has applied the following constructions
`
`42.
`
`in the Toyota IPR, and that I should apply these constructions for the
`
`purposes of rendering my opinions:
`
`Integration Subsystem: The Board construed the term as “A subsystem to
`
`perform at least: (1) connecting one or more portable devices or inputs to the
`
`car audio/video system via an interface, (2) processing and handling signals,
`
`audio, and/or video information, (3) allowing a user to control the one or
`
`more portable devices via the car audio/video system, and (4) displaying
`
`data from the one or more portable devices on the car audio/video system.”
`
`Car Audio/Video System: The Board construed the term as “a car audio
`
`system, a car video system, or a car audio and video system.”
`
`Device Presence Signal: The Board construed the term as “a signal
`
`
`
`17
`
`17
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`indicating that a portable device is connected to the car audio/video system
`
`
`
`
`
`
`
`through the integration subsystem”
`
`I have also been instructed that Petitioners have submitted constructions for
`
`the terms:
`
`Multimedia device integration system: No construction (but noting that “the
`
`integration subsystem must perform the ‘integration’ function defined in the
`
`Specification of the ’342 patent.”)
`
`Generated . . . for playing on the car audio/video system: Plain and ordinary
`
`meaning).
`
`I do not believe that construction of these terms is necessary, nor do the
`
`constructions propounded by Petitioners change my analysis.
`
`D. The Clayton Reference
`43. Clayton discloses “embodiments that wirelessly provide content from a
`
`content source to a content player.” Ex. 1002 at Abstract, ¶ [0004]. Clayton
`
`teaches using a wireless interface to receive content from a content source,
`
`and to transmit the content using a wired connection to a content player (i.e.,
`
`a car audio/video system) for playback. Id.
`
`44. Clayton seeks to solve problems that arise from the fact that people use
`
`multiple content sources, such as MP3 players, cellular telephones, personal
`
`computers and car audio systems, to store and, in many cases, play content.
`
`
`
`18
`
`18
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`One of these problems is that “there is no currently fast and convenient way
`
`
`
`
`
`
`
`to transfer content between a user’s multiple devices.” Id. at ¶ [0003].
`
`45. Clayton broadly discloses a system that includes a content service 120 that
`
`provides content to target devices 140 via a network 130 as shown in Figure
`
`1.
`
`Clayton, Figure 1
`
`
`
`46. Content is defined as including “media such as audio, video, text;
`
`multimedia that includes two or more of audio, video, and text; or other
`
`types of data.” Id. at ¶ [0014] (emphasis added). While “content” is defined
`
`as including media such as audio, the disclosure of “audio content” or
`
`“content that is audio” would have been interpreted by one of ordinary skill
`
`in the art to mean audio files, and not the decoded audio itself.
`
`
`
`19
`
`19
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`47. Target devices 140 may include personal computers, cellular telephones, car
`
`
`
`
`
`
`
`radios, home stereos, MP3 players, and other end-user devices. Id. at ¶
`
`[0018]. Target devices are capable of downloading content from the content
`
`service and transmitting data to, and receiving data from, other target
`
`devices. Id. This disclosure is consistent with the understanding that the
`
`“content” disclosed by Clayton was data, such as audio files.
`
`48. Clayton describes embodiments in which a cellular telephone 143 can
`
`connect to a car radio and transfer its content, i.e., its files, to the car radio
`
`for playback on the car radio.
`
`Clayton, Figure 2
`
`
`
`49. Clayton discloses that “[a] portable content device such as the cellular
`
`
`
`20
`
`20
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`telephone 142 may send content to a content player such as the car audio
`
`
`
`
`
`
`
`system 143 via the wireless interface 150 of the cellular telephone 142. In
`
`addition, a wireless adaptor 173 may be used to enable communications
`
`between the cellular telephone 142 and the car audio system 143 for
`
`receiving content and for controlling playback of the content.” Id. at ¶
`
`[0052].
`
`50. The wireless adaptor 173 that is in communication with the car audio system
`
`“allows a suitably enabled portable device to wirelessly integrate with a
`
`component in a content player,” id. at ¶ [0053], which, in the case of Figure
`
`3, is the car radio component of a car audio system 143. See id.
`
`Clayton, Figure 3
`
`21
`
`
`
`
`
`21
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`51. The wireless adaptor 173 “provides an interface that allows the cellular
`
`
`
`
`
`
`
`telephone 142 to transmit stored content to the car audio system 143 to the
`
`car audio system 143 for content playback by the car radio component.” Id.
`
`at ¶ [0055].
`
`52. The description in Figure 3 of Clayton above does not teach where the audio
`
`itself is generated, and one of ordinary skill in the art would have needed
`
`more detail with regard to the underlying software and hardware in the black
`
`box of 173.
`
`53. Clayton does provide additional detail regarding the box 173 as shown in
`
`Figure 4. Box 173 has a decoder 440 that includes “a content decoder 446
`
`that decodes content received from the cellular telephone 142.” Id. at ¶
`
`[0056] (emphasis added). An optional metadata decoder 442 decodes
`
`metadata, such as song tiles and artists names, for display on the car audio
`
`system 143. Id.
`
`
`
`22
`
`22
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`
`Clayton, Figure 4 (annotated)
`
`54. The wireless adaptor has an application profiles stack 420 that includes
`
`“Bluetooth profiles to regulate the transmission of content from the cellular
`
`telephone 142 to the car audio system 143 . . . as streaming audio for stereo
`
`audio playback through the later [sic].” Id. at ¶ [0062-0063]. “The network
`
`manager 430 further controls the content decoder 446 to decode the
`
`streaming audio into a format understood by the car audio system 143 for
`
`output to the I/O interface 450, to which the car audio system 143 is
`
`connected through its own I/O interface.” Id.
`
`55. Thus, because Clayton only describes a system that sends an audio file, and
`
`
`
`23
`
`23
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`not a system that decodes the audio file on the portable device, and thus
`
`
`
`
`
`
`
`generates audio on the portable device, Clayton only describes Case 3,
`
`whereas the claims of the ’342 patent require Case 2.
`
`
`
`Figure 2, annotated version of Figure 1, depicting cases 1-3.
`
`56. Clayton describes an embodiment in which a cellular telephone acting as a
`
`content source transmits audio data to a car stereo, where the stereo acts as
`
`the playback device generating the audio. The wireless interface in
`
`communication with the car stereo receives the streamed audio data, which
`
`
`
`24
`
`24
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`one of ordinary skill in the art would have understood was a stream of audio
`
`
`
`
`
`
`
`data packets, then the interface decodes the data into a format understood by
`
`the car stereo, and outputs the decoded data as an audio stream to the car
`
`stereo system for playing on the car stereo.
`
`E.
`
`The Decisions Denying Institution of IPR2016-00419 and
`Rehearing of that Decision
`
`57. On July 19, 2016, the Board denied institution of the -00419 proceeding on
`
`grounds based on U.S. Patent Pub. No. 2001/0028717 A1 (“Ohmura”). The
`
`Board found that Ohmura did not meet the “audio generated by the portable
`
`device” limitation recited in claim 49 of the ’342 Patent because in Ohmura,
`
`music data is transferred from the portable device to the car audio/video
`
`system, and audio is generated by playing the transferred music file at the
`
`car audio/video system, and not on the portable device. Toyota Motor Corp.
`
`v. Blitzsafe Texas, IPR2016-00419, Paper 13 at 27-28 (July 19, 2016)
`
`(emphasis added).
`
`58.
`
`I agree with the Board’s determination that Ohmura does not teach the
`
`“audio generated by the portable device” limitation because Ohmura is
`
`another type of system that falls in “Case 3” that I describe above.
`
`59.
`
`I understand that Toyota requested rehearing of the decision denying
`
`institution in IPR2016-00419. I understand that Toyota argued that the
`
`Board overlooked that Toyota’s petition relied on an embodiment in Ohmura
`
`
`
`25
`
`25
`
`
`
`IPR2018-00090
`
`PATENT NO. 8,155,342
`
`
`
`
`that discloses “streaming” music data from the portable device to the car
`
`
`
`
`
`
`
`audio/video system, and not storing a music file on the car audio/video
`
`system for later decoding. Paper 14 at 4, 6-7. According to Toyota, this
`
`embodiment disclosed immediately decoding the data in the car audio/video
`
`system without storing the music data first. Id.
`
`60.
`
`I understand that the Board denied the petition for rehearing. Toyota Motor
`
`Corp. v. Blitzsafe Texas, IPR2016-00419, Paper 15 (August 31, 2016). The
`
`Board stated that all claims of the ’342 Patent require “the portable device to
`
`play (i.e., decode) an audio file.” Id. at 5. Ohmura does not meet the “audio
`
`generated on the portable device” limitation because “Ohmura’s CPU 101
`
`receives music data from portable device 200a or 200b that it must decode
`
`into audio for output, via speakers 20, of car audio system 100.” Id. at 6
`
`(emphasis added). I agree with the Board’s determination that (1)
`
`“decoding” by the portable device is required by the claims, and (2) that
`
`Ohmura, as another example of my aforementioned Case 3, does not decode
`
`audio on the portable device.
`
`VI. THE CLAYTON REFERENCE DOES NOT TEACH OR DISCLOSE
`THE “AUDIO GENERATED BY