throbber
Distributed Speech Recognition
`
`Enabling New Speech Driven Services for Mobile Devices:
`
`An overview of the ETSI standards activities for
`Distributed Speech Recognition Front-ends
`
`David Pearce
`
`&
`Chairman ETSI STQ-Aurora DSR Working Group
`
`bdp003@email.mot.com
`
`Motorola Limited
`Jays Close
`Viables Industrial Estate
`Basingstoke
`HANTS
`RG22 4PD
`UK
`
`AVIOS 2000: The Speech Applications Conference, May 22-24, 2000 San Jose, CA, USA
`
`Comcast - Exhibit 1007, page 1
`
`

`

`Enabling New Speech Driven Services for Mobile Devices:
`
`An overview of the ETSI standards activities for Distributed
`Speech Recognition Front-ends
`
`Abstract
`
`In a distributed speech recognition (DSR) architecture the recogniser front-end is located
`in the terminal and is connected over a data network to a remote back-end recognition
`server. DSR provides particular benefits for applications for mobile devices such as
`improved recognition performance compared to using the voice channel and ubiquitous
`access from different networks with a guaranteed level of recognition performance.
`Because it uses the data channel, DSR facilitates the creation of an exciting new set of
`applications combining voice and data. To enable all these benefits in a wide market
`containing a variety of players including terminal manufactures, operators, server
`providers and recognition vendors, a standard for the front-end is needed to ensure
`compatibility between the terminal and the remote recogniser. The STQ-Aurora DSR
`Working Group within ETSI has been actively developing this standard and as a result of
`this work the first DSR standard was published by ETSI in February 2000. This paper
`presents an overview of the standard for the DSR Mel-Cepstrum front-end and
`compression algorithm together with its performance characteristics. The current activity
`in Aurora is to develop a future standard for an Advanced DSR front-end that will give
`half the error rate in noise compared to the Mel-Cepstrum.
`
`1
`
`Introduction to Distributed Speech Recognition
`
`As has already happened in the wireline world, the trend to ever-increasing use of data
`communication is spreading to the mobile wireless world. As part of this, people want the
`ability to access information while on the move and the technologies to enable them to do
`this are now starting to be deployed. The small portable devices that will be used to
`access these data services cry out for improved user interfaces using speech input. At
`present, however, the complexity of medium and large vocabulary speech recognition
`systems are beyond the memory and computational resources of such devices.
`
`Centralised servers can share the computational burden between users and enable the
`easy upgrade of technologies and services provided. Mobile voice networks, however,
`can degrade the performance obtained from centrally deployed recognisers. The
`degradations are a result of both the low bit rate speech coding and channel transmission
`errors. A Distributed Speech Recognition (DSR) system overcomes these problems by
`eliminating the speech channel and instead using an error protected data channel to send a
`parameterised representation of the speech, which is suitable for recognition. The
`processing is distributed between the terminal and the network. The terminal performs
`the feature parameter extraction, or the front-end of the speech recognition system. These
`features are transmitted over a data channel to a remote “back-end” recogniser. The end
`
`AVIOS 2000: The Speech Applications Conference, May 22-24, 2000 San Jose, CA, USA
`
`Comcast - Exhibit 1007, page 2
`
`

`

`result is that the transmission channel has minimal impact on the recognition system
`performance and channel invariability is achieved. This performance advantage is good
`for DSR services provided for a particular network e.g. by a network operator and
`additionally so for 3rd party DSR applications that may be accessed over a variety of
`different networks.
`
`In summary the main benefits of DSR are as follows:
`
`•
`
`Improved recognition performance over wireless channels.
`The use of DSR minimises impact of speech codec & channel errors that reduce
`the performance from recognisers accessed over digital mobile speech channels
`(see figure 4 for an example)
`• Ease of integration of combined speech and data applications.
`Many new mobile multimodal applications are envisioned; such as the use of
`speech to access wireless internet content. The use of DSR enables these to
`operate over a single wireless data transport rather than having separate speech
`and data channels.
`• Ubiquitous access with guaranteed recognition performance levels.
`There are currently 4 different major digital mobile systems each using several
`different speech codecs. These all produce different effects on recognition
`performance. Mismatches in the channels used for training and recognition can
`result in severe degradations in performance, while models that have been trained
`over many networks give a compromise performance. DSR on the other hand
`offers the promise of guaranteed level of recognition performance over every
`network. It uses the same front-end and there is no channel distortion coming
`from the speech codec and its behaviour in transmission errors.
`
`2
`
`DSR Standardisation in ETSI – (STQ-Aurora DSR working group)
`
`To enable widespread applications using DSR in the market place, a standard for the
`front-end is needed to ensure compatibility between the terminal and the remote
`recogniser. The Aurora DSR Working Group within ETSI has been actively developing
`this standard over the last two years. To allow the optimisation of the details of the
`feature extraction algorithm a reference database and experimental framework has been
`established. The database is based on the original TIdigits database with controlled
`filtering and simulated noise addition over a range of signal to noise ratios from 20dB to
`–5dB. A reference recogniser configuration using Entropic’s HTK HMM software was
`agreed to investigate changes solely in the front-end. This database has been made
`publicly available via the European Language Resources Association (ELRA). Extensive
`experimentation has been performed using this and other internal databases to agree on
`the feature extraction and to test the compression algorithms. The first DSR standard was
`published by ETSI in February 2000. This standard and some aspects of its performance
`are described in the sections that follow.
`
`The Mel-Cepstrum was chosen for the first standard because of its widespread use
`throughout the speech recognition industry. It was acknowledged, however, that a front-
`
`AVIOS 2000: The Speech Applications Conference, May 22-24, 2000 San Jose, CA, USA
`
`Comcast - Exhibit 1007, page 3
`
`

`

`end that had improved performance in background noise would be desirable. The current
`activity of the Aurora working group is the development of a future standard for an
`Advanced DSR front-end that will give half the error rate in noise compared to the Mel-
`Cepstrum. A call for proposals has been publicised and in October 1999 eight
`organisations submitted candidates to the qualification phase. New evaluation databases
`with real-world noise are being developed to allow comparison between these algorithm
`candidates. A comprehensive set of evaluation criteria has been agreed and the proposers
`will make their final submissions to the selection phase in October 2000. The new
`standard is forecast to be published some time in 2002.
`
`Even after the Advanced standard is produced it is expected that the Mel-Cepstrum
`standard will continue to be used, since noise robustness can also be achieved by
`algorithms used at the server back-end (eg adaptation and model compensation
`techniques).
`
`It is anticipated that the DSR bitstream will be used as a payload in other higher level
`protocols when deployed in specific systems supporting DSR applications. Thus the
`standard does not cover the areas of data transmission or any higher level application
`protocols that may run over them. In this respect it is similar to speech codec standards
`where the codec is specified separately to the systems that use it.
`
`It is expected that the DSR Front-end will be used in a variety of both current and future
`mobile networks and associated protocols. Even so it is beneficial to agree on the
`appropriate combinations of protocols in the chain from the client terminal device to the
`recognition server to support DSR applications. A new sub-group of Aurora called “DSR
`Applications and Protocols” is now being formed to address this. DSR has the power to
`improve the performance of applications like Voice Activated WAP pages, Voice
`Browsing (multi-modal I/O) or Large Directory Assistance. The sub-group will work in
`collaboration with other standards groups such as W3C (internet standards), WAP
`Applications and 3GPP to recommend how to integrate DSR into applications and extend
`existing protocols where necessary.
`
`2
`
`The DSR Mel-Cepstrum Front-end and Compression Standard
`
`Figure 1 shows a block diagram of the processing stages for a DSR front-end. At the
`terminal the speech signal is sampled and parameterised using a Mel-Cepstrum algorithm
`to generate 12 cepstral coefficients together with C0 and a log energy parameter. These
`are then compressed to obtain a lower data rate for transmission. To be suitable for
`today’s wireless networks a data rate of 4800 b/s was chosen as the requirement. The
`compressed parameters are formatted into a defined bitstream for transmission.
`
`The defined bitstream is sent over a wireless or wireline transmission link to the remote
`server where parameters received with transmission errors are detected and the front-end
`parameters are decompressed to reconstitute the DSR Mel-Cepstrum features. These are
`passed to the recognition decoder residing on the central server. The recogniser back-end
`is not part of the standard.
`
`AVIOS 2000: The Speech Applications Conference, May 22-24, 2000 San Jose, CA, USA
`
`Comcast - Exhibit 1007, page 4
`
`

`

`Since the data channels used for the transport of the DSR bitstream may be subject to
`errors (transparent data channels), special attention has been given to make the whole
`system robust to the types of burst errors that occur on wireless channels. To achieve this,
`error detection bits are added in the terminal DSR encoder as part of the bitstream and a
`special error mitigation algorithm is used at the decoder.
`
`When developing the standard the following requirements were met:
`
`• Mel-Cepstrum feature set consisting of 12 cepstral coefficients logE and C0
`• Data transmission rate of 4800 b/s
`• Low computational and memory requirements for implementation in the mobile
`terminals
`• Low latency
`• Robustness to transmission errors
`
`Full details of the algorithms are given in the standards document [1] and can be accessed
`from the ETSI standards web site [see further information].
`
`Figure 1: Block diagram of DSR system
`
`Terminal DSR Front-end
`
`Parameterisation
`
`Mel-Cepstrum
`
`Compression
`
`Split VQ
`
`Frame structure &
`error protection
`
`Wireless data channel – 4.8kbit/s
`
`Server DSR Back-end
`
`Error detection
`& mitigation
`
`Decompression
`
`Recognition
`decoder
`
`3
`
`Optimisation of the Mel-Cepstrum Parameterisation
`
`Nokia submitted the original proposal for the Mel-Cepstrum parameterisation to be used
`in the standard. The feature vector consists of 14 components composed of 12 cepstral
`coefficients (C1 to C12) together with C0 and a log energy parameter. C0 was included to
`
`AVIOS 2000: The Speech Applications Conference, May 22-24, 2000 San Jose, CA, USA
`
`Comcast - Exhibit 1007, page 5
`
`

`

`support algorithms that might need it in the back-end, such as noise adaptation (e.g.
`PMC). The basic building blocks of aspects of the Mel-Cepstrum will be very familiar to
`speech recognition experts using this front-end and the block diagram of the standard is
`shown in figure 2. There were, however, many detailed points concerning its
`implementation that needed to be agreed and there was extensive review and optimisation
`within Aurora. In particular the following aspects were tested and the conclusion reached
`for each is noted:
`
`− Frame rate
`− Frame length
`− Filterbank definition
`− Energy calculation
`− DC removal
`− Liftering
`
`10ms
`25ms
`As proposed
`changed to before preemphasis
`GSM high pass filter
`Removed
`
`Figure 2: Block diagram of Mel-Cepstrum DSR Front-end standard
`
`Input
`speech
`
`ADC
`
`Offcom
`
`Framing
`
`PE
`
`W
`
`FFT
`
`MF
`
`LOG
`
`DCT
`
`Abbreviations:
`
`logE
`
`ADC
`Offcom
`PE
`logE
`W
`FFT
`MF
`LOG
`DCT
`MFCC
`
`analog-to-digital conversion
`offset compensation
`pre-emphasis
`energy measure computation
`windowing
`fast Fourier transform (only magnitude components)
`mel-filtering
`nonlinear transformation
`discrete cosine transform
`mel-frequency cepstral coefficient
`
`Feature Compression
`
`Bit Stream Formatting
`Framing
`
`To transmission channel
`
`4
`
`Compression Algorithm and its Performance Evaluation
`
`Two proposals were made for the compression algorithm and the one eventually selected
`for the standard was from Motorola.
`
`A split vector quantisation (VQ) algorithm is used to obtain a final total data rate of 4800
`bits per second of speech. A codebook of size 64 is used for each pair of cepstral
`coefficients from C1 to C12 and 256 vectors are used for C0 and energy. This results in
`44 bits per speech frame. Since the parameters may be transmitted over error prone
`
`AVIOS 2000: The Speech Applications Conference, May 22-24, 2000 San Jose, CA, USA
`
`Comcast - Exhibit 1007, page 6
`
`

`

`channels error detection bits (4 bits of CRC for each pair of speech frames) have been
`applied to the compressed data. For transmission and decoding the compressed speech
`frames are grouped into multiframes corresponding to 240ms of speech. The format is
`such that the bits corresponding to two frames may be transmitted as soon as they are
`ready. This results in only 10 ms additional latency at the terminal.
`
`For full details see the DSR standard document [1]
`
`The compression algorithms were tested on both small and large vocabulary tasks. The
`two areas that are important in evaluation of the performance of the compression
`algorithm are:
`
`• That it does not produce a degradation in recognition performance compared with
`using the floating point mel-cepstrum parameters directly
`• That the performance holds up in the presence of transmission errors typical of those
`that occur on wireless data channels (see section 5)
`
`4.1
`
`Small vocabulary database
`
`The experimental framework for evaluation of DSR proposals is described in [2]. The
`performance of the quantizer has been evaluated by measuring the performance both with
`and without quantization applied to the test set. Models were trained using logE as the
`energy measure on unquantized data. The results are summarised in table 1 and show
`that the quantizer does not introduce significant performance degradation. The
`performance from a DSR recognition system can be considered transparent to the
`compression scheme proposed.
`
`Table 1: Summary of recognition performance on Aurora noisy TIdigits evaluation
`databases
`
`Training database
`8kHz Multicondition
`8kHz Clean
`16kHz Multicondition
`16kHz Clean
`
`Unquantized test
`85.56
`68.80
`83.59
`62.95
`
`Quantized test
`85.40
`68.27
`84.00
`62.16
`
`4.2
`
`Large Vocabulary Databases
`
`The quantizer has also been tested over a wider range of tasks and recogniser
`configurations. The experiments presented here test the quantizers on large vocabulary
`tasks using sub-word models. The quantiser has been tested both on the Resource
`Management (RM) task and ATIS. Results presented here are for RM where two
`modelling configurations were used:
`
`AVIOS 2000: The Speech Applications Conference, May 22-24, 2000 San Jose, CA, USA
`
`Comcast - Exhibit 1007, page 7
`
`

`

`1) Monophone sub-word models
`
`The first configuration uses simple monophone modelling of 42 phones, each
`having 3 states and 5 mixtures per state. Monophone models for the ATIS3 task
`were trained using the RM1, ATIS2, and ATIS3 speaker independent training
`sets. RM monophone models, however, were trained on only the RM1 speaker
`independent training set. In experiments using monophone models, a bigram
`language model is used.
`
`2) Context dependent triphone sub-word models
`
`This is more indicative of a state-of-the-art modelling approach. Context-
`dependent triphone modelling was performed, resulting in 1708 triphones. Each
`triphone is represented by 3 states, 5 mixtures per state. Triphone models for both
`ATIS3 and RM were trained using the RM1, ATIS2, and ATIS3 speaker
`independent training sets. A trigram language model was also utilised instead of
`the simpler bigram for these experiments.
`
`The results are presented in the table 2:
`
`Table 2: Quantizer evaluation using RM (Sept 92)
`
`8kHz
`
`16kHz
`
`No
`Quantizer
`
`Motorola
`Quantizer
`
`No
`Quantizer
`
`Motorola
`Quantizer
`
`Unquantized
`
`Word %acc
`
`Triphone
`
`Word %corr
`
`Trigram
`
`Sent %corr
`
`Unquantized
`
`Word %acc
`
`Monophone
`
`Word %corr
`
`Bigram
`
`Sent %corr
`
`89.88
`
`90.93
`
`59.00
`
`80.70
`
`83.31
`
`40.00
`
`89.96
`
`91.01
`
`58.33
`
`80.77
`
`83.00
`
`42.00
`
`91.60
`
`92.65
`
`59.67
`
`83.86
`
`85.62
`
`40.67
`
`91.99
`
`92.97
`
`62.33
`
`83.74
`
`85.46
`
`40.67
`
`5
`
`Channel Error Robustness
`
`The algorithm for error mitigation consists of two stages:
`• Detection of speech frames received with errors
`• Substitution of parameters when errors are detected
`
`To detect the speech frames received with errors the 4 error detection bits on each pair of
`frames is used first. Since errors may be missed due to overloading of the CRC a
`heuristic algorithm that looks at the consistency of the parameters in the decoded frames
`is also used. It measures the difference between cepstral coefficients for adjacent frames
`and flags them as errored if it is greater than expected for speech. The thresholds used are
`
`AVIOS 2000: The Speech Applications Conference, May 22-24, 2000 San Jose, CA, USA
`
`Comcast - Exhibit 1007, page 8
`
`

`

`based on measurements of error free speech. If this algorithm was to run continuously
`then the number of misfirings would be too high, therefore it is only applied in presence
`of CRC errors.
`
`When an frame is flagged as having errors then the whole frame is replaced with the a
`copy of the cepstral parameters for the nearest good frame received (occurring before or
`after the frame under consideration).
`
`Robustness to channel errors has been measured according to procedures specified in
`Aurora on the noisy TIdigits task. Figure 3 shows the process for creating a test database
`that has been subject to channel error. The test set for these experiments is the digits at
`20dB SNR with models trained on multicondition 8kHz data. In each of the channel
`experiments the encoded bitstream has had the bit error mask for the corresponding
`channel applied before decoding. The channels tested are for TETRA and GSM. The
`channel tested for TETRA was one at the edge of coverage while for GSM the 3 channels
`tested were those commonly used for GSM codec testing. EP1 represents a good quality
`channel, EP2 a medium one and EP3 a poor channel beyond the normal design target.
`
`Figure 3: Evaluation tests for robustness to channel errors
`
`DSR Encoder
`
`Decoder
`
`Test
`Database
`
`Channel error mask
`
`Errored
`Test
`Database
`
`The results are presented in table 3 and show that for the TETRA channel at the edge of
`coverage and the EP1 and EP2 GSM data channels there is no significant degradation due
`to channel errors. For the GSM EP3 channel there is a 5% drop in performance. It
`should be noted that the EP3 channel is an extreme and represents an exceedingly poor
`channel. Speech audio quality for coded speech on such a channel is very poor. Speech
`recogniser performance based on speech coded with a GSM Enhanced Full Rate (EFR)
`codec and transmitted over an EP3 channel gives a performance of 78.1%. The
`significance of the performance benefits coming from DSR is illustrated in figure 4. For
`GSM channels the figure compares the error rates from DSR with those obtained using a
`GSM speech channel and the EFR codec.
`
`The error robustness of the compression scheme is exceptionally good. Under most
`channel conditions within the designed coverage of a mobile system the degradation will
`be negligible.
`
`AVIOS 2000: The Speech Applications Conference, May 22-24, 2000 San Jose, CA, USA
`
`Comcast - Exhibit 1007, page 9
`
`

`

`Table 3: Error robustness performance over data channels with transmission
`errors
`
`Channel
`Unquantized
`error free quantize
`TETRA TU50 20dB
`GSM EP1
`GSM EP2
`GSM EP3
`
`Noise 1
`97.26
`97.42
`97.17
`97.39
`97.39
`91.33
`
`Noise 2
`96.19
`96.28
`96.01
`96.28
`96.28
`91.41
`
`Noise 3
`97.61
`97.76
`97.26
`97.76
`97.64
`92.72
`
`Noise 4
`98.33
`98.33
`98.18
`98.33
`98.24
`93.55
`
`Average
`97.35
`97.45
`97.16
`97.44
`97.39
`92.25
`
`Figure 4: Performance with channel errors: DSR compared to a mobile speech
`channel
`
`GSM EFR Coded
`
`DSR
`
`100
`
`95
`
`90
`
`85
`
`80
`Baseline error free strong medium weak
`GSM signal strength
`
`WordAccuracy(%)
`
`6
`
`Latency
`
`The total latency introduced over a wireless channel is made up of components from
`• the encoder bitstream frame structure
`• a minimum transmission time over the data channel (note other implementation
`specific delays are ignored here since they will vary from system to system but are
`independent of the bitstream frame structure)
`• the decoder . This is the time from when a speech frame is received to when it can be
`made available to the back-end.
`
`Quantized speech vectors are transmitted as frame pair packets corresponding to 2
`frames. The additional latency introduced by the bitstream structure at the encoder is
`therefore 10ms (Note the Mel-Cepstrum FE parameterisation has a latency of 25ms
`while that for a frame pair is 25ms + 10ms = 35ms so the difference coming from the
`framing is 10ms).
`
`AVIOS 2000: The Speech Applications Conference, May 22-24, 2000 San Jose, CA, USA
`
`Comcast - Exhibit 1007, page 10
`
`

`

`There is a finite transmission time that depends on the data rate of the channel. In
`addition there is a delay as result of the frame structure of the data channel for the
`particular network (i.e. the interleaving in GSM) and implementation delays. For a
`particular channel though these will be constant and independent of the DSR bitstream.
`Once a frame packet pair of 92 bits has been processed and is ready for transmission the
`additional latency from the transmission of these parameters will be 92/data rate of
`channel (b/ms). For typical data channels such as GSM 9.6kbit/s this will be 9.6ms and
`for a TETRA 4.8kbit/s channel it will be 19.2ms.
`
`The decoder can introduce an additional delay. This is the time taken from the receipt of
`the speech frame over the channel to when it is made available to the back-end
`recogniser. If an error free channel can be guaranteed then there would be no additional
`delay from the decoder (processing delay ignored). The error mitigation strategy used in
`the decoder makes use of two frame packet pairs. There is therefore an additional 20ms
`delay introduced by the decoder for error prone channels.
`
`Encoder
`Transmission [92/data rate of channel (b/ms)]
`Decoder - error free transmission
`Decoder - with error mitigation
`Total additional Latency
`
`10ms
`9.6ms GSM9.6 ~ 19.2ms TETRA4.8
`0ms
`20ms
`30ms
`
`7
`
`Terminal i/p characteristics
`
`The goal of DSR is to have the best and consistent performance using server based
`recognisers. A further area where there is variability is in the terminal input
`characteristics including the frequency response of the microphone, analogue interface
`circuitry and the A/D converter. For GSM terminals the constraints on frequency
`response are specified in GSM 03.50 [3]. In Aurora an extensive set of experiments was
`conducted with data simulating the range different frequency responses conforming to
`this specification. A range of combinations of training and test conditions was tried. The
`conclusion was that if models were trained using data that had been filtered through
`G.712 and MIRS filtering (as specified by the ITU) and cepstral mean normalisation was
`used, then consistent performance was obtained for all filter responses conforming to
`03.50. It was therefore concluded that the DSR standard was suitable for use in all
`terminals operating within the ranges of the characteristics as specified in GSM 03.50.
`DSR terminal developers should be aware that reduced recognition performance might be
`obtained if they operate outside the recommended tolerances.
`
`8
`
`Conclusions
`
`The ETSI Aurora working group has completed the preparation of the standard for a DSR
`Mel-Cepstrum front-end and compression algorithm. The details of the Mel-Cepstrum
`feature extraction algorithm have been extensively tested and agreed. The compression
`algorithm produces no degradation in performance and the error robustness of the
`
`AVIOS 2000: The Speech Applications Conference, May 22-24, 2000 San Jose, CA, USA
`
`Comcast - Exhibit 1007, page 11
`
`

`

`compression scheme is exceptionally good. In addition it has low latency and the
`complexity is suitable for mobile handsets.
`
`References
`
`1. “ETSI ES 201 108 v1.1.2 Distributed Speech Recognition; Front-end Feature
`Extraction Algorithm; Compression Algorithm”, April 2000.
`
`2. “Experimental Framework for the Evaluation and Verification of Distributed Speech
`Recognition Front-ends”: version 5, AU/134/98, 8th September 1998
`
`3. GTS GSM 03.50:”Digital cellular telecommunications system (Phase 2+);
`Transmission planning aspects service in the GSM Public Land Mobile Network
`(PLMN) system (GSM 03.50)”.
`
`Further Information
`
`The ETSI DSR standard document can be found at:
`
`http://webapp.etsi.org/WorkProgram/Report_WorkItem.asp?WKI_ID=6400
`
`For ETSI members more information about the Aurora working group is available from
`the ETSI FTP site where there are electronic copies of some of the documents from
`previous meetings is at
`
`http://docbox.etsi.org/Tech-Org/STQ/Document/stq-aurora
`
`To be able to access this area you need to be an ETSI member and obtain a password
`from them
`
`http:/www.etsi.org
`
`will provide you with information about how to do this.
`
`Acknowledgement
`
`My thanks to colleagues at Motorola who have contributed significantly to the
`compression part of this standard and its evaluation particularly Jon Gibbs, Jeff Meunier
`and Yan-Ming Cheng.
`
`Thanks also to all the individuals who have contributed to this standard as participants in
`the Aurora Working Group. The main organisations represented there being Alcatel,
`British Telecommunications, Ericsson, France Telecom, Matra-Nortel, Motorola, Nokia,
`OGI, Qualcomm, Siemens, Sony, Texas Instruments.
`
`AVIOS 2000: The Speech Applications Conference, May 22-24, 2000 San Jose, CA, USA
`
`Comcast - Exhibit 1007, page 12
`
`

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket