throbber
Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 1 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 1 of 38
`
`
`
`
`
`
`
`
`EXHIBIT 8
`EXHIBIT 8
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 2 of 38
`ee” STETAA ATTA
`
`US010779080B2
`
`a2) United States Patent
`US 10,779,080 B2
`10) Patent No.:
`
` Burnett (45) Date of Patent: *Sep. 15, 2020
`
`
`(54) DUAL OMNIDIRECTIONAL MICROPHONE
`ARRAY (DOMA)
`
`(71) Applicant: JAWB Acquisition LLC, New York,
`NY (US)
`
`(72)
`
`Inventor: Gregory C. Burnett, Dodge Center,
`MN (US)
`
`(73) Assignee:
`
`JAWB ACQUISITION LLC, New
`York, NY (US)
`
`(*) Notice:
`
`Subject to any disclaimer, the term ofthis
`patent is extended or adjusted under 35
`U.S.C. 154(b) by 0 days.
`This patent is subject to a terminal dis-
`claimer.
`
`(21) Appl. No.: 13/948,160
`
`(22)
`
`Filed:
`
`Jul. 22, 2013
`
`(65)
`
`Prior Publication Data
`US 2014/0177860 Al
`Jun. 26, 2014
`
`Related U.S. Application Data
`(63) Continuation of application No. 12/139,355, filed on
`Jun. 13, 2008, now Pat. No. 8,494,177.
`
`(60) Provisional application No. 60/934,551, filed on Jun.
`13, 2007, provisional application No. 60/953,444,
`filed on Aug. 1, 2007, provisional application No.
`60/954,712,
`filed on Aug. 8, 2007, provisional
`application No. 61/045,377, filed on Apr. 16, 2008.
`
`(51)
`
`Int. Cl.
`HOR 3/00
`GIOL 21/0208
`HO4R 1/40
`HOAR 3/04
`GIOL 21/0216
`
`(2006.01)
`(2013.01)
`(2006.01)
`(2006.01)
`(2013.01)
`
`(52) U.S. Cl.
`CPC oe. HOAR 3/002 (2013.01); GIOL 21/0208
`(2013.01); HO4R 1/406 (2013.01); HO4R
`3/005 (2013.01); HO4R 3/04 (2013.01); GIOL
`2021/02165 (2013.01)
`(58) Field of Classification Search
`CPC ...... G10L 2021/02165; G10L 21/0208; HO4R
`3/005
`USPC weceeeneee 381/92, 94.7; 704/233, E21.004
`See application file for complete search history.
`
`(56)
`
`References Cited
`
`U.S. PATENT DOCUMENTS
`5,473,701 A
`12/1995 Cezanneet al.
`6,473,733 Bl
`10/2002 McArthuret al.
`(Continued)
`
`OTHER PUBLICATIONS
`
`Weiss, Howard; Office Action mailed by U.S. Patent and Trademark
`Office dated Mar. 15, 2012 for U.S. Appl. No. 12/139,355.
`(Continued)
`Primary Examiner — Howard Weiss
`(74) Attorney, Agent, or Firm — Hard IP LLC
`
`ABSTRACT
`67)
`A dual omnidirectional microphone array noise suppression
`is described. Compared to conventional arrays and algo-
`rithms, which seek to reduce noise by nulling out noise
`sources, the array of an embodiment is used to form two
`distinct virtual directional microphones which are config-
`uredto have very similar noise responses andvery dissimilar
`speech responses. The only null formed is one used to
`remove the speech of the user from V,. The two virtual
`microphones maybe paired with an adaptivefilter algorithm
`and VAD algorithm to significantly reduce the noise without
`distorting the speech, significantly improving the SNR ofthe
`desired speech over conventional noise suppression systems.
`
`20 Claims, 17 Drawing Sheets
`
`104
`
`Voicing Information
`
`100
`
`101
`
`)
`SIGNAL
`s(n)
`
`100
`
`
`
`Cleaned Speech
`
`
`
`Noise Removal
`
`
`
`
`
`
`
`>
`
`(ce)
`NOISE
`n(n)
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 3 of 38
`Page 3 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22
`
`US 10,779,080 B2
`
`Page 2
`
`(56)
`
`References Cited
`U.S. PATENT DOCUMENTS
`
`7,706,549 B2
`4/2010 Zhang et al.
`8,068,619 B2* 11/2011 Zhang etal. ow. 381/92
`
`w. 381/92
`8,494,177 B2*
`7/2013 Burnett
`....
`8/2013 Burnett wo. 381/92
`8,503,691 B2*
`
`OTHER PUBLICATIONS
`
`Weiss, Howard; Office Action mailed by U.S. Patent and Trademark
`Office dated Jun. 13, 2008 for U.S. Appl. No. 12/139,355.
`
`* cited by examiner
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 4 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 4 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 1 of 17
`
`US 10,779,080 B2
`
`
`
`WOTCWIOJU]SUIIIOA
`
`v0l
`
`(«*»)|001
`
`‘TVNDIS
`
`(u)s
`
`101
`
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 5 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 5 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`US 10,779,080 B2
`
`Sheet 2 of 17
`
`
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 6 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 6 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 3 of 17
`
`US 10,779,080 B2
`
`
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 7 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 7 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 4 of 17
`
`US 10,779,080 B2
`
`
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 8 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 8 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 5 of 17
`
`US 10,779,080 B2
`
`
`
`FIG.6
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 9 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 9 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 6 of 17
`
`US 10,779,080 B2
`
`Receive acoustic signals at a first physical
`microphone and a second physical microphone.
`
`Output first microphone signal from first physical
`microphone and second microphonesignal from
`second physical microphone.
`
`Form first virtual microphone using the first combination
`offirst microphone signal and second microphonesignal.
`
`Form second virtual microphone using second combination
`of first microphone signal and second microphonesignal.
`
`Generate denoised output signals having less
`acoustic noise than received acoustic signals.
`700—*
`FIG.7
`
`Form physical microphone array includingfirst
`physical microphone and second physical microphone.
`
`signals from physical microphonearray.
`
`Form virtual microphonearray includingfirst virtual
`microphone and secondvirtual microphone using
`
`800 —~
`
`FIG.8
`
`102
`
`704
`
`706
`
`708
`
`710
`
`802
`
`804
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 10 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 10 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 7 of 17
`
`US 10,779,080 B2
`
`Linear response of V2 to a speech source at 0.10 meters
`Linear response 0
`
`f V2 to a noise source at 1 meters
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 11 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 11 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 8 of 17
`
`US 10,779,080 B2
`
`Linear response of V1 to a speech source at 0.10 meters
`
`14411Vrme
`
`FIG.11
`Linear response of V1 to a noise source at | meters
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 12 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 12 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 9 of 17
`
`US 10,779,080 B2
`
`Linear response of V1 to a speech source at 0.1 meters
`
`90
`
`2
`
`180 |---}---+---|----+------ 4
` V
`
`—
`
` “F300 Hz
`
`/ 4000Hz~~
`
`240
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 13 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 13 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 10 of 17
`
`US 10,779,080 B2
`
`
`
`Response(dB)
`
`Frequency responseat 0 degrees
`
`Cardioid speech
`response
`
`!
`
`nooracponwnoroconebevccoeeecodenms WI. gpeethy --feevnnveeecbevcneh
`!
`|
`response
`!
`
`0
`
`1000
`
`2000
`
`3000
`
`4000
`
`5000
`
`6000
`
`7000
`
`8000
`
`Frequency (Hz)
`
`FIG.14
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 14 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 14 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 11 of 17
`
`US 10,779,080 B2
`
`
`
`
`
`V1/V2 for speech versus B assuming d, = 0.1m
`
` V1 (top, dashed) and V2 speech response vs. B assuming d, = 0.1m
`
`0.9
`
`0.7
`
`0.8
`B
`FIG.15
`
`
`
`Response(dB)
`
`
`
`V1/V2forspeech(dB)
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 15 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 15 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 12 of 17
`
`US 10,779,080 B2
`
`
`
`
`
`0.05
`
`O01
`
`015
`
`O38
`025
`O02
`Actual d, (meters)
`FIG.17
`
`035
`
`O04
`
`045
`
`05
`
`B factor vs. actual d, assuming d, = 0.1m and theta = 0
`
`
`B versus theta assuming d, = 0.1m
`
`
`
`80
`
`60
`
`40
`
`20
`0
`-20
`theta (degrees)
`FIG.18
`
`40
`
`60
`
`#80
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 16 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 16 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 13 of 17
`
`US 10,779,080 B2
`
`(dB)
`Amplitude
`(degrees)
`Phase
`
`0
`
`1000
`
`2000
`
`3000
`
`4000
`
`5000
`
`6000
`
`7000
`
`8000
`
`0
`
`1000
`
`2000
`
`3000
`
`4000
`
`5000
`
`6000
`
`7000
`
`8000
`
`Frequency (Hz)
`
`FIG.19
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 17 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 17 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 14 of 17
`
`US 10,779,080 B2
`
`(dB)
`Amplitude
` Phase
`(degrees)
`
`“40
`
`1000
`
`§=2000
`
`3000
`
`4000
`
`5000
`
`6000
`
`7000
`
`8000
`
`0
`
`1000
`
`2000
`
`3000
`
`4000
`
`5000
`
`6000
`
`7000
`
`8000
`
`Frequency (Hz)
`
`FIG.20
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 18 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 18 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 15 of 17
`
`US 10,779,080 B2
`
`J 2
`
`0fn
`
`oe
`
`
`
`Amplitude(dB)
`
`
`
`
`
`
`
`
`
`Cancellation with dl = 1, thetal = 0, d2 = 1, and theta2 = 30
`40
`
`
`
`
`
`
`
`0 1000+2000 3000 4000 5000 6000 7000 8000
`
`
`
`Phase(degrees)
`
`
`
`60
`
`0
`
`1000
`
`2000
`
`3000
`
`©4000
`
`5000
`
`6000
`
`7000
`
`8000
`
`Frequency (Hz)
`
`FIG.21
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 19 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 19 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 16 of 17
`
`US 10,779,080 B2
`
`
`
`Amplitude(dB)
`
`
`
`0
`
`1000
`
`2000
`
`3000
`
`4000
`
`5000
`
`6000
`
`7000
`
`8000
`
`Cancellation with dl = 1, thetal = 0, d2 = 1, and theta2 = 45
`
`
`
` Phase(degrees)
`
`
`
`0
`
`1000
`
`2000
`
`3000
`
`4000
`
`5000
`
`6000
`
`7000
`
`8000
`
`Frequency (Hz)
`
`FIG.22
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 20 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 20 of 38
`
`U.S. Patent
`
`Sep. 15, 2020
`
`Sheet 17 of 17
`
`US 10,779,080 B2
`
`Original V1 (top) and cleaned V1 (bottom) with simplified VAD (dashed) in noise
`
`Cleaned
`
`Noisy
`
`0
`
`0.5
`
`1
`1.5
`Time(samples at 8 kHz/sec)
`
`2
`
`2.5
`
`FIG.23
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 21 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 21 of 38
`
`US 10,779,080 B2
`
`1
`DUAL OMNIDIRECTIONAL MICROPHONE
`
`ARRAY (DOMA)
`
`CROSS-REFERENCE TO RELATED
`APPLICATIONS
`
`This application is continuation of U.S. Nonprovisional
`patent application Ser. No. 12/139,355, filed Jun. 13, 2008,
`now USS. Pat. No. 8,494,177, entitled “Dual Omnidirec-
`tional Microphone Array (DOMA),” which claims the ben-
`efit of U.S. Provisional Patent Application No. 60/934,551,
`filed Jun. 13, 2007, U.S. Provisional Patent Application No.
`60/953,444, filed Aug. 1, 2007, U.S. Provisional Patent
`Application No. 60/954,712, filed Aug. 8, 2007, and U.S.
`Provisional Patent Application No. 61/045,377, filed Apr.
`16, 2008, all of which are incorporated by reference herein
`in their entirety for all purposes.
`
`TECHNICAL FIELD
`
`The disclosure herein relates generally to noise suppres-
`sion. In particular, this disclosure relates to noise suppres-
`sion systems, devices, and methods for use in acoustic
`applications.
`
`BACKGROUND
`
`Conventional adaptive noise suppression algorithms have
`been around for some time. These conventional algorithms
`have used two or more microphones to sample both an
`(unwanted) acoustic noise field and the (desired) speech of
`a user. The noise relationship between the microphonesis
`then determined using an adaptive filter (such as Least-
`Mean-Squares
`as
`described in Haykin & Widrow,
`ISBN#0471215708, Wiley, 2002, but any adaptive or sta-
`tionary system identification algorithm may be used) and
`that relationship used to filter the noise from the desired
`signal.
`Most conventional noise suppression systems currently in
`use for speech communication systems are based on a
`single-microphone
`spectral
`subtraction technique
`first
`develop in the 1970’s and described, for example, by S. F.
`Boll in “Suppression of Acoustic Noise in Speech using
`Spectral Subtraction,” IEEE Trans. on ASSP, pp. 113-120,
`1979. These techniques have been refined overthe years, but
`the basic principles of operation have remained the same.
`See, for example, U.S. Pat. No. 5,687,243 of McLaughlin, et
`al., and U.S. Pat. No. 4,811,404 of Vilmur, et al. There have
`also been several attempts at multimicrophone noise sup-
`pression systems, such as those outlined in U.S. Pat. No.
`5,406,622 of Silverberg et al. and U.S. Pat. No. 5,463,694 of
`Bradley et al. Multi-microphone systems have not been very
`successful for a variety of reasons, the most compelling
`being poornoise cancellation performance and/or significant
`speech distortion. Primarily, conventional multi-microphone
`systems attempt to increase the SNR ofthe user’s speech by
`“steering” the nulls of the system to the strongest noise
`sources. This approach is limited in the number of noise
`sources removed by the numberof available nulls.
`The Jawbone earpiece (referred to as the “Jawbone),
`introduced in December 2006 by AliphCom of San Fran-
`cisco, Calif., was thefirst known commercial product to use
`a pair of physical directional microphones (instead of omni-
`directional microphones) to reduce environmental acoustic
`noise. The technology supporting the Jawboneis currently
`described under one or more of U.S. Pat. No. 7,246,058 by
`Burnett and/or U.S. patent application Ser. Nos. 10/400,282,
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`2
`10/667,207, and/or 10/769,302. Generally, multi-micro-
`phone techniques make use of an acoustic-based Voice
`Activity Detector (VAD)to determine the background noise
`characteristics, where “voice” is generally understood to
`include human voiced speech, unvoiced speech, or a com-
`bination of voiced and unvoiced speech. The Jawbone
`improved on this by using a microphone-based sensor to
`construct a VAD signal using directly detected speech vibra-
`tions in the user’s cheek. This allowed the Jawbone to
`
`aggressively remove noise when the user was not producing
`speech. However, the Jawbone uses a directional micro-
`phonearray.
`
`INCORPORATION BY REFERENCE
`
`Each patent, patent application, and/or publication men-
`tioned in this specification is herein incorporated by refer-
`ence in its entirety to the same extent as if each individual
`patent, patent application, and/or publication was specifi-
`cally and individually indicated to be incorporated by ref-
`erence.
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`
`FIG. 1 is a two-microphone adaptive noise suppression
`system, under an embodiment.
`FIG. 2 is an array and speech source (S) configuration,
`under an embodiment. The microphones are separated by a
`distance approximately equal to 2d,, and the speech source
`is located a distance ds away from the midpoint of the array
`at an angle 0. The system is axially symmetric so only d, and
`8 need be specified.
`FIG. 3 is a block diagram for a first order gradient
`microphone using two omnidirectional elements O, and O,,
`under an embodiment.
`
`FIG. 4 is a block diagram for a DOMAincluding two
`physical microphones configured to form two virtual micro-
`phones V, and V,, under an embodiment.
`FIG. 5 is a block diagram for a DOMAincluding two
`physical microphones configured to form N virtual micro-
`phones V, through V,,, where N is any numbergreater than
`one, under an embodiment.
`FIG. 6 is an example of a headset or head-worn device
`that includes the DOMA, as described herein, under an
`embodiment.
`
`FIG. 7 is a flow diagram for denoising acoustic signals
`using the DOMA, under an embodiment.
`FIG. 8 is a flow diagram for forming the DOMA, under
`an embodiment.
`FIG.9 is a plot of linear response of virtual microphone
`V, to a 1 kHz speech sourceat a distance of 0.1 m, under an
`embodiment. The null is at 0 degrees, where the speech is
`normally located.
`FIG. 10 is a plot of linear response of virtual microphone
`V, toa 1 kHz noise source at a distance of 1.0 m, under an
`embodiment. There is no null and all noise sources are
`detected.
`
`FIG.11 is a plot of linear response of virtual microphone
`V, toa 1 kHz speech source at a distance of 0.1 m, under an
`embodiment. There is no null and the response for speech is
`greater than that shown in FIG. 9.
`FIG.12 is a plot of linear response of virtual microphone
`V, toa 1 kHz noise source at a distance of 1.0 m, under an
`embodiment. There is no null and the response is very
`similar to V, shown in FIG. 10.
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 22 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 22 of 38
`
`US 10,779,080 B2
`
`3
`FIG.13 is a plot of linear response of virtual microphone
`V, to a speech source at a distance of 0.1 m for frequencies
`of 100, 500, 1000, 2000, 3000, and 4000 Hz, under an
`embodiment.
`FIG. 14 is a plot showing comparison of frequency
`responses for speech for the array of an embodiment andfor
`a conventional cardioid microphone.
`FIG. 15 is a plot showing speech response for V, (top,
`dashed) and V2 (bottom, solid) versus B with ds assumed to
`be 0.1 m, under an embodiment. The spatial null in V2 is
`relatively broad.
`FIG. 16 is a plot showing a ratio of V,/V, speech
`responses shown in FIG. 10 versus B, under an embodiment.
`Theratio is above 10 dB for all 0.8<B<1.1. This meansthat
`
`the physical 6 of the system need not be exactly modeled for
`good performance.
`FIG.17 is a plot of B versusactual ds assuming that ds=10
`cm and theta=0, under an embodiment.
`FIG. 18 is a plot of B versus theta with ds=10 cm and
`assuming ds=10 cm, under an embodiment.
`FIG. 19 is a plot of amplitude (top) and phase (bottom)
`response of N(s) with B=1 and D=-7.2 usec, under an
`embodiment. The resulting phase difference clearly affects
`high frequencies more than low.
`FIG. 20 is a plot of amplitude (top) and phase (bottom)
`response of N(s) with B=1.2 and D=-7.2 usec, under an
`embodiment. Non-unity B affects the entire frequency range.
`FIG. 21 is a plot of amplitude (top) and phase (bottom)
`response of the effect on the speech cancellation in V, due
`to a mistake in the location of the speech source with q1=0
`degrees and q2=30 degrees, under an embodiment. The
`cancellation remains below -10 dB for frequencies below 6
`kHz.
`
`FIG. 22 is a plot of amplitude (top) and phase (bottom)
`response of the effect on the speech cancellation in V2 due
`to a mistake in the location of the speech source with q1=0
`degrees and q2=45 degrees, under an embodiment. The
`cancellation is below -10 dB only for frequencies below
`about 2.8 kHz and a reduction in performance is expected.
`FIG. 23 shows experimental results for a 2do=19 mm
`array using a linear B of 0.83 on a Bruel and Kjaer Head and
`Torso Simulator (HATS) in very loud (~85 dBA) music/
`speech noise environment, under an embodiment. The noise
`has been reduced by about 25 dB and the speech hardly
`affected, with no noticeable distortion.
`
`DETAILED DESCRIPTION
`
`A dual omnidirectional microphone array (DOMA) that
`provides improved noise suppression is described herein.
`Compared to conventional arrays and algorithms, which
`seek to reduce noise by nulling out noise sources, the array
`of an embodiment
`is used to form two distinct virtual
`directional microphones which are configured to have very
`similar noise
`responses
`and very dissimilar
`speech
`responses. The only null formed by the DOMAis one used
`to remove the speech of the user from V2*The twovirtual
`microphones of an embodiment can be paired with an
`adaptive filter algorithm and/or VAD algorithm to signifi-
`cantly reduce the noise without distorting the speech, sig-
`nificantly improving the SNR of the desired speech over
`conventional noise suppression systems. The embodiments
`described herein are stable in operation, flexible with respect
`to virtual microphonepattern choice, and have proven to be
`robust with respect to speech source-to-array distance and
`orientation as well as temperature and calibration tech-
`niques.
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`4
`In the following description, numerousspecific details are
`introduced to provide a thorough understanding of, and
`enabling description for, embodiments of the DOMA. One
`skilled in the relevant art, however, will recognize that these
`embodiments can be practiced without one or more of the
`specific details, or with other components, systems, etc. In
`other instances, well-known structures or operations are not
`shown, or are not described in detail, to avoid obscuring
`aspects of the disclosed embodiments.
`Unless otherwise specified, the following terms have the
`corresponding meanings in addition to any meaning or
`understanding they may convey to one skilled in theart.
`The term “bleedthrough” meansthe undesired presence of
`noise during speech.
`The term “denoising” means removing unwanted noise
`from Micl, and also refers to the amount of reduction of
`noise energy in a signal in decibels (dB).
`The term “devoicing” means removing/distorting the
`desired speech from Mic.
`The term “directional microphone (DM)” meansa physi-
`cal directional microphonethat is vented on both sides of the
`sensing diaphragm.
`The term “Mic1 (M1)” meansa general designation for an
`adaptive noise suppression system microphonethat usually
`contains more speech than noise.
`The term “Mic2 (M2)” meansa general designation for an
`adaptive noise suppression system microphonethat usually
`contains more noise than speech.
`The term “noise” means unwanted environmental acous-
`tic noise.
`
`The term “null” means a zero or minima in the spatial
`response of a physical or virtual directional microphone.
`The term “0,” means a first physical omnidirectional
`microphone used to form a microphonearray.
`The term “0,” means a second physical omnidirectional
`microphone used to form a microphonearray.
`The term “speech” means desired speech of the user.
`The term “Skin Surface Microphone (SSM)”is a micro-
`phone used in an earpiece (e.g.,
`the Jawbone earpiece
`available from Aliph of San Francisco, Calif.) to detect
`speech vibrations on the user’s skin.
`The term “V,” means the virtual directional “speech”
`microphone, which has no nulls.
`The term “V,” means the virtual directional “noise”
`microphone, which hasa null for the user’s speech.
`The term “Voice Activity Detection (VAD) signal” means
`a signal indicating when user speech is detected.
`The term “virtual microphones (VM)”or “virtual direc-
`tional microphones” means a microphone constructed using
`two or more omnidirectional microphones and associated
`signal processing.
`FIG. 1 is a two-microphone adaptive noise suppression
`system 100, under an embodiment. The two-microphone
`system 100 including the combination of physical micro-
`phones MIC 1 and MIC 2 along with the processing or
`circuitry components to which the microphones couple
`(described in detail below, but not shown in this figure) is
`referred to herein as the dual omnidirectional microphone
`array (DOMA) 110, but the embodimentis not so limited.
`Referring to FIG. 1, in analyzing the single noise source 101
`and the direct path to the microphones, the total acoustic
`information coming into MIC 1
`(102, which can be an
`physical or virtual microphone) is denoted by m,(n). The
`total acoustic information coming into MIC 2 (103, which
`can also be an physical or virtual microphone) is similarly
`labeled m2(n). In the z (digital frequency) domain, these are
`represented as M,(z) and M,(Z). Then,
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 23 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 23 of 38
`
`M,@)=S()+N(2)
`
`My(2)=N@)+S(2)
`
`with
`
`N2(Z)=N@)H|)
`
`So{Z)=S@)H2@)
`
`so that
`
`M,@)=S@)+N(@)H(2)
`
`M2(Z)-N@)+S(@)HZ)
`
`US 10,779,080 B2
`
`6
`for H,(z) are held constant (and vice versa) andit is assumed
`that the noise level is not high enoughto cause errors in the
`H,(z) calculation.
`they are used to
`After calculating H1(z) and H.(z),
`removethe noise from the signal. If Equation 1 is rewritten
`as
`
`S()=M, @)-N@)Ai)
`
`N@)-M{2)-S@)Ha(Z)
`
`S@)=M,@)-[Mo@)-S@)a2)@),
`
`S@)[1-H2@)1@)]-M@)-M,(@)H(z)
`
`Eq. 1
`
`then N(z) may be substituted as shown to solve for S(z) as
`
`This is the general case for all two microphone systems.
`Equation 1 has four unknownsand only two known rela-
`tionships and therefore cannot be solved explicitly.
`However, there is another way to solve for some of the
`unknownsin Equation 1. The analysis starts with an exami-
`nation of the case where the speech is not being generated,
`that is, where a signal from the VAD subsystem 104 (op-
`tional) equals zero. In this case, s(n)=S(z)=0, and Equation
`1 reduces to
`
`My@)-N@AME)
`
`Moy{z)=N(z)
`
`where the N subscript on the M variables indicate that only
`noise is being received.
`This leads to
`
`20
`
`25
`
`30
`
`My (@) = Moy @)M1 (2)
`
`_ Min)
`MQ) = Mon (2)
`
`Eq. 2
`
`35
`
`The function H,(z) can be calculated using any of the
`available system identification algorithms and the micro-
`phone outputs when the system is certain that only noise is
`being received. The calculation can be done adaptively, so
`that the system can react to changesin the noise.
`A solution is now available for H,(z), one of the
`unknownsin Equation 1. The final unknown, H,(z), can be
`determined by using the instances where speech is being
`produced and the VAD equals one. Whenthis is occurring,
`but the recent (perhaps less than 1 second) history of the
`microphonesindicate low levels of noise, it can be assumed
`that n(s)=N(z)~O. Then Equation 1 reduces to
`
`M,,(2)-S@)
`
`M2.(2)-S@)\H2@),
`which in turn leads to
`
`Mos(z) = Mis(2)Fa(@)
`
`Mos (2)
`Mis(Z)
`
`A(z) =
`
`whichis the inverse of the H,(z) calculation. However, it is
`noted that different inputs are being used (now only the
`speech is occurring whereas before only the noise was
`occurring). While calculating H,(z), the values calculated
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`Mi(z) — Ma(2Ai(@)
`TT n@M@
`
`5) =
`
`If the transfer functions H,(z) and H,(z) can be described
`with sufficient accuracy, then the noise can be completely
`removed andthe original signal recovered. This remains true
`without respect to the amplitude or spectral characteristics of
`the noise. If there is very little or no leakage from the speech
`source into M,, then H,(z)=0 and Equation 3 reduces to
`
`S(z)&M(z)-M2(z)A(z).
`
`Eq. 4
`
`Equation 4 is much simpler to implement and is very
`stable, assuming H,(z) is stable. However,
`if significant
`speech energy is in M,(Z), devoicing can occur. In order to
`construct a well-performing system and use Equation 4,
`consideration is given to the following conditions:
`R1. Availability of a perfect (or at least very good) VAD
`in noisy conditions
`R2. Sufficiently accurate H,(z)
`R3. Very small (ideally zero) H,(z)
`R4. During speech production, H,(z) cannot change sub-
`stantially.
`R5. During noise, H(z) cannot change substantially.
`Condition R1 is easy to satisfy if the SNR of the desired
`speech to the unwanted noise is high enough. “Enough”
`meansdifferent things depending on the method of VAD
`generation. If a VAD vibration sensor is used, as in Burnett
`USS. Pat. No. 7,256,048, accurate VAD in very low SNRs
`(-10 dB orless) is possible. Acoustic-only methods using
`information from O, and O, can also return accurate VADs,
`but are limited to SNRs of ~3 dB or greater for adequate
`performance.
`Condition R5 is normally simple to satisfy because for
`most applications the microphones will not change position
`with respect to the user’s mouth very often or rapidly. In
`those applications where it may happen (such as hands-free
`conferencing systems) it can be satisfied by configuring
`Mic2 so that H,(z)=0.
`Satisfying conditions R2, R3, and R4 are more difficult
`but are possible given the right combination of V, and
`V,*Methods are examined below that have proven to be
`effective in satisfying the above, resulting in excellent noise
`suppression performance and minimal speech removal and
`distortion in an embodiment.
`The DOMA,in various embodiments, can be used with
`the Pathfinder system as the adaptive filter system or noise
`removal. The Pathfinder system, available from AliphCom,
`San Francisco, Calif, is described in detail in other patents
`and patent applications referenced herein. Alternatively, any
`
`

`

`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 24 of 38
`Case 6:21-cv-00984-ADA Document 55-8 Filed 05/25/22 Page 24 of 38
`
`US 10,779,080 B2
`
`7
`adaptive filter or noise removal algorithm can be used with
`the DOMA in one or more various alternative embodiments
`
`8
`a wide variety of virtual microphones (VMs), also referred
`to herein as virtual directional microphones, can berealized.
`While the processing component 402 described in this
`or configurations.
`example includes four processing paths generating two
`When the DOMA is used with the Pathfinder system, the
`virtual microphones or microphonesignals, the embodiment
`Pathfinder system generally provides adaptive noise cancel-
`is not so limited. For example, FIG. 5 is a block diagram for
`lation by combining the two microphonesignals (e.g., Micl,
`Mic2) by filtering and summing in the time domain. The
`a DOMA 500 including two physical microphones config-
`
`adaptive filter generally uses the signal received fromafirst ured to form N virtual microphones V, through V,,, where
`microphone of the DOMA to removenoise from the speech
`N is any number greater than one, under an embodiment.
`received from at least one other microphone of the DOMA,
`Thus, the DOMA can include a processing component 502
`which relies on a slowly varying linear transfer function
`having any numberof processing paths as appropriate to
`
`between the two microphones for sources of noise. Follow- form a numberNof virtual microphones.
`ing processing of the two channels of the DOMA,an output
`The DOMAof an embodiment can be coupled or con-
`signal is generated in which the noise content is attenuated
`nected to one or more remote devices. In a system configu-
`with respect to the speech content, as described in detail
`ration, the DOMAoutputs signals to the remote devices. The
`below.
`remote devices include, but are not limitedto, at least one of
`cellular telephones, satellite telephones, portable telephones,
`wireline telephones, Internet telephones, wireless transceiv-
`ers, wireless communication radios, personal digital assis-
`tants (PDAs), personal computers (PCs), headset devices,
`head-worn devices, and earpieces.
`Furthermore,
`the DOMA of an embodiment can be a
`component or subsystem integrated with a host device. In
`this system configuration,
`the DOMA outputs signals to
`components or subsystems of the host device. The host
`device includes, but is not limited to, at least one of cellular
`telephones, satellite telephones, portable telephones, wire-
`line telephones, Internet telephones, wireless transceivers,
`wireless communication radios, personal digital assistants
`(PDAs), personal computers (PCs), headset devices, head-
`worn devices, and earpieces.
`As an example, FIG. 6 is an example of a headset or
`head-worn device 600 that
`includes
`the DOMA,
`as
`described herein, under an embodiment. The headset 600 of
`an embodiment includes a housing having two areas or
`receptacles (not shown) that receive and hold two micro-
`phones (e.g., O, and O,), The headset 600 is generally a
`device that can be worn by a speaker 602, for example, a
`headset or earpiece that positions or holds the microphones
`in the vicinity of the speaker’s mouth. The headset 600 of an
`embodimentplaces a first physical microphone(e.g., physi-
`cal microphoneO, ) ina vicinity of a speaker’s lips. A second
`physical microphone (e.g., physical microphone O,) is
`placed a distance behindthe first physical microphone. The
`distance of an embodimentis in a range of a few centimeters
`behind the first physical microphone or as described herein
`(e.g., described with reference to FIGS. 1-5). The DOMAis
`symmetric and is used in the same configuration or manner
`output from each microphone is coupled to a processing
`as a single close-talk microphone, but is not so limited.
`component 402, or circuitry, and the processing component
`FIG. 7 is a flow diagram for denoising 700 acoustic
`outputs signals representing or corresponding to the virtual
`signals using the DOMA, under an embodiment. The
`microphones V, and V3.
`denoising 700 begins by receiving 702 acoustic signals at a
`In this example system 400, the output of physical micro-
`first physical microphone and a second physical micro-
`phone 201 is coupled to processing component 402 that
`includesa first processing path that includes application of
`phone.In responseto the acoustic signals, a first microphone
`55
`
`a first delay Z,, andafirst gain A,, and a second processing signal is output from the first physical microphone and a
`path that includes application of a second delay Z,, and a
`second microphonesignal is output from the second physi-
`second gain A,,*The output of physical microphone 202 is
`cal microphone 704. A first virtual microphone is formed
`coupled to a third processing path of the processing com-
`706 by gen

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket