throbber
(19) United States
`(12) Patent Application Publication (10) Pub. No.: US 2008/0240458 A1
`(43) Pub. Date:
`Oct. 2, 2008
`Goldstein et al.
`
`US 20080240458A1
`
`(54) METHOD AND DEVICE CONFIGURED FOR
`SOUND SIGNATURE DETECTION
`
`(75) Inventors:
`
`Steven W. Goldstein, Delray
`Beach, FL (US); Mark A.
`Clements, Lilburn, GA (US); Marc
`A. Boillot, Plantation, FL (US)
`
`Correspondence Address:
`GREENBERG TRAURIG, LLP
`2101 L Street, N.W., Suite 1000
`Washington, DC 20037 (US)
`
`(73) Assignee:
`
`PERSONICS HOLDINGS INC.,
`Boca Raton, FL (US)
`
`(21) Appl. No.:
`
`11/966,457
`
`(22) Filed:
`
`Dec. 28, 2007
`
`
`
`Related U.S. Application Data
`(60) Provisional application No. 60/883,013, filed on Dec.
`31, 2006.
`Publication Classification
`
`(51) Int. Cl.
`(2006.01)
`A6IF II/06
`(52) U.S. Cl. .......................................................... 381/72
`(57)
`ABSTRACT
`At least one exemplary embodiment is directed to a method
`for personalized listening which can be used with an earpiece
`is provided that can include capturing ambient Sound from an
`Ambient Sound Microphone (ASM) of an earpiece partially
`or fully occluded in an ear canal, monitoring the ambient
`Sound for a target Sound, and adjusting by way of an Ear Canal
`Receiver (ECR) in the earpiece a delivery of audio to an ear
`canal based on a detected target Sound. A Volume of audio
`content can be adjusted upon the detection of a target Sound,
`and an audible notification can be presented to provide a
`warning.
`
`Exhibit 1022
`Page 01 of 14
`
`

`

`Patent Application Publication
`
`Oct. 2, 2008 Sheet 1 of 5
`
`US 2008/0240458A1
`
`130
`
`Ambient : Sound
`Sound ; Signature :
`Detection :
`- - - - - - - - - - - - -
`
`110
`
`
`
`
`
`Interface
`212
`
`
`
`Audio Content
`(e.g., music, Cell
`phone, voice
`mail)
`
`100
`
`FIG. 2
`
`Exhibit 1022
`Page 02 of 14
`
`

`

`Patent Application Publication
`
`Oct. 2, 2008 Sheet 2 of 5
`
`US 2008/0240458A1
`
`Monitoring and Target Detection
`
`Monitoring the environment for
`target Sounds
`302
`
`
`
`
`
`
`
`
`
`Generating an audible alarm
`within the ear canal identifying
`the detected sound signature
`(e.g., Sound byte, text-to
`Speech, "ambulance
`approaching")
`
`308
`
`Detecting the target sounds
`Within the environment
`based on sound signatures
`304
`
`Adjusting audio delivered to
`to the ear canal in view of a
`detected target sound
`306
`
`
`
`
`
`
`
`
`
`
`
`Sending a message to a mobile
`device identifying the detected
`Sound signature (e.g., "alarm
`Sounding")
`310
`
`
`
`300
`FIG. 3
`
`
`
`Signature
`Sound Pass
`Through
`Mode
`(SSPTM)
`
`
`
`
`
`
`
`
`
`
`
`
`
`Mode Selection
`
`
`
`
`
`
`
`Signature
`Sound
`Replacement
`Mode (SSRM)
`
`Signature
`Sound Boost
`Mode
`(SSBM)
`
`Signature
`Sound
`Attenuation
`Mode
`(SSAM)
`
`
`
`
`
`
`
`
`
`
`
`
`
`Signature
`Sound
`Rejection
`Mode
`(SSRM)
`
`
`
`Exhibit 1022
`Page 03 of 14
`
`

`

`Patent Application Publication
`
`Oct. 2, 2008 Sheet 3 of 5
`
`US 2008/0240458A1
`
`
`
`
`
`520
`
`at 504
`
`508
`
`User Defined
`database
`
`Generate Speec
`h
`Recognition Mod
`es
`506
`
`510
`
`On-line
`database
`
`Generate Sound
`Signature Models
`
`
`
`Input signal from Ambient
`Sound Microphone (ASM)
`522
`Data Buffer
`524
`Signal Conditioning (e.g.,
`Noise Reduction, Noise gate)
`526
`Threshold Decision
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`530
`
`540
`
`Pattern Recognition to detect
`Sound signatures in input
`signal using Models
`
`Speech Recognition to
`recognize speech in input
`signal using Models
`
`Sound Source
`range/bearing
`analysis
`
`Sound Source
`speed
`analysis
`
`Key word
`detection
`(e.g. "help")
`
`542
`
`532
`Target signal
`detector (e.g.,
`horn, alarm)
`
`552
`
`Prioritizing
`
`554
`
`
`
`
`
`Operating
`
`550
`
`Selection
`
`an
`
`F G 5
`
`Exhibit 1022
`Page 04 of 14
`
`

`

`Patent Application Publication
`
`Oct. 2, 2008 Sheet 4 of 5
`
`US 2008/0240458A1
`
`Managing Audio Delivery
`
`Audio Interface receives audio
`content from a media player 602
`
`Audio content delivered to the
`user's ear Canal
`604
`
`
`
`600
`FIG. 6
`
`Signature
`Detected?
`608
`
`
`
`NO
`
`Processor monitors ambient
`Sound in the environment
`captured at the ASM
`
`60
`
`Processor determines a priority
`of the detected signature 610
`varara
`
`Processor selectively manages at
`least a portion of the audio Content
`based on the priority
`612
`
`
`
`Processor sends a message to a
`device operated by the user to
`visually display the notification 616
`
`Processor presents an audible
`notification to the user
`614
`
`
`
`
`
`
`
`Buffer data
`711
`
`Extract features
`712
`
`Compare models
`713
`
`Classify target Sound
`715
`
`
`
`
`
`Distortion Metric
`714.
`
`
`
`
`
`
`
`
`
`Generate alarm
`716
`
`FIG. 7
`
`Target Sounds
`database 718
`
`Exhibit 1022
`Page 05 of 14
`
`

`

`Patent Application Publication
`
`Oct. 2, 2008 Sheet 5 of 5
`
`US 2008/0240458A1
`
`Processor
`
`
`
`PHL
`Filtering
`
`Audio
`Content
`
`800
`FIG. 8
`
`Exhibit 1022
`Page 06 of 14
`
`

`

`US 2008/0240458 A1
`
`Oct. 2, 2008
`
`METHOD AND DEVICE CONFIGURED FOR
`SOUND SIGNATURE DETECTION
`
`CROSS REFERENCE TO RELATED
`APPLICATIONS
`0001. This Application is a Non-Provisional and claims
`the priority benefit of Provisional Application No. 60/883,
`013 filed on Dec. 31, 2006, the entire disclosure of which is
`incorporated herein by reference.
`
`FIELD
`0002 The present invention relates to a device that moni
`tors target (e.g. warning) sounds, and more particularly,
`though not exclusively, to an earpiece and method of operat
`ing an earpiece that detects target Sounds.
`
`BACKGROUND
`0003 Excess noise exposure can generate auditory
`fatigue, possibly comprising a person's listening abilities. On
`a daily basis, people are exposed to various environmental
`Sounds and noises within their environment, Such as the
`Sounds from traffic, construction, and industry. Some of the
`Sounds in the environment may correspond to warnings, such
`as those associated with an alarm or siren. A person that can
`hear the warning sounds can generally react in time to avoid
`danger. In contrast, a person that cannot adequately hear the
`warning Sounds, or whose hearing faculties have been com
`promised due to auditory fatigue, may be susceptible to dan
`ger.
`0004 Environmental noise can mask warning sounds and
`impair a person's judgment. Moreover, when people wear
`headphones to listen to music, or engage in a call using a
`telephone, they can effectively impair their auditory judg
`ment and their ability to discriminate between sounds. With
`Such devices, the person is immersed in the audio experience
`and generally less likely to hear target Sounds within their
`environment. In some cases, the user may even turn up the
`Volume to hear their personal audio over environmental
`noises. This can put the user in a compromising situation
`since they may not be aware of target Sounds in their environ
`ment. It also puts them at high sound exposure risk, which can
`potentially cause long term hearing damage.
`0005. A need therefore exists for enhancing the user's
`ability to hear target sounds in their environment without
`compromising his hearing.
`
`SUMMARY
`0006. At least one exemplary embodiment is directed to a
`method and device for Sound signature detection.
`0007. In at least one exemplary embodiment, an earpiece,
`can include an Ambient Sound Microphone (ASM) config
`ured to capture ambient sound, at least one Ear Canal
`Receiver (ECR) configured to deliver audio to an ear canal,
`and a processor operatively coupled to the ASM and the at
`least one ECR to monitor target Sounds in the ambient sound.
`Target (e.g., warning) sounds can be amplified, attenuated, or
`reproduced and reported to the userby way of the ECR. As an
`example, the target (e.g., warning) sound can be an alarm, a
`horn, a voice, or a noise. The processor can detect Sound
`signatures in the ambient sound to identify the target (e.g.,
`warning) sounds and adjust the audio delivered to the ear
`canal based on detected Sound signatures.
`
`0008. In a second exemplary embodiment, a method for
`personalized listening Suitable for use with an earpiece is
`provided. The method can include capturing ambient Sound
`from an Ambient Sound Microphone (ASM) of an earpiece
`that is partially or fully occluded in an ear canal, monitoring
`the ambient sound for a target Sound, and adjusting by way of
`an Ear Canal Receiver (ECR) in the earpiece a delivery of
`audio to an ear canal based on a detected target Sound. The
`method can include passing, amplifying, attenuating, or
`reproducing the target Sound for delivery to the ear canal.
`0009. In a third exemplary embodiment a method for per
`Sonalized listening Suitable for use with an earpiece can
`include the steps of capturing ambient Sound from an Ambi
`ent Sound Microphone (ASM) of an earpiece that is partially
`or fully occluded in an ear canal, detecting a sound signature
`within the ambient sound that is associated with a target
`Sound, and mixing the target Sound with audio content deliv
`ered to the earpiece in accordance with a priority of the target
`Sound. A direction and speed of a sound source generating the
`target Sound can be determined, and presented as a notifica
`tion to a user of the earpiece. The method can include detect
`ing a spoken utterance in the ambient sound that corresponds
`to a verbal warning or help request.
`0010. In a fourth exemplary embodiment a method for
`Sound signature detection can include capturing ambient
`sound from an Ambient Sound Microphone (ASM) of an
`earpiece, and receiving a directive to learn a Sound signature
`within the ambient sound. The method can include receiving
`a voice command or detecting a user interaction with the
`earpiece to initiate the step of capturing and learning. A Sound
`signature can be generated for a target Sound in the environ
`ment and saved to a memory locally on the earpiece or
`remotely on a server.
`0011. In a fifth exemplary embodiment a method for per
`Sonalized listening can include capturing ambient sound from
`an Ambient Sound Microphone (ASM) of an earpiece that is
`partially or fully occluded in an ear canal, detecting a Sound
`signature within the ambient Sound that is associated with a
`target Sound, and mixing the target Sound with audio content
`delivered to the earpiece in accordance with a priority of the
`target sound and a personalized hearing level (PHL). The
`method can include retrieving from a database learned mod
`els, comparing the sound signature to the learned models, and
`identifying the target Sound from the learned models in view
`of the comparison. Auditory queues in the target Sound can be
`enhanced relative to the audio content based on a spectrum of
`the ambient sound captured at the ASM. A perceived direction
`of a sound source generating the target Sounds can be spatial
`ized using Head Related Transfer Functions (HRTFs).
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`0012 FIG. 1 is a pictorial diagram of an earpiece in accor
`dance with an exemplary embodiment;
`0013 FIG. 2 is a block diagram of the earpiece in accor
`dance with an exemplary embodiment;
`0014 FIG. 3 is a flowchart of a method for ambient sound
`monitoring and target detection in accordance with an exem
`plary embodiment;
`0015 FIG. 4 illustrates earpiece modes in accordance with
`an exemplary embodiment;
`0016 FIG. 5 illustrates a flowchart of a method for sound
`signature detection in accordance with an exemplary embodi
`ment;
`
`Exhibit 1022
`Page 07 of 14
`
`

`

`US 2008/0240458 A1
`
`Oct. 2, 2008
`
`0017 FIG. 6 is a flowchart of a method for managing audio
`delivery based on detected Sound signatures in accordance
`with an exemplary embodiment;
`0018 FIG. 7 is a flowchart for sound signature detection in
`accordance with an exemplary embodiment; and
`0019 FIG. 8 is a pictorial diagram for mixing ambient
`Sounds and target Sounds with audio content in accordance
`with an exemplary embodiment.
`
`DETAILED DESCRIPTION
`0020. The following description of at least one exemplary
`embodiment is merely illustrative in nature and is in no way
`intended to limit the invention, its application, or uses.
`0021
`Processes, techniques, apparatus, and materials as
`known by one of ordinary skill in the relevant art may not be
`discussed in detail but are intended to be part of the enabling
`description where appropriate, for example the fabrication
`and use of transducers. Additionally in at least one exemplary
`embodiment the sampling rate of the transducers can be var
`ied to pick up pulses of Sound, for example less than 50
`milliseconds.
`0022. In all of the examples illustrated and discussed
`herein, any specific values, for example the Sound pressure
`level change, should be interpreted to be illustrative only and
`non-limiting. Thus, other examples of the exemplary embodi
`ments could have different values.
`0023 Note that similar reference numerals and letters
`refer to similar items in the following figures, and thus once
`an item is defined in one figure, it may not be discussed for
`following figures.
`0024 Note that herein when referring to correcting or
`preventing an error or damage (e.g., hearing damage), a
`reduction of the damage or error and/or a correction of the
`damage or error are intended.
`0025. At least one exemplary embodiment of the invention
`is directed to an earpiece for ambient Sound monitoring and
`target detection. Reference is made to FIG. 1 in which an
`earpiece device, generally indicated as earpiece 100, is con
`structed in accordance with at least one exemplary embodi
`ment of the invention. Earpiece 100 includes an Ambient
`Sound Microphone (ASM) 110 to capture ambient sound, an
`Ear Canal Receiver (ECR) 120 to deliver audio to an earcanal
`140, and an ear canal microphone (ECM) 130 to assess a
`sound exposure level within the ear canal. The earpiece 100
`can partially or fully occlude the ear canal 140 to provide
`various degrees of acoustic isolation.
`0026. The earpiece 100 can actively monitor a sound pres
`Sure level both inside and outside an ear canal and enhance
`spatial and timbral sound quality to ensure safe reproduction
`levels. The earpiece 100 in various exemplary embodiments
`can provide listening tests, filter sounds in the environment,
`monitor target Sounds in the environment, present notifica
`tions based on identified target Sounds, adjust audio content
`levels with respect to ambient sound levels, and filter sound in
`accordance with a Personalized Hearing Level (PHL). The
`earpiece 100 is suitable for use with users having healthy or
`abnormal auditory functioning. The earpiece 100 can be an in
`the ear earpiece, behind the ear earpiece, receiver in the ear,
`open-fit device, or any other Suitable earpiece type. Accord
`ingly, the earpiece 100 can be partially or fully occluded in the
`ear canal.
`0027. As part of its operation, the earpiece 100 can gener
`ate an Ear Canal Transfer Function (ECTF) to model the ear
`canal 140 using ECR 120 and ECM 130. The ECTF can be
`
`used to establish a personalized hearing level profile. The
`earpiece 100 can also determine a sealing profile with the
`user's ear to compensate for any sound leakage. In one con
`figuration, the earpiece 100 can provide personalized full
`band width general audio reproduction within the user's ear
`canal via timbral equalization based on the ECTF to account
`for a user's hearing sensitivity. The earpiece 100 also provides
`Sound Pressure Level dosimetry to estimate sound exposure
`of the ear and associated recovery times from excessive Sound
`exposure. This permits the earpiece 100 to safely administer
`and monitor Sound exposure to the ear.
`0028 Referring to FIG. 2, a block diagram of the earpiece
`100 in accordance with an exemplary embodiment is shown.
`As illustrated, the earpiece 100 can include a processor 206
`operatively coupled to the ASM 110, ECR 120, and ECM 130
`via one or more Analog to Digital Converters (ADC) 202 and
`Digital to Analog Converters (DAC) 203. The processor 206
`can monitor the ambient sound captured by the ASM 110 for
`target Sounds in the environment, such as an alarm (e.g., bell,
`emergency vehicle, security system, etc.), siren (e.g. police
`car, ambulance, etc.), Voice (e.g., "help', 'stop', 'police'.
`etc.), or specific noise type (e.g., breaking glass, gunshot,
`etc.). The memory 208 can store sound signatures for previ
`ously learned target sounds from which the processor 206
`refers to for detecting target Sounds. The Sound signatures can
`be resident in the memory 208 or downloaded to the earpiece
`100 via the transceiver 204 during operation as needed. Upon
`detecting a target Sound, the processor 206 can report the
`target to the user via audio delivered from the ECR 120 to the
`ear canal.
`0029. The earpiece 100 can also include an audio interface
`212 operatively coupled to the processor 206 to receive audio
`content, for example from a media player, and deliver the
`audio content to the processor 206. The processor 206 respon
`sive to detecting target Sounds can adjust the audio content
`and the target Sounds delivered to the earcanal. The processor
`206 can actively monitor the sound exposure level inside the
`ear canal and adjust the audio to within a safe and Subjectively
`optimized listening level range. The processor 206 can utilize
`computing technologies Such as a microprocessor, Applica
`tion Specific Integrated Chip (ASIC), and/or digital signal
`processor (DSP) with associated storage memory 208 such a
`Flash, ROM, RAM, SRAM, DRAM or other like technolo
`gies for controlling operations of the earpiece device 100.
`0030 The earpiece 100 can further include a transceiver
`204 that can Support singly or in combination any number of
`wireless access technologies including without limitation
`BluetoothTM, Wireless Fidelity (WiFi), Worldwide Interoper
`ability for Microwave Access (WiMAX), and/or other short
`or long range communication protocols. The transceiver 204
`can also provide Support for dynamic downloading over-the
`air to the earpiece 100. It should be noted also that next
`generation access technologies can also be applied to the
`present disclosure.
`0031. The power supply 210 can utilize common power
`management technologies such as replaceable batteries, Sup
`ply regulation technologies, and charging system technolo
`gies for Supplying energy to the components of the earpiece
`100 and to facilitate portable applications. A motor (not
`shown) can be a single Supply motor driver coupled to the
`power Supply 210 to improve sensory input via haptic vibra
`tion. As an example, the processor 206 can direct the motor to
`vibrate responsive to an action, such as a detection of a target
`Sound or an incoming Voice call.
`
`Exhibit 1022
`Page 08 of 14
`
`

`

`US 2008/0240458 A1
`
`Oct. 2, 2008
`
`0032. The earpiece 100 can further represent a single
`operational device or a family of devices configured in a
`master-slave arrangement, for example, a mobile device and
`an earpiece. In the latter exemplary embodiment, the compo
`nents of the earpiece 100 can be reused in different form
`factors for the master and slave devices.
`0033 FIG. 3 is a flowchart of a method 300 for earpiece
`monitoring and target detection in accordance with an exem
`plary embodiment. The method 300 can be practiced with
`more or less than the number of steps shown and is not limited
`to the order shown. To describe the method 300, reference
`will be made to components of FIG. 2, although it is under
`stood that the method 300 can be implemented in any other
`manner using other suitable components. The method 300
`can be implemented in a single earpiece, a pair of earpieces,
`headphones, or other suitable headset audio delivery device.
`0034. The method 300 can start in a state wherein the
`earpiece 100 has been inserted and powered on. As shown in
`step 302, the processor 206 can monitor the environment for
`target Sounds, such as an alarm, a horn, a voice, or a noise.
`Each of the target sounds can have certain identifiable fea
`tures that characterize the sound. The features can be collec
`tively referred to as a sound signature which can be used for
`recognizing the target Sound. As an example, the Sound sig
`nature may include statistical properties or parametric prop
`erties of the target Sound. For example, a Sound signature can
`describe prominent frequencies with associated amplitude
`and phase information. As another example, the sound signa
`ture can contain principal components identifying the most
`likely recognizable features of a target Sound.
`0035. The processor 206 at step 304 can then detect the
`target Sounds within the environment based on the Sound
`signatures. As will be shown ahead, feature extraction tech
`niques are applied to the ambient sound captured at the ASM
`110 to generate the Sound signatures. Pattern recognition
`approaches are applied based on known Sound signatures to
`detect the target Sounds from their corresponding Sound sig
`natures. More specifically, Sound signatures can then be com
`pared to learned models to identify a corresponding target
`Sound. Notably, the processor 206 can detect sound signa
`tures from the ambient sound regardless of the state of the
`earpiece 100. For example, the earpiece 100 may be in a
`listening state wherein ambient Sound is transparently passed
`to the ECR 120, in a media state wherein audio content is
`delivered from the audio interface 212 to the ECR 120, or in
`an active listening state wherein Sounds in the environment
`are selectively enhanced or suppressed.
`0036. At step 306, the processor 206 can adjust sound
`delivered to the ear canal in view of a detected target sound.
`For instance, if the earpiece is in a listening state, the proces
`Sor 206 can amplify detected target Sounds inaccordance with
`a Personalized Hearing Level (PHL). The PHL establishes
`comfortable and uncomfortable levels of hearing, and can be
`referenced by the processor 206 to set the volume level of the
`target Sound (or ambient sound) so as not to exceed the user's
`preferred listening levels. As another example, if the earpiece
`is in a media State, the processor 206 can attenuate the audio
`content delivered to the ear canal, and amplify the target
`sounds in the ear canal. The PHL can also be used to properly
`mix the volumes of the different sounds. As yet another
`example, if the earpiece 100 is in an active state, the processor
`206 can selectively adjust the volume of the target sounds
`relative to background noises in the environment.
`
`0037. The processor 206 can also compensate for an ear
`seal leakage due to a fitting of the earpiece 100 with the ear
`canal. An ear seal profile can be generated by evaluating
`amplitude and phase difference between the ASM 110 and the
`ECM 202 for known signals produced by the ECR 120. That
`is, the processor 120 can monitor and report transmission
`levels offrequencies through the earcanal 140. The processor
`206 can take into account the ear seal leakage when perform
`ing audio enhancement, or other spectral enhancement tech
`niques, to maintain minimal audibility of the ambient noise
`while audio content is playing.
`0038. Upon detecting a target sound in the ambient sound
`of the user's environment, the processor at step 308 can gen
`erate an audible alarm within the ear canal that identifies the
`detected Sound signature. The audible alarm can be a repro
`duction of the target Sound, an amplification of the target
`Sound (or the entire ambient Sound), a text-to-speech message
`(e.g. synthetic Voice) identifying the target Sound, a haptic
`vibration via a motor in the earpiece 100, oran audio clip. For
`example, the earpiece 100 can play a Sound bite (i.e., audio
`clip) corresponding to the detected target Sound such as an
`ambulance, fire engine, or other environmental sound. As
`another example, the processor 206 can synthesize a Voice to
`describe the detected target Sound (e.g., “ambulance
`approaching”).
`0039 FIG. 4 illustrates earpiece modes in accordance with
`an exemplary embodiment. The earpiece mode can be manu
`ally selected by the user, for example, by pressing a button, or
`automatically selected, for example, when the earpiece 100
`detects it is in an active listen state or in a media state. As
`shown in FIG. 4, the earpiece mode can correspond to Sig
`nature Sound Pass Through Mode (SSPTM), Signature
`Sound Boost Mode (SSBM), Signature Sound Replacement
`Mode (SSRM), Signature Sound Attenuation Mode (SSAM),
`and Signature Sound Replacement Mode (SSRM).
`0040. In SSPTM mode, ambient sound captured at the
`ASM 110 is passed transparently to the ECR 120 for repro
`duction within the ear canal. In this mode, the Sound produced
`in the ear canal Sufficiently matches the ambient sound out
`side the ear canal, thereby providing a “transparency effect.
`That is, the earpiece 100 recreates the sound captured at the
`ASM 110 to overcome occlusion effects of the earpiece 100
`when inserted within the ear. The processor 206 by way of
`sound measured at the ECM 130 adjusts the properties of
`sound delivered to the ear canal so the sound within the
`occluded ear canal is the same as the ambient sound outside
`the ear, as though the earpiece 100 were absent in the ear
`canal. In one configuration, the processor 206 can predict an
`approximation of an equalizing filter to provide the transpar
`ency by comparing an ASM 110 signal and an ECM 130
`signal transfer function.
`0041. In SSBM, target sounds and/or ambient sounds are
`amplified upon the processor 206 detecting a target Sound.
`The target sound can be amplified relative to the normal level
`received, or amplified above an audio content level if audio
`content is being delivered to the ear canal. As noted previ
`ously, the target Sound can also be amplified in accordance
`with a user's PHL to be within safe hearing levels, and within
`subjectively determined listening levels.
`0042. In SSRM, target sounds detected in the environment
`can be replaced with audible warning messages. For example,
`the processor 206 upon detecting a target Sound can generate
`synthetic speech identifying the target Sound (e.g., “ambu
`lance detected'). In such regard, the earpiece 100 audibly
`
`Exhibit 1022
`Page 09 of 14
`
`

`

`US 2008/0240458 A1
`
`Oct. 2, 2008
`
`reports the target sound identified thereby relieving the user
`from having to interpret the target Sound. The synthetic
`speech can be mixed with the ambient sound (e.g., amplified,
`attenuated, cropped, etc.), or played alone with the ambient
`Sound muted.
`0043. In SSAM, sounds other than target sounds can be
`attenuated. For instance, annoying Sounds or noises not asso
`ciated with target Sounds can be Suppressed. For instance, by
`way of a learning session, the user can establish what sounds
`are considered target Sounds (e.g., “ambulance') and which
`Sounds are non-target Sounds (e.g. jackhammer”). The pro
`cessor 206 upon detecting non-target Sounds can thus attenu
`ate these sounds within the occluded or partially occluded ear
`canal.
`0044 FIG. 5 is a flowchart of a method 500 for a method
`for Sound signature detection in accordance with an exem
`plary embodiment. The method 500 can be practiced with
`more or less than the number of steps shown and is not limited
`to the order shown. To describe the method 500, reference
`will be made to components of FIG. 2, although it is under
`stood that the method 500 can be implemented in any other
`manner using other suitable components. The method 500
`can be implemented in a single earpiece, a pair of earpieces,
`headphones, or other suitable headset audio delivery device.
`0045. The method can start at step 502, in which the ear
`piece 100 can enter a learn mode. Notably, the earpiece upon
`completion of a learning mode or previous learning configu
`ration can start instead at step 520. In the learning mode of
`step 502, the earpiece 100 can actively generate and learn
`Sound signatures from ambient Sounds within the environ
`ment. In learning mode, the earpiece 100 can also receive
`previously trained learning models to use for detecting target
`Sounds in the environment. In an active learning mode, the
`user can press a button or otherwise (e.g. Voice recognition)
`initiate a recording of ambient Sounds in the environment. For
`example, the user can upon hearing a new target Sound in the
`environment ("car horn'), activate the earpiece 100 to learn
`the new target Sound. Upon generating a sound signature for
`the new target Sound, it can be stored in the user defined
`database 504. In another arrangement, the earpiece 100 upon
`detecting a unique Sound, characteristic to a target Sound, can
`ask the user if they desire to have the sound signature for the
`unique Sound learned. In Such regard, the earpiece 100
`actively senses sounds and queries the user about their envi
`ronment to learn the Sounds. Moreover, the earpiece can
`organize learned sounds based on environmental context, for
`example, in outdoor (e.g. traffic, car, etc.) or indoor (e.g.,
`restaurant, airport) environments.
`0046. In another learning mode, trained models can be
`retrieved from an on-line database 506 for use in detecting
`target Sounds. The previously learned models can be trans
`mitted on a scheduled basis to the earpiece, or as needed,
`depending on the environmental context. For example, upon
`the earpiece 100 detecting traffic noise, Sound signature mod
`els associated with target Sounds (e.g., ambulance, police car)
`in traffic can be retrieved. In another exemplary embodiment,
`upon the earpiece 100 detecting conversational noise (e.g.
`people talking), Sound signature models for Verbal warnings
`(“help”, “police') can be retrieved. Groups ofsound signature
`models can be retrieved based on the environmental context
`or on user directed action.
`0047. As shown in step 508, the earpiece can also generate
`speech recognition models for target Sounds corresponding to
`voice, such as “help', 'police”, “fire', etc. The speech recog
`
`nition models can be retrieved from the on-line database 506
`or the user defined database 504. In the latter for example, the
`user can say a word or enter a text version of a word to
`associate with a verbal warning Sound. For instance, the user
`can define a set of words of interest along with mappings to
`their meanings, and then use keyword spotting to detect their
`occurrences. If the user enters an environment wherein
`another individual says the same word (e.g., “help') the ear
`piece 100 can inform the user of the verbal warning sound.
`For other acoustic sounds, the earpiece 100 can generate
`sound signature models as shown in step 510. Notably, the
`earpiece 100 itself can generate the Sound signature models,
`or transmit the captured target Sounds to external systems
`(e.g., remote server) that generate the sound signature mod
`els. Such learning can be conducted off-line in a training
`phase, and the earpiece 100 can be uploaded with the new
`learning models.
`0048. It should also be noted that the learning models can
`be updated during use of the earpiece, for example, when the
`earpiece 100 detects target sounds. The detected target sounds
`can be used to adapt the learning models as new target Sound
`variants are encountered. For example, the earpiece 100 upon
`detecting a target Sound, can use the Sound signature of the
`target Sound to update the learned models in accordance with
`the training phase. In Such an exemplary embodiment a first
`learned model is adapted based on new training data collected
`in the environment by the earpiece. In Such regard, for
`example, a new set of “horn’ target sounds could be included
`in real-time training without discarding the other "horn'
`Sounds already captured in the existing model.
`0049. Upon completion of learning, uploading, or
`retrieval of sound signature models, the earpiece 100 can
`monitor and report target Sounds within the environment. As
`shown in step 520, ambient Sounds (e.g. input signal) within
`the environment are captured by the ASM 110. The ambient
`sounds can be digitized by way of the ADC 202 and stored
`temporarily to a data buffer in memory 208 as shown in step
`522. The data buffer holds enough data to allow for generation
`of a sound signature as will be described ahead in FIG. 7.
`0050. In another configuration, the processor 206 can
`implement a “lookahead' analysis system by way of the data
`buffer for reproduction of pre-recorded audio content, using a
`data buffer to offset the reproduction of the audio signal. The
`look-ahead system allows the processor to analyze poten
`tially harmful audio artifacts (e.g. high level onsets, bursts,
`etc.) either received from an external media device, or
`detected with the ambient microphones,

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket