throbber
as United States
`a2) Patent Application Publication co) Pub. No.: US 2002/0178003 A1
`
` Gehrkeet al. (43) Pub. Date: Nov. 28, 2002
`
`
`US 20020178003A1
`
`(75)
`
`(54) METHOD AND APPARATUS FOR
`PROVIDING VOICE RECOGNITION
`SERVICE TO A WIRELESS
`COMMUNICATION DEVICE
`Inventors: James K. Gehrke, Lake in the Hills, IL
`(US); Terry Mansfield,Palatine,IT.
`(US); Richard A. Rose, Hoffman
`Estates, IL (US)
`Correspondence Address:
`MOTOROLA,INC.
`1303 EAST ALGONQUIN ROAD
`IL01/3RD
`SCHAUMBURG,IL 60196
`
`(73) Assignee: MOTOROLA, INC.
`
`(21) Appl. No.:
`(22)
`Filed:
`
`09/802,630
`Mar.9, 2001
`Publication Classification
`
`(SL)
`
`Tint. C17 eee cecccssssecsssneccseneessssneesseseees G10L 15/00
`
`(52) US. Che ce eeececseessesecssssnsceeenssnnecceesnneeeseennnnseeeees 704/246
`
`(57)
`
`ABSTRACT
`
`A wireless communication system employs a method and
`apparatus for providing voice recognition service to a wire-
`less communication device. Voice recognition information
`(e.g., a context model and voice training parameters) is
`generated by a wireless device user and stored in a memory
`(e.g., a SIM card) ofthe wireless device to form one portion
`of a voice recognition processing engine. Another portion of
`the voice recognition processing engine (e.g., a voice rec-
`ognition processor and operating software therefor) is imple-
`mented in a wireless system infrastructure of the wireless
`communication system. The wireless device transmits the
`voice recognition information to the system infrastructure
`preferably upon request for such information by the system
`infrastructure. The system infrastructure then uses both
`portions of the voice recognition processing engine to pro-
`vide voice recognition service to the wireless device andits
`user during operation of the wireless device.
`
`117
`
`119
`
`DEVICE
`
`VOICE
`RECOGNITION
`PROCESSOR
`
`MEMORY
`
`WIRELINE
`COMMUNICATION
`
`1
`
`APPLE 1041
`
`APPLE 1041
`
`1
`
`

`

`Patent Application Publication Nov. 28,2002 Sheet 1 of 4
`
`US 2002/0178003 A1
`
`117
`
`119
`
`104
`
`103 SYSTEM CONTROLLER
`
`
`VOICE
`
`RECOGNITION
`
`PROCESSOR
`
`
`100
`
`b—-——--—- ~~ —-— ~~ a
`
`FIG.
`
`1
`
`17
`
`215
`
`217.
`
`
`
`
`
`Z TRANSMITTER 7RECEIVER
` ANTENNA
`
`DUPLEXER 2035
`
`201
`
`205
`
`SWITCH/
`
`~
`FIG. 2
`
`
`
`20
`
`USER
`
`
`219 MTA
`
`
`
`2
`
`

`

`Patent Application Publication Nov. 28,2002 Sheet 2 of 4
`
`US 2002/0178003 A1
`
`409
`403
`
`
`
`WIRELESS|yoy CONTEXT VOICE
`
`
`peyice
`VERSION ID
`MODEL
`TRAINING
`
`PARAMETERS
`
`
`402
`
`
`0100 0001|SEND MESSAGE <CONJ> <PERSON>|SA G/0/T
`
`AE B/A/T
`
`DIAL <PHONE NUMBER>
`AH B/U/T
`
`
`405
`
`
`
`407
`
`401
`
`
`
`3
`
`

`

`Patent Application Publication Nov. 28,2002 Sheet 3 of 4
`
`US 2002/0178003 A1
`
`IMPLEMENT SECOND PORTION OF VOICE
`RECOGNITION PROCESSING ENGINE IN
`WIRELESS SYSTEM INFRASTRUCTURE
`
`USE BOTH FIRST AND SECOND PORTIONS
`OF VOICE RECOGNITION PROCESSING
`ENGINE TO PROVIDE VOICE RECOGNITION
`
`WIRELESS DEVICE TRANSMIT REQUEST 10 OPERATE IN
`
`FUNCTIONALITY 10 WIRELESS DEVICE
`
`STORE VOICE RECOGNITION
`INFORMATION SPECIFIC 10 WIRELESS
`DEVICE USER IN MEMORY OF
`
`
`
`
`
`
`
`
`WIRELESS SYSTEM TO WIRELESS
`SYSTEM INFRASTRUCTURE
`
`
`
`
`RECEIVE REQUEST FOR VOICE
`RECOGNITION INFORMATION FROM
`
`
`WIRELESS SYSTEM INFRASTRUCTURE
`
`
`
`TRANSMIT VOICE RECOGNITION
`INFORMATION TO WIRELESS SYSTEM
`
`INFRASTRUCTURE
`
`RECEIVE VOICE INSTRUCTION FROM
`WIRELESS DEVICE USER
`
`
`
`
`
`
`WIRELESS SYSTEM INFRASTRUCTURE
`FOR EXECUTION PURSUANT TO VOICE
`
`
`RECOGNITION INFORMATION
`
`4
`
`

`

`Patent Application Publication Nov. 28,2002 Sheet 4 of 4
`
`US 2002/0178003 Al
`
`START
`
`701
`
`RECEIVE REQUEST 10 OPERATE IN
`WIRELESS SYSTEM FROM WIRELESS DEVICE
`
`703
`
`705
`
`
`

`
`
`
`NO
`
`
`
`VOICE RECOGNITION
`INFORMATION ASSOCTATED WITH
`WIRELESS DEVICE PRESENTLY STORED
`IN WIRELESS SYSTEM
`INFRASTRUCTURE?
`
`
`
`
`
`
`
` PRESENT
`CURRENT?
`
`709
`
`711
`
`713
`
`
`
`VERSION OF STORED VOICE
`
`RECOGNITION INFORMATION
`
`
`REQUEST TRANSMISSION OF
`VOICE RECOGNITION INFORMATION
`
`
`RECEIVE NEW/UPDATED VOICE
`
`RECOGNITION INFORMATION FROM
`WIRELESS DEVICE
`
`
`STORE RECEIVED VOICE
`RECOGNITION INFORMATION
`
`Pp
`
`715
`
`RECEIVE DATA MESSAGE CONTAINING
`VOICE INSTRUCTION FROM
`WIRELESS DEVICE
`
`717
`DETERMINE INSTRUCTION CONTAINED IN
`DATA MESSAGE BASED ON STORED
`VOICE RECOGNITION INFORMATION
`
`719
`GENERATE DATA MESSAGE REPRESENTATIVE
`OF DETERMINED INSTRUCTION TO
`FACILITATE EXECUTION OF INSTRUCTION
`
`700
`
`ao
`
`724
`
`FIC.
`
`7
`
`oa
`
`5
`
`

`

`US 2002/0178003 Al
`
`Nov.28, 2002
`
`METHOD AND APPARATUS FOR PROVIDING
`VOICE RECOGNITION SERVICE TO A WIRELESS
`COMMUNICATION DEVICE
`
`FIELD OF THE INVENTION
`
`[0001] The present invention relates generally to wireless
`communication systems and, in particular, to a method and
`apparatus for providing voice recognition service to a wire-
`less communication device operating in a wireless commu-
`nication system.
`
`BACKGROUND OF THE INVENTION
`
`[0002] Wireless communication systems are well known.
`Such systems include, but are not limited to, cellular com-
`munication systems operating in accordance with various
`promulgated radio access technologies, such as Advanced
`Mobile Phone Service (AMPS), Narrowband Advanced
`Mobile Phone Service (NAMPS), United States Digital
`Cellular (USDC), Global Systems for Mobile Communica-
`tions (GSM), and Code Division Multiple Access (CDMA),
`personal communication systems (PCS) operating in accor-
`dance with various radio access technologies, such as
`CDMA,and multi-service systems, such as the “MOTORO-
`LA”’“iDEN” system, that provide many other services in
`addition to person-to-person calling, such as packet data,
`paging, short message service, and wireless Internet access.
`Many PCSoperators are also entering the wireless Internet
`access arena.
`
`[0003] To help facilitate hands-free operation of wireless
`communication devices, such as radiotelephones or two-way
`radios, operating on such systems, some systems and/or
`communication devices provide voice recognition service
`and/or functionality. To provide voice recognition capability,
`a hardware and software voice recognition processing
`engine, such as the IBM Voice Type Application Factory for
`Windows voice recognition processor and accompanying
`software that is commercially available from International
`Business Machines Corporation of Armonk, N.Y., must be
`trained to recognize commands or instructions spoken by
`each user tor which the system or device will be providing
`voice recognition service. Typically, a user-defined vocabu-
`lary (commonly referred to as a “context model”) is estab-
`lished and associated with the user’s speech during a setup
`phase of the voice recognition engine. The size or scope of
`the context modcl that can be supported depends upon how
`the voice recognition engine is implemented.
`
`In the prior art, voice recognition in wireless sys-
`[0004]
`temsis either completely infrastructure-based or completely
`device-based. Thatis,all the voice recognition hardware and
`software resides either in the wireless system infrastructure
`(e.g.,
`in a mobile switching center (MSC) of a cellular
`system) or in the wireless communication device itself.
`When voice recognition is implemented completely in the
`system infrastructure, a high power processing system may
`be employed that is capable of supporting relatively large
`context models for individual wireless device users. Since
`
`the wireless system infrastructure is shared by manyusers or
`subscribers, the cost of providing a high power voice rec-
`ognition processing system is typically recovered through
`incremental service fees charged to many device users.
`Therefore, each user incurs a relatively small expense for
`voice recognition service.
`
`[0005] On the other hand, incorporating a high perfor-
`mance voice recognition processing system (processor and
`memorycapacity) directly into a wireless deviceis typically
`cost-prohibitive. Consequently, lower powervoice recogni-
`tion processing systems are typically incorporated in wire-
`less devices. Such lower power voice recognition systems
`are costly enough (typically ten to twenty percent (10-20%)
`of the cost of the wireless device for the additional memory
`and computational power), reduce battery life, and only
`support a very limited context modelor instruction set. For
`example, a voice recognition system completely incorpo-
`rated in a wireless device typically only facilitates telephone
`calls based on a single format, such as speaking the digits of
`a target telephone number or speaking a moderate number
`(e.g., ten to twenty) of voice-recognizable sound signatures
`(c.g., names) that may be used to represent specific target
`telephone numbers. When sound signatures are accommo-
`dated, each sound signature is identified during voice rec-
`ognilion training and is associated wilh a target telephone
`numberthat is entered into and stored by the wireless device.
`Once the voice recognition system is trained, the user can
`say the nameoridentity of the stored sound signature and an
`instruction from a small instruction set (which quite likely
`includes only a “Call”instruction). For example, the voice
`recognition system, when trained, may recognize “Call
`[Target Name from Stored Set]”. The system, whentrained,
`may also recognize the numbers “Zero” through “Nine” to
`facilitate digit dialing, but that is about the extent of the
`voice recognition service provided by completely device-
`based voice recognition systems due to the wireless device’s
`cost-limited processing capabilities.
`
`[0006] Although each of the two aforementioned voice
`recognition system implementations provides at least some
`voice recognition capability for wireless device users, the
`two implementations suffer from certain undesirable limi-
`tations. For example, although the completely infrastruc-
`ture-based voice recognition system supports a large context
`model for each wireless device user, voice recognition may
`be used by a wireless device user only when the user is
`operating his or her wireless device in the wireless system
`containing the user’s context model. Since the voice recog-
`nition system is completely infrastructure-based, all
`the
`hardware and software, including the context models and
`any user-specific training parameters, are stored in infra-
`structure memory(e.g., in a homelocation register (HLR)or
`some other database associated with the voice recognition
`system). Thus,if a wireless device user roamsto a different
`wireless system, the user cannot use the voice recognition
`feature cven though the new system may support voice
`recognition, unless the user goes through the process of
`training the newvoice recognition system and storing his or
`her context model in the new system. A completely device-
`based voice recognition system enables voice recognition
`functionality to travel with the device, but at
`increased
`device cost and with much more limited voice recognition
`capabilities as compared to an infrastructure-based system.
`
`[0007] Therefore, a need exists for a method and apparatus
`for providing voice recognition service to a wireless com-
`munication device that provide the benefits of both com-
`pletely infrastructure-based and completely device-based
`voice recognition systems, without their respective disad-
`vantages.
`
`6
`
`

`

`US 2002/0178003 Al
`
`Nov.28, 2002
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`
`[0008] FIG. 1 is a block diagram of a wireless commu-
`nication system in accordance with the present invention.
`
`[0009] FIG. 2 is a block diagram of a wireless commu-
`nication device in accordance with a preferred embodiment
`of the present invention.
`
`[0010] FIG. 3 is a block diagram of an arrangement for
`generating and storing voice recognition information in
`accordance with a preferred embodiment of the present
`invention.
`
`(0011] FIG. 4 illustrates an exemplary voice recognition
`information database stored in a memory of a wireless
`system infrastructure in accordance with a preferred
`embodiment of the present invention.
`
`the present invention.
`
`[0013] FIG.6 is a logic flow diagram of steps executed by
`a wireless communication device to enable a wireless sys-
`tem infrastructure to provide voice recognition service to the
`wireless communication device in accordance with a pre-
`ferred embodimentof the present invention.
`
`stored voice recognition information to the infrastructure to
`enable the infrastructure to provide voice recognition service
`to the wireless device. In addition, by only storing a small
`portion of the overall voice recognition processing engine in
`the wireless device,
`the present invention eliminates the
`need for a high power processor in the wireless device to
`support voice recognition functionality. Further, by dividing
`the voice recognition processing engine between the wire-
`less device and the wireless system infrastructure,
`the
`present invention facilitates the use of a much more expan-
`sive user-defined vocabulary (e.g., context model) than does
`wireless device-based voice recognition systems because the
`voice recognition system of the present invention is much
`less processor-limited due to incorporation of the voice
`recognition processor in the infrastructure rather than the
`wireless device. Thus, the present invention provides voice
`recognition functionality that follows a wireless communi-
`[0012] FIG.5is a logic flow diagramof steps executed to
`cation device whereverit goes by utilizing a wireless device
`provide voice recognition functionality to a wireless com-
`that maintains its own voice recognition information (c.g.,
`munication device in accordance with one embodiment of
`context model) andutilizing a wireless system infrastructure
`that maintains the high performance processing necessary to
`facilitate voice recognition service.
`[0017] The present invention can be more fully understood
`with reference to FIGS. 1-7, in which like reference numer-
`als designate like items. FIG. 1 is a block diagram of a
`wireless communication system 100 in accordance with the
`present invention. The wireless communication system 100
`[0014] FIG.7is a logic flow diagram of steps executed by
`includes a wireless system infrastructure 101 and one or
`a wireless system infrastructure to provide voice recognition
`service to a wireless communication device in accordance
`more wireless communication devices 103, 104 (two
`shown). The wireless communication system may be any
`form of wireless system,
`including without limitation, a
`cellular communication system, a PCS system, a multi-
`service system, such as the “MOTOROLA”iDEN”system,
`a two-way radio system, a paging system, a wireless dala
`system, or any other wireless system that supports voice
`recognition as herein described.
`[0018] The wireless system infrastructure 101 includes
`one or more base transceiver sites (BTSs) 106, 107 (two
`shown), a system controller 109, a local or wide area
`network (LAN/WAN) 111 , and one or more memory
`devices 113 that may be separately coupled to the LAN/
`WAN 111 as shownor be distributed in the various infra-
`
`with a preferred embodimentof the present invention.
`
`DETAILED DESCRIPTION OF A PREFERRED
`EMBODIMENT
`
`invention encompasses a
`the present
`[0015] Generally,
`method and apparatus for providing voice recognition ser-
`vice to a wireless communication device. Voice recognition
`information(e.g., a context model and voice training param-
`eters) is generated by a wireless communication device user
`and stored in a memory (e.g., a smart or SIM card) of the
`wireless communication device to form one portion of a
`voice recognition processing engine. Another portion of the
`voice recognition processing engine (e.g., a voice recogni-
`tion processor and operating software therefor) is imple-
`mented in a wireless system infrastructure. The wireless
`communication device transmits the voice recognition infor-
`mation to the wireless system infrastructure preferably upon
`request for such information by the wireless system infra-
`structure. The wireless system infrastructure then uses both
`portions of the voice recognition processing engine to pro-
`vide voice recognition service to the wireless communica-
`tion device and its user during operation of the wireless
`communication device.
`
`[0016] By providing voice recognition functionalityto the
`wireless communication device in this manner, the present
`invention enables voice recognition to be used by a wireless
`communication device on any system that has infrastructure-
`based voice recognition capability, withoul requiring a new
`context model
`to be generated prior to accessing each
`system as is required in the prior art. Thus, when a wireless
`communication device roams from its home system to
`another
`system that
`supports voice recognition (e.g.,
`includes an infrastructure-based voice recognition proces-
`sor), the wireless device need only transmit its previously
`
`structure components (such as memory device 115 in the
`system controller 109). Each BTS 106, 107 is a conventional
`BTSthat includes one or more base transceiverstations that
`
`preferably transmit and receive digital messages over a
`respective wireless communication link 117, 119 (e.g., radio
`frequency (RF) channel). The system controller 109 is
`operably coupled to each BTS 106, 107 via the LAN/WAN
`111 and preferably includes a voice recognition processor
`121 and optional memory 115. The system controller 109 is
`preferably a controller that coordinates or controls commu-
`nication within the entire wireless system 100. For example,
`the system controller 109 may be a central controller of a
`two-way trunked radio system, a mobile switching center
`(MSC)of a cellular, PCS or multi-service system, a dispatch
`application processor (DAP)of a multi-service system, such
`as the “iDEN”system,or a base station controller in a single
`base station system. The voice recognition processor 121
`preferably comprises a microprocessor or another other
`suitable processor that operates in accordance with opera-
`tional or programming instructions (e.g., a software engine)
`stored in memory device 115 or some other memory device
`113. Alternatively, the voice recognition processor 121 may
`
`7
`
`

`

`US 2002/0178003 Al
`
`Nov.28, 2002
`
`be another microprocessor, a microcontroller, a digital signal
`processor (DSP), a state machine, logic circuitry, or any
`other device or group of devices that processes information
`based on operational or programming instructions. One of
`ordinary skill in the art will recognize that when the voice
`recognition processor 121 has one or more of its functions
`performed by a state machineorlogic circuitry, the memory
`115 containing the corresponding operational instructions
`may be embeddedwithin the state machine orlogiccircuitry.
`In the simplest systems, the voice recognition processor 121
`mayreside in a personal computer and the voice recognition
`software engine may run in the background on the personal
`computer, provided that the microprocessor and the memory
`115 are appropriately sized.
`
`[0019] The memorydevices 113, 115 may include one of
`more of various digital storage media, such as any form of
`random access memory (RAM), any form of read only
`memory (ROM), a hard disk, or any other medium for
`storing digital
`information. As mentioned above,
`the
`memory 115 preferably stores operational instructions that,
`when executed, cause the voice recognition processor 121 to
`perform its particular functions. The operations performed
`by the voice recognition processor 121 and the rest of the
`elements of the wireless communication system 100 are
`described in detail below.
`
`[0020] An electronic device 123 may be coupled to the
`wireless system’s LAN/WAN 111 via an appropriate com-
`munication link 125, such asthe Internet (e.g., via a dial-up
`telephoneline, a digital subscriber line (DSL), an integrated
`digital systems network (ISDN) connection, or a cable
`connection) or some other wide area Internet protocol (IP)
`network. Such an electronic device 123 may be an Internet
`appliance, an IP addressable garage door opener, an IP
`addressable television or other entertainment device, or any
`other electronic device that may be operated or controlled
`remotely in accordance with digital or analog control signals
`issued by the wireless system infrastructure 101. As
`described in detail below, such control signals are generated
`in response to voice commandsissuedby a userof a wireless
`communication device 103. One or more wireline commu-
`nication devices 127 (one shown), such as a telephone, an
`audio interface to a computer, a data terminal, or a set top
`box, and/or any other means to send and receive audio
`commands, may also be coupled to the wircless system’s
`LAN/WAN 111 via an appropriate communication link 129
`(e.g., via the public switched telephone network (PSTN), the
`Internet, or some other network) to facilitate a communica-
`tion between a user of the wireline device 127 and the user
`of the wireless device 103 having voice recognition func-
`tionality.
`[0021] A preferred embodiment of a wireless communi-
`cation device 103 having voice recognition functionality in
`accordance with the present invention is illustrated in block
`diagram form in FIG. 2. The wireless device 103 includes,
`inter alia, an antenna 201, an antenna switch/duplexer 203,
`a transmitter 205, a receiver 207, a processor 209, memory
`211 for storing operating instructions executable by the
`processor 209 and for storing other information (e.g., voice
`recognition information and wireless device identification
`information) as described in more detail below, a user
`interface 213, a display 215, and a data port 217.
`[0022] The wireless device 103 may be any two-way
`communication device capable of communicating in a wire-
`
`less communication system 100. Thus, the wireless device
`103 may be a two-wayradio, a radiotelephone, a two-way
`pager, a wireless data terminal, a laptop computer, a palmtop
`computer, a personal digital assistant (PDA), or any other
`two-way device having wireless capabilities.
`
`[0023] The antenna 201 may include a single antenna
`element or multiple antenna elements (e.g., an array). The
`antenna switch/duplexer 203 may be a known PIN diode or
`other switch to implement an antenna switch for half-duplex
`operation or a knownarrangementof filters to implement a
`duplexer for full duplex operation.
`
`{0024] The transmitter 205 and the receiver 207 include
`appropriate circuitry to enable digital or analog transmis-
`sions over a wireless communication link 117. For example,
`the transmitter 205 and the receiver 207 may be imple-
`mented as an appropriate wireless modem, or as conven-
`tional transmitting and receiving components in a two-way
`wireless device. In the eventthat the transmitter 205 and the
`
`receiver 207 are implemented as a wireless modem, the
`wireless modem may be located on a Personal Computer
`Memory Card International Association (PCMCIA) card
`that may be inserted into a computing device, such as a
`laptop or palmtop computer or PDA,to facilitate wireless
`communications. Wireless modemsare well known; thus no
`further discussion of them will be presented except
`to
`facilitate an understanding of the present invention.
`
`[0025] The processor 209 may be a microprocessor, a
`microcontroller, a digital signal processor (DSP), a state
`machine, logic circuitry, or any other device or group of
`devices that processes information based on operational or
`programming instructions. One of ordinary skill in the art
`will recognize that when the processor 209 has one or more
`of its functions performed by a state machine or logic
`circuitry, the memory containing the corresponding opera-
`tional
`instructions may be embedded within the state
`machine or logic circuitry. The memory 211 may include
`one of more of various digital storage media, such as RAM,
`ROM,flash memory, a smart card, a subscriber identity
`module (SIM) card, a floppy disk, a compact disk read only
`memory (CD-ROM), a hard disk drive, a digital versatile
`disk (DVD), flash memory or any other medium ordevice(s)
`for storing digital information. Thus, the memory 211 may
`be embedded within the wireless device 103, may be
`inserted into or otherwise operably coupled to the wireless
`device 103 bythe wireless device user, or both (e.g., certain
`information may be stored in embedded ROM, while other
`information may be stored on an insertable SIM card). As
`mentioned above, the memory 211 preferably stores oper-
`ating instructions that, when executed, cause the processor
`209 to perform its particular functions. In addition,
`the
`memory 211 preferably includes one or more memory
`locations 219, 220 (¢.g., registers or sets of registers) that
`store a small portion of a voice recognition processing
`engine, as described in detail below, to enable the wireless
`device to reccive voice recognition service in multiple
`wireless systems. The operations performed bythe proces-
`sor 209 and the rest of the elements of the wireless com-
`munication device 103 are described in detail below.
`
`[0026] The userinterface 213 preferably includes a micro-
`phone to receive voice instructions issued by the wireless
`device user and may also include other conventional user
`interface elements, such as a keyboard, a keypad, a mouse
`
`8
`
`

`

`US 2002/0178003 Al
`
`Nov.28, 2002
`
`or rollerball, a thumbwheel, a touchscreen, a touchpad, or
`any other device for allowing the user of the wireless device
`103 to make a selection or instruct the device 103 to take
`
`some action. The display 215 may be any conventional
`cathode ray tube (CRT) display,
`liquid crystal display
`(LCD), or other display. In addition, when audio display is
`desired, the display 215 preferably includes an audio display
`device, such as one or more speakers. Although not shown
`in FIG. 2, the wireless device 103 may further include an
`alerting device, such as a tone generator that produces an
`audiblealert or an electrically actuatable vibration device, to
`alert the device that a message or a communication has been
`received that may require the user’s attention. The data port
`217 preferably comprises a conventional data port, such as
`a wired or wireless serial port or equivalent.
`[0027] FIG. 3 is a block diagram of an arrangement for
`generating voice recognition information and storing the
`voice recognition information in the wireless communica-
`tion device memory 211 in accordance with a preferred
`embodiment of the present
`invention. As illustrated,
`the
`arrangement includes a voice recognition information (VRI)
`generation node 301 and a communication link 303 coupling
`the VRI generation node 301 to the wireless device memory
`211 on orin which the voice recognition informationis to be
`stored. When the voice recognition information is to be
`stored in embedded memory211 of the wireless device 103,
`the communication link 303 is coupled to the data port 215
`of the wireless device 103. On the other hand, when the
`voice recognition information is to be stored in or on a
`memory device 211 that
`is insertable into or otherwise
`operably coupleable to the wireless device 103, the com-
`munication link 303 is coupled to an appropriate drive 304
`for writing data to the particular memory device 211. The
`VRI generation node 301 is preferably coupled by an
`appropriate communication link 305 (e.g., the Internet) to
`the LAN/WAN (e.g., LAN/WAN 111) of the wireless
`device’s home wireless system infrastructure 101 to allow
`the VRI generation node 301 to communicate with the voice
`recognition processor 121 as described in more detail below.
`[0028] The VRI generation node 301 preferably comprises
`a computer (e.g., a personal computer, a workstation, a
`laptop or notebook computer, or a local server) or similar
`data device executing a software program that provides a
`user-friendly graphical user interface (GUI) to enable the
`wireless device user to generate unique voice recognition
`information to be used in providing voice recognition func-
`tionality to the wireless device 103. In a preferred embadi-
`ment, the voice recognition information includes a context
`model and voice training parameters. The context model is
`a user-defined, unique, personal vocabulary that includes a
`set of instructions and operandsthat are to be automatically
`recognized by the infrastructure’s voice recognition proces-
`sor 121 upon receipt of an instruction and operand(s) from
`the wireless device 103. The context model may include
`instructions that, inter alia, allow the user of the wireless
`device to control operation of the wireless device 103 (e.g.,
`turn the device 103 off, or turn features of the device 103 on
`and off), control operation of a remotely located electronic
`device 123 (e.g., control operation of the wireless device
`user’s residential garage door opener, sprinkler system,
`security system, or other [P-addressed device), retrieve
`information stored in the wireless device 103 (e.g., retrieve
`stored telephone numbers or other contact
`information),
`establish a communication in a wireless communication
`
`initiate a telephone call with one or more
`system (e.g.,
`wireless and/or wireline communication devices 104, 127),
`and control, to some extent, operationof the infrastructure’s
`voice recognition processor 121 (e.g., activate or wake-up
`the voice recognition processor 121).
`
`[0029] An exemplary context model may include the
`following instruction set and operands:
`
`
`
`<Action>::= SEND MESSAGE<conj> <Person>
`| DIAL <PhoneNumber>
`| CALL <Person> <conj> <Path>
`| PLAY MESSAGE<conj> <Person>
`| OPEN <conj> <Doors> DOOR
`| CLOSE <conj> <Doors> DOOR
`| DISPLAY MESSAGES
`| CANCEL
`| TURN ON <Devices>
`| TURN OFF <Devices>
`| STANDBY MODE
`<conj>::= ON
`TO
`FROM
`ON THE
`THE
`<PhoneNumber>::= <Singles> <Singles> <Singles> <Singles> <Singles>
`<Singles> <Singles> <Singles> <Singles> <Singles> <Singles>
`<Singles>::= OH
`ZERO
`ONE
`TWO
`THREE
`FOUR
`FIVE
`SIX
`SEVEN
`LIGHT
`NINE
`HUNDRED
`<Person>::= MOM
`| DAD
`| PIZZA
`| BABY SITTER
`| [Other Names, Nicknames or Places]
`<Path>::= RADIO
`| CELL PHONE
`| PHONE
`<Doors>::= FRONT
`| GARAGE
`| LEFT GARAGE
`| RIGHT GARAGE
`<Devices>::= SECURITY SYSTEM
`| OVEN
`| SPRINKLER SYSTEM
`
`[0030] One of ordinary skill in the art will recognize and
`appreciate that various other context models may be readily
`generated to coincide with the particular requirementsof the
`wireless device user.
`
`In addition to a context model, the voice recogni-
`[0031]
`tion information preferably includes training parameters
`related to a voice of the wireless device user. The voice
`
`training parameters include data for adapting the infrastruc-
`ture’s voice recognilion processor to the voice characteris-
`tics of the wireless device user. For example,
`training
`parameters may include the following phonemesrepresent-
`
`9
`
`

`

`US 2002/0178003 Al
`
`Nov.28, 2002
`
`ing English sounds in accordance with IBM’s Voice Type
`Application Factory for Windowsor anyother user-defined
`phonemes:
`
`wireless device’s audio processing, and providesthe training
`parameters to the VRI generation node 301 via the infra-
`structure’s LAN/WAN 111 and communication link 305.
`
`AA
`AO
`AY
`CH
`DH
`EY
`GD
`XxX
`K
`M
`Ow
`PD
`SH
`TD
`UH
`Ww
`7H
`
`cfost
`b/ough/t
`bfi/te
`/chfurch
`/th/ey
`b/aist
`ta/g/
`ros/es/
`{k/ick
`/m/om
`bfoa/t
`shi/p/
`/sh/oe
`se/t/
`b/oofk
`/w/et
`mea/s/ure
`
`AE
`AX
`B
`D
`EH
`F
`Till
`IY
`KD
`N
`OY
`R
`SIL
`TH
`UW
`Y
`
`b/ast
`th/e/
`/ofob
`/dfad
`b/eft
`/ffire
`fh/ay
`b/ea/t
`comi/c/
`/nfon
`b/oy/
`it/ed
`(silence)
`/thfief
`b/oo/t
`fy/et
`
`AH
`AXR
`BD
`DD
`ER
`G
`Til
`JH
`L
`NG
`P
`Ss
`T
`TS
`Vv
`Z
`
`b/u/t
`summ/er/
`tu/bje
`delete/d/
`bfir/d
`/g/ag
`b/ift
`/jfadge
`fed
`si/ng/
`{p/op
`/sfis
`it/o
`i/ts/
`iwiery
`/z/oo
`
`‘Training parameters may additionally include
`[0032]
`modifications or corrections to such phonemes to account
`for (a) dialect,
`inflection, or other characteristics of the
`wireless device user’s voice, (b) processing (¢.g., speech
`encoding) performed by the wireless device 103to facilitate
`transmission over the wireless link 117, and/or (c) audio-
`modifying characteristics of the wireless link 117itself. For
`example, the training parameters may include the frequency
`ranges associated with various individuals in accordance
`with the well-known Markov speech models to enable the
`voice recognition processor to optimize performance based
`on the gender, age, or particular speech patterns of the
`wireless device user. Alternatively or additionally, the train-
`ing parameters mayinclude correction factors to account for
`the audio characteristics of the wireless link 117 or speech
`encoding performed by the wireless device 103 to obtain a
`desired transmission quality. For example, correction factors
`may be used to modify the Markov speech models to match
`the speech models to the characteristics of the sound signa-
`ture (e.g., phonemes) of the wireless device user as such
`sound signature is actually processed by the wireless device
`103 and received over the wireless link 117.
`
`the wireless device
`In a preferred embodiment,
`[0033]
`user uses the VRI generation node 301 and the wireless
`device 103 to generate his or her unique voice recognition
`information and store the generated voice recognition infor-
`mation in one

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket