throbber
(19) United States
`(12) Patent Application Publication (10) Pub. No.: US 2010/0085216 A1
`(43) Pub. Date:
`Apr. 8, 2010
`MS
`
`US 2010.0085216A1
`
`(54) VIBRATION BASED USER INPUT FOR
`MOBILE DEVICES
`
`AJITH MS, Bangalore (IN)
`(76) Inventor:
`Correspondence Address:
`Raj Abhyanker LLP
`1580 W. El Camino Real, Suite 8
`Mountain View, CA 94040 (US)
`
`(21) Appl. No.:
`
`12/243,976
`
`(22) Filed:
`
`Oct. 2, 2008
`
`Publication Classification
`
`(51) Int. Cl.
`HO3K 7/94
`
`(2006.01)
`
`(52) U.S. Cl. .......................................................... 341?20
`
`ABSTRACT
`(57)
`A method, system, and apparatus of a vibration based user
`input for mobile devices are disclosed. In one embodiment, a
`method of controlling an electronic device includes receiving
`an impact signal, (e.g., a vibration generated by a user con
`trolled impact with a Surface mechanically coupled to a hous
`ing of the electronic device). The method further includes
`identifying a user command to the electronic device based on
`the impact signal, and performing a predesignated action
`(e.g., a mute, a power on, a power off, a Volume increase, a
`Volume decrease, a music track change, a call redirect, a call
`directed to Voicemail, etc.) based on the user command. The
`user command may be identified using a number of user
`controlled impacts and an interval between a prior user con
`trolled impact and a later user controlled impact.
`
`ELECTRONIC
`DEVICE 250
`
`ill,
`
`PRIOR TO USER GENERATED IMPACT
`
`ELECTRONIC
`DEVICE 250
`
`Hill,
`
`USER GENERATED IMPACT CREATES VIBRATIONS
`
`ELECTRONIC
`DEVICE 250
`
`O >
`
`fe
`
`ELECTRONICELECTRONIC
`
`DEVICE 250 DEVICE 252
`
`WBRATIONS ARE RECEIVED BY ELECTRONIC DEVICE
`
`ELECTRONIC DEVICE PERFORMS ANACTION
`
`APPLE 1052
`
`1
`
`

`

`Patent Application Publication
`
`Apr. 8, 2010 Sheet 1 of 5
`
`US 2010/008521.6 A1
`
`
`
`
`
`5ÕT ETT GOW
`
`
`
`SDNISSE OO}}c}
`
`OHNO}}_LOETE`N
`
`2
`
`

`

`Patent Application Publication
`
`Apr. 8, 2010 Sheet 2 of 5
`
`US 2010/008521.6 A1
`
`3
`
`

`

`Patent Application Publication
`
`Apr. 8, 2010 Sheet 3 of 5
`
`US 2010/008521.6 A1
`
`350
`
`308
`
`PROCESSOR 30
`
`INSTRUCTIONS 324
`
`MAIN MEMORY 304
`
`INSTRUCTIONS 324
`
`
`
`STATIC MEMORY 306
`
`INSTRUCTIONS 324
`
`NETWORK iNTERFACE
`DEVICE 320
`
`VIDEO DISPLAY 31
`
`ALPHA-NUMERIC
`INPUT DEVICE 312
`
`CURSOR CONTROL
`DEVICE 314
`
`DRIVE UNIT 316
`
`MACHINE READABLE
`MEDIUM 322
`
`INSTRUCTIONS 324
`
`SIGNAL GENERATION DEVICE 318
`
`FIGURE 3
`
`4
`
`

`

`Patent Application Publication
`
`Apr. 8, 2010 Sheet 4 of 5
`
`US 2010/008521.6 A1
`
`RECEIVE AN IMPACT SIGNAL
`
`402
`
`404
`
`DISTINGUISH AN IMPACT SIGNAL FROMAN OTHER SIGNAL BY A FREOUENCY OF
`THE VIBRATION CAUSED BY THE USER CONTROLLED IMPACT
`
`DETERMINE THE PREDESIGNATED ACTION BASED ON A CONTEXT OF THE
`ELECTRONIC DEVICE
`
`408
`
`DETERMINE THE PREDESIGNATED ACTION BASED ON A CONTEXT OF THE
`ELECTRONIC DEVICE
`
`410
`
`PERFORMAPREDESIGNATED ACTION BASED ON THE USER COMMAND
`
`FIGURE 4
`
`5
`
`

`

`Patent Application Publication
`
`Apr. 8, 2010 Sheet 5 of 5
`
`US 2010/008521.6 A1
`
`502
`
`
`
`
`
`FORMA VIBRATION SENSING MODULE IN AELECTRONIC DEVICE TO RECEIVE AN
`IMPACT SIGNAL COMPRISED OF A WIBRATION GENERATED BY A USER
`CONTROLLED IMPACT WITH A SURFACE MECHANICALLY COUPLED TO A
`VIBRATION RECEIVER OF THE ELECTRONIC DEVICE
`
`
`
`PLACEA DIFFERENTIATIONMODULE IN THE ELECTRONIC DEVICE TO IDENTIFY
`THE AUDITORY IMPACT SIGNAL ASA COMMAND TO THE ELECTRONIC DEVICE
`
`504
`
`CREATE ATRANSLATION MODULE IN THE ELECTRONIC DEVICE TO TRANSLATE
`THE COMMAND INTO A BINARY NUMBER
`
`508
`
`FORMA CENTRAL PROCESSING MODULE IN THE ELECTRONIC DEVICE TO
`EXECUTE A PREDESIGNATED ACTION BASED ON THE WALUE OF THE BINARY
`NUMBER
`
`506
`
`FIGURE 5
`
`6
`
`

`

`US 2010/008521.6 A1
`
`Apr. 8, 2010
`
`VIBRATION BASED USER INPUT FOR
`MOBILE DEVICES
`
`FIELD OF TECHNOLOGY
`0001. This disclosure relates generally to an enterprise
`method, a technical field of software and/or hardware tech
`nology and, in one example embodiment, to a vibration based
`user input for mobile devices.
`
`BACKGROUND
`0002. A control for an electronic device (e.g., a mobile
`device, a laptop, a car navigation system, a stress sensor, a car
`audio, an underwater sensor, etc.) may require a particular
`user input (e.g., a touch, a movement, a sound, etc.) to control
`a function (e.g., a mute, a Voicemail, a Volume, a track control,
`a call disconnect, a speed, a door control, etc.) of an electronic
`device.
`0003. The particular user input may involve a direct con
`tact with a limited area surface (e.g., a button, a touchpad,
`etc.). The limited area Surface may require user to view the
`limited area surface to interact with it. In addition, the limited
`area Surface may also require a user to extract the electronic
`device from a container (e.g., a pocket, a wallet, a purse, a
`protective case, a cover, etc.). These requirements may cause
`a delay (e.g., a time to extract the device from the container, a
`time to locate the limited area surface, etc.) and/or may cause
`a user to decide not to access the device. The delay or the
`failure to access the device may result in a disturbance (e.g.,
`a phone ring) and/or an undesired result (e.g., a missed call, a
`transfer to Voicemail, a different music track, a distraction, a
`loss of vehicle control, a missed direction, etc.). The undes
`ired result may further result in an additional delay and/or
`inefficiency (e.g., an access to Voicemail, a u-turn to find a
`missed location, a repeat of an already heard audio segment,
`etc.)
`0004. The limited area surface may also contain a moving
`part (e.g., a button mechanism) and/oran unsealed edge (e.g.,
`a crack next to a touchpad edge and/or a button edge). The
`moving part may allow a foreign Substance (e.g., a fluid
`and/or a dirt) to enter a housing for the electronic device. The
`foreign Substance may accelerate a breakdown of the moving
`part and/or a component of the electronic device (e.g., a
`processor, a memory, a hard drive, a speaker, a microphone,
`etc.). The breakdown of the moving part and/or the compo
`nent of the electronic device may prevent the electronic
`device from operating, which may result in a further delay
`and/or inefficiency (e.g., a repair cost, a replacement cost,
`etc.).
`0005. In addition, the particular user input may involve a
`Sound greater than a threshold Volume. By exceeding the
`threshold Volume, the sound may cause a disturbance. The
`audible sound may also be easily generated by environmental
`Sources, which may cause the electronic device to be con
`trolled in undesired ways (e.g., a call disconnect, a mute, a
`Video end, a power on, a power off, etc.). These undesired
`effects may also result in a delay and/or inefficiency.
`
`SUMMARY
`0006. A method, system, and apparatus of a vibration
`based user input for mobile devices are disclosed. In one
`aspect, a method of controlling an electronic device includes
`receiving an impact signal (e.g., a vibration generated by a
`user controlled impact with a Surface mechanically coupled
`
`to a housing of the electronic device), identifying a user
`command to the electronic device based on the impact signal,
`and performing a predesignated action based on the user
`command.
`0007. The user command may be identified using a num
`ber of user controlled impacts and an interval between a prior
`user controlled impact and a later user controlled impact. The
`method may include determining the predesignated action
`based on a context of the electronic device. The method may
`also include a user controlled impact generated when a user
`causes a physical object to contact a second physical object
`above a threshold velocity and a threshold force. The first
`physical object may be coupled to a user's hand (e.g., a user's
`finger, a user's arm, a user's elbow, a stylus, a pen, a pencil,
`etc.) and/or foot (e.g., a toe, a heel, a shoe, a knee, etc.). The
`second physical object (e.g., a table top, an automobile dash
`board, a bridge Support structure, a purse, a wallet, a mobile
`device protective cover, etc.) may be mechanically coupled to
`a housing of the electronic device. The second physical object
`may include a housing of the electronic device. The housing
`for the electronic device may include a sealed exterior, and the
`electronic device may be submersible.
`0008. The method may include distinguishing an impact
`signal from an other signal by a frequency of the vibration
`caused by the user controlled impact. The vibration receiver
`may be a microphone. The user command may be identified
`using a frequency, an amplitude, a signal duration, a fre
`quency spectrum, a signal energy, a signal rising edge, a
`signal falling edge, a signal start, a signal end, a mechanically
`transmitted vibration, a duration between impacts, a number
`of impacts, and/or a device context to identify the impact
`signal as an impact command. The user command may be
`identified by any characteristic that differentiates between
`signals.
`0009. The predesignated action may include a mute, a
`power on, a power off, a Volume increase, a Volume decrease,
`a program end, a program beginning, a music track change, a
`call redirect, a call directed to voicemail, an instruction to
`transmit data, an advance to a next program, an instruction to
`receive data, an instruction to execute a mechanical action, an
`image capture, an auditory signal capture, a recording start, a
`recording end, a display change, an indicator light change, a
`Sound generation, a vibration generation, a locking, an
`unlocking, a speakerphone control, an emergency call, and/or
`a car navigation instruction. The predesignated action may
`further include control of an underwater electronic device
`(e.g., a sensor, a camera, etc.) The predesignated action may
`also include any other action controllable by an electronic
`device.
`0010. In another aspect, a system to control an electronic
`device includes a vibration sensing module to receive an
`impact signal including a vibration generated by one user
`controlled impact with a surface mechanically coupled to the
`electronic device, an interpretation module to identify the
`auditory impact signal as a command to the electronic device,
`and a central processing module to execute a predesignated
`action based on the command.
`0011. The user command may be identified using a num
`ber (e.g., a quantity, a total number) of user controlled impacts
`and an interval (e.g., a delay, a period of time) between a prior
`user controlled impact and a later user controlled impact. The
`predesignated action executed by the central processing mod
`ule may be determined based on a context of the electronic
`device when the impact signal may be received. The impact
`
`7
`
`

`

`US 2010/008521.6 A1
`
`Apr. 8, 2010
`
`signal may be generated when a user causes a physical object
`to contact a second physical object above a threshold Velocity.
`The first physical object may be coupled to a user's hand
`and/or foot, and the second physical object may be mechani
`cally coupled to a housing of the electronic device.
`0012. In yet another aspect, a method includes forming a
`vibration sensing module in a electronic device to receive an
`impact signal, in which the impact signal includes a vibration
`generated by a user controlled impact with a Surface mechani
`cally coupled to a vibration receiver of the electronic device.
`The method further includes placing a differentiation module
`in the electronic device to identify the auditory impact signal
`as a command to the electronic device, creating a translation
`module in the electronic device to translate the command into
`a binary number, and forming a central processing module in
`the electronic device to execute a predesignated action based
`on the value of the binary number.
`0013 The user command may be identified using a num
`ber (e.g., a quantity) of user controlled impacts and/or an
`interval between a prior user controlled impact and a later
`user controlled impact. The central processing module may
`determine the predesignated action based on a context of the
`electronic device when the vibration sensing module may
`receive a vibration generated by a user controlled impact. The
`impact signal may be generated when a user causes a physical
`object to contact a second physical object above a threshold
`velocity.
`0014. The methods, systems, and apparatuses disclosed
`herein may be implemented in any means for achieving Vari
`ous aspects, and may be executed in a form of a machine
`readable medium embodying a set of instructions that, when
`executed by a machine, causes the machine to performany of
`the operations disclosed herein. Other features will be appar
`ent from the accompanying drawings and from the detailed
`description that follows.
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`00.15 Example embodiments are illustrated by way of
`example and not limitation in the figures of the accompanying
`drawings, in which like references indicate similar elements
`and in which:
`0016 FIG. 1 is a system view of a electronic device,
`according to one embodiment.
`0017 FIG. 2 is a system view illustrating receiving a
`vibration from a user generated impact and performing an
`action, according to one embodiment.
`0018 FIG. 3 is a diagrammatic system view of a data
`processing system in which any of the embodiments dis
`closed herein may be performed, according to one embodi
`ment.
`0019 FIG. 4 is a process flow illustrating receiving the
`user controlled impact and performing an action, according to
`one embodiment.
`0020 FIG. 5 is a process flow illustrating forming various
`modules for receiving a vibration from a user generated
`impact and performing an action, according to one embodi
`ment.
`0021. Other features of the present embodiments will be
`apparent from the accompanying drawings and from the
`detailed description that follows.
`
`DETAILED DESCRIPTION
`0022. A method, system, and apparatus of a vibration
`based user input for mobile devices are disclosed. Although
`
`the present embodiments have been described with reference
`to specific example embodiments, it will be evident that vari
`ous modifications and changes may be made to these embodi
`ments without departing from the broader spirit and scope of
`the various embodiments.
`0023. In one embodiment, a method of controlling an elec
`tronic device 150 includes receiving an impact signal (e.g., a
`vibration generated by a user controlled impact), identifying
`a user command to the electronic device 150 based on the
`impact signal, and performing a predesignated action based
`on the user command.
`0024. In another embodiment, a system to control an elec
`tronic device 150 includes a vibration sensing module 104 to
`receive an impact signal including a vibration generated by a
`user controlled impact with a Surface mechanically coupled
`to the electronic device 150 (e.g., as illustrated in FIG. 2). The
`system further includes an interpretation module 100 to iden
`tify the auditory impact signal as a command to the electronic
`device 150, and a central processing module 106 to execute a
`predesignated action based on the command.
`0025. In yet another embodiment, a method includes
`forming a vibration sensing module (e.g., the vibration sens
`ing module 104 of FIG. 1) in an electronic device (e.g., the
`electronic device 150, 250, and/or 252) to receive an impact
`signal that includes a vibration generated by a user controlled
`impact with a surface mechanically coupled to a vibration
`receiver of the electronic device. The method further includes
`placing a differentiation module (e.g., the differentiation
`module 102 of FIG. 1) in the electronic device 150 to identify
`the auditory impact signal as a command to the electronic
`device 250, creating a translation module in the electronic
`device (e.g., the electronic device 150, 250, and/or 252) to
`translate the command into a binary number, and forming a
`central processing module 106 in the electronic device to
`execute a predesignated action based on the value of the
`binary number.
`0026 FIG. 1 is a system view of the electronic device (e.g.,
`the electronic device 150, 250, and/or 252), according to one
`embodiment. In particular, FIG. 1 illustrates an interpretation
`module 100 (e.g., a translation module), a differentiation
`module 102, a vibration sensing module 104, a central pro
`cessing module 106, and an electronic device 150, according
`to one embodiment.
`0027. The interpretation module 100 (e.g., a translation
`module) may identify the auditory impact signal (e.g., vibra
`tions, a tap, etc.) as a command to the electronic device. The
`interpretation module 100 may be a hardware unit and/or
`software unit that may convert the vibration patterns into
`binary numbers to provide to the processor (e.g., the central
`processing module 106). The interpretation module 100 may
`also be configured to receive and/or interpret binary numbers
`(e.g. 0s and 1s) coming from the previous stage (e.g., the
`differentiation module 102). The interpretation module 100
`may also be configured to measure the time gap between two
`or more consecutive vibration inputs. The interpretation mod
`ule 100 may also include a timeout logic that may indicate the
`end of a particular vibration pattern (e.g., the output of the
`vibration sensing module 104).
`0028. The differentiation module 102 may be a circuit
`(e.g., a squelch circuit) that may differentiate between an
`impact (e.g., a tap) deliberately generated by a user and a
`background noise and/or a random impact. The differentia
`tion module 102 may isolate strong vibrations and may trans
`mit them to an other module (e.g., the translation module).
`
`8
`
`

`

`US 2010/008521.6 A1
`
`Apr. 8, 2010
`
`The differentiation module 102 may output a “0” which may
`be generated when the differentiation module 102 receives a
`low amplitude vibration (e.g., a vibration below a threshold
`value) and/or an alternate vibration (e.g., a vibration distin
`guishable from a user generated impact). The differentiation
`module 102 may output a “1” which may be generated when
`the differentiation module 102 receives a strong vibration
`(e.g., a vibration above a threshold value) and/or an identifi
`able user generated vibration. The “0” may represent a low
`logic level, and the “1” may represent a high logic level.
`0029. The vibration sensing module 104 may receive an
`impact signal that includes the vibration generated by the user
`controlled impact with a surface mechanically coupled to the
`electronic device 150. The vibration sensing module 104 may
`include a microphone that receives the impact signal. The
`microphone may already be present in a mobile device (e.g.,
`a phone, a pda, a mobile headset, a laptop, etc.) and/or a
`sensor (e.g., a stress sensor, an embedded strain sensor, etc.).
`0030 The central processing module 106 may execute a
`predesignated action based on the command. The predesig
`nated action may be determined using a look up table, which
`may relate a vibration pattern to a predesignated action. The
`predesignated action may be determined based on a context of
`the electronic device 150 when the impact signal may be
`received by the central processing module 106.
`0031. The electronic device 150 may be a communication
`device (e.g., a wireless network card, a cellular phone, etc.)
`that may use network of specialized base stations known as
`cell sites and/or a portable electronic device (e.g., an mp3
`player, a laptop, a CD player, a movie player, a PDA etc.). The
`electronic device 150 may be watertight and may be sub
`merged while receiving an impact signal and performing a
`predesignated action.
`0032. In an example embodiment, the electronic device
`150 may include the vibration sensing module 104, the dif
`ferentiation module 102, the interpretation module 100, and
`the central processing module 106. The vibration sensing
`module 104 may sense the vibrations and may communicate
`the output to the differentiation module 102. The interpreta
`tion module 100 may take and/or receive input signals (e.g.,
`the logic high and/or the logic low signals) from the differ
`entiation module 102. The interpretation module 100 may
`provide the output signals to the central processing module
`106.
`0033. In one embodiment, the impact signal (e.g., a vibra
`tion generated by a user controlled impact) may be received,
`and the signal may be received with a vibration sensing mod
`ule 104. The user may identify (e.g., using an interpretation
`module 100) the command to the electronic device 150 based
`on the impact signal. The predesignated action may be per
`formed (e.g., using a central processing module 106) based on
`the user command. The user command may be identified
`using a number and/or a quantity of user controlled impacts.
`The user command may be identified using an interval
`between a prior user controlled impact and a later user con
`trolled impact.
`0034. The predesignated action may be determined (e.g.,
`using the central processing module 106) based on a context
`of the electronic device 150. The user controlled impact may
`occur whena user causes a physical object to contact a second
`physical object above a threshold velocity and a threshold
`force.
`0035. A system may control an electronic device (e.g., the
`electronic device 150,250, and/or 252). The electronic device
`
`may include a vibration sensing module 104. The vibration
`sensing module 104 may receive an impact signal, and the
`impact signal may include a vibration generated by one or
`more user controlled impacts with a Surface mechanically
`coupled to the electronic device.
`0036. The interpretation module 100 (e.g., a translation
`module) may identify the impact signal (e.g., an auditory
`signal, a vibration transferred through a mechanical coupling,
`etc.) as a command to the electronic device 150. The central
`processing module 106 may execute a predesignated action
`based on the command. The user command may be identified
`using a number of user controlled impacts. The user com
`mand may be identified by an interval between a prior user
`controlled impact and a later user controlled impact. The first
`physical object may be coupled to a user's hand and/or foot,
`and the second physical object may be mechanically coupled
`to a housing of the electronic device 150.
`0037 FIG. 2 is a system view illustrating receiving a
`vibration from a user generated impact and performing an
`action, according to one embodiment. In particular, FIG. 2
`illustrates an electronic device 250, and an electronic device
`252, according to one embodiment.
`0038. The electronic device 250 may be a communication
`device (e.g., a wireless, cellular phone, etc.) and/or a portable
`electronic device (e.g., a pda, a laptop, an mp3 player, a
`multimedia player, a cd player, etc.). The electronic device
`250 may perform a predesignated action after receiving the
`command, including a power on, a power off, a Volume
`increase, a Volume decrease, a program end, a program begin
`ning, a music track change, a call redirect, a call directed to
`Voicemail, an instruction to transmit data, an advance to a next
`program, an instruction to receive data, an instruction to
`execute a mechanical action, and/or an image capture. The
`predesignated action may include an auditory signal capture,
`a recording start, a recording end, a display change, an indi
`cator light change, a Sound generation, a vibration generation,
`a locking, an unlocking, a speakerphone control, an emer
`gency call, and/or a car navigation instruction, etc. The pre
`designated action may relate to a watertight electronic device
`(e.g., a waterproof mp3 player, an underwater camera, etc.)
`The user command may be identified using a number of user
`controlled impacts and/or an interval between a prior user
`controlled impact and a later user controlled impact.
`0039. The electronic device 252 may be a mobile device
`(e.g., a wireless phone, a cellular connection) and/or a por
`table electronic device. The electronic device 252 may per
`form a predesignated action when a command is received.
`0040. In an example embodiment, the electronic device
`250 may receive a vibration from the impact of the finger of
`the user, and the electronic device 250 may perform the pre
`designated action after the vibration has been received. The
`impact generated by the user's finger may create vibrations,
`which may be received by the electronic device 250. The
`electronic device (e.g., the electronic device 150, 250, and/or
`252) may perform the predesignated action after the vibra
`tions are received.
`0041. In one embodiment, the first physical object may be
`coupled to a user's hand (e.g., a user's finger, a user's arm, a
`user's elbow, a stylus, a pen, a pencil, etc.) and/or foot (e.g., a
`toe, a heel, a shoe, a knee, etc.). The second physical object
`(e.g., a table top, an automobile dashboard, a bridge Support
`structure, a purse, a wallet, a mobile device protective cover,
`etc.) may be mechanically coupled to a housing of the elec
`tronic device 250. The first physical object may be a user's
`
`9
`
`

`

`US 2010/008521.6 A1
`
`Apr. 8, 2010
`
`finger and/or toe. The second physical object may include a
`housing of the electronic device 250. The impact signal may
`be distinguished from an other signal by a frequency of the
`vibration caused by the user controlled impact.
`0042. The vibration receiver may be a microphone, and
`the user command may be identified using a frequency, an
`amplitude, a signal duration, a frequency spectrum, a signal
`energy, a signal rising edge, and a signal falling edge. The user
`command may further be identified by a signal start, a signal
`end, a mechanically transmitted vibration, a duration between
`impacts, a number of impacts, and a device context (e.g., a
`power condition, an incoming call, a Volume on, etc.).
`0043. The predesignated action may include a mute, a
`power on, a power off, a Volume increase, a Volume decrease,
`a program end, a program beginning, a music track change, a
`call redirect, a call directed to voicemail, an instruction to
`transmit data, an advance to a next program, an instruction to
`receive data, an instruction to execute a mechanical action, an
`image capture, an auditory signal capture, a recording start, a
`recording end, a display change, an indicator light change, a
`Sound generation, a vibration generation, a locking, an
`unlocking, a speakerphone control, an emergency call, and/or
`a car navigation instruction.
`0044) The vibration sensing module 104 may receive an
`impact signal that includes a vibration generated by a user
`controlled impact. The impact may involve a Surface
`mechanically coupled to a vibration receiver of the electronic
`device (e.g., the electronic device 150,250, 252). The differ
`entiation module 102 may identify the auditory impact signal
`as a command to the electronic device 250 of FIG. 2. The
`translation module may be created in the electronic device
`250 that may translate the command into a binary number.
`The central processing module (e.g., the central processing
`module 106 of FIG.1) may beformed in the electronic device
`252 of FIG. 2. The central processing module 106 may
`execute a predesignated action based on the value of the
`binary number.
`0045. The user command may be identified using a num
`ber of user controlled impacts and/or an interval between a
`prior user controlled impact and a later user controlled
`impact. The central processing module 106 may determine
`the predesignated action based on a context of the electronic
`device (e.g., the electronic device 150,250, and/or 252) when
`the vibration sensing module 104 receives a vibration gener
`ated by a user controlled impact. The impact signal may be
`generated when a user causes a physical object to contact a
`second physical object above a threshold velocity and/or
`force. The predesignated action may be executed by the cen
`tral processing module 106, and the predesignated action may
`be determined based on a context of the electronic device 252
`at the time the impact signal is received.
`0046 FIG. 3 is a diagrammatic system view of a data
`processing system in which any of the embodiments dis
`closed herein may be performed, according to one embodi
`ment. In particular, the diagrammatic system view 300 of
`FIG. 3 illustrates a processor 302, a main memory 304, a
`static memory 306, a bus 308, a video display 310, an alpha
`numeric input device 312, a cursor control device 314, a drive
`unit 313, a signal generation device 318, a network interface
`device 320, a machine readable medium 322, instructions
`324, and a network 326, according to one embodiment.
`0047. The diagrammatic system view 300 may indicate a
`personal computer and/or the data processing system in
`which one or more operations disclosed herein are performed.
`
`The processor 302 may be a microprocessor, a state machine,
`an application specific integrated circuit, a field program
`mable gate array, etc. (e.g., an Intel(R) Pentium(R) processor).
`The main memory 304 may be a dynamic random access
`memory and/or a primary memory of a computer system.
`0048. The static memory 306 may be a hard drive, a flash
`drive, and/or other memory information associated with the
`data processing system. The bus 308 may be an interconnec
`tion between various circuits and/or structures of the data
`processing system. The video display 310 may provide a
`graphical representation of information on the data process
`ing system. The alpha-numeric input device 312 may be a
`keypad, a keyboard and/or any other input device of text (e.g.,
`a special device to aid the physically handicapped).
`0049. The cursor control device 314 may be a pointing
`device such as a mouse. The drive unit 316 may be the hard
`drive, a storage system, and/or other longer term storage
`subsystem. The signal generation device 318 may be a bios
`and/or a functional operating system of the data processing
`system. The network interface device 320 may be a device
`that performs interface functions such as code conversion,
`protocol conversion and/or buffering required for communi
`cation to and from the network 326. The machine readable
`medium 322 may provide instructions on which any of the
`methods disclosed herein may be performed. The instructions
`324 may provide Source code and/or data code to the proces
`sor 302 to enable any one or more operations disclosed
`herein.
`0050 FIG. 4 is a process flow illustrating receiving the
`user controlled impact and performing an action, according to
`one embodiment. In operation 402, an impact signal (e.g., a
`vibration generated by one user controlled impact with a
`Surface mechanically coupled to a housing of the electronic
`device 150) may be received. The impact signal may be
`received using the vibration sensing module 104 of FIG. 1.
`0051. In operation 404, an impact signal may be distin
`guished from an other signal by a frequency of the vibration
`caused by the user controlled impact. The vibration receiver
`may be a microphone.
`0052. In operation 406, a user may identify the command
`to the electronic device 150 based on the impact signal. The
`user command may be identified using a frequency, an ampli
`tude, a signal duration, a frequency spectrum, a signal energy,
`a signal rising edge, a signal falling edge, a signal start, a
`signal end, a mechanically transmitted vibration, a duration
`between impacts, and/or a number of impacts and/or a device
`context to identify the impact signal as an impact command.
`0053. In operation 408, the predesignated action may be
`determined based on a context (e.g., an electronic device
`operation, a power status, a ringer operation, an open com
`munication with another person using another electronic
`device, etc.) of the electronic device 150. The user controlled
`impact may occur when a user may cause a physical object to
`contact a second physical object above a threshold Velocity
`and/or a threshold force.
`0054 The first physical object may be coupled to a user's
`hand (e.g., a user's finger, a user's arm, a user's elbow, a
`stylus, a pen, a pencil, etc.) and/or foot (e.g., a toe, a heel, a
`shoe, a knee, etc.). The first physical object may be a user's
`finger and/or toe. The second physical object (e.g., a table top,
`an automobile dashboard, a bridge Support structure, a purse,
`a wallet, a mobile device protective cover, etc.) may be
`mechanically coupled to and/or include a housing of the
`electronic device 150.
`
`10
`
`

`

`US 2010/008521.6 A1
`
`Apr. 8, 2010
`
`0055. In operation 410, a predesignated action may be
`performed based on the user command. The u

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket