`(12) Patent Application Publication (10) Pub. No.: US 2010/0085216 A1
`(43) Pub. Date:
`Apr. 8, 2010
`MS
`
`US 2010.0085216A1
`
`(54) VIBRATION BASED USER INPUT FOR
`MOBILE DEVICES
`
`AJITH MS, Bangalore (IN)
`(76) Inventor:
`Correspondence Address:
`Raj Abhyanker LLP
`1580 W. El Camino Real, Suite 8
`Mountain View, CA 94040 (US)
`
`(21) Appl. No.:
`
`12/243,976
`
`(22) Filed:
`
`Oct. 2, 2008
`
`Publication Classification
`
`(51) Int. Cl.
`HO3K 7/94
`
`(2006.01)
`
`(52) U.S. Cl. .......................................................... 341?20
`
`ABSTRACT
`(57)
`A method, system, and apparatus of a vibration based user
`input for mobile devices are disclosed. In one embodiment, a
`method of controlling an electronic device includes receiving
`an impact signal, (e.g., a vibration generated by a user con
`trolled impact with a Surface mechanically coupled to a hous
`ing of the electronic device). The method further includes
`identifying a user command to the electronic device based on
`the impact signal, and performing a predesignated action
`(e.g., a mute, a power on, a power off, a Volume increase, a
`Volume decrease, a music track change, a call redirect, a call
`directed to Voicemail, etc.) based on the user command. The
`user command may be identified using a number of user
`controlled impacts and an interval between a prior user con
`trolled impact and a later user controlled impact.
`
`ELECTRONIC
`DEVICE 250
`
`ill,
`
`PRIOR TO USER GENERATED IMPACT
`
`ELECTRONIC
`DEVICE 250
`
`Hill,
`
`USER GENERATED IMPACT CREATES VIBRATIONS
`
`ELECTRONIC
`DEVICE 250
`
`O >
`
`fe
`
`ELECTRONICELECTRONIC
`
`DEVICE 250 DEVICE 252
`
`WBRATIONS ARE RECEIVED BY ELECTRONIC DEVICE
`
`ELECTRONIC DEVICE PERFORMS ANACTION
`
`APPLE 1052
`
`1
`
`
`
`Patent Application Publication
`
`Apr. 8, 2010 Sheet 1 of 5
`
`US 2010/008521.6 A1
`
`
`
`
`
`5ÕT ETT GOW
`
`
`
`SDNISSE OO}}c}
`
`OHNO}}_LOETE`N
`
`2
`
`
`
`Patent Application Publication
`
`Apr. 8, 2010 Sheet 2 of 5
`
`US 2010/008521.6 A1
`
`3
`
`
`
`Patent Application Publication
`
`Apr. 8, 2010 Sheet 3 of 5
`
`US 2010/008521.6 A1
`
`350
`
`308
`
`PROCESSOR 30
`
`INSTRUCTIONS 324
`
`MAIN MEMORY 304
`
`INSTRUCTIONS 324
`
`
`
`STATIC MEMORY 306
`
`INSTRUCTIONS 324
`
`NETWORK iNTERFACE
`DEVICE 320
`
`VIDEO DISPLAY 31
`
`ALPHA-NUMERIC
`INPUT DEVICE 312
`
`CURSOR CONTROL
`DEVICE 314
`
`DRIVE UNIT 316
`
`MACHINE READABLE
`MEDIUM 322
`
`INSTRUCTIONS 324
`
`SIGNAL GENERATION DEVICE 318
`
`FIGURE 3
`
`4
`
`
`
`Patent Application Publication
`
`Apr. 8, 2010 Sheet 4 of 5
`
`US 2010/008521.6 A1
`
`RECEIVE AN IMPACT SIGNAL
`
`402
`
`404
`
`DISTINGUISH AN IMPACT SIGNAL FROMAN OTHER SIGNAL BY A FREOUENCY OF
`THE VIBRATION CAUSED BY THE USER CONTROLLED IMPACT
`
`DETERMINE THE PREDESIGNATED ACTION BASED ON A CONTEXT OF THE
`ELECTRONIC DEVICE
`
`408
`
`DETERMINE THE PREDESIGNATED ACTION BASED ON A CONTEXT OF THE
`ELECTRONIC DEVICE
`
`410
`
`PERFORMAPREDESIGNATED ACTION BASED ON THE USER COMMAND
`
`FIGURE 4
`
`5
`
`
`
`Patent Application Publication
`
`Apr. 8, 2010 Sheet 5 of 5
`
`US 2010/008521.6 A1
`
`502
`
`
`
`
`
`FORMA VIBRATION SENSING MODULE IN AELECTRONIC DEVICE TO RECEIVE AN
`IMPACT SIGNAL COMPRISED OF A WIBRATION GENERATED BY A USER
`CONTROLLED IMPACT WITH A SURFACE MECHANICALLY COUPLED TO A
`VIBRATION RECEIVER OF THE ELECTRONIC DEVICE
`
`
`
`PLACEA DIFFERENTIATIONMODULE IN THE ELECTRONIC DEVICE TO IDENTIFY
`THE AUDITORY IMPACT SIGNAL ASA COMMAND TO THE ELECTRONIC DEVICE
`
`504
`
`CREATE ATRANSLATION MODULE IN THE ELECTRONIC DEVICE TO TRANSLATE
`THE COMMAND INTO A BINARY NUMBER
`
`508
`
`FORMA CENTRAL PROCESSING MODULE IN THE ELECTRONIC DEVICE TO
`EXECUTE A PREDESIGNATED ACTION BASED ON THE WALUE OF THE BINARY
`NUMBER
`
`506
`
`FIGURE 5
`
`6
`
`
`
`US 2010/008521.6 A1
`
`Apr. 8, 2010
`
`VIBRATION BASED USER INPUT FOR
`MOBILE DEVICES
`
`FIELD OF TECHNOLOGY
`0001. This disclosure relates generally to an enterprise
`method, a technical field of software and/or hardware tech
`nology and, in one example embodiment, to a vibration based
`user input for mobile devices.
`
`BACKGROUND
`0002. A control for an electronic device (e.g., a mobile
`device, a laptop, a car navigation system, a stress sensor, a car
`audio, an underwater sensor, etc.) may require a particular
`user input (e.g., a touch, a movement, a sound, etc.) to control
`a function (e.g., a mute, a Voicemail, a Volume, a track control,
`a call disconnect, a speed, a door control, etc.) of an electronic
`device.
`0003. The particular user input may involve a direct con
`tact with a limited area surface (e.g., a button, a touchpad,
`etc.). The limited area Surface may require user to view the
`limited area surface to interact with it. In addition, the limited
`area Surface may also require a user to extract the electronic
`device from a container (e.g., a pocket, a wallet, a purse, a
`protective case, a cover, etc.). These requirements may cause
`a delay (e.g., a time to extract the device from the container, a
`time to locate the limited area surface, etc.) and/or may cause
`a user to decide not to access the device. The delay or the
`failure to access the device may result in a disturbance (e.g.,
`a phone ring) and/or an undesired result (e.g., a missed call, a
`transfer to Voicemail, a different music track, a distraction, a
`loss of vehicle control, a missed direction, etc.). The undes
`ired result may further result in an additional delay and/or
`inefficiency (e.g., an access to Voicemail, a u-turn to find a
`missed location, a repeat of an already heard audio segment,
`etc.)
`0004. The limited area surface may also contain a moving
`part (e.g., a button mechanism) and/oran unsealed edge (e.g.,
`a crack next to a touchpad edge and/or a button edge). The
`moving part may allow a foreign Substance (e.g., a fluid
`and/or a dirt) to enter a housing for the electronic device. The
`foreign Substance may accelerate a breakdown of the moving
`part and/or a component of the electronic device (e.g., a
`processor, a memory, a hard drive, a speaker, a microphone,
`etc.). The breakdown of the moving part and/or the compo
`nent of the electronic device may prevent the electronic
`device from operating, which may result in a further delay
`and/or inefficiency (e.g., a repair cost, a replacement cost,
`etc.).
`0005. In addition, the particular user input may involve a
`Sound greater than a threshold Volume. By exceeding the
`threshold Volume, the sound may cause a disturbance. The
`audible sound may also be easily generated by environmental
`Sources, which may cause the electronic device to be con
`trolled in undesired ways (e.g., a call disconnect, a mute, a
`Video end, a power on, a power off, etc.). These undesired
`effects may also result in a delay and/or inefficiency.
`
`SUMMARY
`0006. A method, system, and apparatus of a vibration
`based user input for mobile devices are disclosed. In one
`aspect, a method of controlling an electronic device includes
`receiving an impact signal (e.g., a vibration generated by a
`user controlled impact with a Surface mechanically coupled
`
`to a housing of the electronic device), identifying a user
`command to the electronic device based on the impact signal,
`and performing a predesignated action based on the user
`command.
`0007. The user command may be identified using a num
`ber of user controlled impacts and an interval between a prior
`user controlled impact and a later user controlled impact. The
`method may include determining the predesignated action
`based on a context of the electronic device. The method may
`also include a user controlled impact generated when a user
`causes a physical object to contact a second physical object
`above a threshold velocity and a threshold force. The first
`physical object may be coupled to a user's hand (e.g., a user's
`finger, a user's arm, a user's elbow, a stylus, a pen, a pencil,
`etc.) and/or foot (e.g., a toe, a heel, a shoe, a knee, etc.). The
`second physical object (e.g., a table top, an automobile dash
`board, a bridge Support structure, a purse, a wallet, a mobile
`device protective cover, etc.) may be mechanically coupled to
`a housing of the electronic device. The second physical object
`may include a housing of the electronic device. The housing
`for the electronic device may include a sealed exterior, and the
`electronic device may be submersible.
`0008. The method may include distinguishing an impact
`signal from an other signal by a frequency of the vibration
`caused by the user controlled impact. The vibration receiver
`may be a microphone. The user command may be identified
`using a frequency, an amplitude, a signal duration, a fre
`quency spectrum, a signal energy, a signal rising edge, a
`signal falling edge, a signal start, a signal end, a mechanically
`transmitted vibration, a duration between impacts, a number
`of impacts, and/or a device context to identify the impact
`signal as an impact command. The user command may be
`identified by any characteristic that differentiates between
`signals.
`0009. The predesignated action may include a mute, a
`power on, a power off, a Volume increase, a Volume decrease,
`a program end, a program beginning, a music track change, a
`call redirect, a call directed to voicemail, an instruction to
`transmit data, an advance to a next program, an instruction to
`receive data, an instruction to execute a mechanical action, an
`image capture, an auditory signal capture, a recording start, a
`recording end, a display change, an indicator light change, a
`Sound generation, a vibration generation, a locking, an
`unlocking, a speakerphone control, an emergency call, and/or
`a car navigation instruction. The predesignated action may
`further include control of an underwater electronic device
`(e.g., a sensor, a camera, etc.) The predesignated action may
`also include any other action controllable by an electronic
`device.
`0010. In another aspect, a system to control an electronic
`device includes a vibration sensing module to receive an
`impact signal including a vibration generated by one user
`controlled impact with a surface mechanically coupled to the
`electronic device, an interpretation module to identify the
`auditory impact signal as a command to the electronic device,
`and a central processing module to execute a predesignated
`action based on the command.
`0011. The user command may be identified using a num
`ber (e.g., a quantity, a total number) of user controlled impacts
`and an interval (e.g., a delay, a period of time) between a prior
`user controlled impact and a later user controlled impact. The
`predesignated action executed by the central processing mod
`ule may be determined based on a context of the electronic
`device when the impact signal may be received. The impact
`
`7
`
`
`
`US 2010/008521.6 A1
`
`Apr. 8, 2010
`
`signal may be generated when a user causes a physical object
`to contact a second physical object above a threshold Velocity.
`The first physical object may be coupled to a user's hand
`and/or foot, and the second physical object may be mechani
`cally coupled to a housing of the electronic device.
`0012. In yet another aspect, a method includes forming a
`vibration sensing module in a electronic device to receive an
`impact signal, in which the impact signal includes a vibration
`generated by a user controlled impact with a Surface mechani
`cally coupled to a vibration receiver of the electronic device.
`The method further includes placing a differentiation module
`in the electronic device to identify the auditory impact signal
`as a command to the electronic device, creating a translation
`module in the electronic device to translate the command into
`a binary number, and forming a central processing module in
`the electronic device to execute a predesignated action based
`on the value of the binary number.
`0013 The user command may be identified using a num
`ber (e.g., a quantity) of user controlled impacts and/or an
`interval between a prior user controlled impact and a later
`user controlled impact. The central processing module may
`determine the predesignated action based on a context of the
`electronic device when the vibration sensing module may
`receive a vibration generated by a user controlled impact. The
`impact signal may be generated when a user causes a physical
`object to contact a second physical object above a threshold
`velocity.
`0014. The methods, systems, and apparatuses disclosed
`herein may be implemented in any means for achieving Vari
`ous aspects, and may be executed in a form of a machine
`readable medium embodying a set of instructions that, when
`executed by a machine, causes the machine to performany of
`the operations disclosed herein. Other features will be appar
`ent from the accompanying drawings and from the detailed
`description that follows.
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`00.15 Example embodiments are illustrated by way of
`example and not limitation in the figures of the accompanying
`drawings, in which like references indicate similar elements
`and in which:
`0016 FIG. 1 is a system view of a electronic device,
`according to one embodiment.
`0017 FIG. 2 is a system view illustrating receiving a
`vibration from a user generated impact and performing an
`action, according to one embodiment.
`0018 FIG. 3 is a diagrammatic system view of a data
`processing system in which any of the embodiments dis
`closed herein may be performed, according to one embodi
`ment.
`0019 FIG. 4 is a process flow illustrating receiving the
`user controlled impact and performing an action, according to
`one embodiment.
`0020 FIG. 5 is a process flow illustrating forming various
`modules for receiving a vibration from a user generated
`impact and performing an action, according to one embodi
`ment.
`0021. Other features of the present embodiments will be
`apparent from the accompanying drawings and from the
`detailed description that follows.
`
`DETAILED DESCRIPTION
`0022. A method, system, and apparatus of a vibration
`based user input for mobile devices are disclosed. Although
`
`the present embodiments have been described with reference
`to specific example embodiments, it will be evident that vari
`ous modifications and changes may be made to these embodi
`ments without departing from the broader spirit and scope of
`the various embodiments.
`0023. In one embodiment, a method of controlling an elec
`tronic device 150 includes receiving an impact signal (e.g., a
`vibration generated by a user controlled impact), identifying
`a user command to the electronic device 150 based on the
`impact signal, and performing a predesignated action based
`on the user command.
`0024. In another embodiment, a system to control an elec
`tronic device 150 includes a vibration sensing module 104 to
`receive an impact signal including a vibration generated by a
`user controlled impact with a Surface mechanically coupled
`to the electronic device 150 (e.g., as illustrated in FIG. 2). The
`system further includes an interpretation module 100 to iden
`tify the auditory impact signal as a command to the electronic
`device 150, and a central processing module 106 to execute a
`predesignated action based on the command.
`0025. In yet another embodiment, a method includes
`forming a vibration sensing module (e.g., the vibration sens
`ing module 104 of FIG. 1) in an electronic device (e.g., the
`electronic device 150, 250, and/or 252) to receive an impact
`signal that includes a vibration generated by a user controlled
`impact with a surface mechanically coupled to a vibration
`receiver of the electronic device. The method further includes
`placing a differentiation module (e.g., the differentiation
`module 102 of FIG. 1) in the electronic device 150 to identify
`the auditory impact signal as a command to the electronic
`device 250, creating a translation module in the electronic
`device (e.g., the electronic device 150, 250, and/or 252) to
`translate the command into a binary number, and forming a
`central processing module 106 in the electronic device to
`execute a predesignated action based on the value of the
`binary number.
`0026 FIG. 1 is a system view of the electronic device (e.g.,
`the electronic device 150, 250, and/or 252), according to one
`embodiment. In particular, FIG. 1 illustrates an interpretation
`module 100 (e.g., a translation module), a differentiation
`module 102, a vibration sensing module 104, a central pro
`cessing module 106, and an electronic device 150, according
`to one embodiment.
`0027. The interpretation module 100 (e.g., a translation
`module) may identify the auditory impact signal (e.g., vibra
`tions, a tap, etc.) as a command to the electronic device. The
`interpretation module 100 may be a hardware unit and/or
`software unit that may convert the vibration patterns into
`binary numbers to provide to the processor (e.g., the central
`processing module 106). The interpretation module 100 may
`also be configured to receive and/or interpret binary numbers
`(e.g. 0s and 1s) coming from the previous stage (e.g., the
`differentiation module 102). The interpretation module 100
`may also be configured to measure the time gap between two
`or more consecutive vibration inputs. The interpretation mod
`ule 100 may also include a timeout logic that may indicate the
`end of a particular vibration pattern (e.g., the output of the
`vibration sensing module 104).
`0028. The differentiation module 102 may be a circuit
`(e.g., a squelch circuit) that may differentiate between an
`impact (e.g., a tap) deliberately generated by a user and a
`background noise and/or a random impact. The differentia
`tion module 102 may isolate strong vibrations and may trans
`mit them to an other module (e.g., the translation module).
`
`8
`
`
`
`US 2010/008521.6 A1
`
`Apr. 8, 2010
`
`The differentiation module 102 may output a “0” which may
`be generated when the differentiation module 102 receives a
`low amplitude vibration (e.g., a vibration below a threshold
`value) and/or an alternate vibration (e.g., a vibration distin
`guishable from a user generated impact). The differentiation
`module 102 may output a “1” which may be generated when
`the differentiation module 102 receives a strong vibration
`(e.g., a vibration above a threshold value) and/or an identifi
`able user generated vibration. The “0” may represent a low
`logic level, and the “1” may represent a high logic level.
`0029. The vibration sensing module 104 may receive an
`impact signal that includes the vibration generated by the user
`controlled impact with a surface mechanically coupled to the
`electronic device 150. The vibration sensing module 104 may
`include a microphone that receives the impact signal. The
`microphone may already be present in a mobile device (e.g.,
`a phone, a pda, a mobile headset, a laptop, etc.) and/or a
`sensor (e.g., a stress sensor, an embedded strain sensor, etc.).
`0030 The central processing module 106 may execute a
`predesignated action based on the command. The predesig
`nated action may be determined using a look up table, which
`may relate a vibration pattern to a predesignated action. The
`predesignated action may be determined based on a context of
`the electronic device 150 when the impact signal may be
`received by the central processing module 106.
`0031. The electronic device 150 may be a communication
`device (e.g., a wireless network card, a cellular phone, etc.)
`that may use network of specialized base stations known as
`cell sites and/or a portable electronic device (e.g., an mp3
`player, a laptop, a CD player, a movie player, a PDA etc.). The
`electronic device 150 may be watertight and may be sub
`merged while receiving an impact signal and performing a
`predesignated action.
`0032. In an example embodiment, the electronic device
`150 may include the vibration sensing module 104, the dif
`ferentiation module 102, the interpretation module 100, and
`the central processing module 106. The vibration sensing
`module 104 may sense the vibrations and may communicate
`the output to the differentiation module 102. The interpreta
`tion module 100 may take and/or receive input signals (e.g.,
`the logic high and/or the logic low signals) from the differ
`entiation module 102. The interpretation module 100 may
`provide the output signals to the central processing module
`106.
`0033. In one embodiment, the impact signal (e.g., a vibra
`tion generated by a user controlled impact) may be received,
`and the signal may be received with a vibration sensing mod
`ule 104. The user may identify (e.g., using an interpretation
`module 100) the command to the electronic device 150 based
`on the impact signal. The predesignated action may be per
`formed (e.g., using a central processing module 106) based on
`the user command. The user command may be identified
`using a number and/or a quantity of user controlled impacts.
`The user command may be identified using an interval
`between a prior user controlled impact and a later user con
`trolled impact.
`0034. The predesignated action may be determined (e.g.,
`using the central processing module 106) based on a context
`of the electronic device 150. The user controlled impact may
`occur whena user causes a physical object to contact a second
`physical object above a threshold velocity and a threshold
`force.
`0035. A system may control an electronic device (e.g., the
`electronic device 150,250, and/or 252). The electronic device
`
`may include a vibration sensing module 104. The vibration
`sensing module 104 may receive an impact signal, and the
`impact signal may include a vibration generated by one or
`more user controlled impacts with a Surface mechanically
`coupled to the electronic device.
`0036. The interpretation module 100 (e.g., a translation
`module) may identify the impact signal (e.g., an auditory
`signal, a vibration transferred through a mechanical coupling,
`etc.) as a command to the electronic device 150. The central
`processing module 106 may execute a predesignated action
`based on the command. The user command may be identified
`using a number of user controlled impacts. The user com
`mand may be identified by an interval between a prior user
`controlled impact and a later user controlled impact. The first
`physical object may be coupled to a user's hand and/or foot,
`and the second physical object may be mechanically coupled
`to a housing of the electronic device 150.
`0037 FIG. 2 is a system view illustrating receiving a
`vibration from a user generated impact and performing an
`action, according to one embodiment. In particular, FIG. 2
`illustrates an electronic device 250, and an electronic device
`252, according to one embodiment.
`0038. The electronic device 250 may be a communication
`device (e.g., a wireless, cellular phone, etc.) and/or a portable
`electronic device (e.g., a pda, a laptop, an mp3 player, a
`multimedia player, a cd player, etc.). The electronic device
`250 may perform a predesignated action after receiving the
`command, including a power on, a power off, a Volume
`increase, a Volume decrease, a program end, a program begin
`ning, a music track change, a call redirect, a call directed to
`Voicemail, an instruction to transmit data, an advance to a next
`program, an instruction to receive data, an instruction to
`execute a mechanical action, and/or an image capture. The
`predesignated action may include an auditory signal capture,
`a recording start, a recording end, a display change, an indi
`cator light change, a Sound generation, a vibration generation,
`a locking, an unlocking, a speakerphone control, an emer
`gency call, and/or a car navigation instruction, etc. The pre
`designated action may relate to a watertight electronic device
`(e.g., a waterproof mp3 player, an underwater camera, etc.)
`The user command may be identified using a number of user
`controlled impacts and/or an interval between a prior user
`controlled impact and a later user controlled impact.
`0039. The electronic device 252 may be a mobile device
`(e.g., a wireless phone, a cellular connection) and/or a por
`table electronic device. The electronic device 252 may per
`form a predesignated action when a command is received.
`0040. In an example embodiment, the electronic device
`250 may receive a vibration from the impact of the finger of
`the user, and the electronic device 250 may perform the pre
`designated action after the vibration has been received. The
`impact generated by the user's finger may create vibrations,
`which may be received by the electronic device 250. The
`electronic device (e.g., the electronic device 150, 250, and/or
`252) may perform the predesignated action after the vibra
`tions are received.
`0041. In one embodiment, the first physical object may be
`coupled to a user's hand (e.g., a user's finger, a user's arm, a
`user's elbow, a stylus, a pen, a pencil, etc.) and/or foot (e.g., a
`toe, a heel, a shoe, a knee, etc.). The second physical object
`(e.g., a table top, an automobile dashboard, a bridge Support
`structure, a purse, a wallet, a mobile device protective cover,
`etc.) may be mechanically coupled to a housing of the elec
`tronic device 250. The first physical object may be a user's
`
`9
`
`
`
`US 2010/008521.6 A1
`
`Apr. 8, 2010
`
`finger and/or toe. The second physical object may include a
`housing of the electronic device 250. The impact signal may
`be distinguished from an other signal by a frequency of the
`vibration caused by the user controlled impact.
`0042. The vibration receiver may be a microphone, and
`the user command may be identified using a frequency, an
`amplitude, a signal duration, a frequency spectrum, a signal
`energy, a signal rising edge, and a signal falling edge. The user
`command may further be identified by a signal start, a signal
`end, a mechanically transmitted vibration, a duration between
`impacts, a number of impacts, and a device context (e.g., a
`power condition, an incoming call, a Volume on, etc.).
`0043. The predesignated action may include a mute, a
`power on, a power off, a Volume increase, a Volume decrease,
`a program end, a program beginning, a music track change, a
`call redirect, a call directed to voicemail, an instruction to
`transmit data, an advance to a next program, an instruction to
`receive data, an instruction to execute a mechanical action, an
`image capture, an auditory signal capture, a recording start, a
`recording end, a display change, an indicator light change, a
`Sound generation, a vibration generation, a locking, an
`unlocking, a speakerphone control, an emergency call, and/or
`a car navigation instruction.
`0044) The vibration sensing module 104 may receive an
`impact signal that includes a vibration generated by a user
`controlled impact. The impact may involve a Surface
`mechanically coupled to a vibration receiver of the electronic
`device (e.g., the electronic device 150,250, 252). The differ
`entiation module 102 may identify the auditory impact signal
`as a command to the electronic device 250 of FIG. 2. The
`translation module may be created in the electronic device
`250 that may translate the command into a binary number.
`The central processing module (e.g., the central processing
`module 106 of FIG.1) may beformed in the electronic device
`252 of FIG. 2. The central processing module 106 may
`execute a predesignated action based on the value of the
`binary number.
`0045. The user command may be identified using a num
`ber of user controlled impacts and/or an interval between a
`prior user controlled impact and a later user controlled
`impact. The central processing module 106 may determine
`the predesignated action based on a context of the electronic
`device (e.g., the electronic device 150,250, and/or 252) when
`the vibration sensing module 104 receives a vibration gener
`ated by a user controlled impact. The impact signal may be
`generated when a user causes a physical object to contact a
`second physical object above a threshold velocity and/or
`force. The predesignated action may be executed by the cen
`tral processing module 106, and the predesignated action may
`be determined based on a context of the electronic device 252
`at the time the impact signal is received.
`0046 FIG. 3 is a diagrammatic system view of a data
`processing system in which any of the embodiments dis
`closed herein may be performed, according to one embodi
`ment. In particular, the diagrammatic system view 300 of
`FIG. 3 illustrates a processor 302, a main memory 304, a
`static memory 306, a bus 308, a video display 310, an alpha
`numeric input device 312, a cursor control device 314, a drive
`unit 313, a signal generation device 318, a network interface
`device 320, a machine readable medium 322, instructions
`324, and a network 326, according to one embodiment.
`0047. The diagrammatic system view 300 may indicate a
`personal computer and/or the data processing system in
`which one or more operations disclosed herein are performed.
`
`The processor 302 may be a microprocessor, a state machine,
`an application specific integrated circuit, a field program
`mable gate array, etc. (e.g., an Intel(R) Pentium(R) processor).
`The main memory 304 may be a dynamic random access
`memory and/or a primary memory of a computer system.
`0048. The static memory 306 may be a hard drive, a flash
`drive, and/or other memory information associated with the
`data processing system. The bus 308 may be an interconnec
`tion between various circuits and/or structures of the data
`processing system. The video display 310 may provide a
`graphical representation of information on the data process
`ing system. The alpha-numeric input device 312 may be a
`keypad, a keyboard and/or any other input device of text (e.g.,
`a special device to aid the physically handicapped).
`0049. The cursor control device 314 may be a pointing
`device such as a mouse. The drive unit 316 may be the hard
`drive, a storage system, and/or other longer term storage
`subsystem. The signal generation device 318 may be a bios
`and/or a functional operating system of the data processing
`system. The network interface device 320 may be a device
`that performs interface functions such as code conversion,
`protocol conversion and/or buffering required for communi
`cation to and from the network 326. The machine readable
`medium 322 may provide instructions on which any of the
`methods disclosed herein may be performed. The instructions
`324 may provide Source code and/or data code to the proces
`sor 302 to enable any one or more operations disclosed
`herein.
`0050 FIG. 4 is a process flow illustrating receiving the
`user controlled impact and performing an action, according to
`one embodiment. In operation 402, an impact signal (e.g., a
`vibration generated by one user controlled impact with a
`Surface mechanically coupled to a housing of the electronic
`device 150) may be received. The impact signal may be
`received using the vibration sensing module 104 of FIG. 1.
`0051. In operation 404, an impact signal may be distin
`guished from an other signal by a frequency of the vibration
`caused by the user controlled impact. The vibration receiver
`may be a microphone.
`0052. In operation 406, a user may identify the command
`to the electronic device 150 based on the impact signal. The
`user command may be identified using a frequency, an ampli
`tude, a signal duration, a frequency spectrum, a signal energy,
`a signal rising edge, a signal falling edge, a signal start, a
`signal end, a mechanically transmitted vibration, a duration
`between impacts, and/or a number of impacts and/or a device
`context to identify the impact signal as an impact command.
`0053. In operation 408, the predesignated action may be
`determined based on a context (e.g., an electronic device
`operation, a power status, a ringer operation, an open com
`munication with another person using another electronic
`device, etc.) of the electronic device 150. The user controlled
`impact may occur when a user may cause a physical object to
`contact a second physical object above a threshold Velocity
`and/or a threshold force.
`0054 The first physical object may be coupled to a user's
`hand (e.g., a user's finger, a user's arm, a user's elbow, a
`stylus, a pen, a pencil, etc.) and/or foot (e.g., a toe, a heel, a
`shoe, a knee, etc.). The first physical object may be a user's
`finger and/or toe. The second physical object (e.g., a table top,
`an automobile dashboard, a bridge Support structure, a purse,
`a wallet, a mobile device protective cover, etc.) may be
`mechanically coupled to and/or include a housing of the
`electronic device 150.
`
`10
`
`
`
`US 2010/008521.6 A1
`
`Apr. 8, 2010
`
`0055. In operation 410, a predesignated action may be
`performed based on the user command. The u